Splice Machine Troubleshooting and Best Practices

    Learn about our products

This topic provides troubleshooting guidance for these issues that you may encounter with your Splice Machine database:

Restarting Splice Machine After HMaster Failure

If you run Splice Machine without redundant HMasters, and you lose your HMaster, follow these steps to restart Splice Machine:

  1. Restart the HMaster node
  2. Restart every HRegion Server node

Slow Restart After Forced Shutdown

We have seen a situation where HMaster doesn’t exit when you attempt a shutdown, and a forced shutdown is used. The forced shutdown means that HBase may not be able to flush all data and delete all write-ahead logs (WALs); as a result, it can take longer than usual to restart HBase and Splice Machine.

Splice Machine now sets the HBase Graceful Shutdown Timeout to 10 minutes, which should be plenty of time. If the shutdown is still hanging up after 10 minutes, a forced shutdown is appropriate.

Updating Stored Query Plans after a Splice Machine Update

When you install a new version of your Splice Machine software, you need to make these two calls:

These calls will update the stored metadata query plans and purge the statement cache, which is required because the query plan APIs have changed. This is true for both minor (patch) releases and major new releases.

Increasing Parallelism for Spark Shuffles

You can adjust the minimum parallelism for Spark shuffles by adjusting the value of the splice.olap.shuffle.partitions configuration option.

This option is similar to the spark.sql.shuffle.partitions option, which configures the number of partitions to use when shuffling data for joins or aggregations; however, the spark.sql.shuffle.partitions option is set to allow a lower number of partitions than is optimal for certain operations.

Specifically, increasing the number of shuffle partitions with the splice.olap.shuffle.partitions option is useful when performing operations on small tables that generate large, intermediate datasets; additional, but smaller sized partitions allows us to operate with better parallelism.

The default value of splice.olap.shuffle.partitions is 200.

Increasing Memory Settings for Heavy Analytical Work Loads

If you are running heavy analytical loads or running OLAP jobs on very large tables, you may want to increase these property settings in your hbase-site.xml file:

Property Default Value (MB) Recommendations for Heavy Analytical Loads
splice.olap_server.memory 1024 Set to the same value as HMaster heap size
splice.olap_server.memoryOverhead 512 Set to 10% of splice.olap_server.memory
splice.olap_server.virtualCores 1 vCore 4 vCores
splice.olap_server.external true true

Force Compaction to Run on Local Region Server

Splice Machine attempts to run database compaction jobs on an executor that is co-located with the serving Region Server; if it cannot find a local executor after a period of time, Splice Machine uses whatever executor Spark executor it can get; to force use of a local executor, you can adjust the splice.spark.dynamicAllocation.minExecutors configuration option.

To do so:

  • Set the value of splice.spark.dynamicAllocation.minExecutors to the number of Region Servers in your cluster
  • Set the value of splice.spark.dynamicAllocation.maxExecutors to equal to or greater than that number. Adjust these setting in the Java Config Options section of your HBase Master configuration.

The default option settings are:


For a cluster with 20 Region Servers, you would set these to:


Kerberos Configuration Option

If you’re using Kerberos, you need to add this option to your HBase Master Java Configuration Options:


Resource Management for Backup Jobs

Splice Machine backup jobs use a Map Reduce job to copy HFiles; this process may hang up if the resources required for the Map Reduce job are not available from Yarn. To make sure the resources are available, follow these three configuration steps:

  1. Configure minimum executors for Splice Spark
  2. Verify that adequate vcores are available for Map Reduce tasks
  3. Verify that adequate memory is available for Map Reduce tasks

Configure the minimum number of executors allocated to Splice Spark

You need to make sure that both of the following configuration settings relationships hold true.

(splice.spark.dynamicAllocation.minExecutors + 1) < (yarn.nodemanager.resource.cpu-vcores * number_of_nodes)
(splice.spark.dynamicAllocation.minExecutors * (splice.spark.yarn.executor.memoryOverhead+splice.spark.executor.memory) + splice.spark.yarn.am.memory) < (yarn.nodemanager.resource.memory-mb * number_of_nodes)

The actual minExecutors allocated to Splice Spark may be less than specified in splice.spark.dynamicAllocation.minExecutors because of memory constraints in the container. Once Splice Spark is launched, Yarn will allocate the actual minExecutor value and memory to Splice Spark. You need to verify that enough vcores and memory remain available for Map Reduce tasks.

Verify that adequate vcores are available

The Map Reduce application master requires the following number of vcores:

yarn.app.mapreduce.am.resource.cpu-vcores * splice.backup.parallesim

There must be at least this many additional vcores available to execute Map Reduce tasks:


Thus, the total number of vcores that must be available for Map Reduce jobs is:

yarn.app.mapreduce.am.resource.cpu-vcores * splice.backup.parallesim + max{mapreduce.map.cpu.vcores,mapreduce.reduce.cpu.vcores}

Verify that adequate memory is available

The Map Reduce application master requires this much memory:

yarn.scheduler.minimum-allocation-mb * splice.backup.parallesim

There must be at least this much memory available to execute Map Reduce tasks:


Thus, the total number of memory that must be available for Map Reduce jobs is:

yarn.scheduler.minimum-allocation-mb * (splice.backup.parallesim+1)

Bulk Import of Very Large Datasets with Spark 2.2

When using Splice Machine with Spark 2.2 with Cloudera, bulk import of very large datasets can fail due to direct memory usage. Use the following settings to resolve this issue:

Update Shuffle-to-Mem Setting

Modify the following setting in the Cloudera Manager’s Java Configuration Options for HBase Master:


Update the YARN User Classpath

Modify the following settings in the Cloudera Manager’s YARN (MR2 Included) Service Environment Advanced Configuration Snippet (Safety Valve):