I have configured standalone spark cluster connected to Cassandra cluster with 1 master, 1 slave and Thrift server which is used as JDBC connector for Tableau application. Slave appears in workers list anyway when I launch any query worker does not seem to be used as executor (0 cores used). All workload is executed master executor. Also in Thrift web console I observe that only one executor is active.
Basically I expect distributed workload on both executors of spark cluster to achieve higher performance.
From master logs:
2019-03-26 15:36:52 INFO Master:54 - I have been elected leader! New state: ALIVE 2019-03-26 15:37:00 INFO Master:54 - Registering worker worker-ip:37678 with 16 cores, 61.8 GB RAM
From worker logs:
2019-03-26 15:37:00 INFO Worker:54 - Successfully registered with master spark://master-hostname:7077
My spark-defaults.conf is:
spark.driver.memory=50g
spark.driver.maxResultSize=4g
spark.sql.thriftServer.incrementalCollect=false
spark.sql.shuffle.partition=17
spark.sql.autoBroadcastJoinThreshold=10485760
spark.sql.inMemoryColumnarStorage.compressed=true
spark.sql.inMemoryColumnarStorage.batchSize=10000
spark.cores.max=32
spark.executor.cores=16
spark.memory.offHeap.enabled=true
spark.memory.offHeap.size=1g
pic1 workers
pic2 executors
Any help highly appreciated.