I setup a Hortonworks Hadoop cluster:
- Hortonworks version is 2.3.2.
- 1 NameNode, 1 Secondary NameNode, 10 DataNode
- Spark 1.4.1 and deployed on all data node.
- YARN is installed.
When I run a spark program, the executor is only run on 4 nodes but not whole data nodes.
How to estimate amount of spark executor on such Hadoop cluster?