I have a Spark cluster with 17 executors in total. I have integrated Spark 2.1 with Kafka and reading the data from topic like:
val df = spark
.readStream
.format("kafka")
.options("kafka.bootstrap.servers","localhost:9092")
.options("subscribe","test")
.load
Now I want to know that when I'll Submit my spark application in cluster mode how many executors (out of the total 17) will be assigned to listen to a Kafka topic and creating micro-batches in Structured Streaming.
Also, how can I limit the size of a micro-batch in Structured Streaming when reading from Kafka?