Hello people and happy new year ;) !
I am bulding a lambda architecture with Apache Spark, HDFS and Elastichsearch. In the following picture, here what I am trying to do:
So far, I have written the source code in java for my spark streaming and spark applications. I read in the spark documentation that spark can be run in a Mesos or YARN clutser. As indicated in the picture, I have already a hadoop cluster. Is it possible to run my spark streaming and spark application within the same hadoop cluster ? If yes, is there any particular configuration to do (for instance the number of nodes, RAM...). Or do I have to add a hadoop cluster specialy for spark streaming ?
I hope my explanation is clear.
Yassir