We have tried to submit a job in Spark on Yarn, which will import data from HDFS to Apache Ignite. So, we need to specify the Ignite configuration file path for Spark containers.
The examples on the Ignite website only define the path like "conf/cache.xml" then Spark drivers and executors "magically" find the file, but I don't understand how Spark executors find it.
We have tried several ways, none worked:
Specify a full path like "file:///disk1/conf/cache.xml" in the code
Upload the config file to HDFS and specify it like "hdfs:///hdfs_root/conf/cache.xml"
Specify the full path in spark-defaults.conf, in the parameters spark.{driver,executor}.extraClassPath
Do we have to put the Ignite config file in every Yarn node for Ignite to work with Spark on Yarn ? Is there any better approach for this ?