I am trying to run this code:
rddCollected=rddCollect.mapValues(lambda x: (x,1))
rddCollected.collect()
rddCollectJoin=rddCollected.reduceByKey(lambda x,y:(x[0]+y[0],x[1]+y[1]))
--rddCollected is running fine with collect but rddCollectJoin is not working and giving below error .
Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.runJob. : org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 58.0 failed 1 times, most recent failure: Lost task 0.0 in stage 58.0 (TID 78, localhost, executor driver): java.io.FileNotFoundException: C:\Users\lenovo\AppData\Local\Temp\blockmgr-431169ff-717a-4728-b9b2-c2ed1b4b5b20\0c\temp_shuffle_d089dc45-014d-4d07-b0c0-ee917ad1b501 (The system cannot find the path specified)
Java version is 1.8- I had java 10 but I downsized it to 8 as there were issues with 10 can anyone help?