I am using JupyterHub on AWS EMR cluster. I am using EMR version 5.16
I submitted a spark application using a pyspark3 notebook. My application is trying to write 1TB data to s3. I am using autoscaling feature of the EMR to scale us the task node.
Hardware configurations: 1.Master node:32 GB RAM with 16 cores 2.Core node:32 GB RAM with 16 cores 3.Task node:16 GB with 8 cores each. (Task nodes scales up 15)
I have observed that Spark application gets killed after running for 50 to 60 minutes. I tried debugging: 1. My cluster still had scope for scaling up. So it is not an issue with a shortage of resources. 2. Livy session also gets killed. 3. In the job log, I saw error message RECVD TERM SIGNAL "Shutdown hook received"
Please note: 1. I have kept :spark.dynamicAllocation.enabled=true" 2. I am using the yarn fair scheduler with user impersonation in Jupiter hub
Can you please help me in understanding the problem and solution for it?