7
votes

I am new to Spark and Python. I have installed python 3.5.1 and Spark-1.6.0-bin-hadoop2.4 on windows.

I am getting below error when I execute sc = SparkContext("local", "Simple App") from python shell..

Can you please help?

from pyspark import SparkConf, SparkContext

sc = SparkContext("local", "Simple App")

Traceback (most recent call last):

File "", line 1, in

sc = SparkContext("local", "Simple App")

File "C:\spark-1.6.0-bin-hadoop2.4\python\pyspark\context.py", line 112, in init

SparkContext._ensure_initialized(self, gateway=gateway)

File "C:\spark-1.6.0-bin-hadoop2.4\python\pyspark\context.py", line 245, in _ensure_initialized

SparkContext._gateway = gateway or launch_gateway()

File "C:\spark-1.6.0-bin-hadoop2.4\python\pyspark\java_gateway.py", line 79, in launch_gateway

proc = Popen(command, stdin=PIPE, env=env)

File "C:\Python35-32\lib\subprocess.py", line 950, in init restore_signals, start_new_session)

File "C:\Python35-32\lib\subprocess.py", line 1220, in _execute_child startupinfo)

FileNotFoundError: [WinError 2] The system cannot find the file specified

3
hi,did you try with sc = SparkContext("local", "Simple\ App") ?Hossein Vatani
did you solve this?Laurens Koppenol

3 Answers

2
votes

You have to set the SPARK_HOME correctly. I have debugged the python scripts and verified. This would work.

2
votes

Check your address to make sure it is written correctly. In my case, I had the address as:

"C:/Users/nekooeimehr/AppData/Local/Programs/Python/Python35-32/spark-1.6.2-bin-hadoop2.4"

while the correct address is:

"C:/Users/nekooeimehr/AppData/Local/Programs/Python/Python35-32/spark-1.6.2-bin-hadoop2.4/spark-1.6.2-bin-hadoop2.4"

0
votes

If the verified the system environment variables and it is still not working, check if your jvm (64 or 32)bit version is compatible with your machine.