0
votes

I need help because I don't know if the Jupyter notebook Kernel are usable in a Spark cluster.

In my local Spark I use this and I don't have problems.

I am using this Kernel for PySpark : https://github.com/Anchormen/pyspark-jupyter-kernels

I am using a Standalone Spark cluster with three nodes without Yarn.

Best regard.

1

1 Answers

0
votes

You can connect to your spark cluster standalone using the master IP with the python kernel.

import pyspark 
sc = pyspark.SparkContext(master='spark://<public-ip>:7077', appName='<your_app_name>')

References