Python 3.x 如何使用Jupyter笔记本运行Spark(Anaconda3,Ubuntu)
我确实下载了anaconda3、spark1.6.1、java、git等 我设置了这样的环境Python 3.x 如何使用Jupyter笔记本运行Spark(Anaconda3,Ubuntu),python-3.x,apache-spark,pyspark,jupyter-notebook,jupyter,Python 3.x,Apache Spark,Pyspark,Jupyter Notebook,Jupyter,我确实下载了anaconda3、spark1.6.1、java、git等 我设置了这样的环境 $ nano .bashrc export PATH="/home/moon/anaconda3/bin:$PATH" export SCALA_HOME=/usr/local/src/scala/scala-2.12.1 export PATH=$SCALA_HOME/bin:$PATH export SPARK_PATH=~/spark-1.6.1-bin-hadoop2.6 export PA
$ nano .bashrc
export PATH="/home/moon/anaconda3/bin:$PATH"
export SCALA_HOME=/usr/local/src/scala/scala-2.12.1
export PATH=$SCALA_HOME/bin:$PATH
export SPARK_PATH=~/spark-1.6.1-bin-hadoop2.6
export PATH=$SPARK_PATH/bin:$PATH
export ANACONDA_ROOT=/usr/home/moon/anaconda3
export PYSPARK_PYTHON=$ANACONDA_ROOT/bin/python3
export PYSPARK_DRIVER_PYTHON="jupyter"
export PYSPARK_DRIVER_PYTHON_OPTS="notebook"
$. .bashrc
$cd spark-1.6.1-bin-hadoop2.6
$bin/pyspark
然后我可以连接jupyter笔记本电脑(python3)但我无法运行spark ex)sc,sc.textFile 如何集成spark和jupyter请添加以下内容
PYSPARK_DRIVER_PYTHON=ipython PYSPARK_DRIVER_PYTHON_OPTS="notebook" ./bin/pyspark
另请检查以下内容:
我遇到了一些与anaconda python 3的兼容性问题,我切换到了anaconda python 2.7