Pyspark:尝试连接到Java服务器时出错
我正在JupyterNotebook中开始我的PypSpark课程:Pyspark:尝试连接到Java服务器时出错,java,apache-spark,pyspark,apache-spark-sql,Java,Apache Spark,Pyspark,Apache Spark Sql,我正在JupyterNotebook中开始我的PypSpark课程: spark = (SparkSession.builder.appName('myApp') .config("spark.executor.memory", "4g") .config("spark.executor.instances", "20") .config("spark.
spark = (SparkSession.builder.appName('myApp')
.config("spark.executor.memory", "4g")
.config("spark.executor.instances", "20")
.config("spark.master", "yarn")
.config("spark.driver.memory", "8g")
.config("spark.ui.port", "4373")
.config("spark.driver.maxResultSize", "1g")
.enableHiveSupport()
.getOrCreate())
但后来出现了这个错误:
Py4JNetworkError: An error occurred while trying
to connect to the Java server (127.0.0.1:34032)
有时,SparkSession
会很好地启动,但稍后当我处理脚本时,它会显示错误。
我想知道如何修复它