Warning: file_get_contents(/data/phpspider/zhask/data//catemap/2/scala/17.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
使用spark/scala从hdfs读取csv_Scala_Apache Spark_Hadoop_Hdfs - Fatal编程技术网

使用spark/scala从hdfs读取csv

使用spark/scala从hdfs读取csv,scala,apache-spark,hadoop,hdfs,Scala,Apache Spark,Hadoop,Hdfs,我正在使用spark 2.3.0和hadoop 2.9.1 我正在尝试用spark加载位于hdfs中的CSV文件 scala> val dataframe = spark.read.format("com.databricks.spark.csv").option("header","true").schema(schema).load("hdfs://127.0.0.1:50075/filesHDFS/data.csv") 但我得到了以下错误: 2018-11-14 11:47:58

我正在使用spark 2.3.0和hadoop 2.9.1 我正在尝试用spark加载位于hdfs中的CSV文件

scala> val dataframe = spark.read.format("com.databricks.spark.csv").option("header","true").schema(schema).load("hdfs://127.0.0.1:50075/filesHDFS/data.csv")
但我得到了以下错误:

2018-11-14 11:47:58 WARN  FileStreamSink:66 - Error while looking for metadata directory.
java.io.IOException: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: Protocol message end-group tag did not match expected tag.; Host Details : local host is: "Desktop-Presario-CQ42-Notebook-PC/127.0.0.1"; destination host is: "localhost":50070;

不要使用127.0.0.1,而是使用默认的FS名称。您可以在属性fs.defaultFS下的core-site.xml文件中找到它

它应该能解决你的问题