Python 在google colab上的spark中加载tsv文件时出错 file_loc='T3.tsv' T3_df=spark.read.csv(文件_loc,标题=False,sep=“\t”)
我在读取文件时收到输入/输出错误,称为Py4JJavaError-- 调用o29.csv时出错: org.apache.spark.sparkeexception:作业因阶段失败而中止:阶段0.0中的任务0失败1次,最近的失败:阶段0.0中的任务0.0丢失(TID 0,localhost,executor driver):org.apache.hadoop.fs.FSError:java.io.IOException:org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileInputStream.read处的输入/输出错误(RawLocalFileSystem.java:163)位于java.io.BufferedInputStream.read1(BufferedInputStream.java:284)Python 在google colab上的spark中加载tsv文件时出错 file_loc='T3.tsv' T3_df=spark.read.csv(文件_loc,标题=False,sep=“\t”),python,pyspark,apache-spark-sql,Python,Pyspark,Apache Spark Sql,我在读取文件时收到输入/输出错误,称为Py4JJavaError-- 调用o29.csv时出错: org.apache.spark.sparkeexception:作业因阶段失败而中止:阶段0.0中的任务0失败1次,最近的失败:阶段0.0中的任务0.0丢失(TID 0,localhost,executor driver):org.apache.hadoop.fs.FSError:java.io.IOException:org.apache.hadoop.fs.RawLocalFileSystem
原因:java.io.IOException:java.io.FileInputStream.readBytes(本机方法)处java.io.FileInputStream.read(FileInputStream.java:255)处的输入/输出错误
<b>file_loc = 'T3.tsv'
T3_df = spark.read.csv(file_loc, header=False, sep="\t")</b>