Warning: file_get_contents(/data/phpspider/zhask/data//catemap/3/apache-spark/5.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
Apache spark 在PySpark中读取多个文件?_Apache Spark_Pyspark - Fatal编程技术网

Apache spark 在PySpark中读取多个文件?

Apache spark 在PySpark中读取多个文件?,apache-spark,pyspark,Apache Spark,Pyspark,有没有一种方法可以一次读取多个序列文件 sc.SequnceFile(['filepath_1.seq','filepath_2.seq','filepath_3.seq']) 使用全局模式 sc.SequnceFile('filepath_*.seq') 或逗号分隔字符串: sc.SequnceFile(",".join(['filepath_1.seq', 'filepath_2.seq', 'filepath_3.seq']))

有没有一种方法可以一次读取多个序列文件

sc.SequnceFile(['filepath_1.seq','filepath_2.seq','filepath_3.seq'])
使用全局模式

sc.SequnceFile('filepath_*.seq')
或逗号分隔字符串:

sc.SequnceFile(",".join(['filepath_1.seq', 'filepath_2.seq', 'filepath_3.seq']))