Java 读取Spark中目录的内容

Java 读取Spark中目录的内容,java,apache-spark,Java,Apache Spark,我是spark的新手,只是想知道如何读取目录的内容并对其进行迭代。C#对应代码 Foreach(var path in Directory. EnumerateFiles(directory, *,.. ) {} ; JavaSparkContext jsc=新的JavaSparkContext(sc); javapairdd rdd=jsc.wholeTextFiles(路径); 对于(Tuple2 str:rdd.toArray()){System.out.println(“+; Syst

我是spark的新手,只是想知道如何读取目录的内容并对其进行迭代。C#对应代码

Foreach(var path in Directory. EnumerateFiles(directory, *,.. ) {} ;
JavaSparkContext jsc=新的JavaSparkContext(sc);
javapairdd rdd=jsc.wholeTextFiles(路径);
对于(Tuple2 str:rdd.toArray()){System.out.println(“+;
System.out.println(“文件名”+str.\u 1);
System.out.println(“+;
System.out.println();
System.out.println(“-----------------------------------------”;
System.out.println(“内容”+str.\u 2);
System.out.println(“-----------------------------------------”;
}
希望有帮助,我也有同样的问题

  JavaSparkContext jsc = new JavaSparkContext(sc);
  JavaPairRDD<String,String> rdd = jsc.wholeTextFiles(path);
          for(Tuple2<String, String> str : rdd.toArray()) {           System.out.println("+++++++++++++++++++++++++++++++++++++++++++");
      System.out.println("File name " + str._1);
      System.out.println("+++++++++++++++++++++++++++++++++++++++++++");
      System.out.println();
      System.out.println("-------------------------------------------");
      System.out.println("content " + str._2);
      System.out.println("-------------------------------------------");
  }