Warning: file_get_contents(/data/phpspider/zhask/data//catemap/3/apache-spark/5.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
Apache spark 块推送线程错误:Kafka Spark streaming_Apache Spark_Apache Kafka_Spark Streaming - Fatal编程技术网

Apache spark 块推送线程错误:Kafka Spark streaming

Apache spark 块推送线程错误:Kafka Spark streaming,apache-spark,apache-kafka,spark-streaming,Apache Spark,Apache Kafka,Spark Streaming,运行kafka consumer时发生以下错误: ERROR receiver.BlockGenerator: Error in block pushing thread java.io.NotSerializableException: org.jnetpcap.packet.PcapPacket at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1183) at java.io.ObjectO

运行kafka consumer时发生以下错误:

    ERROR receiver.BlockGenerator: Error in block pushing thread
    java.io.NotSerializableException: org.jnetpcap.packet.PcapPacket
at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1183)
at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1547)
at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1508)
at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1431)
at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1177)
at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:347)
at org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:42)
at org.apache.spark.serializer.SerializationStream$class.writeAll(Serializer.scala:102)
at org.apache.spark.serializer.JavaSerializationStream.writeAll(JavaSerializer.scala:30)
at org.apache.spark.storage.BlockManager.dataSerializeStream(BlockManager.scala:996)
at org.apache.spark.storage.BlockManager.dataSerialize(BlockManager.scala:1005)
at org.apache.spark.storage.MemoryStore.putValues(MemoryStore.scala:79)
at org.apache.spark.storage.BlockManager.doPut(BlockManager.scala:663)
at org.apache.spark.storage.BlockManager.put(BlockManager.scala:574)
build.sbt文件:

name := "testpacket"
version := "1.0"
scalaVersion := "2.10.3"
libraryDependencies += "org.apache.spark" % "spark-core_2.10" % "1.0.2
libraryDependencies += "org.apache.spark" % "spark-streaming_2.10" % "1.0.2"
libraryDependencies += "org.apache.spark" % "spark-streaming-kafka_2.10" % "1.0.2"
libraryDependencies += "javax.servlet" % "javax.servlet-api" % "3.0.1"
resolvers += "Akka Repository" at "http://repo.akka.io/releases/"

错误的原因可能是什么?

我以前在两种情况下遇到过这个问题,因此没有看到您的代码,我无法确切确定问题是什么

  • 您正在RDD数据集中包含不可序列化的类
  • 您正在仅驱动程序类/函数/代码之外导入不可序列化的类
  • 我的猜测是,您正在体验#1,并且将PcapPacket作为RDD的一部分。如果是这种情况,那么您需要创建PcapPacket的可序列化版本,这应该不会太困难,因为基础字节数组支持PcapPacket