Apache spark IntelliJ中的结构化流媒体未向控制台显示数据帧
我正在尝试使用结构化流加载Spark流数据帧,但使用IntelliJ Idea无法在控制台中看到任何输出 我的代码:Apache spark IntelliJ中的结构化流媒体未向控制台显示数据帧,apache-spark,apache-spark-sql,spark-structured-streaming,Apache Spark,Apache Spark Sql,Spark Structured Streaming,我正在尝试使用结构化流加载Spark流数据帧,但使用IntelliJ Idea无法在控制台中看到任何输出 我的代码: import org.apache.spark.sql_ 对象SparkConsumerTest{ def main(参数:数组[字符串]):单位={ System.setProperty(“hadoop.home.dir”,“C:\\hadoop\”) val spark=火花会话 建设者 .appName(“本地测试”) .config(“spark.master”、“本地”
import org.apache.spark.sql_
对象SparkConsumerTest{
def main(参数:数组[字符串]):单位={
System.setProperty(“hadoop.home.dir”,“C:\\hadoop\”)
val spark=火花会话
建设者
.appName(“本地测试”)
.config(“spark.master”、“本地”)
.getOrCreate()
val data_stream=spark.readStream.text(“src/main/resources/data_string.txt”)
val result=data_stream.writeStream.format(“控制台”).start()
我的data_string.txt文件中有什么内容
structured streaming
这是IntelliJ Idea中运行应用程序后的控制台/运行窗口
Using Spark's default log4j profile:
org/apache/spark/log4j-defaults.properties
18/09/07 19:03:33 INFO SparkContext: Running Spark version 2.1.0
18/09/07 19:03:33 WARN NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where
applicable
18/09/07 19:03:33 INFO SecurityManager: Changing view acls to: userID
18/09/07 19:03:33 INFO SecurityManager: Changing modify acls to:
userID
18/09/07 19:03:33 INFO SecurityManager: Changing view acls groups to:
18/09/07 19:03:33 INFO SecurityManager: Changing modify acls groups
to:
18/09/07 19:03:33 INFO SecurityManager: SecurityManager:
authentication disabled; ui acls disabled; users with view
permissions: Set(userID); groups with view permissions: Set(); users
with modify permissions: Set(userID); groups with modify permissions:
Set()
18/09/07 19:03:34 INFO Utils: Successfully started service
'sparkDriver' on port 60845.
18/09/07 19:03:34 INFO SparkEnv: Registering MapOutputTracker
18/09/07 19:03:34 INFO SparkEnv: Registering BlockManagerMaster
18/09/07 19:03:34 INFO BlockManagerMasterEndpoint: Using
org.apache.spark.storage.DefaultTopologyMapper for getting topology
information
18/09/07 19:03:34 INFO BlockManagerMasterEndpoint:
BlockManagerMasterEndpoint up
18/09/07 19:03:34 INFO DiskBlockManager: Created local directory at
C:\Users\userID\AppData\Local\Temp\etc...
18/09/07 19:03:34 INFO MemoryStore: MemoryStore started with capacity
1983.3 MB
18/09/07 19:03:34 INFO SparkEnv: Registering OutputCommitCoordinator
18/09/07 19:03:35 INFO Utils: Successfully started service 'SparkUI'
on port 4040.
18/09/07 19:03:35 INFO SparkUI: Bound SparkUI to
0.0.0.0, and started at "http address"
18/09/07 19:03:35 INFO Executor: Starting executor ID driver on host
localhost
18/09/07 19:03:35 INFO Utils: Successfully started service
'org.apache.spark.network.netty.NettyBlockTransferService' on port
60855.
18/09/07 19:03:35 INFO NettyBlockTransferService: Server created on
"server address"
18/09/07 19:03:35 INFO BlockManager: Using
org.apache.spark.storage.RandomBlockReplicationPolicy for block
replication policy
18/09/07 19:03:35 INFO BlockManagerMaster: Registering BlockManager
BlockManagerId(driver, server address, 60855, None)
18/09/07 19:03:35 INFO BlockManagerMasterEndpoint: Registering block
manager server address with 1983.3 MB RAM, BlockManagerId(driver,
server address, 60855, None)
18/09/07 19:03:35 INFO BlockManagerMaster: Registered BlockManager
BlockManagerId(driver, server address, 60855, None)
18/09/07 19:03:35 INFO BlockManager: Initialized BlockManager:
BlockManagerId(driver,
server address, 60855, None)
18/09/07 19:03:35 INFO SharedState: Warehouse path is
'file:/C:/Users/userid/Documents//SparkTestLocal/spark-warehouse/'.
Process finished with exit code 0
这是因为您没有将等待终止。您需要添加以下内容
result.waittermination()
在第行开始查询后
val result=data\u stream.writeStream.format(“控制台”).start()
希望这能有所帮助。这成功了,谢谢!我想知道为什么结构化流媒体指南的示例中没有更清楚地说明这一点。