Warning: file_get_contents(/data/phpspider/zhask/data//catemap/7/neo4j/3.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
ApacheKylin-多维数据集构建的第一步输出错误_Apache_Hadoop_Hive_Olap_Kylin - Fatal编程技术网

ApacheKylin-多维数据集构建的第一步输出错误

ApacheKylin-多维数据集构建的第一步输出错误,apache,hadoop,hive,olap,kylin,Apache,Hadoop,Hive,Olap,Kylin,我正在尝试使用Apache Kylin构建我的第一个多维数据集,在最后一步出现错误之前,一切都很顺利: java.lang.IllegalStateException: Can't get cube source record count. at com.google.common.base.Preconditions.checkState(Preconditions.java:149) at org.apache.kylin.job.cube.UpdateCubeInfoAfterBuildS

我正在尝试使用Apache Kylin构建我的第一个多维数据集,在最后一步出现错误之前,一切都很顺利:

java.lang.IllegalStateException: Can't get cube source record count.
at com.google.common.base.Preconditions.checkState(Preconditions.java:149)
at org.apache.kylin.job.cube.UpdateCubeInfoAfterBuildStep.doWork(UpdateCubeInfoAfterBuildStep.java:104)
at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
根据此问题https://github.com/kylinap/Kylin/issues/101,出现上述错误是因为Kylin试图在配置单元的输出中找到此模式:“HDFS Read:(\d+)HDFS Write:(\d+)SUCCESS”

多维数据集构建教程的正确输出:github.com/kylinap/Kylin/wiki/Kylin-cube-Build-and-Job-Monitoring-tutorial:

我在Kylin中的输出看起来已损坏:

但是,当我检查配置单元日志时,它看起来正常:

2015-05-27 08:40:13,419 INFO  [main]: ql.Driver (Driver.java:execute(1285)) - Starting command: 
INSERT OVERWRITE TABLE kylin_intermediate_Kubek_19700101000000_2922789940817071255_f23ac1b1_10fe_4112_ac9e_b4e6baf07654 SELECT
FACT_TABLE.DATE
,FACT_TABLE.MONEY_ADVERTISER
,FACT_TABLE.MONEY_PUBLISHER
FROM DEFAULT.ADVSTATS as FACT_TABLE 
...
2015-05-27 08:45:05,132 INFO  [main]: ql.Driver (SessionState.java:printInfo(824)) - MapReduce Jobs Launched: 
2015-05-27 08:45:05,148 INFO  [main]: ql.Driver (SessionState.java:printInfo(824)) - Stage-Stage-1: Map: 1   Cumulative CPU: 17.32 sec   HDFS Read: 44644035 HDFS Write: 2347008 SUCCESS
2015-05-27 08:45:05,153 INFO  [main]: ql.Driver (SessionState.java:printInfo(824)) - Total MapReduce CPU Time Spent: 17 seconds 320 msec
2015-05-27 08:45:05,167 INFO  [main]: ql.Driver (SessionState.java:printInfo(824)) - OK
我正在使用Hortonworks沙盒2.2:

hadoop-2.6.0
hbase-0.98.12
hive-0.14.0
zookeeper-3.4.6

有人能告诉我为什么我的kylin日志预览是这样的吗?首先,这可能是最后一步出错的原因吗?

几天前kylin开发人员承诺绕过此类问题

因为有时候hadoop可能无法获得计数器,即使作业成功 从现在起,当获取多维数据集源失败时,Kylin将发出警告而不是错误


我所要做的就是重建Kylin。

请将此问题及时提交给邮件列表以获得支持: