Warning: file_get_contents(/data/phpspider/zhask/data//catemap/3/apache-spark/6.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
Apache spark “意外”;无法实例化Tachyon文件系统";对于HDFS检查点目录_Apache Spark_Persistence - Fatal编程技术网

Apache spark “意外”;无法实例化Tachyon文件系统";对于HDFS检查点目录

Apache spark “意外”;无法实例化Tachyon文件系统";对于HDFS检查点目录,apache-spark,persistence,Apache Spark,Persistence,对于一个已经运行并执行了几十次的spark程序,突然在以下设置检查点目录的逻辑上发生了一个有趣的文件系统错误: val tempDir = s"alsTest" sc.setCheckpointDir(tempDir) 以下是错误: org.apache.hadoop.fs.FileSystem: Provider tachyon.hadoop.TFS could not be instantiated 以下是完整的stacktrace: Exception in thread "main"

对于一个已经运行并执行了几十次的spark程序,突然在以下设置检查点目录的逻辑上发生了一个有趣的文件系统错误:

val tempDir = s"alsTest"
sc.setCheckpointDir(tempDir)
以下是错误:

org.apache.hadoop.fs.FileSystem: Provider tachyon.hadoop.TFS could not be instantiated
以下是完整的stacktrace:

Exception in thread "main" java.util.ServiceConfigurationError: org.apache.hadoop.fs.FileSystem: Provider tachyon.hadoop.TFS could not be instantiated
    at java.util.ServiceLoader.fail(ServiceLoader.java:232)
    at java.util.ServiceLoader.access$100(ServiceLoader.java:185)
    at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:384)
    at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404)
    at java.util.ServiceLoader$1.next(ServiceLoader.java:480)
    at org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:2400)
    at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2411)
    at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
    at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
    at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
    at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
    at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
    at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
    at org.apache.spark.SparkContext$$anonfun$setCheckpointDir$2.apply(SparkContext.scala:2076)
    at org.apache.spark.SparkContext$$anonfun$setCheckpointDir$2.apply(SparkContext.scala:2074)
    at scala.Option.map(Option.scala:145)
    at org.apache.spark.SparkContext.setCheckpointDir(SparkContext.scala:2074)
    at com.blazedb.spark.ml.AlsTest$.main(AlsTest.scala:331)
    at com.blazedb.spark.ml.AlsTest.main(AlsTest.scala)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at com.intellij.rt.execution.application.AppMain.main(AppMain.java:147)
Caused by: java.lang.ExceptionInInitializerError
    at tachyon.Constants.<clinit>(Constants.java:328)
    at tachyon.hadoop.AbstractTFS.<clinit>(AbstractTFS.java:63)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
    at java.lang.Class.newInstance(Class.java:442)
    at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:380)
    ... 21 more
Caused by: java.lang.RuntimeException: java.net.ConnectException: Permission denied (connect failed)
    at com.google.common.base.Throwables.propagate(Throwables.java:160)
    at tachyon.util.network.NetworkAddressUtils.getLocalIpAddress(NetworkAddressUtils.java:398)
    at tachyon.util.network.NetworkAddressUtils.getLocalHostName(NetworkAddressUtils.java:320)
    at tachyon.conf.TachyonConf.<init>(TachyonConf.java:122)
    at tachyon.conf.TachyonConf.<init>(TachyonConf.java:111)
    at tachyon.Version.<clinit>(Version.java:27)
    ... 29 more
Caused by: java.net.ConnectException: Permission denied (connect failed)
    at java.net.Inet6AddressImpl.isReachable0(Native Method)
    at java.net.Inet6AddressImpl.isReachable(Inet6AddressImpl.java:77)
    at java.net.InetAddress.isReachable(InetAddress.java:502)
    at java.net.InetAddress.isReachable(InetAddress.java:461)
    at tachyon.util.network.NetworkAddressUtils.isValidAddress(NetworkAddressUtils.java:414)
    at tachyon.util.network.NetworkAddressUtils.getLocalIpAddress(NetworkAddressUtils.java:382)
    ... 33 more
那么为什么Spark文件系统类现在要尝试访问堆外的
(tachyon)

Update这变得越来越有趣:即使明确指定
hdfs
URL,也会导致超光速子错误

val tempDir = s"hdfs://$host:8020:alsTest/"
sc.setCheckpointDir(tempDir)

<same error as above>
val tempDir=s“hdfs://$host:8020:alsTest/”
sc.setCheckpointDir(tempDir)

问题在于昨天在我的系统上首次启用的新的VPN软件,当时VPN软件被挂起,然后
HDFS
url再次得到正确解决

val tempDir = s"hdfs://$host:8020:alsTest/"
sc.setCheckpointDir(tempDir)

<same error as above>