Apache spark 连接到Spark Thrift服务器的ODBC配置

Apache spark 连接到Spark Thrift服务器的ODBC配置,apache-spark,hive,spark-thriftserver,Apache Spark,Hive,Spark Thriftserver,这个问题可能会重复,事实上,我已经看到了一些与此相关的问题,但不是完全相同的错误,所以我想看看是否有人有线索 我已经设置了一个运行默认设置的Spark Thrift服务器。Spark版本是2.1,它在纱线上运行(Hadoop 2.7.3) 事实上,我无法设置Simba hive ODBC驱动程序或Microsoft ODBC驱动程序,因此ODBC设置中的测试无法成功 这是我用于Microsoft配置单元ODBC驱动程序的配置: 当我按下测试按钮时,显示的错误消息如下: 在Spark Thri

这个问题可能会重复,事实上,我已经看到了一些与此相关的问题,但不是完全相同的错误,所以我想看看是否有人有线索

我已经设置了一个运行默认设置的Spark Thrift服务器。Spark版本是2.1,它在纱线上运行(Hadoop 2.7.3)

事实上,我无法设置Simba hive ODBC驱动程序或Microsoft ODBC驱动程序,因此ODBC设置中的测试无法成功

这是我用于Microsoft配置单元ODBC驱动程序的配置:

当我按下测试按钮时,显示的错误消息如下:

在Spark Thrift服务器日志中,可以看到以下内容:

17/09/15 17:31:36 INFO ThriftCLIService: Client protocol version: HIVE_CLI_SERVICE_PROTOCOL_V1
17/09/15 17:31:36 INFO SessionState: Created local directory: /tmp/00abf145-2928-4995-81f2-fea578280c42_resources
17/09/15 17:31:36 INFO SessionState: Created HDFS directory: /tmp/hive/test/00abf145-2928-4995-81f2-fea578280c42
17/09/15 17:31:36 INFO SessionState: Created local directory: /tmp/vagrant/00abf145-2928-4995-81f2-fea578280c42
17/09/15 17:31:36 INFO SessionState: Created HDFS directory: /tmp/hive/test/00abf145-2928-4995-81f2-fea578280c42/_tmp_space.db
17/09/15 17:31:36 INFO HiveSessionImpl: Operation log session directory is created: /tmp/vagrant/operation_logs/00abf145-2928-4995-81f2-fea578280c42
17/09/15 17:31:36 INFO SparkExecuteStatementOperation: Running query 'set -v' with 82d7f9a6-f2a6-4ebd-93bb-5c8da1611f84
17/09/15 17:31:36 INFO SparkSqlParser: Parsing command: set -v
17/09/15 17:31:36 INFO SparkExecuteStatementOperation: Result Schema: StructType(StructField(key,StringType,false), StructField(value,StringType,false), StructField(meaning,StringType,false))
如果我使用JDBC驱动程序通过Beeline(正常工作)进行连接,以下是日志:

17/09/15 17:04:24 INFO ThriftCLIService: Client protocol version: HIVE_CLI_SERVICE_PROTOCOL_V8
17/09/15 17:04:24 INFO SessionState: Created HDFS directory: /tmp/hive/test
17/09/15 17:04:24 INFO SessionState: Created local directory: /tmp/c0681d6f-cc0f-40ae-970d-e3ea366aa414_resources
17/09/15 17:04:24 INFO SessionState: Created HDFS directory: /tmp/hive/test/c0681d6f-cc0f-40ae-970d-e3ea366aa414
17/09/15 17:04:24 INFO SessionState: Created local directory: /tmp/vagrant/c0681d6f-cc0f-40ae-970d-e3ea366aa414
17/09/15 17:04:24 INFO SessionState: Created HDFS directory: /tmp/hive/test/c0681d6f-cc0f-40ae-970d-e3ea366aa414/_tmp_space.db
17/09/15 17:04:24 INFO HiveSessionImpl: Operation log session directory is created: /tmp/vagrant/operation_logs/c0681d6f-cc0f-40ae-970d-e3ea366aa414
17/09/15 17:04:24 INFO SparkSqlParser: Parsing command: use default
17/09/15 17:04:25 INFO HiveMetaStore: 1: get_database: default
17/09/15 17:04:25 INFO audit: ugi=vagrant   ip=unknown-ip-addr  cmd=get_database: default   
17/09/15 17:04:25 INFO HiveMetaStore: 1: Opening raw store with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
17/09/15 17:04:25 INFO ObjectStore: ObjectStore, initialize called
17/09/15 17:04:25 INFO Query: Reading in results for query "org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is closing
17/09/15 17:04:25 INFO MetaStoreDirectSql: Using direct SQL, underlying DB is DERBY
17/09/15 17:04:25 INFO ObjectStore: Initialized ObjectStore

通过安装MicrosoftSparkODBC驱动程序而不是Hive驱动程序,我成功地实现了连接。 问题似乎与驱动程序拒绝连接Spark Thrift Server有关,因为根据某些服务器属性发现它不是Hive2服务器。我怀疑Hive2和Spark thrift server之间在连线级别上是否存在实际差异,因为后者是前者的一个端口,在协议级别(thrift)上没有更改,但在任何情况下,解决方案都是移动到此驱动程序,并以与Hive2相同的方式配置它:


我通过安装Microsoft Spark ODBC驱动程序而不是Hive驱动程序成功地连接到了网络。 问题似乎与驱动程序拒绝连接Spark Thrift Server有关,因为根据某些服务器属性发现它不是Hive2服务器。我怀疑Hive2和Spark thrift server之间在连线级别上是否存在实际差异,因为后者是前者的一个端口,在协议级别(thrift)上没有更改,但在任何情况下,解决方案都是移动到此驱动程序,并以与Hive2相同的方式配置它: