Apache kafka Kafka Connect配置单元集成不创建配置单元表

Apache kafka Kafka Connect配置单元集成不创建配置单元表,apache-kafka,hive,apache-kafka-connect,Apache Kafka,Hive,Apache Kafka Connect,我有以下卡夫卡连接器配置: { "name": "some-topic-connector", "config": { "connector.class": "io.confluent.connect.hdfs.HdfsSinkConnector", "tasks.max": "1", "topics":

我有以下卡夫卡连接器配置:

{
  "name": "some-topic-connector",
  "config": {
    "connector.class": "io.confluent.connect.hdfs.HdfsSinkConnector",
    "tasks.max": "1",
    "topics": "some-topic",
    "hdfs.url": "hdfs://hadoopams1",
    "logs.dir": "apps/kafka-connect-preview/some-topic.logs",
    "topics.dir": "apps/kafka-connect-preview/some-topic.db",
    "hadoop.conf.dir": "/etc/hadoop/conf",
    "flush.size": "1000000",
    "rotate.interval.ms": "3600000",
    "rotate.schedule.interval.ms": "86400000",
    "hive.integration": "true",
    "hive.metastore.uris": "thrift://metastore-1.hadoop-1.foobar.com:9083",
    "hive.database": "preview",
    "locale": "en_GB",
    "timezone": "Europe/Berlin",
    "key.converter": "org.apache.kafka.connect.storage.StringConverter",
    "value.converter": "io.confluent.connect.avro.AvroConverter",
    "value.converter.schema.registry.url": "http://schema-registry.preview.foobar.com",
    "schema.compatibility": "BACKWARD",
    "partitioner.class": "io.confluent.connect.storage.partitioner.TimeBasedPartitioner",
    "partition.duration.ms": "86400000",
    "path.format": "'dt'=YYYYMMdd",
    "partition.field.name": "dt"
  }
}
我已验证数据已成功写入HDFS,但由于某些原因,配置单元中的表未被创建。从日志中,我看不到Kafka Connect中的任何错误

我做错了什么?是否缺少一些配置或要求?

如果
logs.dir
topics.dir
已经存在,则
HdfsSinkConnector
不会在配置单元中创建表。例如,如果您决定在连接器创建后的某个时间点启用配置单元集成,则可能会发生这种情况

还有一个解决此问题的方法,但尚未合并

所以

  • 您可以根据上面链接的拉动请求构建自己的
    HdfsSinkConnector
  • 重命名目录,重新创建连接器,直到创建配置单元表,然后将表移回(当然,在生产环境中很难)
  • 或者手动创建表