Apache kafka Kafka Connect配置单元集成不创建配置单元表
我有以下卡夫卡连接器配置:Apache kafka Kafka Connect配置单元集成不创建配置单元表,apache-kafka,hive,apache-kafka-connect,Apache Kafka,Hive,Apache Kafka Connect,我有以下卡夫卡连接器配置: { "name": "some-topic-connector", "config": { "connector.class": "io.confluent.connect.hdfs.HdfsSinkConnector", "tasks.max": "1", "topics":
{
"name": "some-topic-connector",
"config": {
"connector.class": "io.confluent.connect.hdfs.HdfsSinkConnector",
"tasks.max": "1",
"topics": "some-topic",
"hdfs.url": "hdfs://hadoopams1",
"logs.dir": "apps/kafka-connect-preview/some-topic.logs",
"topics.dir": "apps/kafka-connect-preview/some-topic.db",
"hadoop.conf.dir": "/etc/hadoop/conf",
"flush.size": "1000000",
"rotate.interval.ms": "3600000",
"rotate.schedule.interval.ms": "86400000",
"hive.integration": "true",
"hive.metastore.uris": "thrift://metastore-1.hadoop-1.foobar.com:9083",
"hive.database": "preview",
"locale": "en_GB",
"timezone": "Europe/Berlin",
"key.converter": "org.apache.kafka.connect.storage.StringConverter",
"value.converter": "io.confluent.connect.avro.AvroConverter",
"value.converter.schema.registry.url": "http://schema-registry.preview.foobar.com",
"schema.compatibility": "BACKWARD",
"partitioner.class": "io.confluent.connect.storage.partitioner.TimeBasedPartitioner",
"partition.duration.ms": "86400000",
"path.format": "'dt'=YYYYMMdd",
"partition.field.name": "dt"
}
}
我已验证数据已成功写入HDFS,但由于某些原因,配置单元中的表未被创建。从日志中,我看不到Kafka Connect中的任何错误
我做错了什么?是否缺少一些配置或要求?如果logs.dir
和topics.dir
已经存在,则HdfsSinkConnector
不会在配置单元中创建表。例如,如果您决定在连接器创建后的某个时间点启用配置单元集成,则可能会发生这种情况
还有一个解决此问题的方法,但尚未合并
所以
- 您可以根据上面链接的拉动请求构建自己的
HdfsSinkConnector
- 重命名目录,重新创建连接器,直到创建配置单元表,然后将表移回(当然,在生产环境中很难)
- 或者手动创建表