注册Avro模式时出错:“字符串” RestClientException:正在注册的模式与较早的模式不兼容;默认值为0。

时间:2020-02-01 15:34:09

标签: kotlin apache-kafka avro confluent-schema-registry

我正在尝试使用Avro模式向我的经纪人发送消息,但“我总是会出错:

2020-02-01 11:24:37.189 [nioEventLoopGroup-4-1]错误应用程序- 未处理:POST-/ api / orchestration / org.apache.kafka.common.errors.SerializationException:错误 正在注册Avro模式:“字符串”,原因是: io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException: 正在注册的架构与早期的架构不兼容;错误 代码:409

这是我的docker容器:

 connect:
    image: confluentinc/cp-kafka-connect:5.4.0
    hostname: confluentinc-connect
    container_name: confluentinc-connect
    depends_on:
      - zookeeper
      - broker
      - schema-registry
    ports:
      - "8083:8083"
    environment:
      CONNECT_BOOTSTRAP_SERVERS: 'broker:29092'
      CONNECT_REST_ADVERTISED_HOST_NAME: connect
      CONNECT_REST_PORT: 8083
      CONNECT_GROUP_ID: confluentinc-connect
      CONNECT_CONFIG_STORAGE_TOPIC: confluentinc-connect-configs
      CONNECT_CONFIG_STORAGE_REPLICATION_FACTOR: 1
      CONNECT_OFFSET_FLUSH_INTERVAL_MS: 10000
      CONNECT_OFFSET_STORAGE_TOPIC: confluentinc-connect-offsets
      CONNECT_OFFSET_STORAGE_REPLICATION_FACTOR: 1
      CONNECT_STATUS_STORAGE_TOPIC: confluentinc-connect-status
      CONNECT_STATUS_STORAGE_REPLICATION_FACTOR: 1
      CONNECT_KEY_CONVERTER: org.apache.kafka.connect.storage.StringConverter
      CONNECT_KEY_CONVERTER_SCHEMAS_ENABLE: "true"
      CONNECT_VALUE_CONVERTER: io.confluent.connect.avro.AvroConverter
      CONNECT_VALUE_CONVERTER_SCHEMA_REGISTRY_URL: 'http://schema-registry:8081'
      CONNECT_INTERNAL_KEY_CONVERTER: "org.apache.kafka.connect.json.JsonConverter"
      CONNECT_INTERNAL_VALUE_CONVERTER: "org.apache.kafka.connect.json.JsonConverter"
      CONNECT_ZOOKEEPER_CONNECT: 'zookeeper:2181'
      CONNECT_LOG4J_ROOT_LOGLEVEL: "INFO"
      CONNECT_LOG4J_LOGGERS: "org.apache.kafka.connect.runtime.rest=WARN,org.reflections=ERROR"
      CONNECT_CONFIG_STORAGE_REPLICATION_FACTOR: "1"
      CONNECT_OFFSET_STORAGE_REPLICATION_FACTOR: "1"
      CONNECT_STATUS_STORAGE_REPLICATION_FACTOR: "1"
      CONNECT_PRODUCER_INTERCEPTOR_CLASSES: "io.confluent.monitoring.clients.interceptor.MonitoringProducerInterceptor"
      CONNECT_CONSUMER_INTERCEPTOR_CLASSES: "io.confluent.monitoring.clients.interceptor.MonitoringConsumerInterceptor"
      CONNECT_LOG4J_LOGGERS: org.apache.zookeeper=ERROR,org.I0Itec.zkclient=ERROR,org.reflections=ERROR
      CONNECT_PLUGIN_PATH: "/usr/share/java,/usr/share/extras"

我的制片人(用科林写)

 val prop: HashMap<String, Any> = HashMap()
    prop[BOOTSTRAP_SERVERS_CONFIG] = bootstrapServers
    prop[KEY_SERIALIZER_CLASS_CONFIG] = StringSerializer::class.java.name
    prop[VALUE_SERIALIZER_CLASS_CONFIG] = KafkaAvroSerializer::class.java.name
    prop[SCHEMA_REGISTRY_URL] = schemaUrl
    prop[ENABLE_IDEMPOTENCE_CONFIG] = idempotence
    prop[ACKS_CONFIG] = acks.value
    prop[RETRIES_CONFIG] = retries
    prop[MAX_IN_FLIGHT_REQUESTS_PER_CONNECTION] = requestPerConnection
    prop[COMPRESSION_TYPE_CONFIG] = compression.value
    prop[LINGER_MS_CONFIG] = linger
    prop[BATCH_SIZE_CONFIG] = batchSize.value

    return KafkaProducer(prop)

我的Avro模式:

{
    "type": "record",
    "namespace": "com.rjdesenvolvimento",
    "name": "create_client_value",
    "doc": "Avro Schema for Kafka Command",
    "fields": [
        {
            "name": "id",
            "type": "string",
            "logicalType": "uuid",
            "doc": "UUID for indentifaction command"
        },
        {
            "name": "status",
            "type": {
                "name": "status",
                "type": "enum",
                "symbols": [
                    "Open",
                    "Closed",
                    "Processing"
                ],
                "doc": "Can be only: Open, Closed or Processing"
            },
            "doc": "Status of the command"
        },
        {
            "name": "message",
            "type": {
                "type": "record",
                "name": "message",
                "doc": "Avro Schema for insert new client",
                "fields": [
                    {
                        "name": "id",
                        "type": "string",
                        "logicalType": "uuid",
                        "doc": "UUID for indentifaction client transaction"
                    },
                    {
                        "name": "active",
                        "type": "boolean",
                        "doc": "Soft delete for client"
                    },
                    {
                        "name": "name",
                        "type": "string",
                        "doc": "Name of the client"
                    },
                    {
                        "name": "email",
                        "type": "string",
                        "doc": "Email of the client"
                    },
                    {
                        "name": "document",
                        "type": "string",
                        "doc": "CPF or CPNJ of the client"
                    },
                    {
                        "name": "phones",
                        "doc": "A list of phone numbers",
                        "type": {
                            "type": "array",
                            "items": {
                                "name": "phones",
                                "type": "record",
                                "fields": [
                                    {
                                        "name": "id",
                                        "type": "string",
                                        "logicalType": "uuid",
                                        "doc": "UUID for indentifaction of phone transaction"
                                    },
                                    {
                                        "name": "active",
                                        "type": "boolean",
                                        "doc": "Soft delete for phone number"
                                    },
                                    {
                                        "name": "number",
                                        "type": "string",
                                        "doc": "The phone number with this regex +xx xx xxxx xxxx"
                                    }
                                ]
                            }
                        }
                    },
                    {
                        "name": "address",
                        "type": "string",
                        "logicalType": "uuid",
                        "doc": "Adrres is an UUID for a other address-microservice"
                    }
                ]
            }
        }
    ]
}

我的帖子:

{       
      "id" : "9ec818da-6ee0-4634-9ed8-c085248cae12",
        "status" : "Open",
        "message": {
            "id" : "9ec818da-6ee0-4634-9ed8-c085248cae12",
            "active" : true,
             "name": "name",
             "email": "email@com",
             "document": "document",
             "phones": [
                 {
                     "id" : "9ec818da-6ee0-4634-9ed8-c085248cae12",
                        "active" : true,
                     "number": "+xx xx xxxx xxxx"
                 },
                    {
                     "id" : "9ec818da-6ee0-4634-9ed8-c085248cae12",
                        "active" : true,
                     "number": "+xx xx xxxx xxxx"
                 }
             ],
             "address": "9ec818da-6ee0-4634-9ed8-c085248cae12"  
        }   
}

我在做什么错? github项目:https://github.com/rodrigodevelms/kafka-registry

更新=====

简而言之: 我没有使用Gradle Avro插件生成类。 在此示例中,我的POST发送了一个Client对象。在使用中,它会按如下方式组装一个Command型对象:

id:相同的客户端ID

状态:打开

消息:发送的POST。

因此我将其发送到KAFKA,并在连接(jdbc sink postgres)中将其作为字段。仅将消息的属性(客户端)列入白名单,而没有获得命令ID或状态。 / p>

on github the only classes that matter to understand the code are:

1 -https://github.com/rodrigodevelms/kafka-registry/blob/master/kafka/src/main/kotlin/com/rjdesenvolvimento/messagebroker/producer/Producer.kt

2- https://github.com/rodrigodevelms/kafka-registry/blob/master/kafka/src/main/kotlin/com/rjdesenvolvimento/messagebroker/commnad/Command.kt

3- https://github.com/rodrigodevelms/kafka-registry/blob/master/src/client/Controller.kt

4 -https://github.com/rodrigodevelms/kafka-registry/blob/master/src/client/Service.kt

5-docker-compose.yml,insert-client-value.avsc,postgresql.json,

如果我将avro方案的兼容模式设置为“ none”,则可以发送消息,但是会显示一些未知字符,如下图所示。

enter image description here

1 个答案:

答案 0 :(得分:4)

我怀疑您正在尝试执行多项操作,并且在之前的尝试之后没有清理状态。全新安装中不应出现该错误

正在注册的架构与早期的架构不兼容

您的数据已更改,注册表中的架构与您要发送的架构不兼容。

您可以向http://registry:8081/subjects/[name]/发送HTTP DELETE请求以删除架构的所有版本,然后可以重新启动连接器