Key Issue :confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Node -1 disconnected
Kafka is Confluent 7.5.3 with same suit for Connect , Zookeeper and Mongo is latest 7.0 . everything is setup over SSL . and I am using Kafka-connect-mongodb old version plus new version 1.11.2 both jars without luck
broker: image: confluentinc/cp-enterprise-kafka:7.5.3 connect: image: confluentinc/cp-kafka-connect:7.5.3
Connect CLASSPATH: /usr/share/java/monitoring-interceptors/monitoring-interceptors-7.5.3.jar
The issue is when Mongo Sink connector API is invoked I am getting error as below :
> [2024-03-07 00:05:50,832] DEBUG Bulk writing 4 document(s) into collection [mydb.products] (com.mongodb.kafka.connect.sink.MongoSinkTask)> [2024-03-07 00:05:51,061] INFO [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Node -1 disconnected. (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,064] INFO [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Cancelled in-flight API_VERSIONS request with correlation id 0 due to node -1 being disconnected (elapsed time since creation: 263ms, elapsed time since send: 263ms, request timeout: 30000ms) (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,064] WARN [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Bootstrap broker broker.local:19092 (id: -1 rack: null) disconnected (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,142] INFO Cluster created with settings {hosts=[mongo1:27017, mongo2:27017, mongo3:27017], mode=MULTIPLE, requiredClusterType=UNKNOWN, serverSelectionTimeout='30000 ms', maxWaitQueueSize=500} (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,142] INFO Adding discovered server mongo1:27017 to client view of cluster (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,151] INFO Adding discovered server mongo2:27017 to client view of cluster (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,152] INFO Adding discovered server mongo3:27017 to client view of cluster (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,161] INFO Cluster description not yet available. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,179] INFO Opened connection [connectionId{localValue:2}] to mongo3:27017 (org.mongodb.driver.connection)> [2024-03-07 00:05:51,179] INFO Opened connection [connectionId{localValue:3}] to mongo2:27017 (org.mongodb.driver.connection)> [2024-03-07 00:05:51,179] INFO Opened connection [connectionId{localValue:1}] to mongo1:27017 (org.mongodb.driver.connection)> [2024-03-07 00:05:51,184] INFO Monitor thread successfully connected to server with description ServerDescription{address=mongo1:27017, type=REPLICA_SET_GHOST, state=CONNECTED, ok=true, version=ServerVersion{versionList=[7, 0, 6]}, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=3293648, setName='null', canonicalAddress=null, hosts=[], passives=[], arbiters=[], primary='null', tagSet=TagSet{[]}, electionId=null, setVersion=null, lastWriteDate=null, lastUpdateTimeNanos=704151849030593} (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,184] INFO Monitor thread successfully connected to server with description ServerDescription{address=mongo2:27017, type=REPLICA_SET_GHOST, state=CONNECTED, ok=true, version=ServerVersion{versionList=[7, 0, 6]}, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=3288036, setName='null', canonicalAddress=null, hosts=[], passives=[], arbiters=[], primary='null', tagSet=TagSet{[]}, electionId=null, setVersion=null, lastWriteDate=null, lastUpdateTimeNanos=704151849030588} (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,184] INFO Monitor thread successfully connected to server with description ServerDescription{address=mongo3:27017, type=REPLICA_SET_GHOST, state=CONNECTED, ok=true, version=ServerVersion{versionList=[7, 0, 6]}, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=3390299, setName='null', canonicalAddress=null, hosts=[], passives=[], arbiters=[], primary='null', tagSet=TagSet{[]}, electionId=null, setVersion=null, lastWriteDate=null, lastUpdateTimeNanos=704151849035558} (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,733] INFO [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Node -1 disconnected. (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,734] INFO [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Cancelled in-flight API_VERSIONS request with correlation id 1 due to node -1 being disconnected (elapsed time since creation: 167ms, elapsed time since send: 167ms, request timeout: 30000ms) (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,734] WARN [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Bootstrap broker broker.local:19092 (id: -1 rack: null) disconnected (org.apache.kafka.clients.NetworkClient)
After executing the shell script to initiate docker compose I see objection on Hostname/Ip doesn't match as below:
Connecting to: mongodb://127.0.0.1:27017/?directConnection=true&serverSelectionTimeoutMS=2000&tls=true&tlsCAFile=%2Fdata%2Fssl%2Fmongodb-ca.crt&tlsCertificateKeyFile=%2Fdata%2Fssl%2Fmongodb.pem&appName=mongosh+2.1.5MongoServerSelectionError: Hostname/IP does not match certificate's altnames: IP: 127.0.0.1 is not in the cert's list:bduser@willow-spr10:~/dpaul/kafka-docker-ssl/mongo$
The docker-compose looks like :
---version: "3"services: zookeeper: image: confluentinc/cp-zookeeper:7.5.3 container_name: zookeeper networks: kafka: aliases: - zookeeper.local mongo_network: environment: ZOOKEEPER_SERVER_ID: 1 ZOOKEEPER_CLIENT_PORT: 22181 ZOOKEEPER_TICK_TIME: 2000 ZOOKEEPER_LOG4J_ROOT_LOGLEVEL: ERROR ZOOKEEPER_LOG4J_LOGLEVEL: ERROR volumes: - zk-data:/var/lib/zookeeper/data - zk-txn-logs:/var/lib/zookeeper/log broker: image: confluentinc/cp-enterprise-kafka:7.5.3 container_name: broker networks: kafka: aliases: - broker.local mongo_network: depends_on: - zookeeper ports: - "19092:19092" - "9092:9092" environment: KAFKA_LOG4J_ROOT_LOGLEVEL: ERROR KAFKA_LOG4J_LOGLEVEL: ERROR KAFKA_BROKER_ID: 1 KAFKA_ZOOKEEPER_CONNECT: zookeeper.local:22181 KAFKA_LISTENERS: SSL://broker.local:19092 KAFKA_ADVERTISED_LISTENERS: SSL://broker.local:19092 KAFKA_SSL_KEYSTORE_FILENAME: kafka.broker.keystore.jks KAFKA_SSL_KEYSTORE_CREDENTIALS: broker_keystore_creds KAFKA_SSL_KEY_CREDENTIALS: broker_sslkey_creds KAFKA_SSL_TRUSTSTORE_FILENAME: kafka.broker.truststore.jks KAFKA_SSL_TRUSTSTORE_CREDENTIALS: broker_truststore_creds KAFKA_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM: "" KAFKA_LISTENER_NAME_INTERNAL_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM: "" KAFKA_SSL_CLIENT_AUTH: requested KAFKA_SECURITY_INTER_BROKER_PROTOCOL: SSL KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 1 KAFKA_GROUP_INITIAL_REBALANCE_DELAY_MS: 0 KAFKA_METRIC_REPORTERS: io.confluent.metrics.reporter.ConfluentMetricsReporter CONFLUENT_METRICS_REPORTER_BOOTSTRAP_SERVERS: broker.local:19092 CONFLUENT_METRICS_REPORTER_ZOOKEEPER_CONNECT: zookeeper.local:22181 CONFLUENT_METRICS_REPORTER_TOPIC_REPLICAS: 1 CONFLUENT_METRICS_REPORTER_SECURITY_PROTOCOL: SSL CONFLUENT_METRICS_REPORTER_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.metrics.truststore.jks CONFLUENT_METRICS_REPORTER_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.metrics.keystore.jks CONFLUENT_METRICS_REPORTER_SSL_TRUSTSTORE_PASSWORD: awesomekafka CONFLUENT_METRICS_REPORTER_SSL_KEYSTORE_PASSWORD: awesomekafka CONFLUENT_METRICS_REPORTER_SSL_KEY_PASSWORD: awesomekafka CONFLUENT_METRICS_ENABLE: "true" CONFLUENT_SUPPORT_CUSTOMER_ID: anonymous volumes: - kafka-data:/var/lib/kafka/data - ./secrets:/etc/kafka/secrets connect: image: confluentinc/cp-kafka-connect:7.5.3 hostname: connect container_name: connect depends_on: - zookeeper - broker ports: - "8083:8083" networks: kafka: aliases: - connect.local mongo_network: environment: CONNECT_REST_ACCESS_CONTROL_ALLOW_METHODS: 'GET,POST,PUT,DELETE,OPTIONS' CONNECT_REST_ACCESS_CONTROL_ALLOW_ORIGIN: '*' CONNECT_BOOTSTRAP_SERVERS: SSL://broker.local:19092 # CONNECT_LISTENERS: 'https://0.0.0.0:8083' CONNECT_REST_ADVERTISED_HOST_NAME: connect CONNECT_REST_PORT: 8083 CONNECT_GROUP_ID: compose-connect-group CONNECT_CONFIG_STORAGE_TOPIC: docker-connect-configs CONNECT_CONFIG_STORAGE_REPLICATION_FACTOR: 1 CONNECT_OFFSET_FLUSH_INTERVAL_MS: 10000 CONNECT_OFFSET_STORAGE_TOPIC: docker-connect-offsets CONNECT_OFFSET_STORAGE_REPLICATION_FACTOR: 1 CONNECT_STATUS_STORAGE_TOPIC: docker-connect-status CONNECT_STATUS_STORAGE_REPLICATION_FACTOR: 1 CONNECT_INTERNAL_KEY_CONVERTER: "org.apache.kafka.connect.json.JsonConverter" CONNECT_INTERNAL_VALUE_CONVERTER: "org.apache.kafka.connect.json.JsonConverter" CONNECT_KEY_CONVERTER: org.apache.kafka.connect.storage.StringConverter CONNECT_LOG4J_ROOT_LOGLEVEL: "INFO" CONNECT_LOG4J_LOGGERS: "org.apache.kafka.connect.runtime.rest=WARN,org.reflections=ERROR,com.mongodb.kafka=DEBUG" CONNECT_PLUGIN_PATH: '/usr/share/confluent-hub-components/kafka-connect-mongodb/lib/mongo-kafka-connect-1.11.2-confluent.jar,/usr/share/java, /usr/share/confluent-hub-components' CONNECT_ZOOKEEPER_CONNECT: zookeeper.local:22181 # CONNECT_SSL_CLIENT_AUTH: 'required' CONNECT_SSL_KEY_PASSWORD: awesomekafka CONNECT_SECURITY_PROTOCOL: SSL CONNECT_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.truststore.jks CONNECT_SSL_TRUSTSTORE_PASSWORD: awesomekafka CONNECT_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.keystore.jks CONNECT_SSL_KEYSTORE_PASSWORD: awesomekafka CONNECT_PRODUCER_SECURITY_PROTOCOL: SSL CONNECT_PRODUCER_BOOTSTRAP_SERVERS: SSL://broker.local:19092 CONNECT_PRODUCER_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.truststore.jks CONNECT_PRODUCER_SSL_TRUSTSTORE_PASSWORD: awesomekafka CONNECT_PRODUCER_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.keystore.jks CONNECT_PRODUCER_SSL_KEYSTORE_PASSWORD: awesomekafka CONNECT_CONSUMER_SECURITY_PROTOCOL: SSL CONNECT_CONSUMER_BOOTSTRAP_SERVERS: SSL://broker.local:19092 CONNECT_CONSUMER_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.truststore.jks CONNECT_CONSUMER_SSL_TRUSTSTORE_PASSWORD: awesomekafka CONNECT_CONSUMER_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.keystore.jks CONNECT_CONSUMER_SSL_KEYSTORE_PASSWORD: awesomekafka # Assumes image is based on confluentinc/kafka-connect-datagen:latest which is pulling 5.2.2 Connect image CLASSPATH: /usr/share/java/monitoring-interceptors/monitoring-interceptors-7.5.3.jar CONNECT_PRODUCER_INTERCEPTOR_CLASSES: "io.confluent.monitoring.clients.interceptor.MonitoringProducerInterceptor" CONNECT_CONSUMER_INTERCEPTOR_CLASSES: "io.confluent.monitoring.clients.interceptor.MonitoringConsumerInterceptor" command: "bash -c 'if [ ! -d /usr/share/confluent-hub-components/confluentinc-kafka-connect-datagen ]; then echo \"WARNING: Did not find directory for kafka-connect-datagen (did you remember to run: docker-compose up -d --build ?)\"; fi ; /etc/confluent/docker/run'" # command: # "bash -c echo \"Installing Connector\" ; confluent-hub install --no-prompt confluentinc/kafka-connect-datagen:0.1.6 ; echo \"Launching Kafka Connect worker\" ; /etc/confluent/docker/run &sleep infinity" volumes: - ./kafka-connect-mongodb:/usr/share/confluent-hub-components/kafka-connect-mongodb - ./secrets:/etc/kafka/secrets #MongoDB Replica Set mongo1: hostname: mongo1 image: mongo container_name: mongo1 ports: - 27017:27017 restart: always entrypoint: [ "/usr/bin/mongod", "--config", "/data/config/ssl.conf", "--bind_ip_all", "--replSet", "dbrs" ] networks: - mongo_network volumes: - ./.db/mongo1:/data/db - ./wait-for-mongodb.sh:/scripts/wait-for-mongodb.sh - ./init.sh:/scripts/init.sh - ./ssl:/data/ssl - ./config:/data/config links: - mongo2 - mongo3 mongo2: hostname: mongo2 image: mongo container_name: mongo2 ports: - 27018:27017 restart: always entrypoint: [ "/usr/bin/mongod", "--bind_ip_all", "--replSet", "dbrs" ] networks: - mongo_network volumes: - ./.db/mongo2:/data/db - ./ssl:/data/ssl - ./wait-for-mongodb.sh:/scripts/wait-for-mongodb.sh mongo3: hostname: mongo3 image: mongo container_name: mongo3 ports: - 27019:27017 restart: always entrypoint: [ "/usr/bin/mongod", "--bind_ip_all", "--replSet", "dbrs" ] networks: - mongo_network volumes: - ./.db/mongo3:/data/db - ./ssl:/data/ssl - ./wait-for-mongodb.sh:/scripts/wait-for-mongodb.sh kafka-tools: build: context: ./kafka-tools depends_on: - broker container_name: kafka-tools networks: kafka: aliases: - kafka-tools.local mongo_network: environment: KAFKA_LOG4J_ROOT_LOGLEVEL: ERROR KAFKA_SECURITY_PROTOCOL: SSL KAFKA_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM: "" KAFKA_HEAP_OPTS: "-Xmx1G -Xms1G" KAFKA_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.kafka-tools.truststore.jks KAFKA_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.kafka-tools.keystore.jks KAFKA_SSL_TRUSTSTORE_PASSWORD: awesomekafka KAFKA_SSL_KEYSTORE_PASSWORD: awesomekafka KAFKA_SSL_KEY_PASSWORD: awesomekafka volumes: - ./secrets:/etc/kafka/secrets - ./config:/etc/kafka/configvolumes: zk-data: external: true zk-txn-logs: external: true kafka-data: external: truenetworks: kafka: localnet: attachable: true mongo_network: driver: bridge
Below is mong-sink connector json looks like :
> > {> "name": "mongo-sink",> "config": {> "connector.class": "com.mongodb.kafka.connect.MongoSinkConnector",> "tasks.max": "1",> "topics": "mytopic",> "connection.uri": "mongodb://mongo1:27017,mongo2:27017,mongo3:27017",> "database": "mydb",> "collection": "products",> "key.converter": "org.apache.kafka.connect.storage.StringConverter",> "value.converter": "org.apache.kafka.connect.storage.StringConverter",> "key.converter.schemas.enable": "false",> "value.converter.schemas.enable": "false",> "mongo.errors.tolerance": "all",> "mongo.errors.log.enable": "true",> "errors.log.include.messages": "true",> "writemodel.strategy": "com.mongodb.kafka.connect.sink.writemodel.strategy.ReplaceOneDefaultStrategy",> "name": "mongo-sink"> },> "tasks": [],> "type": "sink"> }
Please note I tried changing this to below without luck :
"connection.uri": "mongodb://mongo1:27017,mongo2:27017,mongo3:27017/?ssl=true",
Mongo SSL config looks like below :
config/ssl.conf as below :
> net:> ssl:> mode: allowSSL> PEMKeyFile: "/data/ssl/mongodb.pem"> CAFile: "/data/ssl/mongodb-cert.crt"> storage:> dbPath: "/data/db"> net:> bindIp: 127.0.0.1> port: 27017
I am pretty sure mongodb replicaset is perfectly working but mongo sink connector task having issue as denoted here :
> [2024-03-07 00:06:21,165] ERROR Error on mongodb operation (com.mongodb.kafka.connect.sink.MongoSinkTask)> com.mongodb.MongoTimeoutException: Timed out after 30000 ms while waiting to connect. Client view of cluster state is {type=UNKNOWN, servers=[{address=mongo3:27017, type=REPLICA_SET_GHOST, roundTripTime=0.7 ms, state=CONNECTED}, {address=mongo2:27017, type=REPLICA_SET_GHOST, roundTripTime=0.6 ms, state=CONNECTED}, {address=mongo1:27017, type=REPLICA_SET_GHOST, roundTripTime=0.7 ms, state=CONNECTED}]> at com.mongodb.internal.connection.BaseCluster.getDescription(BaseCluster.java:182)> at com.mongodb.internal.connection.AbstractMultiServerCluster.getDescription(AbstractMultiServerCluster.java:52)> at com.mongodb.client.internal.MongoClientDelegate.getConnectedClusterDescription(MongoClientDelegate.java:136)> at com.mongodb.client.internal.MongoClientDelegate.createClientSession(MongoClientDelegate.java:94)> at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.getClientSession(MongoClientDelegate.java:249)> at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:190)> at com.mongodb.client.internal.MongoCollectionImpl.executeBulkWrite(MongoCollectionImpl.java:468)> at com.mongodb.client.internal.MongoCollectionImpl.bulkWrite(MongoCollectionImpl.java:448)> at com.mongodb.kafka.connect.sink.MongoSinkTask.processSinkRecords(MongoSinkTask.java:177)> at com.mongodb.kafka.connect.sink.MongoSinkTask.lambda$put$2(MongoSinkTask.java:117)> at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)> at com.mongodb.kafka.connect.sink.MongoSinkTask.lambda$put$3(MongoSinkTask.java:116)> at java.base/java.util.HashMap.forEach(HashMap.java:1337)> at com.mongodb.kafka.connect.sink.MongoSinkTask.put(MongoSinkTask.java:114)> at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:593)> at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:342)> at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:242)> at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:211)> at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:204)> at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:259)> at org.apache.kafka.connect.runtime.isolation.Plugins.lambda$withClassLoader$1(Plugins.java:181)> at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)> at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)> at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)> at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)> at java.base/java.lang.Thread.run(Thread.java:829)> [2024-03-07 00:06:21,168] ERROR Writing 4 document(s) into collection [mydb.products] failed -> remaining retries (3) (com.mongodb.kafka.connect.sink.MongoSinkTask)
I have tested with curl https://broker.local:19092 and curl -v https://mongo1:27017 from connect node and response is positive
> > * Trying 192.168.xx.x...> * TCP_NODELAY set> * Connected to broker.local (192.168.xx.xx) port 19092 (#0)> * ALPN, offering h2> * ALPN, offering http/1.1> * successfully set certificate verify locations:> * CAfile: /etc/pki/tls/certs/ca-bundle.crt> CApath: none> * TLSv1.3 (OUT), TLS handshake, Client hello (1):> * TLSv1.3 (IN), TLS handshake, Server hello (2):> * TLSv1.3 (IN), TLS handshake, [no content] (0):> * TLSv1.3 (IN), TLS handshake, Encrypted Extensions (8):> * TLSv1.3 (IN), TLS handshake, Request CERT (13):> * TLSv1.3 (IN), TLS handshake, Certificate (11):> * TLSv1.3 (OUT), TLS alert, unknown CA (560):
config parameters initialized from mongo connect container docker logs as below :
[2024-03-07 00:26:37,270] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [SSL://broker.local:19092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = compose-connect-group--offsets client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = compose-connect-group group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 45000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = [hidden] ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = /etc/kafka/secrets/kafka.connect.keystore.jks ssl.keystore.password = [hidden] ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = /etc/kafka/secrets/kafka.connect.truststore.jks ssl.truststore.password = [hidden] ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer> (org.apache.kafka.clients.consumer.ConsumerConfig)