Quantcast
Channel: Recent Questions - Stack Overflow
Viewing all articles
Browse latest Browse all 12111

Kafka Mongo Sink Issue - Monitoring.interceptor.connector-consumer-mongo-sink node disconnected

$
0
0

Key Issue :confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Node -1 disconnected

Kafka is Confluent 7.5.3 with same suit for Connect , Zookeeper and Mongo is latest 7.0 . everything is setup over SSL . and I am using Kafka-connect-mongodb old version plus new version 1.11.2 both jars without luck

   broker:    image: confluentinc/cp-enterprise-kafka:7.5.3   connect:    image: confluentinc/cp-kafka-connect:7.5.3

Connect CLASSPATH: /usr/share/java/monitoring-interceptors/monitoring-interceptors-7.5.3.jar

The issue is when Mongo Sink connector API is invoked I am getting error as below :

> [2024-03-07 00:05:50,832] DEBUG Bulk writing 4 document(s) into collection [mydb.products] (com.mongodb.kafka.connect.sink.MongoSinkTask)> [2024-03-07 00:05:51,061] INFO [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Node -1 disconnected. (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,064] INFO [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Cancelled in-flight API_VERSIONS request with correlation id 0 due to node -1 being disconnected (elapsed time since creation: 263ms, elapsed time since send: 263ms, request timeout: 30000ms) (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,064] WARN [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Bootstrap broker broker.local:19092 (id: -1 rack: null) disconnected (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,142] INFO Cluster created with settings {hosts=[mongo1:27017, mongo2:27017, mongo3:27017], mode=MULTIPLE, requiredClusterType=UNKNOWN, serverSelectionTimeout='30000 ms', maxWaitQueueSize=500} (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,142] INFO Adding discovered server mongo1:27017 to client view of cluster (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,151] INFO Adding discovered server mongo2:27017 to client view of cluster (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,152] INFO Adding discovered server mongo3:27017 to client view of cluster (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,161] INFO Cluster description not yet available. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,179] INFO Opened connection [connectionId{localValue:2}] to mongo3:27017 (org.mongodb.driver.connection)> [2024-03-07 00:05:51,179] INFO Opened connection [connectionId{localValue:3}] to mongo2:27017 (org.mongodb.driver.connection)> [2024-03-07 00:05:51,179] INFO Opened connection [connectionId{localValue:1}] to mongo1:27017 (org.mongodb.driver.connection)> [2024-03-07 00:05:51,184] INFO Monitor thread successfully connected to server with description ServerDescription{address=mongo1:27017, type=REPLICA_SET_GHOST, state=CONNECTED, ok=true, version=ServerVersion{versionList=[7, 0, 6]}, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=3293648, setName='null', canonicalAddress=null, hosts=[], passives=[], arbiters=[], primary='null', tagSet=TagSet{[]}, electionId=null, setVersion=null, lastWriteDate=null, lastUpdateTimeNanos=704151849030593} (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,184] INFO Monitor thread successfully connected to server with description ServerDescription{address=mongo2:27017, type=REPLICA_SET_GHOST, state=CONNECTED, ok=true, version=ServerVersion{versionList=[7, 0, 6]}, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=3288036, setName='null', canonicalAddress=null, hosts=[], passives=[], arbiters=[], primary='null', tagSet=TagSet{[]}, electionId=null, setVersion=null, lastWriteDate=null, lastUpdateTimeNanos=704151849030588} (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,184] INFO Monitor thread successfully connected to server with description ServerDescription{address=mongo3:27017, type=REPLICA_SET_GHOST, state=CONNECTED, ok=true, version=ServerVersion{versionList=[7, 0, 6]}, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=3390299, setName='null', canonicalAddress=null, hosts=[], passives=[], arbiters=[], primary='null', tagSet=TagSet{[]}, electionId=null, setVersion=null, lastWriteDate=null, lastUpdateTimeNanos=704151849035558} (org.mongodb.driver.cluster)> [2024-03-07 00:05:51,733] INFO [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Node -1 disconnected. (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,734] INFO [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Cancelled in-flight API_VERSIONS request with correlation id 1 due to node -1 being disconnected (elapsed time since creation: 167ms, elapsed time since send: 167ms, request timeout: 30000ms) (org.apache.kafka.clients.NetworkClient)> [2024-03-07 00:05:51,734] WARN [Producer clientId=confluent.monitoring.interceptor.connector-consumer-mongo-sink-0] Bootstrap broker broker.local:19092 (id: -1 rack: null) disconnected (org.apache.kafka.clients.NetworkClient)

After executing the shell script to initiate docker compose I see objection on Hostname/Ip doesn't match as below:

Connecting to:          mongodb://127.0.0.1:27017/?directConnection=true&serverSelectionTimeoutMS=2000&tls=true&tlsCAFile=%2Fdata%2Fssl%2Fmongodb-ca.crt&tlsCertificateKeyFile=%2Fdata%2Fssl%2Fmongodb.pem&appName=mongosh+2.1.5MongoServerSelectionError: Hostname/IP does not match certificate's altnames: IP: 127.0.0.1 is not in the cert's list:bduser@willow-spr10:~/dpaul/kafka-docker-ssl/mongo$

The docker-compose looks like :


    ---version: "3"services:  zookeeper:    image: confluentinc/cp-zookeeper:7.5.3    container_name: zookeeper    networks:       kafka:          aliases:            - zookeeper.local       mongo_network:    environment:      ZOOKEEPER_SERVER_ID: 1      ZOOKEEPER_CLIENT_PORT: 22181      ZOOKEEPER_TICK_TIME: 2000      ZOOKEEPER_LOG4J_ROOT_LOGLEVEL: ERROR      ZOOKEEPER_LOG4J_LOGLEVEL: ERROR    volumes:      - zk-data:/var/lib/zookeeper/data      - zk-txn-logs:/var/lib/zookeeper/log  broker:    image: confluentinc/cp-enterprise-kafka:7.5.3    container_name: broker    networks:       kafka:          aliases:            - broker.local       mongo_network:    depends_on:      - zookeeper    ports:      - "19092:19092"      - "9092:9092"    environment:      KAFKA_LOG4J_ROOT_LOGLEVEL: ERROR      KAFKA_LOG4J_LOGLEVEL: ERROR      KAFKA_BROKER_ID: 1      KAFKA_ZOOKEEPER_CONNECT: zookeeper.local:22181      KAFKA_LISTENERS: SSL://broker.local:19092      KAFKA_ADVERTISED_LISTENERS: SSL://broker.local:19092      KAFKA_SSL_KEYSTORE_FILENAME: kafka.broker.keystore.jks      KAFKA_SSL_KEYSTORE_CREDENTIALS: broker_keystore_creds      KAFKA_SSL_KEY_CREDENTIALS: broker_sslkey_creds      KAFKA_SSL_TRUSTSTORE_FILENAME: kafka.broker.truststore.jks      KAFKA_SSL_TRUSTSTORE_CREDENTIALS: broker_truststore_creds      KAFKA_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM: ""      KAFKA_LISTENER_NAME_INTERNAL_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM: ""      KAFKA_SSL_CLIENT_AUTH: requested      KAFKA_SECURITY_INTER_BROKER_PROTOCOL: SSL      KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 1      KAFKA_GROUP_INITIAL_REBALANCE_DELAY_MS: 0      KAFKA_METRIC_REPORTERS: io.confluent.metrics.reporter.ConfluentMetricsReporter      CONFLUENT_METRICS_REPORTER_BOOTSTRAP_SERVERS: broker.local:19092      CONFLUENT_METRICS_REPORTER_ZOOKEEPER_CONNECT: zookeeper.local:22181      CONFLUENT_METRICS_REPORTER_TOPIC_REPLICAS: 1      CONFLUENT_METRICS_REPORTER_SECURITY_PROTOCOL: SSL      CONFLUENT_METRICS_REPORTER_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.metrics.truststore.jks      CONFLUENT_METRICS_REPORTER_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.metrics.keystore.jks      CONFLUENT_METRICS_REPORTER_SSL_TRUSTSTORE_PASSWORD: awesomekafka      CONFLUENT_METRICS_REPORTER_SSL_KEYSTORE_PASSWORD: awesomekafka      CONFLUENT_METRICS_REPORTER_SSL_KEY_PASSWORD: awesomekafka      CONFLUENT_METRICS_ENABLE: "true"      CONFLUENT_SUPPORT_CUSTOMER_ID: anonymous    volumes:      - kafka-data:/var/lib/kafka/data      - ./secrets:/etc/kafka/secrets  connect:    image: confluentinc/cp-kafka-connect:7.5.3    hostname: connect    container_name: connect    depends_on:      - zookeeper      - broker    ports:      - "8083:8083"    networks:       kafka:          aliases:            - connect.local       mongo_network:    environment:      CONNECT_REST_ACCESS_CONTROL_ALLOW_METHODS: 'GET,POST,PUT,DELETE,OPTIONS'      CONNECT_REST_ACCESS_CONTROL_ALLOW_ORIGIN: '*'      CONNECT_BOOTSTRAP_SERVERS: SSL://broker.local:19092      # CONNECT_LISTENERS: 'https://0.0.0.0:8083'      CONNECT_REST_ADVERTISED_HOST_NAME: connect      CONNECT_REST_PORT: 8083      CONNECT_GROUP_ID: compose-connect-group      CONNECT_CONFIG_STORAGE_TOPIC: docker-connect-configs      CONNECT_CONFIG_STORAGE_REPLICATION_FACTOR: 1      CONNECT_OFFSET_FLUSH_INTERVAL_MS: 10000      CONNECT_OFFSET_STORAGE_TOPIC: docker-connect-offsets      CONNECT_OFFSET_STORAGE_REPLICATION_FACTOR: 1      CONNECT_STATUS_STORAGE_TOPIC: docker-connect-status      CONNECT_STATUS_STORAGE_REPLICATION_FACTOR: 1      CONNECT_INTERNAL_KEY_CONVERTER: "org.apache.kafka.connect.json.JsonConverter"      CONNECT_INTERNAL_VALUE_CONVERTER: "org.apache.kafka.connect.json.JsonConverter"      CONNECT_KEY_CONVERTER: org.apache.kafka.connect.storage.StringConverter      CONNECT_LOG4J_ROOT_LOGLEVEL: "INFO"      CONNECT_LOG4J_LOGGERS: "org.apache.kafka.connect.runtime.rest=WARN,org.reflections=ERROR,com.mongodb.kafka=DEBUG"      CONNECT_PLUGIN_PATH: '/usr/share/confluent-hub-components/kafka-connect-mongodb/lib/mongo-kafka-connect-1.11.2-confluent.jar,/usr/share/java, /usr/share/confluent-hub-components'      CONNECT_ZOOKEEPER_CONNECT: zookeeper.local:22181      # CONNECT_SSL_CLIENT_AUTH: 'required'      CONNECT_SSL_KEY_PASSWORD: awesomekafka      CONNECT_SECURITY_PROTOCOL: SSL      CONNECT_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.truststore.jks      CONNECT_SSL_TRUSTSTORE_PASSWORD: awesomekafka      CONNECT_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.keystore.jks      CONNECT_SSL_KEYSTORE_PASSWORD: awesomekafka      CONNECT_PRODUCER_SECURITY_PROTOCOL: SSL      CONNECT_PRODUCER_BOOTSTRAP_SERVERS: SSL://broker.local:19092      CONNECT_PRODUCER_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.truststore.jks      CONNECT_PRODUCER_SSL_TRUSTSTORE_PASSWORD: awesomekafka      CONNECT_PRODUCER_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.keystore.jks      CONNECT_PRODUCER_SSL_KEYSTORE_PASSWORD: awesomekafka      CONNECT_CONSUMER_SECURITY_PROTOCOL: SSL          CONNECT_CONSUMER_BOOTSTRAP_SERVERS: SSL://broker.local:19092      CONNECT_CONSUMER_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.truststore.jks        CONNECT_CONSUMER_SSL_TRUSTSTORE_PASSWORD: awesomekafka      CONNECT_CONSUMER_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.connect.keystore.jks      CONNECT_CONSUMER_SSL_KEYSTORE_PASSWORD: awesomekafka      # Assumes image is based on confluentinc/kafka-connect-datagen:latest which is pulling 5.2.2 Connect image      CLASSPATH: /usr/share/java/monitoring-interceptors/monitoring-interceptors-7.5.3.jar      CONNECT_PRODUCER_INTERCEPTOR_CLASSES: "io.confluent.monitoring.clients.interceptor.MonitoringProducerInterceptor"      CONNECT_CONSUMER_INTERCEPTOR_CLASSES: "io.confluent.monitoring.clients.interceptor.MonitoringConsumerInterceptor"    command: "bash -c 'if [ ! -d /usr/share/confluent-hub-components/confluentinc-kafka-connect-datagen ]; then echo \"WARNING: Did not find directory for kafka-connect-datagen (did you remember to run: docker-compose up -d --build ?)\"; fi ; /etc/confluent/docker/run'"      # command:       #   "bash -c echo \"Installing Connector\" ;  confluent-hub install --no-prompt confluentinc/kafka-connect-datagen:0.1.6 ; echo \"Launching Kafka Connect worker\" ; /etc/confluent/docker/run &sleep infinity"    volumes:      - ./kafka-connect-mongodb:/usr/share/confluent-hub-components/kafka-connect-mongodb      - ./secrets:/etc/kafka/secrets   #MongoDB Replica Set  mongo1:    hostname: mongo1    image: mongo    container_name: mongo1    ports:      - 27017:27017    restart: always    entrypoint: [ "/usr/bin/mongod", "--config", "/data/config/ssl.conf", "--bind_ip_all", "--replSet", "dbrs" ]    networks:      - mongo_network    volumes:      - ./.db/mongo1:/data/db      - ./wait-for-mongodb.sh:/scripts/wait-for-mongodb.sh      - ./init.sh:/scripts/init.sh      - ./ssl:/data/ssl      - ./config:/data/config    links:      - mongo2      - mongo3  mongo2:    hostname: mongo2    image: mongo    container_name: mongo2    ports:      - 27018:27017    restart: always    entrypoint: [ "/usr/bin/mongod", "--bind_ip_all", "--replSet", "dbrs" ]    networks:      - mongo_network    volumes:      - ./.db/mongo2:/data/db      - ./ssl:/data/ssl      - ./wait-for-mongodb.sh:/scripts/wait-for-mongodb.sh  mongo3:    hostname: mongo3    image: mongo    container_name: mongo3    ports:      - 27019:27017    restart: always    entrypoint: [ "/usr/bin/mongod", "--bind_ip_all", "--replSet", "dbrs" ]    networks:      - mongo_network    volumes:      - ./.db/mongo3:/data/db      - ./ssl:/data/ssl      - ./wait-for-mongodb.sh:/scripts/wait-for-mongodb.sh  kafka-tools:    build:      context: ./kafka-tools    depends_on:      - broker    container_name: kafka-tools    networks:       kafka:          aliases:            - kafka-tools.local       mongo_network:    environment:      KAFKA_LOG4J_ROOT_LOGLEVEL: ERROR      KAFKA_SECURITY_PROTOCOL: SSL      KAFKA_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM: ""      KAFKA_HEAP_OPTS: "-Xmx1G -Xms1G"      KAFKA_SSL_TRUSTSTORE_LOCATION: /etc/kafka/secrets/kafka.kafka-tools.truststore.jks      KAFKA_SSL_KEYSTORE_LOCATION: /etc/kafka/secrets/kafka.kafka-tools.keystore.jks      KAFKA_SSL_TRUSTSTORE_PASSWORD: awesomekafka      KAFKA_SSL_KEYSTORE_PASSWORD: awesomekafka      KAFKA_SSL_KEY_PASSWORD: awesomekafka    volumes:      - ./secrets:/etc/kafka/secrets      - ./config:/etc/kafka/configvolumes:  zk-data:    external: true  zk-txn-logs:    external: true  kafka-data:    external: truenetworks:  kafka:  localnet:    attachable: true  mongo_network:    driver: bridge

Below is mong-sink connector json looks like :

> > {>   "name": "mongo-sink",>   "config": {>     "connector.class": "com.mongodb.kafka.connect.MongoSinkConnector",>     "tasks.max": "1",>     "topics": "mytopic",>     "connection.uri": "mongodb://mongo1:27017,mongo2:27017,mongo3:27017",>     "database": "mydb",>     "collection": "products",>     "key.converter": "org.apache.kafka.connect.storage.StringConverter",>     "value.converter": "org.apache.kafka.connect.storage.StringConverter",>     "key.converter.schemas.enable": "false",>     "value.converter.schemas.enable": "false",>     "mongo.errors.tolerance": "all",>     "mongo.errors.log.enable": "true",>     "errors.log.include.messages": "true",>     "writemodel.strategy": "com.mongodb.kafka.connect.sink.writemodel.strategy.ReplaceOneDefaultStrategy",>     "name": "mongo-sink">   },>   "tasks": [],>   "type": "sink"> }

Please note I tried changing this to below without luck :

"connection.uri": "mongodb://mongo1:27017,mongo2:27017,mongo3:27017/?ssl=true",

Mongo SSL config looks like below :

config/ssl.conf as below :

> net:>   ssl:>     mode: allowSSL>     PEMKeyFile: "/data/ssl/mongodb.pem">     CAFile: "/data/ssl/mongodb-cert.crt"> storage:>   dbPath: "/data/db"> net:>   bindIp: 127.0.0.1>   port: 27017

I am pretty sure mongodb replicaset is perfectly working but mongo sink connector task having issue as denoted here :

> [2024-03-07 00:06:21,165] ERROR Error on mongodb operation (com.mongodb.kafka.connect.sink.MongoSinkTask)> com.mongodb.MongoTimeoutException: Timed out after 30000 ms while waiting to connect. Client view of cluster state is {type=UNKNOWN, servers=[{address=mongo3:27017, type=REPLICA_SET_GHOST, roundTripTime=0.7 ms, state=CONNECTED}, {address=mongo2:27017, type=REPLICA_SET_GHOST, roundTripTime=0.6 ms, state=CONNECTED}, {address=mongo1:27017, type=REPLICA_SET_GHOST, roundTripTime=0.7 ms, state=CONNECTED}]>         at com.mongodb.internal.connection.BaseCluster.getDescription(BaseCluster.java:182)>         at com.mongodb.internal.connection.AbstractMultiServerCluster.getDescription(AbstractMultiServerCluster.java:52)>         at com.mongodb.client.internal.MongoClientDelegate.getConnectedClusterDescription(MongoClientDelegate.java:136)>         at com.mongodb.client.internal.MongoClientDelegate.createClientSession(MongoClientDelegate.java:94)>         at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.getClientSession(MongoClientDelegate.java:249)>         at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:190)>         at com.mongodb.client.internal.MongoCollectionImpl.executeBulkWrite(MongoCollectionImpl.java:468)>         at com.mongodb.client.internal.MongoCollectionImpl.bulkWrite(MongoCollectionImpl.java:448)>         at com.mongodb.kafka.connect.sink.MongoSinkTask.processSinkRecords(MongoSinkTask.java:177)>         at com.mongodb.kafka.connect.sink.MongoSinkTask.lambda$put$2(MongoSinkTask.java:117)>         at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)>         at com.mongodb.kafka.connect.sink.MongoSinkTask.lambda$put$3(MongoSinkTask.java:116)>         at java.base/java.util.HashMap.forEach(HashMap.java:1337)>         at com.mongodb.kafka.connect.sink.MongoSinkTask.put(MongoSinkTask.java:114)>         at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:593)>         at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:342)>         at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:242)>         at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:211)>         at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:204)>         at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:259)>         at org.apache.kafka.connect.runtime.isolation.Plugins.lambda$withClassLoader$1(Plugins.java:181)>         at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)>         at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)>         at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)>         at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)>         at java.base/java.lang.Thread.run(Thread.java:829)> [2024-03-07 00:06:21,168] ERROR Writing 4 document(s) into collection [mydb.products] failed -> remaining retries (3) (com.mongodb.kafka.connect.sink.MongoSinkTask)

I have tested with curl https://broker.local:19092 and curl -v https://mongo1:27017 from connect node and response is positive

> > *   Trying 192.168.xx.x...> * TCP_NODELAY set> * Connected to broker.local (192.168.xx.xx) port 19092 (#0)> * ALPN, offering h2> * ALPN, offering http/1.1> * successfully set certificate verify locations:> *   CAfile: /etc/pki/tls/certs/ca-bundle.crt>   CApath: none> * TLSv1.3 (OUT), TLS handshake, Client hello (1):> * TLSv1.3 (IN), TLS handshake, Server hello (2):> * TLSv1.3 (IN), TLS handshake, [no content] (0):> * TLSv1.3 (IN), TLS handshake, Encrypted Extensions (8):> * TLSv1.3 (IN), TLS handshake, Request CERT (13):> * TLSv1.3 (IN), TLS handshake, Certificate (11):> * TLSv1.3 (OUT), TLS alert, unknown CA (560):

config parameters initialized from mongo connect container docker logs as below :

   [2024-03-07 00:26:37,270] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [SSL://broker.local:19092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = compose-connect-group--offsets client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = compose-connect-group group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 45000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = [hidden] ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = /etc/kafka/secrets/kafka.connect.keystore.jks ssl.keystore.password = [hidden] ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = /etc/kafka/secrets/kafka.connect.truststore.jks ssl.truststore.password = [hidden] ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer>  (org.apache.kafka.clients.consumer.ConsumerConfig)

Viewing all articles
Browse latest Browse all 12111

Trending Articles



<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>