RuntimeError: Coordinator stopped with threads still running: Thread-2 Thread-3 Thread-1 Thread-4

博客介绍了TensorFlow在创建队列并启动多进程操作同一队列时的报错原因,即某线程退出时队列不会自动关闭,入队操作持续进行导致报错。解决方法是在执行相关方法前关闭队列,如使用sess.run(q.close()) 。

参考网址

https://github.com/tensorflow/tensorflow/issues/2130

  • 报错原因:创建tf.FIFOQueue队列并且启动多个进程对同一个队列进行操作时,往往是通过调用tf.Coordinator类的should_stop,request_stopjoin三个方法进行停止;
  • 当某一个线程退出时,则should_stop返回True时,停止当前线程;
  • 通常使用tf.FIFOQueue先入先出队列,should_top不会自动返回true,关闭当前线程,只有当调用request_stop停止其他线程,才会在下一个step去判断should_stop的返回值,但此时当前线程的入队操作依然在进行,等待数据入队,因此在Coordinator stopped会报错Coordinator stopped with threads still running: Thread-2 Thread-3 Thread-1 Thread-4
  • 解决方法:在执行coord.request_stop()方法之前,sess.run(queue.close())把队列关闭就可以;有时可能需要给队列指定参数cancel_pending_enqueues=True,如sess.run(queue.close(cancel_pending_enqueues=True))

“queue.close()”, but that actually returns an op which needs to be run to do anything. You need to do sess.run(q.close()). Since your first queue is not closed, your “batch” queue is waiting forever for something to be added to the first queue.

Furthermore, this wait is happening in C++ mutex, so stop_grace_period_secs is useless – the queue runner thread checks for “stop_requested” between session run calls, but because dequeue op never returns, it’s stuck inside session.run forever.

def create_session():
  """Resets local session, returns new InteractiveSession"""

  config = tf.ConfigProto(log_device_placement=True)
  config.gpu_options.per_process_gpu_memory_fraction=0.3 # don't hog all vRAM
  config.operation_timeout_in_ms=5000   # terminate on long hangs
  sess = tf.InteractiveSession("", config=config)
  return sess

tf.reset_default_graph()
q = tf.FIFOQueue(4, tf.string)
enqueue_val = tf.placeholder(dtype=tf.string)
enqueue_op = q.enqueue(enqueue_val)
size_op = q.size()
dequeue_op = q.dequeue()
sess = create_session()
def enqueueit(val):
  sess.run([enqueue_op], feed_dict={enqueue_val:val})
  print "queue1 size: ", sess.run(size_op)
enqueueit("1")
enqueueit("2")
enqueueit("3")

dequeue_op.set_shape([])
queue2 = tf.train.batch([dequeue_op], batch_size=1, num_threads=1, capacity=1)
threads = tf.train.start_queue_runners()

def dequeueit():
  print "queue1 size: ", sess.run(size_op)
  print "queue2 size before: ", sess.run("batch/fifo_queue_Size:0")
  print "result: ", sess.run(queue2)
  print "queue2 size after: ", sess.run("batch/fifo_queue_Size:0")

dequeueit()
dequeueit()
dequeueit()
#solution here
# Ask the threads to stop and wait until they do it
sess.run(q.close(cancel_pending_enqueues=True))
coord.request_stop()
coord.join(threads, stop_grace_period_secs=5)
#sess.close()
[root@kafka1 hejia]# /usr/local/kafka_2.13-3.7.2/bin/kafka-server-start.sh /usr/local/kafka_2.13-3.7.2/config/server.properties [2025-09-01 17:10:47,478] INFO Registered kafka:type=kafka.Log4jController MBean (kafka.utils.Log4jControllerRegistration$) [2025-09-01 17:10:47,911] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) [2025-09-01 17:10:48,125] INFO Registered signal handlers for TERM, INT, HUP (org.apache.kafka.common.utils.LoggingSignalHandler) [2025-09-01 17:10:48,127] INFO starting (kafka.server.KafkaServer) [2025-09-01 17:10:48,128] INFO Connecting to zookeeper on 192.168.14.124:2181,192.168.14.125:2181,192.168.14.126:2181 (kafka.server.KafkaServer) [2025-09-01 17:10:48,145] INFO [ZooKeeperClient Kafka server] Initializing a new session to 192.168.14.124:2181,192.168.14.125:2181,192.168.14.126:2181. (kafka.zookeeper.ZooKeeperClient) [2025-09-01 17:10:48,151] INFO Client environment:zookeeper.version=3.8.4-9316c2a7a97e1666d8f4593f34dd6fc36ecc436c, built on 2024-02-12 22:16 UTC (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,152] INFO Client environment:host.name=kafka1 (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,152] INFO Client environment:java.version=1.8.0_171 (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,152] INFO Client environment:java.vendor=Oracle Corporation (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,152] INFO Client environment:java.home=/usr/local/jdk1.8.0_171/jre (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,152] INFO Client environment:java.class.path=/usr/local/jdk1.8.0_171/jre/lib/ext:/usr/local/jdk1.8.0_171/lib/tools.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/activation-1.1.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/aopalliance-repackaged-2.6.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/argparse4j-0.7.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/audience-annotations-0.12.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/caffeine-2.9.3.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/checker-qual-3.19.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/commons-beanutils-1.9.4.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/commons-cli-1.4.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/commons-collections-3.2.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/commons-digester-2.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/commons-io-2.14.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/commons-lang3-3.8.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/commons-logging-1.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/commons-validator-1.7.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/connect-api-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/connect-basic-auth-extension-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/connect-json-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/connect-mirror-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/connect-mirror-client-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/connect-runtime-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/connect-transforms-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/error_prone_annotations-2.10.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/hk2-api-2.6.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/hk2-locator-2.6.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/hk2-utils-2.6.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jackson-annotations-2.16.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jackson-core-2.16.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jackson-databind-2.16.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jackson-dataformat-csv-2.16.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jackson-datatype-jdk8-2.16.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jackson-jaxrs-base-2.16.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jackson-jaxrs-json-provider-2.16.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jackson-module-jaxb-annotations-2.16.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jackson-module-scala_2.13-2.16.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jakarta.activation-api-1.2.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jakarta.annotation-api-1.3.5.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jakarta.inject-2.6.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jakarta.validation-api-2.0.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jakarta.ws.rs-api-2.1.6.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jakarta.xml.bind-api-2.3.3.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/javassist-3.29.2-GA.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/javax.activation-api-1.2.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/javax.annotation-api-1.3.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/javax.servlet-api-3.1.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/javax.ws.rs-api-2.1.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jaxb-api-2.3.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jersey-client-2.39.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jersey-common-2.39.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jersey-container-servlet-2.39.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jersey-container-servlet-core-2.39.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jersey-hk2-2.39.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jersey-server-2.39.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-client-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-continuation-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-http-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-io-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-security-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-server-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-servlet-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-servlets-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-util-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jetty-util-ajax-9.4.56.v20240826.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jline-3.25.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jopt-simple-5.0.4.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jose4j-0.9.4.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/jsr305-3.0.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka_2.13-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-clients-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-group-coordinator-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-log4j-appender-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-metadata-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-raft-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-server-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-server-common-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-shell-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-storage-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-storage-api-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-streams-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-streams-examples-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-streams-scala_2.13-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-streams-test-utils-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-tools-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/kafka-tools-api-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/lz4-java-1.8.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/maven-artifact-3.8.8.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/metrics-core-2.2.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/metrics-core-4.1.12.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/netty-buffer-4.1.115.Final.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/netty-codec-4.1.115.Final.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/netty-common-4.1.115.Final.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/netty-handler-4.1.115.Final.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/netty-resolver-4.1.115.Final.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/netty-transport-4.1.115.Final.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/netty-transport-classes-epoll-4.1.115.Final.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/netty-transport-native-epoll-4.1.115.Final.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/netty-transport-native-unix-common-4.1.115.Final.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/opentelemetry-proto-1.0.0-alpha.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/osgi-resource-locator-1.0.3.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/paranamer-2.8.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/pcollections-4.0.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/plexus-utils-3.3.1.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/protobuf-java-3.25.5.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/reflections-0.10.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/reload4j-1.2.25.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/rocksdbjni-7.9.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/scala-collection-compat_2.13-2.10.0.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/scala-java8-compat_2.13-1.0.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/scala-library-2.13.12.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/scala-logging_2.13-3.9.4.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/scala-reflect-2.13.12.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/slf4j-api-1.7.36.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/slf4j-reload4j-1.7.36.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/snappy-java-1.1.10.5.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/swagger-annotations-2.2.8.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/trogdor-3.7.2.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/zookeeper-3.8.4.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/zookeeper-jute-3.8.4.jar:/usr/local/kafka_2.13-3.7.2/bin/../libs/zstd-jni-1.5.6-4.jar (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,153] INFO Client environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,153] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,153] INFO Client environment:java.compiler=<NA> (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,154] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,154] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,154] INFO Client environment:os.version=3.10.0-1160.el7.x86_64 (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,154] INFO Client environment:user.name=root (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,154] INFO Client environment:user.home=/root (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,154] INFO Client environment:user.dir=/home/hejia (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,154] INFO Client environment:os.memory.free=1003MB (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,154] INFO Client environment:os.memory.max=1024MB (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,154] INFO Client environment:os.memory.total=1024MB (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,157] INFO Initiating client connection, connectString=192.168.14.124:2181,192.168.14.125:2181,192.168.14.126:2181 sessionTimeout=18000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@6150c3ec (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:48,161] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket) [2025-09-01 17:10:48,170] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) [2025-09-01 17:10:48,172] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient) [2025-09-01 17:10:48,174] INFO Opening socket connection to server /192.168.14.126:2181. (org.apache.zookeeper.ClientCnxn) [2025-09-01 17:10:48,178] INFO Socket connection established, initiating session, client: /192.168.14.124:42150, server: /192.168.14.126:2181 (org.apache.zookeeper.ClientCnxn) [2025-09-01 17:10:48,184] INFO Session establishment complete on server /192.168.14.126:2181, session id = 0x3000359362800e9, negotiated timeout = 18000 (org.apache.zookeeper.ClientCnxn) [2025-09-01 17:10:48,187] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient) [2025-09-01 17:10:48,627] INFO Cluster ID = eQFewTrDRa6Csx8shnSzQQ (kafka.server.KafkaServer) [2025-09-01 17:10:48,709] INFO KafkaConfig values: advertised.listeners = null alter.config.policy.class.name = null alter.log.dirs.replication.quota.window.num = 11 alter.log.dirs.replication.quota.window.size.seconds = 1 authorizer.class.name = auto.create.topics.enable = true auto.include.jmx.reporter = true auto.leader.rebalance.enable = true background.threads = 10 broker.heartbeat.interval.ms = 2000 broker.id = 0 broker.id.generation.enable = true broker.rack = null broker.session.timeout.ms = 9000 client.quota.callback.class = null compression.type = producer connection.failed.authentication.delay.ms = 100 connections.max.idle.ms = 600000 connections.max.reauth.ms = 0 control.plane.listener.name = null controlled.shutdown.enable = true controlled.shutdown.max.retries = 3 controlled.shutdown.retry.backoff.ms = 5000 controller.listener.names = null controller.quorum.append.linger.ms = 25 controller.quorum.election.backoff.max.ms = 1000 controller.quorum.election.timeout.ms = 1000 controller.quorum.fetch.timeout.ms = 2000 controller.quorum.request.timeout.ms = 2000 controller.quorum.retry.backoff.ms = 20 controller.quorum.voters = [] controller.quota.window.num = 11 controller.quota.window.size.seconds = 1 controller.socket.timeout.ms = 30000 create.topic.policy.class.name = null default.replication.factor = 1 delegation.token.expiry.check.interval.ms = 3600000 delegation.token.expiry.time.ms = 86400000 delegation.token.master.key = null delegation.token.max.lifetime.ms = 604800000 delegation.token.secret.key = null delete.records.purgatory.purge.interval.requests = 1 delete.topic.enable = true early.start.listeners = null eligible.leader.replicas.enable = false fetch.max.bytes = 57671680 fetch.purgatory.purge.interval.requests = 1000 group.consumer.assignors = [org.apache.kafka.coordinator.group.assignor.UniformAssignor, org.apache.kafka.coordinator.group.assignor.RangeAssignor] group.consumer.heartbeat.interval.ms = 5000 group.consumer.max.heartbeat.interval.ms = 15000 group.consumer.max.session.timeout.ms = 60000 group.consumer.max.size = 2147483647 group.consumer.min.heartbeat.interval.ms = 5000 group.consumer.min.session.timeout.ms = 45000 group.consumer.session.timeout.ms = 45000 group.coordinator.new.enable = false group.coordinator.rebalance.protocols = [classic] group.coordinator.threads = 1 group.initial.rebalance.delay.ms = 0 group.max.session.timeout.ms = 1800000 group.max.size = 2147483647 group.min.session.timeout.ms = 6000 initial.broker.registration.timeout.ms = 60000 inter.broker.listener.name = null inter.broker.protocol.version = 3.7-IV4 kafka.metrics.polling.interval.secs = 10 kafka.metrics.reporters = [] leader.imbalance.check.interval.seconds = 300 leader.imbalance.per.broker.percentage = 10 listener.security.protocol.map = PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL listeners = PLAINTEXT://192.168.14.124:9092 log.cleaner.backoff.ms = 15000 log.cleaner.dedupe.buffer.size = 134217728 log.cleaner.delete.retention.ms = 86400000 log.cleaner.enable = true log.cleaner.io.buffer.load.factor = 0.9 log.cleaner.io.buffer.size = 524288 log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308 log.cleaner.max.compaction.lag.ms = 9223372036854775807 log.cleaner.min.cleanable.ratio = 0.5 log.cleaner.min.compaction.lag.ms = 0 log.cleaner.threads = 1 log.cleanup.policy = [delete] log.dir = /tmp/kafka-logs log.dirs = /data/kafka/log log.flush.interval.messages = 9223372036854775807 log.flush.interval.ms = null log.flush.offset.checkpoint.interval.ms = 60000 log.flush.scheduler.interval.ms = 9223372036854775807 log.flush.start.offset.checkpoint.interval.ms = 60000 log.index.interval.bytes = 4096 log.index.size.max.bytes = 10485760 log.local.retention.bytes = -2 log.local.retention.ms = -2 log.message.downconversion.enable = true log.message.format.version = 3.0-IV1 log.message.timestamp.after.max.ms = 9223372036854775807 log.message.timestamp.before.max.ms = 9223372036854775807 log.message.timestamp.difference.max.ms = 9223372036854775807 log.message.timestamp.type = CreateTime log.preallocate = false log.retention.bytes = -1 log.retention.check.interval.ms = 300000 log.retention.hours = 168 log.retention.minutes = null log.retention.ms = null log.roll.hours = 168 log.roll.jitter.hours = 0 log.roll.jitter.ms = null log.roll.ms = null log.segment.bytes = 1073741824 log.segment.delete.delay.ms = 60000 max.connection.creation.rate = 2147483647 max.connections = 2147483647 max.connections.per.ip = 2147483647 max.connections.per.ip.overrides = max.incremental.fetch.session.cache.slots = 1000 message.max.bytes = 1048588 metadata.log.dir = null metadata.log.max.record.bytes.between.snapshots = 20971520 metadata.log.max.snapshot.interval.ms = 3600000 metadata.log.segment.bytes = 1073741824 metadata.log.segment.min.bytes = 8388608 metadata.log.segment.ms = 604800000 metadata.max.idle.interval.ms = 500 metadata.max.retention.bytes = 104857600 metadata.max.retention.ms = 604800000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 min.insync.replicas = 1 node.id = 0 num.io.threads = 8 num.network.threads = 3 num.partitions = 1 num.recovery.threads.per.data.dir = 1 num.replica.alter.log.dirs.threads = null num.replica.fetchers = 1 offset.metadata.max.bytes = 4096 offsets.commit.required.acks = -1 offsets.commit.timeout.ms = 5000 offsets.load.buffer.size = 5242880 offsets.retention.check.interval.ms = 600000 offsets.retention.minutes = 10080 offsets.topic.compression.codec = 0 offsets.topic.num.partitions = 50 offsets.topic.replication.factor = 1 offsets.topic.segment.bytes = 104857600 password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding password.encoder.iterations = 4096 password.encoder.key.length = 128 password.encoder.keyfactory.algorithm = null password.encoder.old.secret = null password.encoder.secret = null principal.builder.class = class org.apache.kafka.common.security.authenticator.DefaultKafkaPrincipalBuilder process.roles = [] producer.id.expiration.check.interval.ms = 600000 producer.id.expiration.ms = 86400000 producer.purgatory.purge.interval.requests = 1000 queued.max.request.bytes = -1 queued.max.requests = 500 quota.window.num = 11 quota.window.size.seconds = 1 remote.log.index.file.cache.total.size.bytes = 1073741824 remote.log.manager.task.interval.ms = 30000 remote.log.manager.task.retry.backoff.max.ms = 30000 remote.log.manager.task.retry.backoff.ms = 500 remote.log.manager.task.retry.jitter = 0.2 remote.log.manager.thread.pool.size = 10 remote.log.metadata.custom.metadata.max.bytes = 128 remote.log.metadata.manager.class.name = org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemoteLogMetadataManager remote.log.metadata.manager.class.path = null remote.log.metadata.manager.impl.prefix = rlmm.config. remote.log.metadata.manager.listener.name = null remote.log.reader.max.pending.tasks = 100 remote.log.reader.threads = 10 remote.log.storage.manager.class.name = null remote.log.storage.manager.class.path = null remote.log.storage.manager.impl.prefix = rsm.config. remote.log.storage.system.enable = false replica.fetch.backoff.ms = 1000 replica.fetch.max.bytes = 1048576 replica.fetch.min.bytes = 1 replica.fetch.response.max.bytes = 10485760 replica.fetch.wait.max.ms = 500 replica.high.watermark.checkpoint.interval.ms = 5000 replica.lag.time.max.ms = 30000 replica.selector.class = null replica.socket.receive.buffer.bytes = 65536 replica.socket.timeout.ms = 30000 replication.quota.window.num = 11 replication.quota.window.size.seconds = 1 request.timeout.ms = 30000 reserved.broker.max.id = 1000 sasl.client.callback.handler.class = null sasl.enabled.mechanisms = [GSSAPI] sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.principal.to.local.rules = [DEFAULT] sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism.controller.protocol = GSSAPI sasl.mechanism.inter.broker.protocol = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null sasl.server.callback.handler.class = null sasl.server.max.receive.size = 524288 security.inter.broker.protocol = PLAINTEXT security.providers = null server.max.startup.time.ms = 9223372036854775807 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 socket.listen.backlog.size = 50 socket.receive.buffer.bytes = 102400 socket.request.max.bytes = 104857600 socket.send.buffer.bytes = 102400 ssl.allow.dn.changes = false ssl.allow.san.changes = false ssl.cipher.suites = [] ssl.client.auth = none ssl.enabled.protocols = [TLSv1.2] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.principal.mapping.rules = DEFAULT ssl.protocol = TLSv1.2 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS telemetry.max.bytes = 1048576 transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000 transaction.max.timeout.ms = 900000 transaction.partition.verification.enable = true transaction.remove.expired.transaction.cleanup.interval.ms = 3600000 transaction.state.log.load.buffer.size = 5242880 transaction.state.log.min.isr = 1 transaction.state.log.num.partitions = 50 transaction.state.log.replication.factor = 1 transaction.state.log.segment.bytes = 104857600 transactional.id.expiration.ms = 604800000 unclean.leader.election.enable = false unstable.api.versions.enable = false unstable.metadata.versions.enable = false zookeeper.clientCnxnSocket = null zookeeper.connect = 192.168.14.124:2181,192.168.14.125:2181,192.168.14.126:2181 zookeeper.connection.timeout.ms = 18000 zookeeper.max.in.flight.requests = 10 zookeeper.metadata.migration.enable = false zookeeper.metadata.migration.min.batch.size = 200 zookeeper.session.timeout.ms = 18000 zookeeper.set.acl = false zookeeper.ssl.cipher.suites = null zookeeper.ssl.client.enable = false zookeeper.ssl.crl.enable = false zookeeper.ssl.enabled.protocols = null zookeeper.ssl.endpoint.identification.algorithm = HTTPS zookeeper.ssl.keystore.location = null zookeeper.ssl.keystore.password = null zookeeper.ssl.keystore.type = null zookeeper.ssl.ocsp.enable = false zookeeper.ssl.protocol = TLSv1.2 zookeeper.ssl.truststore.location = null zookeeper.ssl.truststore.password = null zookeeper.ssl.truststore.type = null (kafka.server.KafkaConfig) [2025-09-01 17:10:48,784] INFO [ThrottledChannelReaper-Fetch]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:48,785] INFO [ThrottledChannelReaper-Produce]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:48,789] INFO [ThrottledChannelReaper-Request]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:48,791] INFO [ThrottledChannelReaper-ControllerMutation]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:48,953] INFO Loading logs from log dirs ArraySeq(/data/kafka/log) (kafka.log.LogManager) [2025-09-01 17:10:48,960] INFO No logs found to be loaded in /data/kafka/log (kafka.log.LogManager) [2025-09-01 17:10:48,982] INFO Loaded 0 logs in 28ms (kafka.log.LogManager) [2025-09-01 17:10:48,985] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager) [2025-09-01 17:10:48,987] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager) [2025-09-01 17:10:49,700] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner$CleanerThread) [2025-09-01 17:10:49,723] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) [2025-09-01 17:10:49,754] INFO [MetadataCache brokerId=0] Updated cache from existing None to latest Features(version=3.7-IV4, finalizedFeatures={}, finalizedFeaturesEpoch=0). (kafka.server.metadata.ZkMetadataCache) [2025-09-01 17:10:49,807] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Starting (kafka.server.NodeToControllerRequestThread) [2025-09-01 17:10:50,663] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) [2025-09-01 17:10:50,694] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer) [2025-09-01 17:10:50,719] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Starting (kafka.server.NodeToControllerRequestThread) [2025-09-01 17:10:50,797] INFO [ExpirationReaper-0-Produce]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:50,928] INFO [ExpirationReaper-0-Fetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:50,929] INFO [ExpirationReaper-0-DeleteRecords]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:50,931] INFO [ExpirationReaper-0-ElectLeader]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:50,939] INFO [ExpirationReaper-0-RemoteFetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:50,963] INFO [LogDirFailureHandler]: Starting (kafka.server.ReplicaManager$LogDirFailureHandler) [2025-09-01 17:10:50,970] INFO [AddPartitionsToTxnSenderThread-0]: Starting (kafka.server.AddPartitionsToTxnManager) [2025-09-01 17:10:51,183] INFO Creating /brokers/ids/0 (is it secure? false) (kafka.zk.KafkaZkClient) [2025-09-01 17:10:51,209] ERROR Error while creating ephemeral at /brokers/ids/0, node already exists and owner &#39;0x200038df42600ef&#39; does not match current session &#39;0x3000359362800e9&#39; (kafka.zk.KafkaZkClient$CheckedEphemeral) [2025-09-01 17:10:51,235] ERROR [KafkaServer id=0] Fatal error during KafkaServer startup. Prepare to shutdown (kafka.server.KafkaServer) org.apache.zookeeper.KeeperException$NodeExistsException: KeeperErrorCode = NodeExists at org.apache.zookeeper.KeeperException.create(KeeperException.java:126) at kafka.zk.KafkaZkClient$CheckedEphemeral.getAfterNodeExists(KafkaZkClient.scala:2238) at kafka.zk.KafkaZkClient$CheckedEphemeral.create(KafkaZkClient.scala:2176) at kafka.zk.KafkaZkClient.checkedEphemeralCreate(KafkaZkClient.scala:2143) at kafka.zk.KafkaZkClient.registerBroker(KafkaZkClient.scala:110) at kafka.server.KafkaServer.startup(KafkaServer.scala:404) at kafka.Kafka$.main(Kafka.scala:112) at kafka.Kafka.main(Kafka.scala) [2025-09-01 17:10:51,238] INFO [KafkaServer id=0] shutting down (kafka.server.KafkaServer) [2025-09-01 17:10:51,239] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Stopping socket server request processors (kafka.network.SocketServer) [2025-09-01 17:10:51,245] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Stopped socket server request processors (kafka.network.SocketServer) [2025-09-01 17:10:51,248] INFO [ReplicaManager broker=0] Shutting down (kafka.server.ReplicaManager) [2025-09-01 17:10:51,250] INFO [LogDirFailureHandler]: Shutting down (kafka.server.ReplicaManager$LogDirFailureHandler) [2025-09-01 17:10:51,250] INFO [LogDirFailureHandler]: Stopped (kafka.server.ReplicaManager$LogDirFailureHandler) [2025-09-01 17:10:51,250] INFO [LogDirFailureHandler]: Shutdown completed (kafka.server.ReplicaManager$LogDirFailureHandler) [2025-09-01 17:10:51,251] INFO [ReplicaFetcherManager on broker 0] shutting down (kafka.server.ReplicaFetcherManager) [2025-09-01 17:10:51,252] INFO [ReplicaFetcherManager on broker 0] shutdown completed (kafka.server.ReplicaFetcherManager) [2025-09-01 17:10:51,252] INFO [ReplicaAlterLogDirsManager on broker 0] shutting down (kafka.server.ReplicaAlterLogDirsManager) [2025-09-01 17:10:51,252] INFO [ReplicaAlterLogDirsManager on broker 0] shutdown completed (kafka.server.ReplicaAlterLogDirsManager) [2025-09-01 17:10:51,253] INFO [ExpirationReaper-0-Fetch]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,253] INFO [ExpirationReaper-0-Fetch]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,254] INFO [ExpirationReaper-0-Fetch]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,254] INFO [ExpirationReaper-0-RemoteFetch]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,255] INFO [ExpirationReaper-0-RemoteFetch]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,255] INFO [ExpirationReaper-0-RemoteFetch]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,255] INFO [ExpirationReaper-0-Produce]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,255] INFO [ExpirationReaper-0-Produce]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,256] INFO [ExpirationReaper-0-Produce]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,256] INFO [ExpirationReaper-0-DeleteRecords]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,256] INFO [ExpirationReaper-0-DeleteRecords]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,256] INFO [ExpirationReaper-0-DeleteRecords]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,257] INFO [ExpirationReaper-0-ElectLeader]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,257] INFO [ExpirationReaper-0-ElectLeader]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,257] INFO [ExpirationReaper-0-ElectLeader]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2025-09-01 17:10:51,264] INFO [AddPartitionsToTxnSenderThread-0]: Shutting down (kafka.server.AddPartitionsToTxnManager) [2025-09-01 17:10:51,265] INFO [AddPartitionsToTxnSenderThread-0]: Stopped (kafka.server.AddPartitionsToTxnManager) [2025-09-01 17:10:51,265] INFO [AddPartitionsToTxnSenderThread-0]: Shutdown completed (kafka.server.AddPartitionsToTxnManager) [2025-09-01 17:10:51,267] INFO [ReplicaManager broker=0] Shut down completely (kafka.server.ReplicaManager) [2025-09-01 17:10:51,267] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Shutting down (kafka.server.NodeToControllerRequestThread) [2025-09-01 17:10:51,268] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Stopped (kafka.server.NodeToControllerRequestThread) [2025-09-01 17:10:51,268] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Shutdown completed (kafka.server.NodeToControllerRequestThread) [2025-09-01 17:10:51,269] INFO Node to controller channel manager for alter-partition shutdown (kafka.server.NodeToControllerChannelManagerImpl) [2025-09-01 17:10:51,269] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Shutting down (kafka.server.NodeToControllerRequestThread) [2025-09-01 17:10:51,270] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Stopped (kafka.server.NodeToControllerRequestThread) [2025-09-01 17:10:51,270] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Shutdown completed (kafka.server.NodeToControllerRequestThread) [2025-09-01 17:10:51,270] INFO Node to controller channel manager for forwarding shutdown (kafka.server.NodeToControllerChannelManagerImpl) [2025-09-01 17:10:51,271] INFO Shutting down. (kafka.log.LogManager) [2025-09-01 17:10:51,273] INFO [kafka-log-cleaner-thread-0]: Shutting down (kafka.log.LogCleaner$CleanerThread) [2025-09-01 17:10:51,275] INFO [kafka-log-cleaner-thread-0]: Stopped (kafka.log.LogCleaner$CleanerThread) [2025-09-01 17:10:51,275] INFO [kafka-log-cleaner-thread-0]: Shutdown completed (kafka.log.LogCleaner$CleanerThread) [2025-09-01 17:10:51,336] INFO Shutdown complete. (kafka.log.LogManager) [2025-09-01 17:10:51,490] INFO [feature-zk-node-event-process-thread]: Shutting down (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) [2025-09-01 17:10:51,490] INFO [feature-zk-node-event-process-thread]: Stopped (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) [2025-09-01 17:10:51,490] INFO [feature-zk-node-event-process-thread]: Shutdown completed (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) [2025-09-01 17:10:51,491] INFO [ZooKeeperClient Kafka server] Closing. (kafka.zookeeper.ZooKeeperClient) [2025-09-01 17:10:51,623] INFO Session: 0x3000359362800e9 closed (org.apache.zookeeper.ZooKeeper) [2025-09-01 17:10:51,624] INFO EventThread shut down for session: 0x3000359362800e9 (org.apache.zookeeper.ClientCnxn) [2025-09-01 17:10:51,745] INFO [ZooKeeperClient Kafka server] Closed. (kafka.zookeeper.ZooKeeperClient) [2025-09-01 17:10:51,748] INFO [ThrottledChannelReaper-Fetch]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,758] INFO [ThrottledChannelReaper-Fetch]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,778] INFO [ThrottledChannelReaper-Fetch]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,778] INFO [ThrottledChannelReaper-Produce]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,792] INFO [ThrottledChannelReaper-Produce]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,792] INFO [ThrottledChannelReaper-Produce]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,792] INFO [ThrottledChannelReaper-Request]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,792] INFO [ThrottledChannelReaper-Request]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,792] INFO [ThrottledChannelReaper-Request]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,806] INFO [ThrottledChannelReaper-ControllerMutation]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,806] INFO [ThrottledChannelReaper-ControllerMutation]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,806] INFO [ThrottledChannelReaper-ControllerMutation]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper) [2025-09-01 17:10:51,830] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Shutting down socket server (kafka.network.SocketServer) [2025-09-01 17:10:52,317] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Shutdown completed (kafka.network.SocketServer) [2025-09-01 17:10:52,363] INFO Metrics scheduler closed (org.apache.kafka.common.metrics.Metrics) [2025-09-01 17:10:52,363] INFO Metrics reporters closed (org.apache.kafka.common.metrics.Metrics) [2025-09-01 17:10:52,415] INFO Broker and topic stats closed (kafka.server.BrokerTopicStats) [2025-09-01 17:10:52,554] INFO App info kafka.server for 0 unregistered (org.apache.kafka.common.utils.AppInfoParser) [2025-09-01 17:10:52,610] INFO [KafkaServer id=0] shut down completed (kafka.server.KafkaServer) [2025-09-01 17:10:52,610] ERROR Exiting Kafka due to fatal exception during startup. (kafka.Kafka$) org.apache.zookeeper.KeeperException$NodeExistsException: KeeperErrorCode = NodeExists at org.apache.zookeeper.KeeperException.create(KeeperException.java:126) at kafka.zk.KafkaZkClient$CheckedEphemeral.getAfterNodeExists(KafkaZkClient.scala:2238) at kafka.zk.KafkaZkClient$CheckedEphemeral.create(KafkaZkClient.scala:2176) at kafka.zk.KafkaZkClient.checkedEphemeralCreate(KafkaZkClient.scala:2143) at kafka.zk.KafkaZkClient.registerBroker(KafkaZkClient.scala:110) at kafka.server.KafkaServer.startup(KafkaServer.scala:404) at kafka.Kafka$.main(Kafka.scala:112) at kafka.Kafka.main(Kafka.scala) [2025-09-01 17:10:52,953] INFO [KafkaServer id=0] shutting down (kafka.server.KafkaServer)
最新发布
09-02
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值