{\rtf1\ansi\ansicpg1252\cocoartf2513
\cocoatextscaling0\cocoaplatform0{\fonttbl\f0\fmodern\fcharset0 Courier;}
{\colortbl;\red255\green255\blue255;\red36\green36\blue36;\red245\green245\blue245;}
{\*\expandedcolortbl;;\cssrgb\c18824\c18824\c18824;\cssrgb\c96863\c96863\c96863;}
\margl1440\margr1440\vieww28420\viewh17660\viewkind0
\deftab720
\pard\pardeftab720\partightenfactor0

\f0\fs29\fsmilli14667 \cf2 \cb3 \expnd0\expndtw0\kerning0
\outl0\strokewidth0 \strokec2 [2021-07-14 16:05:15,485] INFO KafkaConfig values: \
	advertised.listeners = null\
	alter.config.policy.class.name = null\
	alter.log.dirs.replication.quota.window.num = 11\
	alter.log.dirs.replication.quota.window.size.seconds = 1\
	authorizer.class.name = \
	auto.create.topics.enable = true\
	auto.leader.rebalance.enable = true\
	background.threads = 10\
	broker.heartbeat.interval.ms = 2000\
	broker.id = 0\
	broker.id.generation.enable = true\
	broker.rack = null\
	broker.session.timeout.ms = 9000\
	client.quota.callback.class = null\
	compression.type = producer\
	connection.failed.authentication.delay.ms = 100\
	connections.max.idle.ms = 600000\
	connections.max.reauth.ms = 0\
	control.plane.listener.name = null\
	controlled.shutdown.enable = true\
	controlled.shutdown.max.retries = 3\
	controlled.shutdown.retry.backoff.ms = 5000\
	controller.listener.names = null\
	controller.quorum.append.linger.ms = 25\
	controller.quorum.election.backoff.max.ms = 1000\
	controller.quorum.election.timeout.ms = 1000\
	controller.quorum.fetch.timeout.ms = 2000\
	controller.quorum.request.timeout.ms = 2000\
	controller.quorum.retry.backoff.ms = 20\
	controller.quorum.voters = []\
	controller.quota.window.num = 11\
	controller.quota.window.size.seconds = 1\
	controller.socket.timeout.ms = 30000\
	create.topic.policy.class.name = null\
	default.replication.factor = 1\
	delegation.token.expiry.check.interval.ms = 3600000\
	delegation.token.expiry.time.ms = 86400000\
	delegation.token.master.key = null\
	delegation.token.max.lifetime.ms = 604800000\
	delegation.token.secret.key = null\
	delete.records.purgatory.purge.interval.requests = 1\
	delete.topic.enable = true\
	fetch.max.bytes = 57671680\
	fetch.purgatory.purge.interval.requests = 1000\
	group.initial.rebalance.delay.ms = 0\
	group.max.session.timeout.ms = 1800000\
	group.max.size = 2147483647\
	group.min.session.timeout.ms = 0\
	initial.broker.registration.timeout.ms = 60000\
	inter.broker.listener.name = null\
	inter.broker.protocol.version = 3.0-IV1\
	kafka.metrics.polling.interval.secs = 10\
	kafka.metrics.reporters = []\
	leader.imbalance.check.interval.seconds = 300\
	leader.imbalance.per.broker.percentage = 10\
	listener.security.protocol.map = PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL\
	listeners = PLAINTEXT://localhost:0\
	log.cleaner.backoff.ms = 15000\
	log.cleaner.dedupe.buffer.size = 2097152\
	log.cleaner.delete.retention.ms = 86400000\
	log.cleaner.enable = true\
	log.cleaner.io.buffer.load.factor = 0.9\
	log.cleaner.io.buffer.size = 524288\
	log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308\
	log.cleaner.max.compaction.lag.ms = 9223372036854775807\
	log.cleaner.min.cleanable.ratio = 0.5\
	log.cleaner.min.compaction.lag.ms = 0\
	log.cleaner.threads = 1\
	log.cleanup.policy = [delete]\
	log.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804\
	log.dirs = null\
	log.flush.interval.messages = 9223372036854775807\
	log.flush.interval.ms = null\
	log.flush.offset.checkpoint.interval.ms = 60000\
	log.flush.scheduler.interval.ms = 9223372036854775807\
	log.flush.start.offset.checkpoint.interval.ms = 60000\
	log.index.interval.bytes = 4096\
	log.index.size.max.bytes = 10485760\
	log.message.downconversion.enable = true\
	log.message.format.version = 3.0-IV1\
	log.message.timestamp.difference.max.ms = 9223372036854775807\
	log.message.timestamp.type = CreateTime\
	log.preallocate = false\
	log.retention.bytes = -1\
	log.retention.check.interval.ms = 300000\
	log.retention.hours = 168\
	log.retention.minutes = null\
	log.retention.ms = null\
	log.roll.hours = 168\
	log.roll.jitter.hours = 0\
	log.roll.jitter.ms = null\
	log.roll.ms = null\
	log.segment.bytes = 1073741824\
	log.segment.delete.delay.ms = 60000\
	max.connection.creation.rate = 2147483647\
	max.connections = 2147483647\
	max.connections.per.ip = 2147483647\
	max.connections.per.ip.overrides = \
	max.incremental.fetch.session.cache.slots = 1000\
	message.max.bytes = 1000000\
	metadata.log.dir = null\
	metadata.log.max.record.bytes.between.snapshots = 20971520\
	metadata.log.segment.bytes = 1073741824\
	metadata.log.segment.min.bytes = 8388608\
	metadata.log.segment.ms = 604800000\
	metadata.max.retention.bytes = -1\
	metadata.max.retention.ms = 604800000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	min.insync.replicas = 1\
	node.id = -1\
	num.io.threads = 8\
	num.network.threads = 3\
	num.partitions = 1\
	num.recovery.threads.per.data.dir = 1\
	num.replica.alter.log.dirs.threads = null\
	num.replica.fetchers = 1\
	offset.metadata.max.bytes = 4096\
	offsets.commit.required.acks = -1\
	offsets.commit.timeout.ms = 5000\
	offsets.load.buffer.size = 5242880\
	offsets.retention.check.interval.ms = 600000\
	offsets.retention.minutes = 10080\
	offsets.topic.compression.codec = 0\
	offsets.topic.num.partitions = 5\
	offsets.topic.replication.factor = 1\
	offsets.topic.segment.bytes = 104857600\
	password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding\
	password.encoder.iterations = 4096\
	password.encoder.key.length = 128\
	password.encoder.keyfactory.algorithm = null\
	password.encoder.old.secret = null\
	password.encoder.secret = null\
	principal.builder.class = class org.apache.kafka.common.security.authenticator.DefaultKafkaPrincipalBuilder\
	process.roles = []\
	producer.purgatory.purge.interval.requests = 1000\
	queued.max.request.bytes = -1\
	queued.max.requests = 500\
	quota.window.num = 11\
	quota.window.size.seconds = 1\
	remote.log.index.file.cache.total.size.bytes = 1073741824\
	remote.log.manager.task.interval.ms = 30000\
	remote.log.manager.task.retry.backoff.max.ms = 30000\
	remote.log.manager.task.retry.backoff.ms = 500\
	remote.log.manager.task.retry.jitter = 0.2\
	remote.log.manager.thread.pool.size = 10\
	remote.log.metadata.manager.class.name = null\
	remote.log.metadata.manager.class.path = null\
	remote.log.metadata.manager.impl.prefix = null\
	remote.log.metadata.manager.listener.name = null\
	remote.log.reader.max.pending.tasks = 100\
	remote.log.reader.threads = 10\
	remote.log.storage.manager.class.name = null\
	remote.log.storage.manager.class.path = null\
	remote.log.storage.manager.impl.prefix = null\
	remote.log.storage.system.enable = false\
	replica.fetch.backoff.ms = 1000\
	replica.fetch.max.bytes = 1048576\
	replica.fetch.min.bytes = 1\
	replica.fetch.response.max.bytes = 10485760\
	replica.fetch.wait.max.ms = 500\
	replica.high.watermark.checkpoint.interval.ms = 5000\
	replica.lag.time.max.ms = 30000\
	replica.selector.class = null\
	replica.socket.receive.buffer.bytes = 65536\
	replica.socket.timeout.ms = 30000\
	replication.quota.window.num = 11\
	replication.quota.window.size.seconds = 1\
	request.timeout.ms = 30000\
	reserved.broker.max.id = 1000\
	sasl.client.callback.handler.class = null\
	sasl.enabled.mechanisms = [GSSAPI]\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.principal.to.local.rules = [DEFAULT]\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism.controller.protocol = GSSAPI\
	sasl.mechanism.inter.broker.protocol = GSSAPI\
	sasl.server.callback.handler.class = null\
	security.inter.broker.protocol = PLAINTEXT\
	security.providers = null\
	socket.connection.setup.timeout.max.ms = 30000\
	socket.connection.setup.timeout.ms = 10000\
	socket.receive.buffer.bytes = 102400\
	socket.request.max.bytes = 104857600\
	socket.send.buffer.bytes = 102400\
	ssl.cipher.suites = []\
	ssl.client.auth = none\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.certificate.chain = null\
	ssl.keystore.key = null\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.principal.mapping.rules = DEFAULT\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.certificates = null\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000\
	transaction.max.timeout.ms = 900000\
	transaction.remove.expired.transaction.cleanup.interval.ms = 3600000\
	transaction.state.log.load.buffer.size = 5242880\
	transaction.state.log.min.isr = 2\
	transaction.state.log.num.partitions = 5\
	transaction.state.log.replication.factor = 3\
	transaction.state.log.segment.bytes = 104857600\
	transactional.id.expiration.ms = 604800000\
	unclean.leader.election.enable = false\
	zookeeper.clientCnxnSocket = null\
	zookeeper.connect = 127.0.0.1:57920\
	zookeeper.connection.timeout.ms = null\
	zookeeper.max.in.flight.requests = 10\
	zookeeper.session.timeout.ms = 10000\
	zookeeper.set.acl = false\
	zookeeper.ssl.cipher.suites = null\
	zookeeper.ssl.client.enable = false\
	zookeeper.ssl.crl.enable = false\
	zookeeper.ssl.enabled.protocols = null\
	zookeeper.ssl.endpoint.identification.algorithm = HTTPS\
	zookeeper.ssl.keystore.location = null\
	zookeeper.ssl.keystore.password = null\
	zookeeper.ssl.keystore.type = null\
	zookeeper.ssl.ocsp.enable = false\
	zookeeper.ssl.protocol = TLSv1.2\
	zookeeper.ssl.truststore.location = null\
	zookeeper.ssl.truststore.password = null\
	zookeeper.ssl.truststore.type = null\
	zookeeper.sync.time.ms = 2000\
 (kafka.server.KafkaConfig:376)\
[2021-07-14 16:05:15,487] INFO starting (kafka.server.KafkaServer:66)\
[2021-07-14 16:05:15,487] INFO Connecting to zookeeper on 127.0.0.1:57920 (kafka.server.KafkaServer:66)\
[2021-07-14 16:05:15,487] INFO [ZooKeeperClient Kafka server] Initializing a new session to 127.0.0.1:57920. (kafka.zookeeper.ZooKeeperClient:66)\
[2021-07-14 16:05:15,488] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient:66)\
[2021-07-14 16:05:15,491] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient:66)\
[2021-07-14 16:05:15,503] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread:66)\
[2021-07-14 16:05:15,504] INFO Feature ZK node at path: /feature does not exist (kafka.server.FinalizedFeatureChangeListener:66)\
[2021-07-14 16:05:15,505] INFO Cleared cache (kafka.server.FinalizedFeatureCache:66)\
[2021-07-14 16:05:15,506] INFO Cluster ID = EAsHEVW1QoKeh1kQiVxN1w (kafka.server.KafkaServer:66)\
[2021-07-14 16:05:15,506] WARN No meta.properties file under dir /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/meta.properties (kafka.server.BrokerMetadataCheckpoint:70)\
[2021-07-14 16:05:15,508] INFO KafkaConfig values: \
	advertised.listeners = null\
	alter.config.policy.class.name = null\
	alter.log.dirs.replication.quota.window.num = 11\
	alter.log.dirs.replication.quota.window.size.seconds = 1\
	authorizer.class.name = \
	auto.create.topics.enable = true\
	auto.leader.rebalance.enable = true\
	background.threads = 10\
	broker.heartbeat.interval.ms = 2000\
	broker.id = 0\
	broker.id.generation.enable = true\
	broker.rack = null\
	broker.session.timeout.ms = 9000\
	client.quota.callback.class = null\
	compression.type = producer\
	connection.failed.authentication.delay.ms = 100\
	connections.max.idle.ms = 600000\
	connections.max.reauth.ms = 0\
	control.plane.listener.name = null\
	controlled.shutdown.enable = true\
	controlled.shutdown.max.retries = 3\
	controlled.shutdown.retry.backoff.ms = 5000\
	controller.listener.names = null\
	controller.quorum.append.linger.ms = 25\
	controller.quorum.election.backoff.max.ms = 1000\
	controller.quorum.election.timeout.ms = 1000\
	controller.quorum.fetch.timeout.ms = 2000\
	controller.quorum.request.timeout.ms = 2000\
	controller.quorum.retry.backoff.ms = 20\
	controller.quorum.voters = []\
	controller.quota.window.num = 11\
	controller.quota.window.size.seconds = 1\
	controller.socket.timeout.ms = 30000\
	create.topic.policy.class.name = null\
	default.replication.factor = 1\
	delegation.token.expiry.check.interval.ms = 3600000\
	delegation.token.expiry.time.ms = 86400000\
	delegation.token.master.key = null\
	delegation.token.max.lifetime.ms = 604800000\
	delegation.token.secret.key = null\
	delete.records.purgatory.purge.interval.requests = 1\
	delete.topic.enable = true\
	fetch.max.bytes = 57671680\
	fetch.purgatory.purge.interval.requests = 1000\
	group.initial.rebalance.delay.ms = 0\
	group.max.session.timeout.ms = 1800000\
	group.max.size = 2147483647\
	group.min.session.timeout.ms = 0\
	initial.broker.registration.timeout.ms = 60000\
	inter.broker.listener.name = null\
	inter.broker.protocol.version = 3.0-IV1\
	kafka.metrics.polling.interval.secs = 10\
	kafka.metrics.reporters = []\
	leader.imbalance.check.interval.seconds = 300\
	leader.imbalance.per.broker.percentage = 10\
	listener.security.protocol.map = PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL\
	listeners = PLAINTEXT://localhost:0\
	log.cleaner.backoff.ms = 15000\
	log.cleaner.dedupe.buffer.size = 2097152\
	log.cleaner.delete.retention.ms = 86400000\
	log.cleaner.enable = true\
	log.cleaner.io.buffer.load.factor = 0.9\
	log.cleaner.io.buffer.size = 524288\
	log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308\
	log.cleaner.max.compaction.lag.ms = 9223372036854775807\
	log.cleaner.min.cleanable.ratio = 0.5\
	log.cleaner.min.compaction.lag.ms = 0\
	log.cleaner.threads = 1\
	log.cleanup.policy = [delete]\
	log.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804\
	log.dirs = null\
	log.flush.interval.messages = 9223372036854775807\
	log.flush.interval.ms = null\
	log.flush.offset.checkpoint.interval.ms = 60000\
	log.flush.scheduler.interval.ms = 9223372036854775807\
	log.flush.start.offset.checkpoint.interval.ms = 60000\
	log.index.interval.bytes = 4096\
	log.index.size.max.bytes = 10485760\
	log.message.downconversion.enable = true\
	log.message.format.version = 3.0-IV1\
	log.message.timestamp.difference.max.ms = 9223372036854775807\
	log.message.timestamp.type = CreateTime\
	log.preallocate = false\
	log.retention.bytes = -1\
	log.retention.check.interval.ms = 300000\
	log.retention.hours = 168\
	log.retention.minutes = null\
	log.retention.ms = null\
	log.roll.hours = 168\
	log.roll.jitter.hours = 0\
	log.roll.jitter.ms = null\
	log.roll.ms = null\
	log.segment.bytes = 1073741824\
	log.segment.delete.delay.ms = 60000\
	max.connection.creation.rate = 2147483647\
	max.connections = 2147483647\
	max.connections.per.ip = 2147483647\
	max.connections.per.ip.overrides = \
	max.incremental.fetch.session.cache.slots = 1000\
	message.max.bytes = 1000000\
	metadata.log.dir = null\
	metadata.log.max.record.bytes.between.snapshots = 20971520\
	metadata.log.segment.bytes = 1073741824\
	metadata.log.segment.min.bytes = 8388608\
	metadata.log.segment.ms = 604800000\
	metadata.max.retention.bytes = -1\
	metadata.max.retention.ms = 604800000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	min.insync.replicas = 1\
	node.id = -1\
	num.io.threads = 8\
	num.network.threads = 3\
	num.partitions = 1\
	num.recovery.threads.per.data.dir = 1\
	num.replica.alter.log.dirs.threads = null\
	num.replica.fetchers = 1\
	offset.metadata.max.bytes = 4096\
	offsets.commit.required.acks = -1\
	offsets.commit.timeout.ms = 5000\
	offsets.load.buffer.size = 5242880\
	offsets.retention.check.interval.ms = 600000\
	offsets.retention.minutes = 10080\
	offsets.topic.compression.codec = 0\
	offsets.topic.num.partitions = 5\
	offsets.topic.replication.factor = 1\
	offsets.topic.segment.bytes = 104857600\
	password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding\
	password.encoder.iterations = 4096\
	password.encoder.key.length = 128\
	password.encoder.keyfactory.algorithm = null\
	password.encoder.old.secret = null\
	password.encoder.secret = null\
	principal.builder.class = class org.apache.kafka.common.security.authenticator.DefaultKafkaPrincipalBuilder\
	process.roles = []\
	producer.purgatory.purge.interval.requests = 1000\
	queued.max.request.bytes = -1\
	queued.max.requests = 500\
	quota.window.num = 11\
	quota.window.size.seconds = 1\
	remote.log.index.file.cache.total.size.bytes = 1073741824\
	remote.log.manager.task.interval.ms = 30000\
	remote.log.manager.task.retry.backoff.max.ms = 30000\
	remote.log.manager.task.retry.backoff.ms = 500\
	remote.log.manager.task.retry.jitter = 0.2\
	remote.log.manager.thread.pool.size = 10\
	remote.log.metadata.manager.class.name = null\
	remote.log.metadata.manager.class.path = null\
	remote.log.metadata.manager.impl.prefix = null\
	remote.log.metadata.manager.listener.name = null\
	remote.log.reader.max.pending.tasks = 100\
	remote.log.reader.threads = 10\
	remote.log.storage.manager.class.name = null\
	remote.log.storage.manager.class.path = null\
	remote.log.storage.manager.impl.prefix = null\
	remote.log.storage.system.enable = false\
	replica.fetch.backoff.ms = 1000\
	replica.fetch.max.bytes = 1048576\
	replica.fetch.min.bytes = 1\
	replica.fetch.response.max.bytes = 10485760\
	replica.fetch.wait.max.ms = 500\
	replica.high.watermark.checkpoint.interval.ms = 5000\
	replica.lag.time.max.ms = 30000\
	replica.selector.class = null\
	replica.socket.receive.buffer.bytes = 65536\
	replica.socket.timeout.ms = 30000\
	replication.quota.window.num = 11\
	replication.quota.window.size.seconds = 1\
	request.timeout.ms = 30000\
	reserved.broker.max.id = 1000\
	sasl.client.callback.handler.class = null\
	sasl.enabled.mechanisms = [GSSAPI]\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.principal.to.local.rules = [DEFAULT]\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism.controller.protocol = GSSAPI\
	sasl.mechanism.inter.broker.protocol = GSSAPI\
	sasl.server.callback.handler.class = null\
	security.inter.broker.protocol = PLAINTEXT\
	security.providers = null\
	socket.connection.setup.timeout.max.ms = 30000\
	socket.connection.setup.timeout.ms = 10000\
	socket.receive.buffer.bytes = 102400\
	socket.request.max.bytes = 104857600\
	socket.send.buffer.bytes = 102400\
	ssl.cipher.suites = []\
	ssl.client.auth = none\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.certificate.chain = null\
	ssl.keystore.key = null\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.principal.mapping.rules = DEFAULT\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.certificates = null\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000\
	transaction.max.timeout.ms = 900000\
	transaction.remove.expired.transaction.cleanup.interval.ms = 3600000\
	transaction.state.log.load.buffer.size = 5242880\
	transaction.state.log.min.isr = 2\
	transaction.state.log.num.partitions = 5\
	transaction.state.log.replication.factor = 3\
	transaction.state.log.segment.bytes = 104857600\
	transactional.id.expiration.ms = 604800000\
	unclean.leader.election.enable = false\
	zookeeper.clientCnxnSocket = null\
	zookeeper.connect = 127.0.0.1:57920\
	zookeeper.connection.timeout.ms = null\
	zookeeper.max.in.flight.requests = 10\
	zookeeper.session.timeout.ms = 10000\
	zookeeper.set.acl = false\
	zookeeper.ssl.cipher.suites = null\
	zookeeper.ssl.client.enable = false\
	zookeeper.ssl.crl.enable = false\
	zookeeper.ssl.enabled.protocols = null\
	zookeeper.ssl.endpoint.identification.algorithm = HTTPS\
	zookeeper.ssl.keystore.location = null\
	zookeeper.ssl.keystore.password = null\
	zookeeper.ssl.keystore.type = null\
	zookeeper.ssl.ocsp.enable = false\
	zookeeper.ssl.protocol = TLSv1.2\
	zookeeper.ssl.truststore.location = null\
	zookeeper.ssl.truststore.password = null\
	zookeeper.ssl.truststore.type = null\
	zookeeper.sync.time.ms = 2000\
 (kafka.server.KafkaConfig:376)\
[2021-07-14 16:05:15,511] INFO KafkaConfig values: \
	advertised.listeners = null\
	alter.config.policy.class.name = null\
	alter.log.dirs.replication.quota.window.num = 11\
	alter.log.dirs.replication.quota.window.size.seconds = 1\
	authorizer.class.name = \
	auto.create.topics.enable = true\
	auto.leader.rebalance.enable = true\
	background.threads = 10\
	broker.heartbeat.interval.ms = 2000\
	broker.id = 0\
	broker.id.generation.enable = true\
	broker.rack = null\
	broker.session.timeout.ms = 9000\
	client.quota.callback.class = null\
	compression.type = producer\
	connection.failed.authentication.delay.ms = 100\
	connections.max.idle.ms = 600000\
	connections.max.reauth.ms = 0\
	control.plane.listener.name = null\
	controlled.shutdown.enable = true\
	controlled.shutdown.max.retries = 3\
	controlled.shutdown.retry.backoff.ms = 5000\
	controller.listener.names = null\
	controller.quorum.append.linger.ms = 25\
	controller.quorum.election.backoff.max.ms = 1000\
	controller.quorum.election.timeout.ms = 1000\
	controller.quorum.fetch.timeout.ms = 2000\
	controller.quorum.request.timeout.ms = 2000\
	controller.quorum.retry.backoff.ms = 20\
	controller.quorum.voters = []\
	controller.quota.window.num = 11\
	controller.quota.window.size.seconds = 1\
	controller.socket.timeout.ms = 30000\
	create.topic.policy.class.name = null\
	default.replication.factor = 1\
	delegation.token.expiry.check.interval.ms = 3600000\
	delegation.token.expiry.time.ms = 86400000\
	delegation.token.master.key = null\
	delegation.token.max.lifetime.ms = 604800000\
	delegation.token.secret.key = null\
	delete.records.purgatory.purge.interval.requests = 1\
	delete.topic.enable = true\
	fetch.max.bytes = 57671680\
	fetch.purgatory.purge.interval.requests = 1000\
	group.initial.rebalance.delay.ms = 0\
	group.max.session.timeout.ms = 1800000\
	group.max.size = 2147483647\
	group.min.session.timeout.ms = 0\
	initial.broker.registration.timeout.ms = 60000\
	inter.broker.listener.name = null\
	inter.broker.protocol.version = 3.0-IV1\
	kafka.metrics.polling.interval.secs = 10\
	kafka.metrics.reporters = []\
	leader.imbalance.check.interval.seconds = 300\
	leader.imbalance.per.broker.percentage = 10\
	listener.security.protocol.map = PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL\
	listeners = PLAINTEXT://localhost:0\
	log.cleaner.backoff.ms = 15000\
	log.cleaner.dedupe.buffer.size = 2097152\
	log.cleaner.delete.retention.ms = 86400000\
	log.cleaner.enable = true\
	log.cleaner.io.buffer.load.factor = 0.9\
	log.cleaner.io.buffer.size = 524288\
	log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308\
	log.cleaner.max.compaction.lag.ms = 9223372036854775807\
	log.cleaner.min.cleanable.ratio = 0.5\
	log.cleaner.min.compaction.lag.ms = 0\
	log.cleaner.threads = 1\
	log.cleanup.policy = [delete]\
	log.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804\
	log.dirs = null\
	log.flush.interval.messages = 9223372036854775807\
	log.flush.interval.ms = null\
	log.flush.offset.checkpoint.interval.ms = 60000\
	log.flush.scheduler.interval.ms = 9223372036854775807\
	log.flush.start.offset.checkpoint.interval.ms = 60000\
	log.index.interval.bytes = 4096\
	log.index.size.max.bytes = 10485760\
	log.message.downconversion.enable = true\
	log.message.format.version = 3.0-IV1\
	log.message.timestamp.difference.max.ms = 9223372036854775807\
	log.message.timestamp.type = CreateTime\
	log.preallocate = false\
	log.retention.bytes = -1\
	log.retention.check.interval.ms = 300000\
	log.retention.hours = 168\
	log.retention.minutes = null\
	log.retention.ms = null\
	log.roll.hours = 168\
	log.roll.jitter.hours = 0\
	log.roll.jitter.ms = null\
	log.roll.ms = null\
	log.segment.bytes = 1073741824\
	log.segment.delete.delay.ms = 60000\
	max.connection.creation.rate = 2147483647\
	max.connections = 2147483647\
	max.connections.per.ip = 2147483647\
	max.connections.per.ip.overrides = \
	max.incremental.fetch.session.cache.slots = 1000\
	message.max.bytes = 1000000\
	metadata.log.dir = null\
	metadata.log.max.record.bytes.between.snapshots = 20971520\
	metadata.log.segment.bytes = 1073741824\
	metadata.log.segment.min.bytes = 8388608\
	metadata.log.segment.ms = 604800000\
	metadata.max.retention.bytes = -1\
	metadata.max.retention.ms = 604800000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	min.insync.replicas = 1\
	node.id = -1\
	num.io.threads = 8\
	num.network.threads = 3\
	num.partitions = 1\
	num.recovery.threads.per.data.dir = 1\
	num.replica.alter.log.dirs.threads = null\
	num.replica.fetchers = 1\
	offset.metadata.max.bytes = 4096\
	offsets.commit.required.acks = -1\
	offsets.commit.timeout.ms = 5000\
	offsets.load.buffer.size = 5242880\
	offsets.retention.check.interval.ms = 600000\
	offsets.retention.minutes = 10080\
	offsets.topic.compression.codec = 0\
	offsets.topic.num.partitions = 5\
	offsets.topic.replication.factor = 1\
	offsets.topic.segment.bytes = 104857600\
	password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding\
	password.encoder.iterations = 4096\
	password.encoder.key.length = 128\
	password.encoder.keyfactory.algorithm = null\
	password.encoder.old.secret = null\
	password.encoder.secret = null\
	principal.builder.class = class org.apache.kafka.common.security.authenticator.DefaultKafkaPrincipalBuilder\
	process.roles = []\
	producer.purgatory.purge.interval.requests = 1000\
	queued.max.request.bytes = -1\
	queued.max.requests = 500\
	quota.window.num = 11\
	quota.window.size.seconds = 1\
	remote.log.index.file.cache.total.size.bytes = 1073741824\
	remote.log.manager.task.interval.ms = 30000\
	remote.log.manager.task.retry.backoff.max.ms = 30000\
	remote.log.manager.task.retry.backoff.ms = 500\
	remote.log.manager.task.retry.jitter = 0.2\
	remote.log.manager.thread.pool.size = 10\
	remote.log.metadata.manager.class.name = null\
	remote.log.metadata.manager.class.path = null\
	remote.log.metadata.manager.impl.prefix = null\
	remote.log.metadata.manager.listener.name = null\
	remote.log.reader.max.pending.tasks = 100\
	remote.log.reader.threads = 10\
	remote.log.storage.manager.class.name = null\
	remote.log.storage.manager.class.path = null\
	remote.log.storage.manager.impl.prefix = null\
	remote.log.storage.system.enable = false\
	replica.fetch.backoff.ms = 1000\
	replica.fetch.max.bytes = 1048576\
	replica.fetch.min.bytes = 1\
	replica.fetch.response.max.bytes = 10485760\
	replica.fetch.wait.max.ms = 500\
	replica.high.watermark.checkpoint.interval.ms = 5000\
	replica.lag.time.max.ms = 30000\
	replica.selector.class = null\
	replica.socket.receive.buffer.bytes = 65536\
	replica.socket.timeout.ms = 30000\
	replication.quota.window.num = 11\
	replication.quota.window.size.seconds = 1\
	request.timeout.ms = 30000\
	reserved.broker.max.id = 1000\
	sasl.client.callback.handler.class = null\
	sasl.enabled.mechanisms = [GSSAPI]\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.principal.to.local.rules = [DEFAULT]\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism.controller.protocol = GSSAPI\
	sasl.mechanism.inter.broker.protocol = GSSAPI\
	sasl.server.callback.handler.class = null\
	security.inter.broker.protocol = PLAINTEXT\
	security.providers = null\
	socket.connection.setup.timeout.max.ms = 30000\
	socket.connection.setup.timeout.ms = 10000\
	socket.receive.buffer.bytes = 102400\
	socket.request.max.bytes = 104857600\
	socket.send.buffer.bytes = 102400\
	ssl.cipher.suites = []\
	ssl.client.auth = none\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.certificate.chain = null\
	ssl.keystore.key = null\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.principal.mapping.rules = DEFAULT\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.certificates = null\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000\
	transaction.max.timeout.ms = 900000\
	transaction.remove.expired.transaction.cleanup.interval.ms = 3600000\
	transaction.state.log.load.buffer.size = 5242880\
	transaction.state.log.min.isr = 2\
	transaction.state.log.num.partitions = 5\
	transaction.state.log.replication.factor = 3\
	transaction.state.log.segment.bytes = 104857600\
	transactional.id.expiration.ms = 604800000\
	unclean.leader.election.enable = false\
	zookeeper.clientCnxnSocket = null\
	zookeeper.connect = 127.0.0.1:57920\
	zookeeper.connection.timeout.ms = null\
	zookeeper.max.in.flight.requests = 10\
	zookeeper.session.timeout.ms = 10000\
	zookeeper.set.acl = false\
	zookeeper.ssl.cipher.suites = null\
	zookeeper.ssl.client.enable = false\
	zookeeper.ssl.crl.enable = false\
	zookeeper.ssl.enabled.protocols = null\
	zookeeper.ssl.endpoint.identification.algorithm = HTTPS\
	zookeeper.ssl.keystore.location = null\
	zookeeper.ssl.keystore.password = null\
	zookeeper.ssl.keystore.type = null\
	zookeeper.ssl.ocsp.enable = false\
	zookeeper.ssl.protocol = TLSv1.2\
	zookeeper.ssl.truststore.location = null\
	zookeeper.ssl.truststore.password = null\
	zookeeper.ssl.truststore.type = null\
	zookeeper.sync.time.ms = 2000\
 (kafka.server.KafkaConfig:376)\
[2021-07-14 16:05:15,524] INFO [ThrottledChannelReaper-Fetch]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:15,524] INFO [ThrottledChannelReaper-Produce]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:15,524] INFO [ThrottledChannelReaper-Request]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:15,525] INFO [ThrottledChannelReaper-ControllerMutation]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:15,527] INFO Loading logs from log dirs ArraySeq(/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804) (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,527] INFO Attempting recovery for all logs in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804 since no clean shutdown file was found (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,528] INFO Loaded 0 logs in 0ms. (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,528] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,528] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,529] INFO Starting the log cleaner (kafka.log.LogCleaner:66)\
[2021-07-14 16:05:15,533] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner:66)\
[2021-07-14 16:05:15,534] INFO [BrokerToControllerChannelManager broker=0 name=forwarding]: Starting (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:15,549] WARN [AdminClient clientId=adminclient-67] Connection to node 0 (localhost/127.0.0.1:57891) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:766)\
[2021-07-14 16:05:15,580] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas:66)\
[2021-07-14 16:05:15,580] INFO Awaiting socket connections on localhost:57923. (kafka.network.Acceptor:66)\
[2021-07-14 16:05:15,586] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer:66)\
[2021-07-14 16:05:15,588] INFO [BrokerToControllerChannelManager broker=0 name=alterIsr]: Starting (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:15,589] INFO [ExpirationReaper-0-Produce]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:15,591] INFO [ExpirationReaper-0-Fetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:15,593] INFO [ExpirationReaper-0-DeleteRecords]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:15,593] INFO [ExpirationReaper-0-ElectLeader]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:15,594] INFO [LogDirFailureHandler]: Starting (kafka.server.ReplicaManager$LogDirFailureHandler:66)\
[2021-07-14 16:05:15,595] INFO Creating /brokers/ids/0 (is it secure? false) (kafka.zk.KafkaZkClient:66)\
[2021-07-14 16:05:15,595] INFO Stat of the created znode at /brokers/ids/0 is: 25,25,1626303915595,1626303915595,1,0,0,72179899213283328,204,0,25\
 (kafka.zk.KafkaZkClient:66)\
[2021-07-14 16:05:15,595] INFO Registered broker 0 at path /brokers/ids/0 with addresses: PLAINTEXT://localhost:57923, czxid (broker epoch): 25 (kafka.zk.KafkaZkClient:66)\
[2021-07-14 16:05:15,598] INFO [ControllerEventThread controllerId=0] Starting (kafka.controller.ControllerEventManager$ControllerEventThread:66)\
[2021-07-14 16:05:15,598] INFO [ExpirationReaper-0-topic]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:15,599] INFO [ExpirationReaper-0-Heartbeat]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:15,599] INFO [ExpirationReaper-0-Rebalance]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:15,599] INFO [GroupCoordinator 0]: Starting up. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,600] INFO Successfully created /controller_epoch with initial epoch 0 (kafka.zk.KafkaZkClient:66)\
[2021-07-14 16:05:15,600] INFO [GroupCoordinator 0]: Startup complete. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,600] INFO [Controller id=0] 0 successfully elected as the controller. Epoch incremented to 1 and epoch zk version is now 1 (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,601] INFO [TransactionCoordinator id=0] Starting up. (kafka.coordinator.transaction.TransactionCoordinator:66)\
[2021-07-14 16:05:15,601] INFO [Controller id=0] Creating FeatureZNode at path: /feature with contents: FeatureZNode(Enabled,Features\{\}) (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,601] INFO [TransactionCoordinator id=0] Startup complete. (kafka.coordinator.transaction.TransactionCoordinator:66)\
[2021-07-14 16:05:15,603] INFO [Transaction Marker Channel Manager 0]: Starting (kafka.coordinator.transaction.TransactionMarkerChannelManager:66)\
[2021-07-14 16:05:15,603] INFO Feature ZK node created at path: /feature (kafka.server.FinalizedFeatureChangeListener:66)\
[2021-07-14 16:05:15,604] INFO [ExpirationReaper-0-AlterAcls]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:15,604] INFO Updated cache from existing <empty> to latest FinalizedFeaturesAndEpoch(features=Features\{\}, epoch=0). (kafka.server.FinalizedFeatureCache:66)\
[2021-07-14 16:05:15,604] INFO [Controller id=0] Registering handlers (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,605] INFO [/config/changes-event-process-thread]: Starting (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread:66)\
[2021-07-14 16:05:15,605] INFO [Controller id=0] Deleting log dir event notifications (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,606] INFO [Controller id=0] Deleting isr change notifications (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,606] INFO [Controller id=0] Initializing controller context (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,607] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Starting socket server acceptors and processors (kafka.network.SocketServer:66)\
[2021-07-14 16:05:15,607] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Started data-plane acceptor and processor(s) for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer:66)\
[2021-07-14 16:05:15,608] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Started socket server acceptors and processors (kafka.network.SocketServer:66)\
[2021-07-14 16:05:15,608] INFO [Controller id=0] Initialized broker epochs cache: HashMap(0 -> 25) (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,608] INFO [KafkaServer id=0] started (kafka.server.KafkaServer:66)\
[2021-07-14 16:05:15,609] INFO [Controller id=0] Currently active brokers in the cluster: Set(0) (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,609] INFO [RequestSendThread controllerId=0] Starting (kafka.controller.RequestSendThread:66)\
[2021-07-14 16:05:15,609] INFO [Controller id=0] Currently shutting brokers in the cluster: HashSet() (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,609] INFO [Controller id=0] Current list of topics in the cluster: HashSet() (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,609] INFO [Controller id=0] Fetching topic deletions in progress (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,610] INFO [Controller id=0] List of topics to be deleted:  (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,610] INFO [Controller id=0] List of topics ineligible for deletion:  (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,610] INFO [Controller id=0] Initializing topic deletion manager (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,610] INFO [Topic Deletion Manager 0] Initializing manager with initial deletions: Set(), initial ineligible deletions: HashSet() (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:15,610] INFO [Controller id=0] Sending update metadata request (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,610] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,610] INFO [ReplicaStateMachine controllerId=0] Initializing replica state (kafka.controller.ZkReplicaStateMachine:66)\
[2021-07-14 16:05:15,610] INFO [ReplicaStateMachine controllerId=0] Triggering online replica state changes (kafka.controller.ZkReplicaStateMachine:66)\
[2021-07-14 16:05:15,610] INFO [ReplicaStateMachine controllerId=0] Triggering offline replica state changes (kafka.controller.ZkReplicaStateMachine:66)\
[2021-07-14 16:05:15,610] INFO [PartitionStateMachine controllerId=0] Initializing partition state (kafka.controller.ZkPartitionStateMachine:66)\
[2021-07-14 16:05:15,610] INFO [PartitionStateMachine controllerId=0] Triggering online partition state changes (kafka.controller.ZkPartitionStateMachine:66)\
[2021-07-14 16:05:15,610] INFO [Controller id=0] Ready to serve as the new controller with epoch 1 (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,611] INFO [RequestSendThread controllerId=0] Controller 0 connected to localhost:57923 (id: 0 rack: null) for sending state change requests (kafka.controller.RequestSendThread:66)\
[2021-07-14 16:05:15,613] INFO [Controller id=0] Partitions undergoing preferred replica election:  (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,613] INFO [Controller id=0] Partitions that completed preferred replica election:  (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,613] INFO [Controller id=0] Skipping preferred replica election for partitions due to topic deletion:  (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,613] INFO [Controller id=0] Resuming preferred replica election for partitions:  (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,613] INFO [Controller id=0] Starting replica leader election (PREFERRED) for partitions  triggered by ZkTriggered (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,613] INFO [Controller id=0] Starting the controller scheduler (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,615] INFO Creating topic inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation with configuration \{\} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)\
[2021-07-14 16:05:15,618] INFO [Controller id=0] New topics: [Set(inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation,Some(c7eNWcrbTMGx0Nj6B_Hj5g),Map(inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,618] INFO [Controller id=0] New partition creation callback for inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,618] INFO [Controller id=0 epoch=1] Changed partition inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:15,618] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,618] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,620] INFO [Controller id=0 epoch=1] Changed partition inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:15,620] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)\
[2021-07-14 16:05:15,621] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,621] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,621] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 1 from controller 0 for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,621] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0) (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:15,621] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 1 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,626] INFO [LogLoader partition=inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:15,631] INFO Created log for partition inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 with properties \{\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,631] INFO [Partition inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 broker=0] No checkpointed highwatermark is found for partition inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,631] INFO [Partition inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 broker=0] Log loaded for partition inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,632] INFO [Broker id=0] Leader inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:15,634] INFO [Broker id=0] Finished LeaderAndIsr request in 0ms correlationId 1 from controller 0 for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,635] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 2 (state.change.logger:66)\
[2021-07-14 16:05:15,637] INFO [BrokerToControllerChannelManager broker=0 name=forwarding]: Recorded new controller, from now on will use broker localhost:57923 (id: 0 rack: null) (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:15,639] INFO StreamsConfig values: \
	acceptable.recovery.lag = 10000\
	application.id = TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation\
	application.server = \
	bootstrap.servers = [localhost:57923]\
	buffered.records.per.partition = 1000\
	built.in.metrics.version = latest\
	cache.max.bytes.buffering = 10485760\
	client.id = \
	commit.interval.ms = 1\
	connections.max.idle.ms = 540000\
	default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler\
	default.key.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde\
	default.list.key.serde.inner = null\
	default.list.key.serde.type = null\
	default.list.value.serde.inner = null\
	default.list.value.serde.type = null\
	default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler\
	default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp\
	default.value.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde\
	max.task.idle.ms = 0\
	max.warmup.replicas = 2\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	num.standby.replicas = 0\
	num.stream.threads = 2\
	poll.ms = 100\
	probing.rebalance.interval.ms = 600000\
	processing.guarantee = at_least_once\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	replication.factor = -1\
	request.timeout.ms = 40000\
	retries = 0\
	retry.backoff.ms = 100\
	rocksdb.config.setter = null\
	security.protocol = PLAINTEXT\
	send.buffer.bytes = 131072\
	state.cleanup.delay.ms = 600000\
	state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-632129218822105537\
	task.timeout.ms = 300000\
	topology.optimization = none\
	upgrade.from = null\
	window.size.ms = null\
	windowed.inner.class.serde = null\
	windowstore.changelog.additional.retention.ms = 86400000\
 (org.apache.kafka.streams.StreamsConfig:376)\
[2021-07-14 16:05:15,641] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:849)\
[2021-07-14 16:05:15,641] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:850)\
[2021-07-14 16:05:15,642] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:334)\
[2021-07-14 16:05:15,645] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)\
[2021-07-14 16:05:15,647] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:386)\
[2021-07-14 16:05:15,649] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:114)\
[2021-07-14 16:05:15,651] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:334)\
[2021-07-14 16:05:15,655] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)\
[2021-07-14 16:05:15,657] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:386)\
[2021-07-14 16:05:15,660] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:114)\
[2021-07-14 16:05:15,666] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:315)\
[2021-07-14 16:05:15,667] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:548)\
[2021-07-14 16:05:15,667] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Starting (org.apache.kafka.streams.processor.internals.StreamThread:548)\
[2021-07-14 16:05:15,667] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:15,667] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:15,671] INFO Creating topic __consumer_offsets with configuration \{segment.bytes=104857600, cleanup.policy=compact, compression.type=producer\} and initial partition assignment HashMap(0 -> ArrayBuffer(0), 1 -> ArrayBuffer(0), 2 -> ArrayBuffer(0), 3 -> ArrayBuffer(0), 4 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)\
[2021-07-14 16:05:15,675] INFO [Controller id=0] New topics: [Set(__consumer_offsets)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(__consumer_offsets,Some(m1p6T8M-Q9mf6oCvxr5uUA),HashMap(__consumer_offsets-4 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-3 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-2 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-1 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,676] INFO [Controller id=0] New partition creation callback for __consumer_offsets-4,__consumer_offsets-3,__consumer_offsets-2,__consumer_offsets-0,__consumer_offsets-1 (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:15,676] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:15,676] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:15,676] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:15,676] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:15,676] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:15,676] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,676] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,678] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:15,679] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:15,679] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:15,679] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:15,679] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:15,679] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 5 become-leader and 0 become-follower partitions (state.change.logger:66)\
[2021-07-14 16:05:15,679] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,679] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,679] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 3 from controller 0 for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,680] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions HashSet(__consumer_offsets-4, __consumer_offsets-3, __consumer_offsets-2, __consumer_offsets-0, __consumer_offsets-1) (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:15,680] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 3 from controller 0 epoch 1 as part of the become-leader transition for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,683] INFO [LogLoader partition=__consumer_offsets-3, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:15,685] INFO Created log for partition __consumer_offsets-3 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-3 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,685] INFO [Partition __consumer_offsets-3 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-3 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,685] INFO [Partition __consumer_offsets-3 broker=0] Log loaded for partition __consumer_offsets-3 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,685] INFO [Broker id=0] Leader __consumer_offsets-3 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:15,689] INFO [BrokerToControllerChannelManager broker=0 name=alterIsr]: Recorded new controller, from now on will use broker localhost:57923 (id: 0 rack: null) (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:15,689] INFO [LogLoader partition=__consumer_offsets-2, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:15,691] INFO Created log for partition __consumer_offsets-2 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-2 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,691] INFO [Partition __consumer_offsets-2 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-2 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,691] INFO [Partition __consumer_offsets-2 broker=0] Log loaded for partition __consumer_offsets-2 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,691] INFO [Broker id=0] Leader __consumer_offsets-2 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:15,695] INFO [LogLoader partition=__consumer_offsets-4, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:15,697] INFO Created log for partition __consumer_offsets-4 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-4 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,697] INFO [Partition __consumer_offsets-4 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-4 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,698] INFO [Partition __consumer_offsets-4 broker=0] Log loaded for partition __consumer_offsets-4 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,698] INFO [Broker id=0] Leader __consumer_offsets-4 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:15,701] INFO [LogLoader partition=__consumer_offsets-1, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:15,703] INFO Created log for partition __consumer_offsets-1 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-1 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,703] INFO [Partition __consumer_offsets-1 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-1 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,703] INFO [Partition __consumer_offsets-1 broker=0] Log loaded for partition __consumer_offsets-1 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,703] INFO [Broker id=0] Leader __consumer_offsets-1 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:15,707] INFO [LogLoader partition=__consumer_offsets-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:15,709] INFO Created log for partition __consumer_offsets-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-0 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:15,709] INFO [Partition __consumer_offsets-0 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,709] INFO [Partition __consumer_offsets-0 broker=0] Log loaded for partition __consumer_offsets-0 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:15,709] INFO [Broker id=0] Leader __consumer_offsets-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:15,711] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 3 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,711] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-3 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,711] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 2 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,711] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-2 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,712] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-3 in 1 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,712] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 4 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,712] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-2 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,712] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-4 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,712] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 1 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,712] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-1 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,712] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-4 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,712] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 0 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,712] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-1 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,712] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-0 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,712] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-0 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:15,712] INFO [Broker id=0] Finished LeaderAndIsr request in 0ms correlationId 3 from controller 0 for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:15,713] INFO [Broker id=0] Add 5 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 4 (state.change.logger:66)\
[2021-07-14 16:05:15,771] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Processed 0 total records, ran 0 punctuators, and committed 0 total tasks since the last update (org.apache.kafka.streams.processor.internals.StreamThread:827)\
[2021-07-14 16:05:15,772] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Processed 0 total records, ran 0 punctuators, and committed 0 total tasks since the last update (org.apache.kafka.streams.processor.internals.StreamThread:827)\
[2021-07-14 16:05:15,776] INFO [GroupCoordinator 0]: Dynamic member with unknown member id joins group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation in Empty state. Created a new member id TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192 and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,777] INFO [GroupCoordinator 0]: Preparing to rebalance group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation in state PreparingRebalance with old generation 0 (__consumer_offsets-2) (reason: Adding new member TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192 with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,778] INFO [GroupCoordinator 0]: Stabilized group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation generation 1 (__consumer_offsets-2) with 1 members (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,779] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] All members participating in this rebalance: \
35f03575-f7d8-4de4-851d-92c048a6204d: [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:581)\
[2021-07-14 16:05:15,779] INFO Decided on assignment: \{35f03575-f7d8-4de4-851d-92c048a6204d=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]\} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)\
[2021-07-14 16:05:15,779] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Assigned tasks [0_0] including stateful [] to clients as: \
35f03575-f7d8-4de4-851d-92c048a6204d=[activeTasks: ([0_0]) standbyTasks: ([])]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:599)\
[2021-07-14 16:05:15,779] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Client 35f03575-f7d8-4de4-851d-92c048a6204d per-consumer assignment:\
	prev owned active \{\}\
	prev owned standby \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192=[]\}\
	assigned active \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192=[0_0]\}\
	revoking active \{\}\
	assigned standby \{\}\
 (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:808)\
[2021-07-14 16:05:15,779] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:827)\
[2021-07-14 16:05:15,779] INFO [GroupCoordinator 0]: Dynamic member with unknown member id joins group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation in CompletingRebalance state. Created a new member id TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,780] INFO [GroupCoordinator 0]: Assignment received from leader TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192 for group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,781] INFO [GroupCoordinator 0]: Preparing to rebalance group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation in state PreparingRebalance with old generation 1 (__consumer_offsets-2) (reason: Adding new member TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:15,781] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1312)\
[2021-07-14 16:05:15,781] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Handle new assignment with:\
	New active tasks: [0_0]\
	New standby tasks: []\
	Existing active tasks: []\
	Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:263)\
[2021-07-14 16:05:15,782] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:15,871] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] task [0_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:240)\
[2021-07-14 16:05:15,873] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] task [0_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:265)\
[2021-07-14 16:05:15,873] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Restoration took 90 ms for all tasks [0_0] (org.apache.kafka.streams.processor.internals.StreamThread:851)\
[2021-07-14 16:05:15,874] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:16,470] WARN [AdminClient clientId=adminclient-67] Connection to node 0 (localhost/127.0.0.1:57891) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:766)\
[2021-07-14 16:05:17,704] WARN [AdminClient clientId=adminclient-67] Connection to node 0 (localhost/127.0.0.1:57891) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:766)\
[2021-07-14 16:05:18,781] INFO [GroupCoordinator 0]: Stabilized group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation generation 2 (__consumer_offsets-2) with 2 members (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:18,782] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] All members participating in this rebalance: \
35f03575-f7d8-4de4-851d-92c048a6204d: [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb, TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:581)\
[2021-07-14 16:05:18,782] INFO Decided on assignment: \{35f03575-f7d8-4de4-851d-92c048a6204d=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([0_0]) prevStandbyTasks: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 2 assigned: 1]\} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)\
[2021-07-14 16:05:18,782] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Assigned tasks [0_0] including stateful [] to clients as: \
35f03575-f7d8-4de4-851d-92c048a6204d=[activeTasks: ([0_0]) standbyTasks: ([])]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:599)\
[2021-07-14 16:05:18,782] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Removing task 0_0 from TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb active assignment until it is safely revoked in followup rebalance (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:945)\
[2021-07-14 16:05:18,782] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Requesting followup rebalance be scheduled immediately by TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb due to tasks changing ownership. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:893)\
[2021-07-14 16:05:18,783] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Client 35f03575-f7d8-4de4-851d-92c048a6204d per-consumer assignment:\
	prev owned active \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192=[0_0]\}\
	prev owned standby \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb=[], TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192=[]\}\
	assigned active \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb=[0_0]\}\
	revoking active \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb=[0_0]\}\
	assigned standby \{\}\
 (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:808)\
[2021-07-14 16:05:18,783] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Finished unstable assignment of tasks, a followup rebalance will be scheduled. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:824)\
[2021-07-14 16:05:18,783] WARN [Consumer clientId=TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer, groupId=TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation] The following subscribed topics are not assigned to any members: [inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation]  (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:611)\
[2021-07-14 16:05:18,784] INFO [GroupCoordinator 0]: Assignment received from leader TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192 for group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation for generation 2. The group has 2 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:18,784] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] State transition from RUNNING to PARTITIONS_REVOKED (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:18,784] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer] Requested to schedule immediate rebalance for new tasks to be safely revoked from current owner. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1306)\
[2021-07-14 16:05:18,785] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Handle new assignment with:\
	New active tasks: []\
	New standby tasks: []\
	Existing active tasks: []\
	Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:263)\
[2021-07-14 16:05:18,785] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:18,785] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] task [0_0] Suspended RUNNING (org.apache.kafka.streams.processor.internals.StreamTask:1187)\
[2021-07-14 16:05:18,785] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] task [0_0] Suspended running (org.apache.kafka.streams.processor.internals.StreamTask:300)\
[2021-07-14 16:05:18,785] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] partition revocation took 1 ms. (org.apache.kafka.streams.processor.internals.StreamThread:97)\
[2021-07-14 16:05:18,785] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1312)\
[2021-07-14 16:05:18,785] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Handle new assignment with:\
	New active tasks: []\
	New standby tasks: []\
	Existing active tasks: [0_0]\
	Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:263)\
[2021-07-14 16:05:18,785] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] task [0_0] Closing record collector clean (org.apache.kafka.streams.processor.internals.RecordCollectorImpl:268)\
[2021-07-14 16:05:18,785] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] task [0_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:524)\
[2021-07-14 16:05:18,786] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] State transition from PARTITIONS_REVOKED to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:18,786] INFO [GroupCoordinator 0]: Preparing to rebalance group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation in state PreparingRebalance with old generation 2 (__consumer_offsets-2) (reason: Leader TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192 re-joining group during Stable) (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:18,832] WARN [AdminClient clientId=adminclient-67] Connection to node 0 (localhost/127.0.0.1:57891) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:766)\
[2021-07-14 16:05:18,839] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Restoration took 53 ms for all tasks [] (org.apache.kafka.streams.processor.internals.StreamThread:851)\
[2021-07-14 16:05:18,839] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:18,842] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Restoration took 57 ms for all tasks [] (org.apache.kafka.streams.processor.internals.StreamThread:851)\
[2021-07-14 16:05:18,842] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:18,842] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:315)\
[2021-07-14 16:05:18,842] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Triggering the followup rebalance scheduled for 0 ms. (org.apache.kafka.streams.processor.internals.StreamThread:585)\
[2021-07-14 16:05:18,843] INFO [GroupCoordinator 0]: Stabilized group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation generation 3 (__consumer_offsets-2) with 2 members (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:18,843] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:315)\
[2021-07-14 16:05:18,844] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:1063)\
[2021-07-14 16:05:18,844] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:18,844] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:1063)\
[2021-07-14 16:05:18,844] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] All members participating in this rebalance: \
35f03575-f7d8-4de4-851d-92c048a6204d: [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb, TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:581)\
[2021-07-14 16:05:18,844] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:18,844] INFO Decided on assignment: \{35f03575-f7d8-4de4-851d-92c048a6204d=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 2 assigned: 1]\} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)\
[2021-07-14 16:05:18,845] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Assigned tasks [0_0] including stateful [] to clients as: \
35f03575-f7d8-4de4-851d-92c048a6204d=[activeTasks: ([0_0]) standbyTasks: ([])]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:599)\
[2021-07-14 16:05:18,845] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Client 35f03575-f7d8-4de4-851d-92c048a6204d per-consumer assignment:\
	prev owned active \{\}\
	prev owned standby \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb=[], TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192=[]\}\
	assigned active \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer-148e1e16-be75-47e9-9bc9-f9cc64a1b2bb=[0_0]\}\
	revoking active \{\}\
	assigned standby \{\}\
 (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:808)\
[2021-07-14 16:05:18,845] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:827)\
[2021-07-14 16:05:18,846] INFO [GroupCoordinator 0]: Assignment received from leader TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer-5fec55f9-4501-440f-97e3-0b1c68d09192 for group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation for generation 3. The group has 2 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:18,847] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1312)\
[2021-07-14 16:05:18,847] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1312)\
[2021-07-14 16:05:18,847] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Handle new assignment with:\
	New active tasks: []\
	New standby tasks: []\
	Existing active tasks: []\
	Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:263)\
[2021-07-14 16:05:18,847] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Handle new assignment with:\
	New active tasks: [0_0]\
	New standby tasks: []\
	Existing active tasks: []\
	Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:263)\
[2021-07-14 16:05:18,941] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:1077)\
[2021-07-14 16:05:18,942] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:1077)\
[2021-07-14 16:05:18,942] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] task [0_0] Suspended CREATED (org.apache.kafka.streams.processor.internals.StreamTask:1187)\
[2021-07-14 16:05:18,945] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] task [0_0] Closing record collector clean (org.apache.kafka.streams.processor.internals.RecordCollectorImpl:268)\
[2021-07-14 16:05:18,945] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] task [0_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:524)\
[2021-07-14 16:05:18,952] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:18,952] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-2] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:1106)\
[2021-07-14 16:05:18,953] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:18,953] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:1106)\
[2021-07-14 16:05:18,954] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:315)\
[2021-07-14 16:05:18,954] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-35f03575-f7d8-4de4-851d-92c048a6204d] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:1380)\
[2021-07-14 16:05:18,968] INFO [Controller id=0] Starting topic deletion for topics inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:18,968] INFO [Topic Deletion Manager 0] Handling deletion for topics inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:18,968] INFO [Topic Deletion Manager 0] Deletion of topic inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation (re)started (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:18,968] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,968] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,969] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,969] INFO [Broker id=0] Add 0 partitions and deleted 1 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 5 (state.change.logger:66)\
[2021-07-14 16:05:18,970] INFO [GroupCoordinator 0]: Removed 0 offsets associated with deleted partitions: inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:18,972] INFO [Controller id=0 epoch=1] Partition inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 state changed to (Leader:-1,ISR:0,LeaderEpoch:1,ControllerEpoch:1) after removing replica 0 from the ISR as part of transition to OfflineReplica (state.change.logger:66)\
[2021-07-14 16:05:18,972] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,973] INFO [Controller id=0 epoch=1] Sending StopReplica request for 1 replicas to broker 0 (state.change.logger:66)\
[2021-07-14 16:05:18,973] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,973] INFO [Controller id=0 epoch=1] Sending StopReplica request for 1 replicas to broker 0 (state.change.logger:66)\
[2021-07-14 16:05:18,973] INFO [Broker id=0] Handling StopReplica request correlationId 6 from controller 0 for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,973] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0) (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:18,973] INFO [ReplicaAlterLogDirsManager on broker 0] Removed fetcher for partitions Set(inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0) (kafka.server.ReplicaAlterLogDirsManager:66)\
[2021-07-14 16:05:18,974] INFO [Broker id=0] Handling StopReplica request correlationId 7 from controller 0 for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,974] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0) (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:18,975] INFO [ReplicaAlterLogDirsManager on broker 0] Removed fetcher for partitions Set(inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0) (kafka.server.ReplicaAlterLogDirsManager:66)\
[2021-07-14 16:05:18,976] INFO Log for partition inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0 is renamed to /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation-0.16ebaa7d63ab4968b83aa468015bd9cb-delete and is scheduled for deletion (kafka.log.LogManager:66)\
[2021-07-14 16:05:18,979] INFO [Controller id=0] Starting topic deletion for topics inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation,__consumer_offsets (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:18,979] INFO [Topic Deletion Manager 0] Handling deletion for topics inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation,__consumer_offsets (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:18,979] INFO [Topic Deletion Manager 0] Deletion of topic __consumer_offsets (re)started (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:18,979] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,980] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,980] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,983] INFO [Broker id=0] Add 0 partitions and deleted 5 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 8 (state.change.logger:66)\
[2021-07-14 16:05:18,983] INFO [GroupCoordinator 0]: Removed 0 offsets associated with deleted partitions: __consumer_offsets-0, __consumer_offsets-3, __consumer_offsets-4, __consumer_offsets-1, __consumer_offsets-2. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:18,985] INFO [Controller id=0 epoch=1] Partition __consumer_offsets-4 state changed to (Leader:-1,ISR:0,LeaderEpoch:1,ControllerEpoch:1) after removing replica 0 from the ISR as part of transition to OfflineReplica (state.change.logger:66)\
[2021-07-14 16:05:18,985] INFO [Controller id=0 epoch=1] Partition __consumer_offsets-3 state changed to (Leader:-1,ISR:0,LeaderEpoch:1,ControllerEpoch:1) after removing replica 0 from the ISR as part of transition to OfflineReplica (state.change.logger:66)\
[2021-07-14 16:05:18,985] INFO [Controller id=0 epoch=1] Partition __consumer_offsets-2 state changed to (Leader:-1,ISR:0,LeaderEpoch:1,ControllerEpoch:1) after removing replica 0 from the ISR as part of transition to OfflineReplica (state.change.logger:66)\
[2021-07-14 16:05:18,985] INFO [Controller id=0 epoch=1] Partition __consumer_offsets-0 state changed to (Leader:-1,ISR:0,LeaderEpoch:1,ControllerEpoch:1) after removing replica 0 from the ISR as part of transition to OfflineReplica (state.change.logger:66)\
[2021-07-14 16:05:18,985] INFO [Controller id=0 epoch=1] Partition __consumer_offsets-1 state changed to (Leader:-1,ISR:0,LeaderEpoch:1,ControllerEpoch:1) after removing replica 0 from the ISR as part of transition to OfflineReplica (state.change.logger:66)\
[2021-07-14 16:05:18,985] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,986] INFO [Controller id=0 epoch=1] Sending StopReplica request for 5 replicas to broker 0 (state.change.logger:66)\
[2021-07-14 16:05:18,987] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,987] INFO [Broker id=0] Handling StopReplica request correlationId 9 from controller 0 for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,987] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(__consumer_offsets-0, __consumer_offsets-3, __consumer_offsets-4, __consumer_offsets-1, __consumer_offsets-2) (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:18,987] INFO [ReplicaAlterLogDirsManager on broker 0] Removed fetcher for partitions Set(__consumer_offsets-0, __consumer_offsets-3, __consumer_offsets-4, __consumer_offsets-1, __consumer_offsets-2) (kafka.server.ReplicaAlterLogDirsManager:66)\
[2021-07-14 16:05:18,987] INFO [Controller id=0 epoch=1] Sending StopReplica request for 5 replicas to broker 0 (state.change.logger:66)\
[2021-07-14 16:05:18,987] INFO [Broker id=0] Handling StopReplica request correlationId 10 from controller 0 for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,988] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(__consumer_offsets-0, __consumer_offsets-3, __consumer_offsets-4, __consumer_offsets-1, __consumer_offsets-2) (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:18,988] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,988] INFO [ReplicaAlterLogDirsManager on broker 0] Removed fetcher for partitions Set(__consumer_offsets-0, __consumer_offsets-3, __consumer_offsets-4, __consumer_offsets-1, __consumer_offsets-2) (kafka.server.ReplicaAlterLogDirsManager:66)\
[2021-07-14 16:05:18,988] INFO [Topic Deletion Manager 0] Handling deletion for topics inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation,__consumer_offsets (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:18,988] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:18,990] INFO Log for partition __consumer_offsets-0 is renamed to /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-0.666e3f0804b842aaa29ef475bf071796-delete and is scheduled for deletion (kafka.log.LogManager:66)\
[2021-07-14 16:05:18,991] INFO Log for partition __consumer_offsets-3 is renamed to /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-3.91cc144f0bbf4a688fea571fb74f22c4-delete and is scheduled for deletion (kafka.log.LogManager:66)\
[2021-07-14 16:05:18,992] INFO Log for partition __consumer_offsets-4 is renamed to /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-4.550e01397c0c4a6eb03adb6f76e36102-delete and is scheduled for deletion (kafka.log.LogManager:66)\
[2021-07-14 16:05:18,993] INFO Log for partition __consumer_offsets-1 is renamed to /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-1.0c196ac85fce4f1bb17ddc389fd995d2-delete and is scheduled for deletion (kafka.log.LogManager:66)\
[2021-07-14 16:05:18,994] INFO [Topic Deletion Manager 0] Deletion of topic inputTaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation successfully completed (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:18,994] INFO Log for partition __consumer_offsets-2 is renamed to /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-2.2d3fb5b3509f4e6da4fffed0009bfac1-delete and is scheduled for deletion (kafka.log.LogManager:66)\
[2021-07-14 16:05:18,995] INFO [Controller id=0] New topics: [Set()], deleted topics: [HashSet()], new partition replica assignment [Set()] (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:18,995] INFO [Controller id=0] Starting topic deletion for topics __consumer_offsets (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:18,995] INFO [Topic Deletion Manager 0] Handling deletion for topics __consumer_offsets (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:18,999] INFO [GroupCoordinator 0]: Resigned as the group coordinator for partition 0 in epoch None (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:18,999] INFO [GroupMetadataManager brokerId=0] Scheduling unloading of offsets and group metadata from __consumer_offsets-0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,000] INFO [GroupCoordinator 0]: Resigned as the group coordinator for partition 3 in epoch None (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,000] INFO [GroupMetadataManager brokerId=0] Finished unloading __consumer_offsets-0 for coordinator epoch None. Removed 0 cached offsets and 0 cached groups. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,000] INFO [GroupMetadataManager brokerId=0] Scheduling unloading of offsets and group metadata from __consumer_offsets-3 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,000] INFO [GroupCoordinator 0]: Resigned as the group coordinator for partition 4 in epoch None (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,000] INFO [GroupMetadataManager brokerId=0] Scheduling unloading of offsets and group metadata from __consumer_offsets-4 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,000] INFO [GroupMetadataManager brokerId=0] Finished unloading __consumer_offsets-3 for coordinator epoch None. Removed 0 cached offsets and 0 cached groups. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,000] INFO [GroupCoordinator 0]: Resigned as the group coordinator for partition 1 in epoch None (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,000] INFO [GroupMetadataManager brokerId=0] Scheduling unloading of offsets and group metadata from __consumer_offsets-1 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,000] INFO [GroupMetadataManager brokerId=0] Finished unloading __consumer_offsets-4 for coordinator epoch None. Removed 0 cached offsets and 0 cached groups. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,000] INFO [GroupCoordinator 0]: Resigned as the group coordinator for partition 2 in epoch None (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,000] INFO [GroupMetadataManager brokerId=0] Finished unloading __consumer_offsets-1 for coordinator epoch None. Removed 0 cached offsets and 0 cached groups. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,000] INFO [GroupMetadataManager brokerId=0] Scheduling unloading of offsets and group metadata from __consumer_offsets-2 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,001] INFO [GroupCoordinator 0]: Unloading group metadata for TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectEndOffsetInformation with generation 3 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,001] INFO [GroupMetadataManager brokerId=0] Finished unloading __consumer_offsets-2 for coordinator epoch None. Removed 0 cached offsets and 1 cached groups. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,001] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,001] INFO [Topic Deletion Manager 0] Handling deletion for topics __consumer_offsets (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:19,002] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,017] INFO [Topic Deletion Manager 0] Deletion of topic __consumer_offsets successfully completed (kafka.controller.TopicDeletionManager:66)\
[2021-07-14 16:05:19,018] INFO [Controller id=0] New topics: [Set()], deleted topics: [HashSet()], new partition replica assignment [Set()] (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:19,094] INFO Creating topic inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation with configuration \{\} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)\
[2021-07-14 16:05:19,098] INFO [Controller id=0] New topics: [Set(inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation,Some(GQ_2RiOrQsanfjPaj-WOnA),Map(inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:19,098] INFO [Controller id=0] New partition creation callback for inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:19,098] INFO [Controller id=0 epoch=1] Changed partition inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:19,098] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,098] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,101] INFO [Controller id=0 epoch=1] Changed partition inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:19,101] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)\
[2021-07-14 16:05:19,101] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,101] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,101] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 11 from controller 0 for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,102] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0) (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:19,102] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 11 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,106] INFO [LogLoader partition=inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:19,108] INFO Created log for partition inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 with properties \{\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:19,109] INFO [Partition inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 broker=0] No checkpointed highwatermark is found for partition inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,109] INFO [Partition inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 broker=0] Log loaded for partition inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,109] INFO [Broker id=0] Leader inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:19,111] INFO [Broker id=0] Finished LeaderAndIsr request in 0ms correlationId 11 from controller 0 for 1 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,113] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 12 (state.change.logger:66)\
[2021-07-14 16:05:19,117] INFO StreamsConfig values: \
	acceptable.recovery.lag = 10000\
	application.id = TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation\
	application.server = \
	bootstrap.servers = [localhost:57923]\
	buffered.records.per.partition = 1000\
	built.in.metrics.version = latest\
	cache.max.bytes.buffering = 10485760\
	client.id = \
	commit.interval.ms = 1\
	connections.max.idle.ms = 540000\
	default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler\
	default.key.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde\
	default.list.key.serde.inner = null\
	default.list.key.serde.type = null\
	default.list.value.serde.inner = null\
	default.list.value.serde.type = null\
	default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler\
	default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp\
	default.value.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde\
	max.task.idle.ms = 0\
	max.warmup.replicas = 2\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	num.standby.replicas = 0\
	num.stream.threads = 2\
	poll.ms = 100\
	probing.rebalance.interval.ms = 600000\
	processing.guarantee = at_least_once\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	replication.factor = -1\
	request.timeout.ms = 40000\
	retries = 0\
	retry.backoff.ms = 100\
	rocksdb.config.setter = null\
	security.protocol = PLAINTEXT\
	send.buffer.bytes = 131072\
	state.cleanup.delay.ms = 600000\
	state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-2490334526672851046\
	task.timeout.ms = 300000\
	topology.optimization = none\
	upgrade.from = null\
	window.size.ms = null\
	windowed.inner.class.serde = null\
	windowstore.changelog.additional.retention.ms = 86400000\
 (org.apache.kafka.streams.StreamsConfig:376)\
[2021-07-14 16:05:19,120] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:849)\
[2021-07-14 16:05:19,120] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:850)\
[2021-07-14 16:05:19,121] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:334)\
[2021-07-14 16:05:19,126] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)\
[2021-07-14 16:05:19,129] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:386)\
[2021-07-14 16:05:19,131] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:114)\
[2021-07-14 16:05:19,133] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:334)\
[2021-07-14 16:05:19,135] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)\
[2021-07-14 16:05:19,137] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:386)\
[2021-07-14 16:05:19,139] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:114)\
[2021-07-14 16:05:19,141] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:315)\
[2021-07-14 16:05:19,142] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:548)\
[2021-07-14 16:05:19,142] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Starting (org.apache.kafka.streams.processor.internals.StreamThread:548)\
[2021-07-14 16:05:19,142] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,142] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,148] INFO Creating topic __consumer_offsets with configuration \{segment.bytes=104857600, cleanup.policy=compact, compression.type=producer\} and initial partition assignment HashMap(0 -> ArrayBuffer(0), 1 -> ArrayBuffer(0), 2 -> ArrayBuffer(0), 3 -> ArrayBuffer(0), 4 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)\
[2021-07-14 16:05:19,151] INFO [Controller id=0] New topics: [Set(__consumer_offsets)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(__consumer_offsets,Some(pzi34XcLTj-tFJknc8PX8g),HashMap(__consumer_offsets-4 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-3 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-2 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-1 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:19,152] INFO [Controller id=0] New partition creation callback for __consumer_offsets-4,__consumer_offsets-3,__consumer_offsets-2,__consumer_offsets-0,__consumer_offsets-1 (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:19,152] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:19,152] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:19,152] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:19,152] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:19,152] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2021-07-14 16:05:19,152] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,152] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,157] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:19,157] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:19,157] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:19,157] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:19,157] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2021-07-14 16:05:19,157] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 5 become-leader and 0 become-follower partitions (state.change.logger:66)\
[2021-07-14 16:05:19,158] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,158] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,158] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 13 from controller 0 for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,160] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions HashSet(__consumer_offsets-4, __consumer_offsets-3, __consumer_offsets-2, __consumer_offsets-0, __consumer_offsets-1) (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:19,161] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 13 from controller 0 epoch 1 as part of the become-leader transition for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,165] INFO [LogLoader partition=__consumer_offsets-3, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:19,168] INFO Created log for partition __consumer_offsets-3 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-3 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:19,169] INFO [Partition __consumer_offsets-3 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-3 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,169] INFO [Partition __consumer_offsets-3 broker=0] Log loaded for partition __consumer_offsets-3 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,169] INFO [Broker id=0] Leader __consumer_offsets-3 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:19,176] INFO [LogLoader partition=__consumer_offsets-2, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:19,179] INFO Created log for partition __consumer_offsets-2 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-2 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:19,179] INFO [Partition __consumer_offsets-2 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-2 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,180] INFO [Partition __consumer_offsets-2 broker=0] Log loaded for partition __consumer_offsets-2 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,180] INFO [Broker id=0] Leader __consumer_offsets-2 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:19,185] INFO [LogLoader partition=__consumer_offsets-4, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:19,189] INFO Created log for partition __consumer_offsets-4 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-4 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:19,189] INFO [Partition __consumer_offsets-4 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-4 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,189] INFO [Partition __consumer_offsets-4 broker=0] Log loaded for partition __consumer_offsets-4 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,189] INFO [Broker id=0] Leader __consumer_offsets-4 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:19,194] INFO [LogLoader partition=__consumer_offsets-1, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:19,196] INFO Created log for partition __consumer_offsets-1 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-1 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:19,196] INFO [Partition __consumer_offsets-1 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-1 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,196] INFO [Partition __consumer_offsets-1 broker=0] Log loaded for partition __consumer_offsets-1 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,196] INFO [Broker id=0] Leader __consumer_offsets-1 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:19,201] INFO [LogLoader partition=__consumer_offsets-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804] Loading producer state till offset 0 with message format version 2 (kafka.log.Log$:1982)\
[2021-07-14 16:05:19,204] INFO Created log for partition __consumer_offsets-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-860228248547597742/log4042677100740363804/__consumer_offsets-0 with properties \{cleanup.policy=compact, compression.type="producer", segment.bytes=104857600\} (kafka.log.LogManager:66)\
[2021-07-14 16:05:19,204] INFO [Partition __consumer_offsets-0 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,204] INFO [Partition __consumer_offsets-0 broker=0] Log loaded for partition __consumer_offsets-0 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2021-07-14 16:05:19,204] INFO [Broker id=0] Leader __consumer_offsets-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2021-07-14 16:05:19,206] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 3 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,206] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-3 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,206] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 2 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,206] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-2 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,206] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 4 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,206] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-4 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,206] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-3 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,206] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 1 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,207] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-1 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,207] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-2 in 1 milliseconds for epoch 0, of which 1 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,207] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 0 in epoch 0 (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,208] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-0 for epoch 0 (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,208] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-4 in 2 milliseconds for epoch 0, of which 2 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,208] INFO [Broker id=0] Finished LeaderAndIsr request in 0ms correlationId 13 from controller 0 for 5 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,208] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-1 in 1 milliseconds for epoch 0, of which 1 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,208] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-0 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)\
[2021-07-14 16:05:19,209] INFO [Broker id=0] Add 5 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 14 (state.change.logger:66)\
[2021-07-14 16:05:19,244] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Processed 0 total records, ran 0 punctuators, and committed 0 total tasks since the last update (org.apache.kafka.streams.processor.internals.StreamThread:827)\
[2021-07-14 16:05:19,244] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Processed 0 total records, ran 0 punctuators, and committed 0 total tasks since the last update (org.apache.kafka.streams.processor.internals.StreamThread:827)\
[2021-07-14 16:05:19,249] INFO [GroupCoordinator 0]: Dynamic member with unknown member id joins group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation in Empty state. Created a new member id TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer-9fd49692-9508-461e-8259-8a17f0688555 and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,249] INFO [GroupCoordinator 0]: Dynamic member with unknown member id joins group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation in Empty state. Created a new member id TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2-consumer-69352491-8bbe-41ff-a3aa-54c74dcd8a3b and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,249] INFO [GroupCoordinator 0]: Preparing to rebalance group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation in state PreparingRebalance with old generation 0 (__consumer_offsets-4) (reason: Adding new member TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer-9fd49692-9508-461e-8259-8a17f0688555 with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,250] INFO [GroupCoordinator 0]: Stabilized group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation generation 1 (__consumer_offsets-4) with 2 members (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,251] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer] All members participating in this rebalance: \
657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8: [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer-9fd49692-9508-461e-8259-8a17f0688555, TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2-consumer-69352491-8bbe-41ff-a3aa-54c74dcd8a3b]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:581)\
[2021-07-14 16:05:19,252] INFO Decided on assignment: \{657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 2 assigned: 1]\} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)\
[2021-07-14 16:05:19,252] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer] Assigned tasks [0_0] including stateful [] to clients as: \
657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8=[activeTasks: ([0_0]) standbyTasks: ([])]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:599)\
[2021-07-14 16:05:19,252] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer] Client 657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8 per-consumer assignment:\
	prev owned active \{\}\
	prev owned standby \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer-9fd49692-9508-461e-8259-8a17f0688555=[], TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2-consumer-69352491-8bbe-41ff-a3aa-54c74dcd8a3b=[]\}\
	assigned active \{TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer-9fd49692-9508-461e-8259-8a17f0688555=[0_0]\}\
	revoking active \{\}\
	assigned standby \{\}\
 (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:808)\
[2021-07-14 16:05:19,252] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:827)\
[2021-07-14 16:05:19,253] INFO [GroupCoordinator 0]: Assignment received from leader TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer-9fd49692-9508-461e-8259-8a17f0688555 for group TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation for generation 1. The group has 2 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,254] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1312)\
[2021-07-14 16:05:19,254] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1312)\
[2021-07-14 16:05:19,255] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Handle new assignment with:\
	New active tasks: []\
	New standby tasks: []\
	Existing active tasks: []\
	Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:263)\
[2021-07-14 16:05:19,255] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Handle new assignment with:\
	New active tasks: [0_0]\
	New standby tasks: []\
	Existing active tasks: []\
	Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:263)\
[2021-07-14 16:05:19,255] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,256] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,344] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] task [0_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:240)\
[2021-07-14 16:05:19,345] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] task [0_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:265)\
[2021-07-14 16:05:19,345] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Restoration took 89 ms for all tasks [0_0] (org.apache.kafka.streams.processor.internals.StreamThread:851)\
[2021-07-14 16:05:19,345] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,345] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Restoration took 90 ms for all tasks [] (org.apache.kafka.streams.processor.internals.StreamThread:851)\
[2021-07-14 16:05:19,345] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,346] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:315)\
[2021-07-14 16:05:19,675] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:315)\
[2021-07-14 16:05:19,677] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:1063)\
[2021-07-14 16:05:19,678] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,678] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:1063)\
[2021-07-14 16:05:19,678] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,752] WARN [AdminClient clientId=adminclient-67] Connection to node 0 (localhost/127.0.0.1:57891) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:766)\
[2021-07-14 16:05:19,754] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:1077)\
[2021-07-14 16:05:19,758] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,758] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-2] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:1106)\
[2021-07-14 16:05:19,776] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:1077)\
[2021-07-14 16:05:19,776] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] task [0_0] Suspended RUNNING (org.apache.kafka.streams.processor.internals.StreamTask:1187)\
[2021-07-14 16:05:19,777] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] task [0_0] Suspended running (org.apache.kafka.streams.processor.internals.StreamTask:300)\
[2021-07-14 16:05:19,777] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] task [0_0] Closing record collector clean (org.apache.kafka.streams.processor.internals.RecordCollectorImpl:268)\
[2021-07-14 16:05:19,777] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] task [0_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:524)\
[2021-07-14 16:05:19,781] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:229)\
[2021-07-14 16:05:19,781] INFO stream-thread [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:1106)\
[2021-07-14 16:05:19,783] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:315)\
[2021-07-14 16:05:19,783] INFO stream-client [TaskMetadataTest_TaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-657ae7fb-272b-4cc3-8ea5-f108ddb8c6a8] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:1380)\
[2021-07-14 16:05:19,784] INFO [KafkaServer id=0] shutting down (kafka.server.KafkaServer:66)\
[2021-07-14 16:05:19,784] INFO [KafkaServer id=0] Starting controlled shutdown (kafka.server.KafkaServer:66)\
[2021-07-14 16:05:19,785] INFO [Controller id=0] Shutting down broker 0 (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:19,785] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2021-07-14 16:05:19,786] INFO [KafkaServer id=0] Controlled shutdown succeeded (kafka.server.KafkaServer:66)\
[2021-07-14 16:05:19,786] INFO [/config/changes-event-process-thread]: Shutting down (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread:66)\
[2021-07-14 16:05:19,787] INFO [/config/changes-event-process-thread]: Stopped (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread:66)\
[2021-07-14 16:05:19,787] INFO [/config/changes-event-process-thread]: Shutdown completed (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread:66)\
[2021-07-14 16:05:19,787] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Stopping socket server request processors (kafka.network.SocketServer:66)\
[2021-07-14 16:05:19,788] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Stopped socket server request processors (kafka.network.SocketServer:66)\
[2021-07-14 16:05:19,788] INFO [data-plane Kafka Request Handler on Broker 0], shutting down (kafka.server.KafkaRequestHandlerPool:66)\
[2021-07-14 16:05:19,789] INFO [data-plane Kafka Request Handler on Broker 0], shut down completely (kafka.server.KafkaRequestHandlerPool:66)\
[2021-07-14 16:05:19,789] INFO [ExpirationReaper-0-AlterAcls]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:19,843] INFO [ExpirationReaper-0-AlterAcls]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:19,843] INFO [ExpirationReaper-0-AlterAcls]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:19,843] INFO [KafkaApi-0] Shutdown complete. (kafka.server.KafkaApis:66)\
[2021-07-14 16:05:19,844] INFO [ExpirationReaper-0-topic]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:19,845] INFO [ExpirationReaper-0-topic]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:19,845] INFO [ExpirationReaper-0-topic]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:19,845] INFO [TransactionCoordinator id=0] Shutting down. (kafka.coordinator.transaction.TransactionCoordinator:66)\
[2021-07-14 16:05:19,845] INFO [Transaction State Manager 0]: Shutdown complete (kafka.coordinator.transaction.TransactionStateManager:66)\
[2021-07-14 16:05:19,846] INFO [Transaction Marker Channel Manager 0]: Shutting down (kafka.coordinator.transaction.TransactionMarkerChannelManager:66)\
[2021-07-14 16:05:19,846] INFO [Transaction Marker Channel Manager 0]: Stopped (kafka.coordinator.transaction.TransactionMarkerChannelManager:66)\
[2021-07-14 16:05:19,846] INFO [Transaction Marker Channel Manager 0]: Shutdown completed (kafka.coordinator.transaction.TransactionMarkerChannelManager:66)\
[2021-07-14 16:05:19,846] INFO [TransactionCoordinator id=0] Shutdown complete. (kafka.coordinator.transaction.TransactionCoordinator:66)\
[2021-07-14 16:05:19,846] INFO [GroupCoordinator 0]: Shutting down. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:19,846] INFO [ExpirationReaper-0-Heartbeat]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,040] INFO [ExpirationReaper-0-Heartbeat]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,040] INFO [ExpirationReaper-0-Heartbeat]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,041] INFO [ExpirationReaper-0-Rebalance]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,057] INFO [ExpirationReaper-0-Rebalance]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,057] INFO [ExpirationReaper-0-Rebalance]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,058] INFO [GroupCoordinator 0]: Shutdown complete. (kafka.coordinator.group.GroupCoordinator:66)\
[2021-07-14 16:05:20,058] INFO [ReplicaManager broker=0] Shutting down (kafka.server.ReplicaManager:66)\
[2021-07-14 16:05:20,058] INFO [LogDirFailureHandler]: Shutting down (kafka.server.ReplicaManager$LogDirFailureHandler:66)\
[2021-07-14 16:05:20,058] INFO [LogDirFailureHandler]: Shutdown completed (kafka.server.ReplicaManager$LogDirFailureHandler:66)\
[2021-07-14 16:05:20,058] INFO [ReplicaFetcherManager on broker 0] shutting down (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:20,058] INFO [LogDirFailureHandler]: Stopped (kafka.server.ReplicaManager$LogDirFailureHandler:66)\
[2021-07-14 16:05:20,058] INFO [ReplicaFetcherManager on broker 0] shutdown completed (kafka.server.ReplicaFetcherManager:66)\
[2021-07-14 16:05:20,059] INFO [ReplicaAlterLogDirsManager on broker 0] shutting down (kafka.server.ReplicaAlterLogDirsManager:66)\
[2021-07-14 16:05:20,059] INFO [ReplicaAlterLogDirsManager on broker 0] shutdown completed (kafka.server.ReplicaAlterLogDirsManager:66)\
[2021-07-14 16:05:20,059] INFO [ExpirationReaper-0-Fetch]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,073] INFO [ExpirationReaper-0-Fetch]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,073] INFO [ExpirationReaper-0-Fetch]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,074] INFO [ExpirationReaper-0-Produce]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,247] INFO [ExpirationReaper-0-Produce]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,247] INFO [ExpirationReaper-0-Produce]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,247] INFO [ExpirationReaper-0-DeleteRecords]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,449] INFO [ExpirationReaper-0-DeleteRecords]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,449] INFO [ExpirationReaper-0-DeleteRecords]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,450] INFO [ExpirationReaper-0-ElectLeader]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,617] INFO [Controller id=0] Processing automatic preferred replica leader election (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:20,654] INFO [ExpirationReaper-0-ElectLeader]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,654] INFO [ExpirationReaper-0-ElectLeader]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)\
[2021-07-14 16:05:20,657] INFO [ReplicaManager broker=0] Shut down completely (kafka.server.ReplicaManager:66)\
[2021-07-14 16:05:20,657] INFO [BrokerToControllerChannelManager broker=0 name=alterIsr]: Shutting down (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:20,658] INFO [BrokerToControllerChannelManager broker=0 name=alterIsr]: Stopped (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:20,658] INFO [BrokerToControllerChannelManager broker=0 name=alterIsr]: Shutdown completed (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:20,659] INFO Broker to controller channel manager for alterIsr shutdown (kafka.server.BrokerToControllerChannelManagerImpl:66)\
[2021-07-14 16:05:20,659] INFO [BrokerToControllerChannelManager broker=0 name=forwarding]: Shutting down (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:20,659] INFO [BrokerToControllerChannelManager broker=0 name=forwarding]: Stopped (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:20,659] INFO [BrokerToControllerChannelManager broker=0 name=forwarding]: Shutdown completed (kafka.server.BrokerToControllerRequestThread:66)\
[2021-07-14 16:05:20,659] INFO Broker to controller channel manager for forwarding shutdown (kafka.server.BrokerToControllerChannelManagerImpl:66)\
[2021-07-14 16:05:20,659] INFO Shutting down. (kafka.log.LogManager:66)\
[2021-07-14 16:05:20,660] INFO Shutting down the log cleaner. (kafka.log.LogCleaner:66)\
[2021-07-14 16:05:20,660] INFO [kafka-log-cleaner-thread-0]: Shutting down (kafka.log.LogCleaner:66)\
[2021-07-14 16:05:20,660] INFO [kafka-log-cleaner-thread-0]: Stopped (kafka.log.LogCleaner:66)\
[2021-07-14 16:05:20,660] INFO [kafka-log-cleaner-thread-0]: Shutdown completed (kafka.log.LogCleaner:66)\
[2021-07-14 16:05:20,662] INFO [ProducerStateManager partition=inputTaskMetadataIntegrationTestshouldReportCorrectCommittedOffsetInformation-0] Wrote producer snapshot at offset 3 with 0 producer ids in 0 ms. (kafka.log.ProducerStateManager:66)\
[2021-07-14 16:05:20,690] INFO [ProducerStateManager partition=__consumer_offsets-4] Wrote producer snapshot at offset 4 with 0 producer ids in 0 ms. (kafka.log.ProducerStateManager:66)\
[2021-07-14 16:05:20,707] INFO Shutdown complete. (kafka.log.LogManager:66)\
[2021-07-14 16:05:20,708] INFO [ControllerEventThread controllerId=0] Shutting down (kafka.controller.ControllerEventManager$ControllerEventThread:66)\
[2021-07-14 16:05:20,708] INFO [ControllerEventThread controllerId=0] Shutdown completed (kafka.controller.ControllerEventManager$ControllerEventThread:66)\
[2021-07-14 16:05:20,708] INFO [ControllerEventThread controllerId=0] Stopped (kafka.controller.ControllerEventManager$ControllerEventThread:66)\
[2021-07-14 16:05:20,709] INFO [PartitionStateMachine controllerId=0] Stopped partition state machine (kafka.controller.ZkPartitionStateMachine:66)\
[2021-07-14 16:05:20,709] INFO [ReplicaStateMachine controllerId=0] Stopped replica state machine (kafka.controller.ZkReplicaStateMachine:66)\
[2021-07-14 16:05:20,709] INFO [RequestSendThread controllerId=0] Shutting down (kafka.controller.RequestSendThread:66)\
[2021-07-14 16:05:20,709] INFO [RequestSendThread controllerId=0] Stopped (kafka.controller.RequestSendThread:66)\
[2021-07-14 16:05:20,709] INFO [RequestSendThread controllerId=0] Shutdown completed (kafka.controller.RequestSendThread:66)\
[2021-07-14 16:05:20,710] INFO [Controller id=0] Resigned (kafka.controller.KafkaController:66)\
[2021-07-14 16:05:20,710] INFO [feature-zk-node-event-process-thread]: Shutting down (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread:66)\
[2021-07-14 16:05:20,711] INFO [feature-zk-node-event-process-thread]: Stopped (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread:66)\
[2021-07-14 16:05:20,711] INFO [feature-zk-node-event-process-thread]: Shutdown completed (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread:66)\
[2021-07-14 16:05:20,711] INFO [ZooKeeperClient Kafka server] Closing. (kafka.zookeeper.ZooKeeperClient:66)\
[2021-07-14 16:05:20,818] INFO [ZooKeeperClient Kafka server] Closed. (kafka.zookeeper.ZooKeeperClient:66)\
[2021-07-14 16:05:20,818] INFO [ThrottledChannelReaper-Fetch]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:20,985] WARN [AdminClient clientId=adminclient-67] Connection to node 0 (localhost/127.0.0.1:57891) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:766)\
[2021-07-14 16:05:21,540] INFO [ThrottledChannelReaper-Fetch]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:21,540] INFO [ThrottledChannelReaper-Fetch]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:21,540] INFO [ThrottledChannelReaper-Produce]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:22,201] WARN [AdminClient clientId=adminclient-67] Connection to node 0 (localhost/127.0.0.1:57891) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:766)\
[2021-07-14 16:05:22,543] INFO [ThrottledChannelReaper-Produce]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:22,543] INFO [ThrottledChannelReaper-Produce]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:22,543] INFO [ThrottledChannelReaper-Request]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:23,435] WARN [AdminClient clientId=adminclient-67] Connection to node 0 (localhost/127.0.0.1:57891) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:766)\
[2021-07-14 16:05:23,544] INFO [ThrottledChannelReaper-Request]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:23,544] INFO [ThrottledChannelReaper-Request]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:23,544] INFO [ThrottledChannelReaper-ControllerMutation]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:24,464] WARN [AdminClient clientId=adminclient-67] Connection to node 0 (localhost/127.0.0.1:57891) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:766)\
[2021-07-14 16:05:24,546] INFO [ThrottledChannelReaper-ControllerMutation]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:24,546] INFO [ThrottledChannelReaper-ControllerMutation]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)\
[2021-07-14 16:05:24,546] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Shutting down socket server (kafka.network.SocketServer:66)\
[2021-07-14 16:05:24,575] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Shutdown completed (kafka.network.SocketServer:66)\
[2021-07-14 16:05:24,579] INFO Broker and topic stats closed (kafka.server.BrokerTopicStats:66)\
[2021-07-14 16:05:24,580] INFO [KafkaServer id=0] shut down completed (kafka.server.KafkaServer:66)\
}