[2020-05-27 17:55:59,288] INFO zookeeper.snapshot.trust.empty : false (org.apache.zookeeper.server.persistence.FileTxnSnapLog:115)
[2020-05-27 17:55:59,289] INFO zookeeper.snapshotSizeFactor = 0.33 (org.apache.zookeeper.server.ZKDatabase:117)
[2020-05-27 17:55:59,289] INFO minSessionTimeout set to 1600 (org.apache.zookeeper.server.ZooKeeperServer:938)
[2020-05-27 17:55:59,289] INFO maxSessionTimeout set to 16000 (org.apache.zookeeper.server.ZooKeeperServer:947)
[2020-05-27 17:55:59,289] INFO Created server with tickTime 800 minSessionTimeout 1600 maxSessionTimeout 16000 datadir /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-6519267964250400743/version-2 snapdir /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-7049686353859294089/version-2 (org.apache.zookeeper.server.ZooKeeperServer:166)
[2020-05-27 17:55:59,289] INFO Configuring NIO connection handler with 10s sessionless connection timeout, 2 selector thread(s), 24 worker threads, and 64 kB direct buffers. (org.apache.zookeeper.server.NIOServerCnxnFactory:673)
[2020-05-27 17:55:59,290] INFO binding to port /127.0.0.1:0 (org.apache.zookeeper.server.NIOServerCnxnFactory:686)
[2020-05-27 17:55:59,291] INFO Snapshotting: 0x0 to /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-7049686353859294089/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog:404)
[2020-05-27 17:55:59,292] INFO Snapshotting: 0x0 to /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-7049686353859294089/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog:404)
[2020-05-27 17:55:59,296] INFO KafkaConfig values:
advertised.host.name = null
advertised.listeners = null
advertised.port = null
alter.config.policy.class.name = null
alter.log.dirs.replication.quota.window.num = 11
alter.log.dirs.replication.quota.window.size.seconds = 1
authorizer.class.name =
auto.create.topics.enable = true
auto.leader.rebalance.enable = true
background.threads = 10
broker.id = 0
broker.id.generation.enable = true
broker.rack = null
client.quota.callback.class = null
compression.type = producer
connection.failed.authentication.delay.ms = 100
connections.max.idle.ms = 600000
connections.max.reauth.ms = 0
control.plane.listener.name = null
controlled.shutdown.enable = true
controlled.shutdown.max.retries = 3
controlled.shutdown.retry.backoff.ms = 5000
controller.socket.timeout.ms = 30000
create.topic.policy.class.name = null
default.replication.factor = 1
delegation.token.expiry.check.interval.ms = 3600000
delegation.token.expiry.time.ms = 86400000
delegation.token.master.key = null
delegation.token.max.lifetime.ms = 604800000
delete.records.purgatory.purge.interval.requests = 1
delete.topic.enable = true
fetch.max.bytes = 57671680
fetch.purgatory.purge.interval.requests = 1000
group.initial.rebalance.delay.ms = 0
group.max.session.timeout.ms = 1800000
group.max.size = 2147483647
group.min.session.timeout.ms = 0
host.name = localhost
inter.broker.listener.name = null
inter.broker.protocol.version = 2.6-IV0
kafka.metrics.polling.interval.secs = 10
kafka.metrics.reporters = []
leader.imbalance.check.interval.seconds = 300
leader.imbalance.per.broker.percentage = 10
listener.security.protocol.map = PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL
listeners = null
log.cleaner.backoff.ms = 15000
log.cleaner.dedupe.buffer.size = 2097152
log.cleaner.delete.retention.ms = 86400000
log.cleaner.enable = true
log.cleaner.io.buffer.load.factor = 0.9
log.cleaner.io.buffer.size = 524288
log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308
log.cleaner.max.compaction.lag.ms = 9223372036854775807
log.cleaner.min.cleanable.ratio = 0.5
log.cleaner.min.compaction.lag.ms = 0
log.cleaner.threads = 1
log.cleanup.policy = [delete]
log.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770
log.dirs = null
log.flush.interval.messages = 9223372036854775807
log.flush.interval.ms = null
log.flush.offset.checkpoint.interval.ms = 60000
log.flush.scheduler.interval.ms = 9223372036854775807
log.flush.start.offset.checkpoint.interval.ms = 60000
log.index.interval.bytes = 4096
log.index.size.max.bytes = 10485760
log.message.downconversion.enable = true
log.message.format.version = 2.6-IV0
log.message.timestamp.difference.max.ms = 9223372036854775807
log.message.timestamp.type = CreateTime
log.preallocate = false
log.retention.bytes = -1
log.retention.check.interval.ms = 300000
log.retention.hours = 168
log.retention.minutes = null
log.retention.ms = null
log.roll.hours = 168
log.roll.jitter.hours = 0
log.roll.jitter.ms = null
log.roll.ms = null
log.segment.bytes = 1073741824
log.segment.delete.delay.ms = 60000
max.connections = 2147483647
max.connections.per.ip = 2147483647
max.connections.per.ip.overrides =
max.incremental.fetch.session.cache.slots = 1000
message.max.bytes = 1000000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
min.insync.replicas = 1
num.io.threads = 8
num.network.threads = 3
num.partitions = 1
num.recovery.threads.per.data.dir = 1
num.replica.alter.log.dirs.threads = null
num.replica.fetchers = 1
offset.metadata.max.bytes = 4096
offsets.commit.required.acks = -1
offsets.commit.timeout.ms = 5000
offsets.load.buffer.size = 5242880
offsets.retention.check.interval.ms = 600000
offsets.retention.minutes = 10080
offsets.topic.compression.codec = 0
offsets.topic.num.partitions = 5
offsets.topic.replication.factor = 1
offsets.topic.segment.bytes = 104857600
password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding
password.encoder.iterations = 4096
password.encoder.key.length = 128
password.encoder.keyfactory.algorithm = null
password.encoder.old.secret = null
password.encoder.secret = null
port = 0
principal.builder.class = null
producer.purgatory.purge.interval.requests = 1000
queued.max.request.bytes = -1
queued.max.requests = 500
quota.consumer.default = 9223372036854775807
quota.producer.default = 9223372036854775807
quota.window.num = 11
quota.window.size.seconds = 1
replica.fetch.backoff.ms = 1000
replica.fetch.max.bytes = 1048576
replica.fetch.min.bytes = 1
replica.fetch.response.max.bytes = 10485760
replica.fetch.wait.max.ms = 500
replica.high.watermark.checkpoint.interval.ms = 5000
replica.lag.time.max.ms = 30000
replica.selector.class = null
replica.socket.receive.buffer.bytes = 65536
replica.socket.timeout.ms = 30000
replication.quota.window.num = 11
replication.quota.window.size.seconds = 1
request.timeout.ms = 30000
reserved.broker.max.id = 1000
sasl.client.callback.handler.class = null
sasl.enabled.mechanisms = [GSSAPI]
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.principal.to.local.rules = [DEFAULT]
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism.inter.broker.protocol = GSSAPI
sasl.server.callback.handler.class = null
security.inter.broker.protocol = PLAINTEXT
security.providers = null
socket.receive.buffer.bytes = 102400
socket.request.max.bytes = 104857600
socket.send.buffer.bytes = 102400
ssl.cipher.suites = []
ssl.client.auth = none
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.principal.mapping.rules = DEFAULT
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000
transaction.max.timeout.ms = 900000
transaction.remove.expired.transaction.cleanup.interval.ms = 3600000
transaction.state.log.load.buffer.size = 5242880
transaction.state.log.min.isr = 2
transaction.state.log.num.partitions = 5
transaction.state.log.replication.factor = 3
transaction.state.log.segment.bytes = 104857600
transactional.id.expiration.ms = 604800000
unclean.leader.election.enable = false
zookeeper.clientCnxnSocket = null
zookeeper.connect = 127.0.0.1:56471
zookeeper.connection.timeout.ms = null
zookeeper.max.in.flight.requests = 10
zookeeper.session.timeout.ms = 10000
zookeeper.set.acl = false
zookeeper.ssl.cipher.suites = null
zookeeper.ssl.client.enable = false
zookeeper.ssl.crl.enable = false
zookeeper.ssl.enabled.protocols = null
zookeeper.ssl.endpoint.identification.algorithm = HTTPS
zookeeper.ssl.keystore.location = null
zookeeper.ssl.keystore.password = null
zookeeper.ssl.keystore.type = null
zookeeper.ssl.ocsp.enable = false
zookeeper.ssl.protocol = TLSv1.2
zookeeper.ssl.truststore.location = null
zookeeper.ssl.truststore.password = null
zookeeper.ssl.truststore.type = null
zookeeper.sync.time.ms = 2000
(kafka.server.KafkaConfig:347)
[2020-05-27 17:55:59,299] INFO starting (kafka.server.KafkaServer:66)
[2020-05-27 17:55:59,300] INFO Connecting to zookeeper on 127.0.0.1:56471 (kafka.server.KafkaServer:66)
[2020-05-27 17:55:59,301] INFO [ZooKeeperClient Kafka server] Initializing a new session to 127.0.0.1:56471. (kafka.zookeeper.ZooKeeperClient:66)
[2020-05-27 17:55:59,301] INFO Initiating client connection, connectString=127.0.0.1:56471 sessionTimeout=10000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@29733763 (org.apache.zookeeper.ZooKeeper:868)
[2020-05-27 17:55:59,301] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket:237)
[2020-05-27 17:55:59,302] INFO zookeeper.request.timeout value is 0. feature enabled= (org.apache.zookeeper.ClientCnxn:1653)
[2020-05-27 17:55:59,303] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient:66)
[2020-05-27 17:55:59,304] INFO Opening socket connection to server localhost/127.0.0.1:56471. Will not attempt to authenticate using SASL (unknown error) (org.apache.zookeeper.ClientCnxn:1112)
[2020-05-27 17:55:59,305] INFO Socket connection established, initiating session, client: /127.0.0.1:56473, server: localhost/127.0.0.1:56471 (org.apache.zookeeper.ClientCnxn:959)
[2020-05-27 17:55:59,306] INFO Creating new log file: log.1 (org.apache.zookeeper.server.persistence.FileTxnLog:218)
[2020-05-27 17:55:59,308] INFO Session establishment complete on server localhost/127.0.0.1:56471, sessionid = 0x100592733730000, negotiated timeout = 10000 (org.apache.zookeeper.ClientCnxn:1394)
[2020-05-27 17:55:59,308] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient:66)
[2020-05-27 17:55:59,341] INFO Cluster ID = XNmsyx4kR5in9tlu1-LxwQ (kafka.server.KafkaServer:66)
[2020-05-27 17:55:59,342] WARN No meta.properties file under dir /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/meta.properties (kafka.server.BrokerMetadataCheckpoint:70)
[2020-05-27 17:55:59,350] INFO KafkaConfig values:
advertised.host.name = null
advertised.listeners = null
advertised.port = null
alter.config.policy.class.name = null
alter.log.dirs.replication.quota.window.num = 11
alter.log.dirs.replication.quota.window.size.seconds = 1
authorizer.class.name =
auto.create.topics.enable = true
auto.leader.rebalance.enable = true
background.threads = 10
broker.id = 0
broker.id.generation.enable = true
broker.rack = null
client.quota.callback.class = null
compression.type = producer
connection.failed.authentication.delay.ms = 100
connections.max.idle.ms = 600000
connections.max.reauth.ms = 0
control.plane.listener.name = null
controlled.shutdown.enable = true
controlled.shutdown.max.retries = 3
controlled.shutdown.retry.backoff.ms = 5000
controller.socket.timeout.ms = 30000
create.topic.policy.class.name = null
default.replication.factor = 1
delegation.token.expiry.check.interval.ms = 3600000
delegation.token.expiry.time.ms = 86400000
delegation.token.master.key = null
delegation.token.max.lifetime.ms = 604800000
delete.records.purgatory.purge.interval.requests = 1
delete.topic.enable = true
fetch.max.bytes = 57671680
fetch.purgatory.purge.interval.requests = 1000
group.initial.rebalance.delay.ms = 0
group.max.session.timeout.ms = 1800000
group.max.size = 2147483647
group.min.session.timeout.ms = 0
host.name = localhost
inter.broker.listener.name = null
inter.broker.protocol.version = 2.6-IV0
kafka.metrics.polling.interval.secs = 10
kafka.metrics.reporters = []
leader.imbalance.check.interval.seconds = 300
leader.imbalance.per.broker.percentage = 10
listener.security.protocol.map = PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL
listeners = null
log.cleaner.backoff.ms = 15000
log.cleaner.dedupe.buffer.size = 2097152
log.cleaner.delete.retention.ms = 86400000
log.cleaner.enable = true
log.cleaner.io.buffer.load.factor = 0.9
log.cleaner.io.buffer.size = 524288
log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308
log.cleaner.max.compaction.lag.ms = 9223372036854775807
log.cleaner.min.cleanable.ratio = 0.5
log.cleaner.min.compaction.lag.ms = 0
log.cleaner.threads = 1
log.cleanup.policy = [delete]
log.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770
log.dirs = null
log.flush.interval.messages = 9223372036854775807
log.flush.interval.ms = null
log.flush.offset.checkpoint.interval.ms = 60000
log.flush.scheduler.interval.ms = 9223372036854775807
log.flush.start.offset.checkpoint.interval.ms = 60000
log.index.interval.bytes = 4096
log.index.size.max.bytes = 10485760
log.message.downconversion.enable = true
log.message.format.version = 2.6-IV0
log.message.timestamp.difference.max.ms = 9223372036854775807
log.message.timestamp.type = CreateTime
log.preallocate = false
log.retention.bytes = -1
log.retention.check.interval.ms = 300000
log.retention.hours = 168
log.retention.minutes = null
log.retention.ms = null
log.roll.hours = 168
log.roll.jitter.hours = 0
log.roll.jitter.ms = null
log.roll.ms = null
log.segment.bytes = 1073741824
log.segment.delete.delay.ms = 60000
max.connections = 2147483647
max.connections.per.ip = 2147483647
max.connections.per.ip.overrides =
max.incremental.fetch.session.cache.slots = 1000
message.max.bytes = 1000000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
min.insync.replicas = 1
num.io.threads = 8
num.network.threads = 3
num.partitions = 1
num.recovery.threads.per.data.dir = 1
num.replica.alter.log.dirs.threads = null
num.replica.fetchers = 1
offset.metadata.max.bytes = 4096
offsets.commit.required.acks = -1
offsets.commit.timeout.ms = 5000
offsets.load.buffer.size = 5242880
offsets.retention.check.interval.ms = 600000
offsets.retention.minutes = 10080
offsets.topic.compression.codec = 0
offsets.topic.num.partitions = 5
offsets.topic.replication.factor = 1
offsets.topic.segment.bytes = 104857600
password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding
password.encoder.iterations = 4096
password.encoder.key.length = 128
password.encoder.keyfactory.algorithm = null
password.encoder.old.secret = null
password.encoder.secret = null
port = 0
principal.builder.class = null
producer.purgatory.purge.interval.requests = 1000
queued.max.request.bytes = -1
queued.max.requests = 500
quota.consumer.default = 9223372036854775807
quota.producer.default = 9223372036854775807
quota.window.num = 11
quota.window.size.seconds = 1
replica.fetch.backoff.ms = 1000
replica.fetch.max.bytes = 1048576
replica.fetch.min.bytes = 1
replica.fetch.response.max.bytes = 10485760
replica.fetch.wait.max.ms = 500
replica.high.watermark.checkpoint.interval.ms = 5000
replica.lag.time.max.ms = 30000
replica.selector.class = null
replica.socket.receive.buffer.bytes = 65536
replica.socket.timeout.ms = 30000
replication.quota.window.num = 11
replication.quota.window.size.seconds = 1
request.timeout.ms = 30000
reserved.broker.max.id = 1000
sasl.client.callback.handler.class = null
sasl.enabled.mechanisms = [GSSAPI]
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.principal.to.local.rules = [DEFAULT]
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism.inter.broker.protocol = GSSAPI
sasl.server.callback.handler.class = null
security.inter.broker.protocol = PLAINTEXT
security.providers = null
socket.receive.buffer.bytes = 102400
socket.request.max.bytes = 104857600
socket.send.buffer.bytes = 102400
ssl.cipher.suites = []
ssl.client.auth = none
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.principal.mapping.rules = DEFAULT
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000
transaction.max.timeout.ms = 900000
transaction.remove.expired.transaction.cleanup.interval.ms = 3600000
transaction.state.log.load.buffer.size = 5242880
transaction.state.log.min.isr = 2
transaction.state.log.num.partitions = 5
transaction.state.log.replication.factor = 3
transaction.state.log.segment.bytes = 104857600
transactional.id.expiration.ms = 604800000
unclean.leader.election.enable = false
zookeeper.clientCnxnSocket = null
zookeeper.connect = 127.0.0.1:56471
zookeeper.connection.timeout.ms = null
zookeeper.max.in.flight.requests = 10
zookeeper.session.timeout.ms = 10000
zookeeper.set.acl = false
zookeeper.ssl.cipher.suites = null
zookeeper.ssl.client.enable = false
zookeeper.ssl.crl.enable = false
zookeeper.ssl.enabled.protocols = null
zookeeper.ssl.endpoint.identification.algorithm = HTTPS
zookeeper.ssl.keystore.location = null
zookeeper.ssl.keystore.password = null
zookeeper.ssl.keystore.type = null
zookeeper.ssl.ocsp.enable = false
zookeeper.ssl.protocol = TLSv1.2
zookeeper.ssl.truststore.location = null
zookeeper.ssl.truststore.password = null
zookeeper.ssl.truststore.type = null
zookeeper.sync.time.ms = 2000
(kafka.server.KafkaConfig:347)
[2020-05-27 17:55:59,355] INFO KafkaConfig values:
advertised.host.name = null
advertised.listeners = null
advertised.port = null
alter.config.policy.class.name = null
alter.log.dirs.replication.quota.window.num = 11
alter.log.dirs.replication.quota.window.size.seconds = 1
authorizer.class.name =
auto.create.topics.enable = true
auto.leader.rebalance.enable = true
background.threads = 10
broker.id = 0
broker.id.generation.enable = true
broker.rack = null
client.quota.callback.class = null
compression.type = producer
connection.failed.authentication.delay.ms = 100
connections.max.idle.ms = 600000
connections.max.reauth.ms = 0
control.plane.listener.name = null
controlled.shutdown.enable = true
controlled.shutdown.max.retries = 3
controlled.shutdown.retry.backoff.ms = 5000
controller.socket.timeout.ms = 30000
create.topic.policy.class.name = null
default.replication.factor = 1
delegation.token.expiry.check.interval.ms = 3600000
delegation.token.expiry.time.ms = 86400000
delegation.token.master.key = null
delegation.token.max.lifetime.ms = 604800000
delete.records.purgatory.purge.interval.requests = 1
delete.topic.enable = true
fetch.max.bytes = 57671680
fetch.purgatory.purge.interval.requests = 1000
group.initial.rebalance.delay.ms = 0
group.max.session.timeout.ms = 1800000
group.max.size = 2147483647
group.min.session.timeout.ms = 0
host.name = localhost
inter.broker.listener.name = null
inter.broker.protocol.version = 2.6-IV0
kafka.metrics.polling.interval.secs = 10
kafka.metrics.reporters = []
leader.imbalance.check.interval.seconds = 300
leader.imbalance.per.broker.percentage = 10
listener.security.protocol.map = PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL
listeners = null
log.cleaner.backoff.ms = 15000
log.cleaner.dedupe.buffer.size = 2097152
log.cleaner.delete.retention.ms = 86400000
log.cleaner.enable = true
log.cleaner.io.buffer.load.factor = 0.9
log.cleaner.io.buffer.size = 524288
log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308
log.cleaner.max.compaction.lag.ms = 9223372036854775807
log.cleaner.min.cleanable.ratio = 0.5
log.cleaner.min.compaction.lag.ms = 0
log.cleaner.threads = 1
log.cleanup.policy = [delete]
log.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770
log.dirs = null
log.flush.interval.messages = 9223372036854775807
log.flush.interval.ms = null
log.flush.offset.checkpoint.interval.ms = 60000
log.flush.scheduler.interval.ms = 9223372036854775807
log.flush.start.offset.checkpoint.interval.ms = 60000
log.index.interval.bytes = 4096
log.index.size.max.bytes = 10485760
log.message.downconversion.enable = true
log.message.format.version = 2.6-IV0
log.message.timestamp.difference.max.ms = 9223372036854775807
log.message.timestamp.type = CreateTime
log.preallocate = false
log.retention.bytes = -1
log.retention.check.interval.ms = 300000
log.retention.hours = 168
log.retention.minutes = null
log.retention.ms = null
log.roll.hours = 168
log.roll.jitter.hours = 0
log.roll.jitter.ms = null
log.roll.ms = null
log.segment.bytes = 1073741824
log.segment.delete.delay.ms = 60000
max.connections = 2147483647
max.connections.per.ip = 2147483647
max.connections.per.ip.overrides =
max.incremental.fetch.session.cache.slots = 1000
message.max.bytes = 1000000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
min.insync.replicas = 1
num.io.threads = 8
num.network.threads = 3
num.partitions = 1
num.recovery.threads.per.data.dir = 1
num.replica.alter.log.dirs.threads = null
num.replica.fetchers = 1
offset.metadata.max.bytes = 4096
offsets.commit.required.acks = -1
offsets.commit.timeout.ms = 5000
offsets.load.buffer.size = 5242880
offsets.retention.check.interval.ms = 600000
offsets.retention.minutes = 10080
offsets.topic.compression.codec = 0
offsets.topic.num.partitions = 5
offsets.topic.replication.factor = 1
offsets.topic.segment.bytes = 104857600
password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding
password.encoder.iterations = 4096
password.encoder.key.length = 128
password.encoder.keyfactory.algorithm = null
password.encoder.old.secret = null
password.encoder.secret = null
port = 0
principal.builder.class = null
producer.purgatory.purge.interval.requests = 1000
queued.max.request.bytes = -1
queued.max.requests = 500
quota.consumer.default = 9223372036854775807
quota.producer.default = 9223372036854775807
quota.window.num = 11
quota.window.size.seconds = 1
replica.fetch.backoff.ms = 1000
replica.fetch.max.bytes = 1048576
replica.fetch.min.bytes = 1
replica.fetch.response.max.bytes = 10485760
replica.fetch.wait.max.ms = 500
replica.high.watermark.checkpoint.interval.ms = 5000
replica.lag.time.max.ms = 30000
replica.selector.class = null
replica.socket.receive.buffer.bytes = 65536
replica.socket.timeout.ms = 30000
replication.quota.window.num = 11
replication.quota.window.size.seconds = 1
request.timeout.ms = 30000
reserved.broker.max.id = 1000
sasl.client.callback.handler.class = null
sasl.enabled.mechanisms = [GSSAPI]
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.principal.to.local.rules = [DEFAULT]
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism.inter.broker.protocol = GSSAPI
sasl.server.callback.handler.class = null
security.inter.broker.protocol = PLAINTEXT
security.providers = null
socket.receive.buffer.bytes = 102400
socket.request.max.bytes = 104857600
socket.send.buffer.bytes = 102400
ssl.cipher.suites = []
ssl.client.auth = none
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.principal.mapping.rules = DEFAULT
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000
transaction.max.timeout.ms = 900000
transaction.remove.expired.transaction.cleanup.interval.ms = 3600000
transaction.state.log.load.buffer.size = 5242880
transaction.state.log.min.isr = 2
transaction.state.log.num.partitions = 5
transaction.state.log.replication.factor = 3
transaction.state.log.segment.bytes = 104857600
transactional.id.expiration.ms = 604800000
unclean.leader.election.enable = false
zookeeper.clientCnxnSocket = null
zookeeper.connect = 127.0.0.1:56471
zookeeper.connection.timeout.ms = null
zookeeper.max.in.flight.requests = 10
zookeeper.session.timeout.ms = 10000
zookeeper.set.acl = false
zookeeper.ssl.cipher.suites = null
zookeeper.ssl.client.enable = false
zookeeper.ssl.crl.enable = false
zookeeper.ssl.enabled.protocols = null
zookeeper.ssl.endpoint.identification.algorithm = HTTPS
zookeeper.ssl.keystore.location = null
zookeeper.ssl.keystore.password = null
zookeeper.ssl.keystore.type = null
zookeeper.ssl.ocsp.enable = false
zookeeper.ssl.protocol = TLSv1.2
zookeeper.ssl.truststore.location = null
zookeeper.ssl.truststore.password = null
zookeeper.ssl.truststore.type = null
zookeeper.sync.time.ms = 2000
(kafka.server.KafkaConfig:347)
[2020-05-27 17:55:59,363] INFO [ThrottledChannelReaper-Fetch]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:55:59,364] INFO [ThrottledChannelReaper-Produce]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:55:59,365] INFO [ThrottledChannelReaper-Request]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:55:59,368] INFO Loading logs. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,369] INFO Logs loading complete in 0 ms. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,371] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,371] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,372] INFO Starting the log cleaner (kafka.log.LogCleaner:66)
[2020-05-27 17:55:59,374] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner:66)
[2020-05-27 17:55:59,454] INFO Awaiting socket connections on localhost:56486. (kafka.network.Acceptor:66)
[2020-05-27 17:55:59,467] INFO [SocketServer brokerId=0] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer:66)
[2020-05-27 17:55:59,473] INFO [ExpirationReaper-0-Produce]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:55:59,474] INFO [ExpirationReaper-0-Fetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:55:59,477] INFO [ExpirationReaper-0-DeleteRecords]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:55:59,478] INFO [ExpirationReaper-0-ElectLeader]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:55:59,481] INFO [LogDirFailureHandler]: Starting (kafka.server.ReplicaManager$LogDirFailureHandler:66)
[2020-05-27 17:55:59,487] INFO Creating /brokers/ids/0 (is it secure? false) (kafka.zk.KafkaZkClient:66)
[2020-05-27 17:55:59,491] INFO Stat of the created znode at /brokers/ids/0 is: 24,24,1590627359490,1590627359490,1,0,0,72155618939699200,190,0,24
(kafka.zk.KafkaZkClient:66)
[2020-05-27 17:55:59,492] INFO Registered broker 0 at path /brokers/ids/0 with addresses: PLAINTEXT://localhost:56486, czxid (broker epoch): 24 (kafka.zk.KafkaZkClient:66)
[2020-05-27 17:55:59,495] INFO [ControllerEventThread controllerId=0] Starting (kafka.controller.ControllerEventManager$ControllerEventThread:66)
[2020-05-27 17:55:59,497] INFO [ExpirationReaper-0-topic]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:55:59,498] INFO [ExpirationReaper-0-Heartbeat]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:55:59,499] INFO Successfully created /controller_epoch with initial epoch 0 (kafka.zk.KafkaZkClient:66)
[2020-05-27 17:55:59,499] INFO [ExpirationReaper-0-Rebalance]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:55:59,501] INFO [Controller id=0] 0 successfully elected as the controller. Epoch incremented to 1 and epoch zk version is now 1 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,501] INFO [Controller id=0] Registering handlers (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,502] INFO [GroupCoordinator 0]: Starting up. (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:55:59,503] INFO [Controller id=0] Deleting log dir event notifications (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,503] INFO [GroupCoordinator 0]: Startup complete. (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:55:59,503] INFO [GroupMetadataManager brokerId=0] Removed 0 expired offsets in 0 milliseconds. (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:55:59,504] INFO [Controller id=0] Deleting isr change notifications (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,505] INFO [Controller id=0] Initializing controller context (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,506] INFO [ProducerId Manager 0]: Acquired new producerId block (brokerId:0,blockStartProducerId:0,blockEndProducerId:999) by writing to Zk with path version 1 (kafka.coordinator.transaction.ProducerIdManager:66)
[2020-05-27 17:55:59,507] INFO [Controller id=0] Initialized broker epochs cache: HashMap(0 -> 24) (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,511] INFO [Controller id=0] Currently active brokers in the cluster: Set(0) (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,511] INFO [RequestSendThread controllerId=0] Starting (kafka.controller.RequestSendThread:66)
[2020-05-27 17:55:59,512] INFO [Controller id=0] Currently shutting brokers in the cluster: HashSet() (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,512] INFO [Controller id=0] Current list of topics in the cluster: HashSet() (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,512] INFO [Controller id=0] Fetching topic deletions in progress (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,513] INFO [Controller id=0] List of topics to be deleted: (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,513] INFO [Controller id=0] List of topics ineligible for deletion: (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,513] INFO [Controller id=0] Initializing topic deletion manager (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,513] INFO [Topic Deletion Manager 0] Initializing manager with initial deletions: Set(), initial ineligible deletions: HashSet() (kafka.controller.TopicDeletionManager:66)
[2020-05-27 17:55:59,513] INFO [Controller id=0] Sending update metadata request (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,513] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,514] INFO [ReplicaStateMachine controllerId=0] Initializing replica state (kafka.controller.ZkReplicaStateMachine:66)
[2020-05-27 17:55:59,514] INFO [ReplicaStateMachine controllerId=0] Triggering online replica state changes (kafka.controller.ZkReplicaStateMachine:66)
[2020-05-27 17:55:59,514] INFO [ReplicaStateMachine controllerId=0] Triggering offline replica state changes (kafka.controller.ZkReplicaStateMachine:66)
[2020-05-27 17:55:59,514] INFO [PartitionStateMachine controllerId=0] Initializing partition state (kafka.controller.ZkPartitionStateMachine:66)
[2020-05-27 17:55:59,514] INFO [PartitionStateMachine controllerId=0] Triggering online partition state changes (kafka.controller.ZkPartitionStateMachine:66)
[2020-05-27 17:55:59,514] INFO [Controller id=0] Ready to serve as the new controller with epoch 1 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,515] INFO [RequestSendThread controllerId=0] Controller 0 connected to localhost:56486 (id: 0 rack: null) for sending state change requests (kafka.controller.RequestSendThread:66)
[2020-05-27 17:55:59,516] INFO [Controller id=0] Partitions undergoing preferred replica election: (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,516] INFO [Controller id=0] Partitions that completed preferred replica election: (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,516] INFO [Controller id=0] Skipping preferred replica election for partitions due to topic deletion: (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,517] INFO [Controller id=0] Resuming preferred replica election for partitions: (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,517] INFO [Controller id=0] Starting replica leader election (PREFERRED) for partitions triggered by ZkTriggered (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,518] INFO [Controller id=0] Starting the controller scheduler (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,531] INFO [TransactionCoordinator id=0] Starting up. (kafka.coordinator.transaction.TransactionCoordinator:66)
[2020-05-27 17:55:59,543] INFO [TransactionCoordinator id=0] Startup complete. (kafka.coordinator.transaction.TransactionCoordinator:66)
[2020-05-27 17:55:59,545] INFO [Transaction Marker Channel Manager 0]: Starting (kafka.coordinator.transaction.TransactionMarkerChannelManager:66)
[2020-05-27 17:55:59,547] INFO [ExpirationReaper-0-AlterAcls]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:55:59,550] INFO [/config/changes-event-process-thread]: Starting (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread:66)
[2020-05-27 17:55:59,553] INFO [SocketServer brokerId=0] Starting socket server acceptors and processors (kafka.network.SocketServer:66)
[2020-05-27 17:55:59,555] INFO [SocketServer brokerId=0] Started data-plane acceptor and processor(s) for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer:66)
[2020-05-27 17:55:59,555] INFO [SocketServer brokerId=0] Started socket server acceptors and processors (kafka.network.SocketServer:66)
[2020-05-27 17:55:59,556] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,556] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,556] INFO Kafka startTimeMs: 1590627359287 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,557] INFO [KafkaServer id=0] started (kafka.server.KafkaServer:66)
[2020-05-27 17:55:59,559] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,561] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,561] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,561] INFO Kafka startTimeMs: 1590627359561 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,572] INFO Creating topic topic-1 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:55:59,578] INFO [Controller id=0] New topics: [Set(topic-1)], deleted topics: [HashSet()], new partition replica assignment [Map(topic-1-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,578] INFO [Controller id=0] New partition creation callback for topic-1-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,578] INFO [Controller id=0 epoch=1] Changed partition topic-1-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,579] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,579] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,585] INFO [Controller id=0 epoch=1] Changed partition topic-1-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,585] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:55:59,586] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,586] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,587] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 1 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,588] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(topic-1-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:55:59,588] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 1 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,593] INFO [Log partition=topic-1-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,594] INFO [Log partition=topic-1-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,595] INFO SessionTrackerImpl exited loop! (org.apache.zookeeper.server.SessionTrackerImpl:163)
[2020-05-27 17:55:59,595] INFO Created log for partition topic-1-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/topic-1-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,596] INFO [Partition topic-1-0 broker=0] No checkpointed highwatermark is found for partition topic-1-0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,596] INFO [Partition topic-1-0 broker=0] Log loaded for partition topic-1-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,596] INFO [Broker id=0] Leader topic-1-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,601] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 2 (state.change.logger:66)
[2020-05-27 17:55:59,605] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,607] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,607] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,607] INFO Kafka startTimeMs: 1590627359607 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,616] INFO Creating topic topic-2 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:55:59,622] INFO [Controller id=0] New topics: [Set(topic-2)], deleted topics: [HashSet()], new partition replica assignment [Map(topic-2-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,622] INFO [Controller id=0] New partition creation callback for topic-2-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,622] INFO [Controller id=0 epoch=1] Changed partition topic-2-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,623] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,623] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,628] INFO [Controller id=0 epoch=1] Changed partition topic-2-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,629] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:55:59,629] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,630] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,630] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 3 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,631] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(topic-2-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:55:59,631] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 3 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,636] INFO [Log partition=topic-2-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,637] INFO [Log partition=topic-2-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,638] INFO Created log for partition topic-2-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/topic-2-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,639] INFO [Partition topic-2-0 broker=0] No checkpointed highwatermark is found for partition topic-2-0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,639] INFO [Partition topic-2-0 broker=0] Log loaded for partition topic-2-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,639] INFO [Broker id=0] Leader topic-2-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,643] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 4 (state.change.logger:66)
[2020-05-27 17:55:59,647] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,649] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,649] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,649] INFO Kafka startTimeMs: 1590627359649 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,659] INFO Creating topic topic-A with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:55:59,664] INFO [Controller id=0] New topics: [Set(topic-A)], deleted topics: [HashSet()], new partition replica assignment [Map(topic-A-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,664] INFO [Controller id=0] New partition creation callback for topic-A-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,665] INFO [Controller id=0 epoch=1] Changed partition topic-A-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,665] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,665] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,670] INFO [Controller id=0 epoch=1] Changed partition topic-A-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,670] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:55:59,671] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,671] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,671] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 5 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,672] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(topic-A-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:55:59,672] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 5 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,677] INFO [Log partition=topic-A-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,678] INFO [Log partition=topic-A-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,679] INFO Created log for partition topic-A-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/topic-A-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,679] INFO [Partition topic-A-0 broker=0] No checkpointed highwatermark is found for partition topic-A-0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,679] INFO [Partition topic-A-0 broker=0] Log loaded for partition topic-A-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,679] INFO [Broker id=0] Leader topic-A-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,683] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 6 (state.change.logger:66)
[2020-05-27 17:55:59,686] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,688] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,689] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,689] INFO Kafka startTimeMs: 1590627359688 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,698] INFO Creating topic topic-C with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:55:59,704] INFO [Controller id=0] New topics: [Set(topic-C)], deleted topics: [HashSet()], new partition replica assignment [Map(topic-C-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,704] INFO [Controller id=0] New partition creation callback for topic-C-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,704] INFO [Controller id=0 epoch=1] Changed partition topic-C-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,704] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,705] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,710] INFO [Controller id=0 epoch=1] Changed partition topic-C-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,710] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:55:59,711] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,711] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 7 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,711] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,712] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(topic-C-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:55:59,712] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 7 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,717] INFO [Log partition=topic-C-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,719] INFO [Log partition=topic-C-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,720] INFO Created log for partition topic-C-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/topic-C-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,721] INFO [Partition topic-C-0 broker=0] No checkpointed highwatermark is found for partition topic-C-0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,721] INFO [Partition topic-C-0 broker=0] Log loaded for partition topic-C-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,721] INFO [Broker id=0] Leader topic-C-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,724] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 8 (state.change.logger:66)
[2020-05-27 17:55:59,728] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,730] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,730] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,730] INFO Kafka startTimeMs: 1590627359730 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,739] INFO Creating topic topic-Y with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:55:59,747] INFO [Controller id=0] New topics: [HashSet(topic-Y)], deleted topics: [HashSet()], new partition replica assignment [Map(topic-Y-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,747] INFO [Controller id=0] New partition creation callback for topic-Y-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,747] INFO [Controller id=0 epoch=1] Changed partition topic-Y-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,748] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,748] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,753] INFO [Controller id=0 epoch=1] Changed partition topic-Y-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,754] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:55:59,754] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,755] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,755] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 9 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,756] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(topic-Y-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:55:59,756] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 9 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,761] INFO [Log partition=topic-Y-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,763] INFO [Log partition=topic-Y-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,763] INFO Created log for partition topic-Y-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/topic-Y-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,764] INFO [Partition topic-Y-0 broker=0] No checkpointed highwatermark is found for partition topic-Y-0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,764] INFO [Partition topic-Y-0 broker=0] Log loaded for partition topic-Y-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,764] INFO [Broker id=0] Leader topic-Y-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,768] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 10 (state.change.logger:66)
[2020-05-27 17:55:59,772] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,773] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,774] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,774] INFO Kafka startTimeMs: 1590627359773 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,783] INFO Creating topic topic-Z with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:55:59,789] INFO [Controller id=0] New topics: [HashSet(topic-Z)], deleted topics: [HashSet()], new partition replica assignment [Map(topic-Z-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,789] INFO [Controller id=0] New partition creation callback for topic-Z-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,790] INFO [Controller id=0 epoch=1] Changed partition topic-Z-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,790] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,791] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,796] INFO [Controller id=0 epoch=1] Changed partition topic-Z-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,797] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:55:59,797] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,797] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,798] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 11 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,798] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(topic-Z-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:55:59,799] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 11 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,804] INFO [Log partition=topic-Z-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,805] INFO [Log partition=topic-Z-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,806] INFO Created log for partition topic-Z-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/topic-Z-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,806] INFO [Partition topic-Z-0 broker=0] No checkpointed highwatermark is found for partition topic-Z-0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,807] INFO [Partition topic-Z-0 broker=0] Log loaded for partition topic-Z-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,807] INFO [Broker id=0] Leader topic-Z-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,810] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 12 (state.change.logger:66)
[2020-05-27 17:55:59,829] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,830] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,830] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,831] INFO Kafka startTimeMs: 1590627359830 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,840] INFO Creating topic fa with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:55:59,846] INFO [Controller id=0] New topics: [HashSet(fa)], deleted topics: [HashSet()], new partition replica assignment [Map(fa-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,846] INFO [Controller id=0] New partition creation callback for fa-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,846] INFO [Controller id=0 epoch=1] Changed partition fa-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,846] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,847] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,854] INFO [Controller id=0 epoch=1] Changed partition fa-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,854] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:55:59,854] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,855] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,855] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 13 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,856] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(fa-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:55:59,856] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 13 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,861] INFO [Log partition=fa-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,862] INFO [Log partition=fa-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,863] INFO Created log for partition fa-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/fa-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,863] INFO [Partition fa-0 broker=0] No checkpointed highwatermark is found for partition fa-0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,864] INFO [Partition fa-0 broker=0] Log loaded for partition fa-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,864] INFO [Broker id=0] Leader fa-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,867] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 14 (state.change.logger:66)
[2020-05-27 17:55:59,885] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,886] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,887] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,887] INFO Kafka startTimeMs: 1590627359886 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,896] INFO Creating topic foo with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:55:59,901] INFO [Controller id=0] New topics: [HashSet(foo)], deleted topics: [HashSet()], new partition replica assignment [Map(foo-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,901] INFO [Controller id=0] New partition creation callback for foo-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,902] INFO [Controller id=0 epoch=1] Changed partition foo-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,902] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,903] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,908] INFO [Controller id=0 epoch=1] Changed partition foo-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,908] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:55:59,908] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,909] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,909] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 15 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,910] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(foo-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:55:59,910] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 15 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:55:59,915] INFO [Log partition=foo-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,917] INFO [Log partition=foo-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,917] INFO Created log for partition foo-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/foo-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,918] INFO [Partition foo-0 broker=0] No checkpointed highwatermark is found for partition foo-0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,918] INFO [Partition foo-0 broker=0] Log loaded for partition foo-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,918] INFO [Broker id=0] Leader foo-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,922] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 16 (state.change.logger:66)
[2020-05-27 17:55:59,926] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,937] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,937] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,937] INFO Kafka startTimeMs: 1590627359937 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:55:59,947] INFO Creating topic partitioned-1 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0), 1 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:55:59,954] INFO [Controller id=0] New topics: [HashSet(partitioned-1)], deleted topics: [HashSet()], new partition replica assignment [Map(partitioned-1-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), partitioned-1-1 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,954] INFO [Controller id=0] New partition creation callback for partitioned-1-0,partitioned-1-1 (kafka.controller.KafkaController:66)
[2020-05-27 17:55:59,954] INFO [Controller id=0 epoch=1] Changed partition partitioned-1-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,954] INFO [Controller id=0 epoch=1] Changed partition partitioned-1-1 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:55:59,955] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,955] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,962] INFO [Controller id=0 epoch=1] Changed partition partitioned-1-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,963] INFO [Controller id=0 epoch=1] Changed partition partitioned-1-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:55:59,963] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 2 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:55:59,963] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 2 partitions (state.change.logger:66)
[2020-05-27 17:55:59,964] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:55:59,964] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 17 from controller 0 for 2 partitions (state.change.logger:66)
[2020-05-27 17:55:59,966] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(partitioned-1-0, partitioned-1-1) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:55:59,966] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 17 from controller 0 epoch 1 as part of the become-leader transition for 2 partitions (state.change.logger:66)
[2020-05-27 17:55:59,972] INFO [Log partition=partitioned-1-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,973] INFO [Log partition=partitioned-1-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,974] INFO Created log for partition partitioned-1-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/partitioned-1-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,975] INFO [Partition partitioned-1-0 broker=0] No checkpointed highwatermark is found for partition partitioned-1-0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,975] INFO [Partition partitioned-1-0 broker=0] Log loaded for partition partitioned-1-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,975] INFO [Broker id=0] Leader partitioned-1-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,982] INFO [Log partition=partitioned-1-1, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:55:59,983] INFO [Log partition=partitioned-1-1, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:55:59,984] INFO Created log for partition partitioned-1-1 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/partitioned-1-1 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:55:59,984] INFO [Partition partitioned-1-1 broker=0] No checkpointed highwatermark is found for partition partitioned-1-1 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,984] INFO [Partition partitioned-1-1 broker=0] Log loaded for partition partitioned-1-1 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:55:59,985] INFO [Broker id=0] Leader partitioned-1-1 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:55:59,989] INFO [Broker id=0] Add 2 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 18 (state.change.logger:66)
[2020-05-27 17:55:59,992] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:55:59,994] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:55:59,994] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:55:59,994] INFO Kafka startTimeMs: 1590627359994 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,003] INFO Creating topic partitioned-2 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0), 1 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:56:00,011] INFO [Controller id=0] New topics: [HashSet(partitioned-2)], deleted topics: [HashSet()], new partition replica assignment [Map(partitioned-2-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), partitioned-2-1 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:56:00,011] INFO [Controller id=0] New partition creation callback for partitioned-2-0,partitioned-2-1 (kafka.controller.KafkaController:66)
[2020-05-27 17:56:00,012] INFO [Controller id=0 epoch=1] Changed partition partitioned-2-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,012] INFO [Controller id=0 epoch=1] Changed partition partitioned-2-1 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,012] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,012] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,019] INFO [Controller id=0 epoch=1] Changed partition partitioned-2-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,019] INFO [Controller id=0 epoch=1] Changed partition partitioned-2-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,019] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 2 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:56:00,020] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 2 partitions (state.change.logger:66)
[2020-05-27 17:56:00,020] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,020] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 19 from controller 0 for 2 partitions (state.change.logger:66)
[2020-05-27 17:56:00,022] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(partitioned-2-0, partitioned-2-1) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:56:00,022] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 19 from controller 0 epoch 1 as part of the become-leader transition for 2 partitions (state.change.logger:66)
[2020-05-27 17:56:00,027] INFO [Log partition=partitioned-2-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,028] INFO [Log partition=partitioned-2-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,029] INFO Created log for partition partitioned-2-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/partitioned-2-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,029] INFO [Partition partitioned-2-0 broker=0] No checkpointed highwatermark is found for partition partitioned-2-0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,030] INFO [Partition partitioned-2-0 broker=0] Log loaded for partition partitioned-2-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,030] INFO [Broker id=0] Leader partitioned-2-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,036] INFO [Log partition=partitioned-2-1, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,037] INFO [Log partition=partitioned-2-1, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,038] INFO Created log for partition partitioned-2-1 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/partitioned-2-1 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,038] INFO [Partition partitioned-2-1 broker=0] No checkpointed highwatermark is found for partition partitioned-2-1 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,038] INFO [Partition partitioned-2-1 broker=0] Log loaded for partition partitioned-2-1 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,038] INFO [Broker id=0] Leader partitioned-2-1 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,042] INFO [Broker id=0] Add 2 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 20 (state.change.logger:66)
[2020-05-27 17:56:00,046] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:56:00,048] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,048] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,048] INFO Kafka startTimeMs: 1590627360048 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,057] INFO Creating topic outputTopic_1 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:56:00,063] INFO [Controller id=0] New topics: [HashSet(outputTopic_1)], deleted topics: [HashSet()], new partition replica assignment [Map(outputTopic_1-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:56:00,063] INFO [Controller id=0] New partition creation callback for outputTopic_1-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:56:00,063] INFO [Controller id=0 epoch=1] Changed partition outputTopic_1-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,064] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,064] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,069] INFO [Controller id=0 epoch=1] Changed partition outputTopic_1-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,070] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:56:00,070] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:56:00,071] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,071] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 21 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:56:00,072] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(outputTopic_1-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:56:00,072] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 21 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:56:00,076] INFO [Log partition=outputTopic_1-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,078] INFO [Log partition=outputTopic_1-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,078] INFO Created log for partition outputTopic_1-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/outputTopic_1-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,079] INFO [Partition outputTopic_1-0 broker=0] No checkpointed highwatermark is found for partition outputTopic_1-0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,079] INFO [Partition outputTopic_1-0 broker=0] Log loaded for partition outputTopic_1-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,079] INFO [Broker id=0] Leader outputTopic_1-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,083] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 22 (state.change.logger:66)
[2020-05-27 17:56:00,087] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:56:00,089] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,089] WARN [AdminClient clientId=adminclient-2] Connection to node 0 (localhost/127.0.0.1:53929) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:757)
[2020-05-27 17:56:00,089] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,089] INFO Kafka startTimeMs: 1590627360089 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,112] INFO Creating topic TEST-TOPIC-1 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:56:00,119] INFO [Controller id=0] New topics: [HashSet(TEST-TOPIC-1)], deleted topics: [HashSet()], new partition replica assignment [Map(TEST-TOPIC-1-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:56:00,119] INFO [Controller id=0] New partition creation callback for TEST-TOPIC-1-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:56:00,119] INFO [Controller id=0 epoch=1] Changed partition TEST-TOPIC-1-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,120] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,120] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,125] INFO [Controller id=0 epoch=1] Changed partition TEST-TOPIC-1-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,125] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:56:00,126] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:56:00,126] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 23 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:56:00,126] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,127] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(TEST-TOPIC-1-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:56:00,128] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 23 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:56:00,132] INFO [Log partition=TEST-TOPIC-1-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,134] INFO [Log partition=TEST-TOPIC-1-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,135] INFO Created log for partition TEST-TOPIC-1-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/TEST-TOPIC-1-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,135] INFO [Partition TEST-TOPIC-1-0 broker=0] No checkpointed highwatermark is found for partition TEST-TOPIC-1-0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,135] INFO [Partition TEST-TOPIC-1-0 broker=0] Log loaded for partition TEST-TOPIC-1-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,135] INFO [Broker id=0] Leader TEST-TOPIC-1-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,139] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 24 (state.change.logger:66)
[2020-05-27 17:56:00,143] INFO StreamsConfig values:
acceptable.recovery.lag = 10000
application.id = regex-source-integration-test
application.server =
bootstrap.servers = [localhost:56486]
buffered.records.per.partition = 1000
built.in.metrics.version = latest
cache.max.bytes.buffering = 0
client.id =
commit.interval.ms = 100
connections.max.idle.ms = 540000
default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler
default.key.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler
default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
default.value.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
max.task.idle.ms = 0
max.warmup.replicas = 2
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
num.standby.replicas = 0
num.stream.threads = 1
partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper
poll.ms = 100
probing.rebalance.interval.ms = 600000
processing.guarantee = at_least_once
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
replication.factor = 1
request.timeout.ms = 40000
retries = 0
retry.backoff.ms = 100
rocksdb.config.setter = null
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
state.cleanup.delay.ms = 600000
state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-4331875499181007495
topology.optimization = none
upgrade.from = null
windowstore.changelog.additional.retention.ms = 86400000
(org.apache.kafka.streams.StreamsConfig:347)
[2020-05-27 17:56:00,144] INFO stream-client [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:694)
[2020-05-27 17:56:00,144] INFO stream-client [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:695)
[2020-05-27 17:56:00,145] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id = regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-admin
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:56:00,147] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,147] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,147] INFO Kafka startTimeMs: 1590627360146 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,147] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:304)
[2020-05-27 17:56:00,148] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = none
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-restore-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = null
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:56:00,151] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,151] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,151] INFO Kafka startTimeMs: 1590627360151 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,152] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)
[2020-05-27 17:56:00,152] INFO ProducerConfig values:
acks = 1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-producer
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
linger.ms = 100
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 1000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 10000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:56:00,155] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,159] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,160] INFO Kafka startTimeMs: 1590627360155 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,160] INFO [Producer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-producer] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:56:00,160] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:353)
[2020-05-27 17:56:00,161] INFO ConsumerConfig values:
allow.auto.create.topics = false
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = regex-source-integration-test
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:56:00,163] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:216)
[2020-05-27 17:56:00,165] WARN The configuration 'admin.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:56:00,165] WARN The configuration 'admin.retries' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:56:00,165] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,166] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,166] INFO Kafka startTimeMs: 1590627360165 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,168] INFO stream-client [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:56:00,168] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:502)
[2020-05-27 17:56:00,169] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:56:00,169] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Subscribed to pattern: 'TEST-TOPIC-\d' (org.apache.kafka.clients.consumer.KafkaConsumer:1025)
[2020-05-27 17:56:00,172] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:56:00,173] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,173] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,174] INFO Kafka startTimeMs: 1590627360173 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,179] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:56:00,184] INFO Creating topic __consumer_offsets with configuration {segment.bytes=104857600, compression.type=producer, cleanup.policy=compact} and initial partition assignment HashMap(0 -> ArrayBuffer(0), 1 -> ArrayBuffer(0), 2 -> ArrayBuffer(0), 3 -> ArrayBuffer(0), 4 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:56:00,184] INFO Creating topic TEST-TOPIC-2 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:56:00,188] INFO [KafkaApi-0] Auto creation of topic __consumer_offsets with 5 partitions and replication factor 1 is successful (kafka.server.KafkaApis:66)
[2020-05-27 17:56:00,191] INFO [Controller id=0] New topics: [HashSet(TEST-TOPIC-2, __consumer_offsets)], deleted topics: [HashSet()], new partition replica assignment [HashMap(__consumer_offsets-4 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), TEST-TOPIC-2-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-3 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-2 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-1 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:56:00,191] INFO [Controller id=0] New partition creation callback for __consumer_offsets-4,TEST-TOPIC-2-0,__consumer_offsets-3,__consumer_offsets-2,__consumer_offsets-0,__consumer_offsets-1 (kafka.controller.KafkaController:66)
[2020-05-27 17:56:00,191] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,191] INFO [Controller id=0 epoch=1] Changed partition TEST-TOPIC-2-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,192] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,192] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,192] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,192] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:56:00,192] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,193] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,202] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,202] INFO [Controller id=0 epoch=1] Changed partition TEST-TOPIC-2-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,202] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,203] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,203] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,203] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:56:00,203] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 6 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:56:00,204] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 6 partitions (state.change.logger:66)
[2020-05-27 17:56:00,205] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 25 from controller 0 for 6 partitions (state.change.logger:66)
[2020-05-27 17:56:00,205] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:56:00,207] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions HashSet(__consumer_offsets-4, TEST-TOPIC-2-0, __consumer_offsets-3, __consumer_offsets-2, __consumer_offsets-0, __consumer_offsets-1) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:56:00,207] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 25 from controller 0 epoch 1 as part of the become-leader transition for 6 partitions (state.change.logger:66)
[2020-05-27 17:56:00,212] INFO [Log partition=__consumer_offsets-3, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,213] INFO [Log partition=__consumer_offsets-3, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,214] INFO Created log for partition __consumer_offsets-3 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/__consumer_offsets-3 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.bytes -> 104857600, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,215] INFO [Partition __consumer_offsets-3 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-3 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,215] INFO [Partition __consumer_offsets-3 broker=0] Log loaded for partition __consumer_offsets-3 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,215] INFO [Broker id=0] Leader __consumer_offsets-3 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,221] INFO [Log partition=__consumer_offsets-2, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,222] INFO [Log partition=__consumer_offsets-2, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,223] INFO Created log for partition __consumer_offsets-2 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/__consumer_offsets-2 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.bytes -> 104857600, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,223] INFO [Partition __consumer_offsets-2 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-2 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,224] INFO [Partition __consumer_offsets-2 broker=0] Log loaded for partition __consumer_offsets-2 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,224] INFO [Broker id=0] Leader __consumer_offsets-2 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,229] INFO [Log partition=TEST-TOPIC-2-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,230] INFO [Log partition=TEST-TOPIC-2-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,231] INFO Created log for partition TEST-TOPIC-2-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/TEST-TOPIC-2-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,232] INFO [Partition TEST-TOPIC-2-0 broker=0] No checkpointed highwatermark is found for partition TEST-TOPIC-2-0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,232] INFO [Partition TEST-TOPIC-2-0 broker=0] Log loaded for partition TEST-TOPIC-2-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,232] INFO [Broker id=0] Leader TEST-TOPIC-2-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,239] INFO [Log partition=__consumer_offsets-4, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,241] INFO [Log partition=__consumer_offsets-4, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,242] INFO Created log for partition __consumer_offsets-4 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/__consumer_offsets-4 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.bytes -> 104857600, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,242] INFO [Partition __consumer_offsets-4 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-4 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,242] INFO [Partition __consumer_offsets-4 broker=0] Log loaded for partition __consumer_offsets-4 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,242] INFO [Broker id=0] Leader __consumer_offsets-4 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,248] INFO [Log partition=__consumer_offsets-1, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,249] INFO [Log partition=__consumer_offsets-1, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,250] INFO Created log for partition __consumer_offsets-1 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/__consumer_offsets-1 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.bytes -> 104857600, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,250] INFO [Partition __consumer_offsets-1 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-1 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,250] INFO [Partition __consumer_offsets-1 broker=0] Log loaded for partition __consumer_offsets-1 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,250] INFO [Broker id=0] Leader __consumer_offsets-1 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,255] INFO [Log partition=__consumer_offsets-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:56:00,257] INFO [Log partition=__consumer_offsets-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:56:00,257] INFO Created log for partition __consumer_offsets-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/__consumer_offsets-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.bytes -> 104857600, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:56:00,257] INFO [Partition __consumer_offsets-0 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,258] INFO [Partition __consumer_offsets-0 broker=0] Log loaded for partition __consumer_offsets-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:56:00,258] INFO [Broker id=0] Leader __consumer_offsets-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:56:00,260] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-3 (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,264] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-2 (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,264] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-3 in 4 milliseconds, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,264] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-4 (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,265] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-1 (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,265] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-2 in 1 milliseconds, of which 1 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,265] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-0 (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,265] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-4 in 0 milliseconds, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,265] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-1 in 0 milliseconds, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,266] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-0 in 1 milliseconds, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager:66)
[2020-05-27 17:56:00,267] INFO [Broker id=0] Add 6 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 26 (state.change.logger:66)
[2020-05-27 17:56:00,272] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-6
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:56:00,273] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:56:00,274] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:56:00,276] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,278] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,278] INFO Kafka startTimeMs: 1590627360276 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,279] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)
[2020-05-27 17:56:00,279] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:56:00,280] INFO [Producer clientId=producer-6] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:56:00,281] INFO [Producer clientId=producer-6] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:56:00,281] INFO [GroupCoordinator 0]: Preparing to rebalance group regex-source-integration-test in state PreparingRebalance with old generation 0 (__consumer_offsets-3) (reason: Adding new member regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer-b1d82918-6566-47df-82fc-c08dd2386041 with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:00,283] INFO [GroupCoordinator 0]: Stabilized group regex-source-integration-test generation 1 (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:00,286] INFO Decided on assignment: {c944b5f3-0d98-4e12-b417-89262da23fd7=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)
[2020-05-27 17:56:00,286] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer] Assigned tasks to clients as
c944b5f3-0d98-4e12-b417-89262da23fd7=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:724)
[2020-05-27 17:56:00,287] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:938)
[2020-05-27 17:56:00,287] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Finished assignment for group at generation 1: {regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer-b1d82918-6566-47df-82fc-c08dd2386041=Assignment(partitions=[TEST-TOPIC-1-0], userDataSize=48)} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:56:00,288] INFO [GroupCoordinator 0]: Assignment received from leader for group regex-source-integration-test for generation 1 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:00,291] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-7
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:56:00,292] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Successfully joined group with generation 1 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:56:00,296] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,297] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,297] INFO Kafka startTimeMs: 1590627360296 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,298] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Updating assignment with
Assigned partitions: [TEST-TOPIC-1-0]
Current owned partitions: []
Added partitions (assigned - owned): [TEST-TOPIC-1-0]
Revoked partitions (owned - assigned): []
(org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)
[2020-05-27 17:56:00,298] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Notifying assignor about the new Assignment(partitions=[TEST-TOPIC-1-0], userDataSize=48) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:56:00,298] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1474)
[2020-05-27 17:56:00,298] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Handle new assignment with:
New active tasks: [0_0]
New standby tasks: []
Existing active tasks: []
Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:176)
[2020-05-27 17:56:00,301] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Adding newly assigned partitions: TEST-TOPIC-1-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:56:00,301] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:56:00,301] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:56:00,302] INFO [GroupCoordinator 0]: Preparing to rebalance group regex-source-integration-test in state PreparingRebalance with old generation 1 (__consumer_offsets-3) (reason: Updating metadata for member regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer-b1d82918-6566-47df-82fc-c08dd2386041) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:00,303] INFO [GroupCoordinator 0]: Stabilized group regex-source-integration-test generation 2 (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:00,303] INFO [Producer clientId=producer-7] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:56:00,304] INFO [Producer clientId=producer-7] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:56:00,305] INFO Decided on assignment: {c944b5f3-0d98-4e12-b417-89262da23fd7=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([0_0]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([TEST-TOPIC-1-0]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)
[2020-05-27 17:56:00,305] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer] Assigned tasks to clients as
c944b5f3-0d98-4e12-b417-89262da23fd7=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([0_0]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([TEST-TOPIC-1-0]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:724)
[2020-05-27 17:56:00,306] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:938)
[2020-05-27 17:56:00,306] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Finished assignment for group at generation 2: {regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer-b1d82918-6566-47df-82fc-c08dd2386041=Assignment(partitions=[TEST-TOPIC-1-0, TEST-TOPIC-2-0], userDataSize=56)} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:56:00,306] INFO [GroupCoordinator 0]: Assignment received from leader for group regex-source-integration-test for generation 2 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:00,309] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Successfully joined group with generation 2 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:56:00,310] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Updating assignment with
Assigned partitions: [TEST-TOPIC-2-0, TEST-TOPIC-1-0]
Current owned partitions: [TEST-TOPIC-1-0]
Added partitions (assigned - owned): [TEST-TOPIC-2-0]
Revoked partitions (owned - assigned): []
(org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)
[2020-05-27 17:56:00,310] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Notifying assignor about the new Assignment(partitions=[TEST-TOPIC-2-0, TEST-TOPIC-1-0], userDataSize=56) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:56:00,310] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1474)
[2020-05-27 17:56:00,310] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Handle new assignment with:
New active tasks: [0_0]
New standby tasks: []
Existing active tasks: [0_0]
Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:176)
[2020-05-27 17:56:00,312] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = true
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = 43954268-6dd4-4885-bcea-acd4c744a7eb
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = true
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 500
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:56:00,312] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Adding newly assigned partitions: TEST-TOPIC-2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:56:00,312] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] State transition from PARTITIONS_ASSIGNED to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:56:00,314] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition TEST-TOPIC-2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:56:00,314] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition TEST-TOPIC-1-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:56:00,314] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:56:00,315] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:56:00,315] INFO Kafka startTimeMs: 1590627360314 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:56:00,316] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition TEST-TOPIC-2-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:56:00,316] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition TEST-TOPIC-1-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:56:00,318] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Subscribed to topic(s): outputTopic_1 (org.apache.kafka.clients.consumer.KafkaConsumer:961)
[2020-05-27 17:56:00,323] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:56:00,324] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:56:00,326] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:56:00,329] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)
[2020-05-27 17:56:00,330] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:56:00,331] INFO [GroupCoordinator 0]: Preparing to rebalance group 43954268-6dd4-4885-bcea-acd4c744a7eb in state PreparingRebalance with old generation 0 (__consumer_offsets-4) (reason: Adding new member consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1-9d54692a-b002-40d1-8025-258fa7cfa7dd with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:00,332] INFO [GroupCoordinator 0]: Stabilized group 43954268-6dd4-4885-bcea-acd4c744a7eb generation 1 (__consumer_offsets-4) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:00,335] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Finished assignment for group at generation 1: {consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1-9d54692a-b002-40d1-8025-258fa7cfa7dd=Assignment(partitions=[outputTopic_1-0])} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:56:00,336] INFO [GroupCoordinator 0]: Assignment received from leader for group 43954268-6dd4-4885-bcea-acd4c744a7eb for generation 1 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:00,338] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Successfully joined group with generation 1 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:56:00,343] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Notifying assignor about the new Assignment(partitions=[outputTopic_1-0]) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:56:00,343] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Adding newly assigned partitions: outputTopic_1-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:56:00,345] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Found no committed offset for partition outputTopic_1-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:56:00,350] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Resetting offset for partition outputTopic_1-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:56:00,374] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] task [0_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:212)
[2020-05-27 17:56:00,375] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition TEST-TOPIC-2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:56:00,376] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] task [0_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:229)
[2020-05-27 17:56:00,376] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:56:00,376] INFO stream-client [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:56:00,413] ERROR Exception caught during Deserialization, taskId: 0_0, topic: TEST-TOPIC-1, partition: 0, offset: 0 (org.apache.kafka.streams.errors.LogAndFailExceptionHandler:39)
java.lang.NullPointerException
at org.apache.kafka.streams.processor.internals.SourceNode.deserializeKey(SourceNode.java:58)
at org.apache.kafka.streams.processor.internals.RecordDeserializer.deserialize(RecordDeserializer.java:65)
at org.apache.kafka.streams.processor.internals.RecordQueue.updateHead(RecordQueue.java:176)
at org.apache.kafka.streams.processor.internals.RecordQueue.addRawRecords(RecordQueue.java:112)
at org.apache.kafka.streams.processor.internals.PartitionGroup.addRawRecords(PartitionGroup.java:185)
at org.apache.kafka.streams.processor.internals.StreamTask.addRecords(StreamTask.java:844)
at org.apache.kafka.streams.processor.internals.StreamThread.addRecordsToTasks(StreamThread.java:836)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:646)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:550)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:509)
[2020-05-27 17:56:00,419] ERROR stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Encountered the following exception during processing and the thread is going to shut down: (org.apache.kafka.streams.processor.internals.StreamThread:529)
org.apache.kafka.streams.errors.StreamsException: Deserialization exception handler is set to fail upon a deserialization error. If you would rather have the streaming pipeline continue after a deserialization error, please set the default.deserialization.exception.handler appropriately.
at org.apache.kafka.streams.processor.internals.RecordDeserializer.deserialize(RecordDeserializer.java:80)
at org.apache.kafka.streams.processor.internals.RecordQueue.updateHead(RecordQueue.java:176)
at org.apache.kafka.streams.processor.internals.RecordQueue.addRawRecords(RecordQueue.java:112)
at org.apache.kafka.streams.processor.internals.PartitionGroup.addRawRecords(PartitionGroup.java:185)
at org.apache.kafka.streams.processor.internals.StreamTask.addRecords(StreamTask.java:844)
at org.apache.kafka.streams.processor.internals.StreamThread.addRecordsToTasks(StreamThread.java:836)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:646)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:550)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:509)
Caused by: java.lang.NullPointerException
at org.apache.kafka.streams.processor.internals.SourceNode.deserializeKey(SourceNode.java:58)
at org.apache.kafka.streams.processor.internals.RecordDeserializer.deserialize(RecordDeserializer.java:65)
... 8 more
[2020-05-27 17:56:00,420] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:56:00,420] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:915)
[2020-05-27 17:56:00,421] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] task [0_0] Prepared dirty close (org.apache.kafka.streams.processor.internals.StreamTask:435)
[2020-05-27 17:56:00,421] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:56:00,421] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] task [0_0] Closed dirty (org.apache.kafka.streams.processor.internals.StreamTask:442)
[2020-05-27 17:56:00,421] INFO [Producer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:56:00,423] INFO [Consumer clientId=regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:56:00,426] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:56:00,426] INFO stream-client [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7] State transition from RUNNING to ERROR (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:56:00,427] ERROR stream-client [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7] All stream threads have died. The instance will be in error state and should be closed. (org.apache.kafka.streams.KafkaStreams:439)
[2020-05-27 17:56:00,427] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:940)
[2020-05-27 17:56:00,427] ERROR Thread StreamsThread threadId: regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1
TaskManager
MetadataState:
Tasks:
died (org.apache.zookeeper.server.NIOServerCnxnFactory:92)
org.apache.kafka.streams.errors.StreamsException: Deserialization exception handler is set to fail upon a deserialization error. If you would rather have the streaming pipeline continue after a deserialization error, please set the default.deserialization.exception.handler appropriately.
at org.apache.kafka.streams.processor.internals.RecordDeserializer.deserialize(RecordDeserializer.java:80)
at org.apache.kafka.streams.processor.internals.RecordQueue.updateHead(RecordQueue.java:176)
at org.apache.kafka.streams.processor.internals.RecordQueue.addRawRecords(RecordQueue.java:112)
at org.apache.kafka.streams.processor.internals.PartitionGroup.addRawRecords(PartitionGroup.java:185)
at org.apache.kafka.streams.processor.internals.StreamTask.addRecords(StreamTask.java:844)
at org.apache.kafka.streams.processor.internals.StreamThread.addRecordsToTasks(StreamThread.java:836)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:646)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:550)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:509)
Caused by: java.lang.NullPointerException
at org.apache.kafka.streams.processor.internals.SourceNode.deserializeKey(SourceNode.java:58)
at org.apache.kafka.streams.processor.internals.RecordDeserializer.deserialize(RecordDeserializer.java:65)
... 8 more
[2020-05-27 17:56:01,114] WARN [AdminClient clientId=adminclient-2] Connection to node 0 (localhost/127.0.0.1:53929) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient:757)
[2020-05-27 17:56:02,029] INFO [AdminClient clientId=adminclient-2] Metadata update failed (org.apache.kafka.clients.admin.internals.AdminMetadataManager:235)
org.apache.kafka.common.errors.TimeoutException: Call(callName=fetchMetadata, deadlineMs=1590627362028, tries=1, nextAllowedTryMs=1590627362129) timed out at 1590627362029 after 1 attempt(s)
Caused by: org.apache.kafka.common.errors.TimeoutException: Timed out waiting for a node assignment.
[2020-05-27 17:56:04,521] INFO [Controller id=0] Processing automatic preferred replica leader election (kafka.controller.KafkaController:66)
[2020-05-27 17:56:10,311] INFO [GroupCoordinator 0]: Member regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer-b1d82918-6566-47df-82fc-c08dd2386041 in group regex-source-integration-test has failed, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:10,312] INFO [GroupCoordinator 0]: Preparing to rebalance group regex-source-integration-test in state PreparingRebalance with old generation 2 (__consumer_offsets-3) (reason: removing member regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1-consumer-b1d82918-6566-47df-82fc-c08dd2386041 on heartbeat expiration) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:56:10,313] INFO [GroupCoordinator 0]: Group regex-source-integration-test with generation 3 is now empty (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,373] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Revoke previously assigned partitions outputTopic_1-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:307)
[2020-05-27 17:57:01,374] INFO [Consumer clientId=consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1, groupId=43954268-6dd4-4885-bcea-acd4c744a7eb] Member consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1-9d54692a-b002-40d1-8025-258fa7cfa7dd sending LeaveGroup request to coordinator localhost:56486 (id: 2147483647 rack: null) due to the consumer is being closed (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:1005)
[2020-05-27 17:57:01,379] INFO [GroupCoordinator 0]: Member[group.instance.id None, member.id consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1-9d54692a-b002-40d1-8025-258fa7cfa7dd] in group 43954268-6dd4-4885-bcea-acd4c744a7eb has left, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,379] INFO [GroupCoordinator 0]: Preparing to rebalance group 43954268-6dd4-4885-bcea-acd4c744a7eb in state PreparingRebalance with old generation 1 (__consumer_offsets-4) (reason: removing member consumer-43954268-6dd4-4885-bcea-acd4c744a7eb-1-9d54692a-b002-40d1-8025-258fa7cfa7dd on LeaveGroup) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,379] INFO [GroupCoordinator 0]: Group 43954268-6dd4-4885-bcea-acd4c744a7eb with generation 2 is now empty (__consumer_offsets-4) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,386] INFO stream-client [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7] State transition from ERROR to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:01,387] INFO stream-thread [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:901)
[2020-05-27 17:57:01,389] INFO stream-client [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:01,389] INFO stream-client [regex-source-integration-test-c944b5f3-0d98-4e12-b417-89262da23fd7] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:986)
[2020-05-27 17:57:01,397] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:01,400] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,400] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,401] INFO Kafka startTimeMs: 1590627421400 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,412] INFO Creating topic outputTopic_2 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:57:01,420] INFO [Controller id=0] New topics: [HashSet(outputTopic_2)], deleted topics: [HashSet()], new partition replica assignment [Map(outputTopic_2-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:01,421] INFO [Controller id=0] New partition creation callback for outputTopic_2-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:01,421] INFO [Controller id=0 epoch=1] Changed partition outputTopic_2-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:57:01,421] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:01,422] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:01,428] INFO [Controller id=0 epoch=1] Changed partition outputTopic_2-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:57:01,428] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:57:01,429] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:01,429] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:01,429] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 27 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:01,430] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(outputTopic_2-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:01,430] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 27 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:01,439] INFO [Log partition=outputTopic_2-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:57:01,441] INFO [Log partition=outputTopic_2-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:57:01,442] INFO Created log for partition outputTopic_2-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/outputTopic_2-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:57:01,443] INFO [Partition outputTopic_2-0 broker=0] No checkpointed highwatermark is found for partition outputTopic_2-0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:01,443] INFO [Partition outputTopic_2-0 broker=0] Log loaded for partition outputTopic_2-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:01,443] INFO [Broker id=0] Leader outputTopic_2-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:57:01,448] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 28 (state.change.logger:66)
[2020-05-27 17:57:01,453] INFO StreamsConfig values:
acceptable.recovery.lag = 10000
application.id = regex-source-integration-test
application.server =
bootstrap.servers = [localhost:56486]
buffered.records.per.partition = 1000
built.in.metrics.version = latest
cache.max.bytes.buffering = 0
client.id =
commit.interval.ms = 100
connections.max.idle.ms = 540000
default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler
default.key.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler
default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
default.value.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
max.task.idle.ms = 0
max.warmup.replicas = 2
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
num.standby.replicas = 0
num.stream.threads = 1
partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper
poll.ms = 100
probing.rebalance.interval.ms = 600000
processing.guarantee = at_least_once
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
replication.factor = 1
request.timeout.ms = 40000
retries = 0
retry.backoff.ms = 100
rocksdb.config.setter = null
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
state.cleanup.delay.ms = 600000
state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-2740709879018537368
topology.optimization = none
upgrade.from = null
windowstore.changelog.additional.retention.ms = 86400000
(org.apache.kafka.streams.StreamsConfig:347)
[2020-05-27 17:57:01,455] INFO stream-client [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:694)
[2020-05-27 17:57:01,455] INFO stream-client [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:695)
[2020-05-27 17:57:01,456] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id = regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-admin
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:01,459] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,460] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,460] INFO Kafka startTimeMs: 1590627421459 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,461] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:304)
[2020-05-27 17:57:01,461] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = none
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-restore-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = null
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:01,465] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,465] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,466] INFO Kafka startTimeMs: 1590627421465 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,466] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)
[2020-05-27 17:57:01,467] INFO ProducerConfig values:
acks = 1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-producer
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
linger.ms = 100
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 1000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 10000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:01,471] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,472] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,472] INFO Kafka startTimeMs: 1590627421471 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,473] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:353)
[2020-05-27 17:57:01,474] INFO ConsumerConfig values:
allow.auto.create.topics = false
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = regex-source-integration-test
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:01,475] INFO [Producer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-producer] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,477] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:216)
[2020-05-27 17:57:01,479] WARN The configuration 'admin.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:01,479] WARN The configuration 'admin.retries' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:01,480] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,480] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,480] INFO Kafka startTimeMs: 1590627421480 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,482] INFO stream-client [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:01,483] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:502)
[2020-05-27 17:57:01,483] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:01,484] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Subscribed to pattern: 'topic-Y|topic-Z|topic-\d|topic-[A-D]' (org.apache.kafka.clients.consumer.KafkaConsumer:1025)
[2020-05-27 17:57:01,486] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-8
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:01,488] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,489] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,490] INFO Kafka startTimeMs: 1590627421488 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,493] INFO [Producer clientId=producer-8] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,493] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,494] INFO [Producer clientId=producer-8] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:01,495] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:57:01,502] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-9
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:01,502] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:01,505] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)
[2020-05-27 17:57:01,505] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:01,506] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,507] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,508] INFO Kafka startTimeMs: 1590627421506 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,509] INFO [GroupCoordinator 0]: Preparing to rebalance group regex-source-integration-test in state PreparingRebalance with old generation 3 (__consumer_offsets-3) (reason: Adding new member regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer-8f66bcdb-f6a2-4dc0-8bc3-71ac95a81b42 with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,510] INFO [GroupCoordinator 0]: Stabilized group regex-source-integration-test generation 4 (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,512] INFO [Producer clientId=producer-9] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,512] INFO Decided on assignment: {d425dc29-c09e-42a0-aba8-9919a5968e38=[activeTasks: ([0_0, 1_0, 2_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 3]} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)
[2020-05-27 17:57:01,513] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer] Assigned tasks to clients as
d425dc29-c09e-42a0-aba8-9919a5968e38=[activeTasks: ([0_0, 1_0, 2_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 3]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:724)
[2020-05-27 17:57:01,513] INFO [Producer clientId=producer-9] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:01,514] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:938)
[2020-05-27 17:57:01,514] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Finished assignment for group at generation 4: {regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer-8f66bcdb-f6a2-4dc0-8bc3-71ac95a81b42=Assignment(partitions=[topic-1-0, topic-2-0, topic-A-0, topic-C-0, topic-Y-0, topic-Z-0], userDataSize=88)} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:57:01,515] INFO [GroupCoordinator 0]: Assignment received from leader for group regex-source-integration-test for generation 4 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,517] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Successfully joined group with generation 4 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:57:01,520] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Updating assignment with
Assigned partitions: [topic-Y-0, topic-Z-0, topic-A-0, topic-1-0, topic-2-0, topic-C-0]
Current owned partitions: []
Added partitions (assigned - owned): [topic-Y-0, topic-Z-0, topic-A-0, topic-1-0, topic-2-0, topic-C-0]
Revoked partitions (owned - assigned): []
(org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)
[2020-05-27 17:57:01,521] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Notifying assignor about the new Assignment(partitions=[topic-Y-0, topic-Z-0, topic-A-0, topic-1-0, topic-2-0, topic-C-0], userDataSize=88) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:57:01,521] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1474)
[2020-05-27 17:57:01,521] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] Handle new assignment with:
New active tasks: [0_0, 1_0, 2_0]
New standby tasks: []
Existing active tasks: []
Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:176)
[2020-05-27 17:57:01,525] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-10
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:01,527] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Adding newly assigned partitions: topic-Y-0, topic-Z-0, topic-A-0, topic-1-0, topic-2-0, topic-C-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:57:01,527] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:01,528] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,529] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,529] INFO Kafka startTimeMs: 1590627421528 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,532] INFO [Producer clientId=producer-10] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,534] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-Y-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,534] INFO [Producer clientId=producer-10] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:01,534] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-Z-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,535] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-1-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,535] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-A-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,536] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,536] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-C-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,538] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition topic-Y-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:01,539] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition topic-Z-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:01,539] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition topic-1-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:01,539] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition topic-A-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:01,539] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition topic-2-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:01,539] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition topic-C-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:01,541] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-11
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:01,545] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,546] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,547] INFO Kafka startTimeMs: 1590627421545 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,550] INFO [Producer clientId=producer-11] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,551] INFO [Producer clientId=producer-11] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:01,559] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-12
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:01,562] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,563] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,563] INFO Kafka startTimeMs: 1590627421562 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,566] INFO [Producer clientId=producer-12] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,567] INFO [Producer clientId=producer-12] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:01,573] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-13
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:01,577] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,578] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,578] INFO Kafka startTimeMs: 1590627421577 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,581] INFO [Producer clientId=producer-13] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,582] INFO [Producer clientId=producer-13] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:01,590] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [0_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:212)
[2020-05-27 17:57:01,590] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [1_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:212)
[2020-05-27 17:57:01,590] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = true
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = 428be263-5641-411b-9a0a-ba94c1728c3a
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = true
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 500
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:01,591] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [2_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:212)
[2020-05-27 17:57:01,592] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-1-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,593] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,593] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [0_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:229)
[2020-05-27 17:57:01,594] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-A-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,594] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,595] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-C-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,595] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,595] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [1_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:229)
[2020-05-27 17:57:01,595] INFO Kafka startTimeMs: 1590627421594 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,596] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-Y-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,597] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition topic-Z-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,597] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [2_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:229)
[2020-05-27 17:57:01,597] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:01,598] INFO stream-client [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:01,599] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Subscribed to topic(s): outputTopic_2 (org.apache.kafka.clients.consumer.KafkaConsumer:961)
[2020-05-27 17:57:01,604] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,604] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:57:01,605] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:01,611] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)
[2020-05-27 17:57:01,611] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:01,612] INFO [GroupCoordinator 0]: Preparing to rebalance group 428be263-5641-411b-9a0a-ba94c1728c3a in state PreparingRebalance with old generation 0 (__consumer_offsets-4) (reason: Adding new member consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2-deb3d8f5-b931-4693-bab1-3160e156c9e6 with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,614] INFO [GroupCoordinator 0]: Stabilized group 428be263-5641-411b-9a0a-ba94c1728c3a generation 1 (__consumer_offsets-4) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,615] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Finished assignment for group at generation 1: {consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2-deb3d8f5-b931-4693-bab1-3160e156c9e6=Assignment(partitions=[outputTopic_2-0])} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:57:01,617] INFO [GroupCoordinator 0]: Assignment received from leader for group 428be263-5641-411b-9a0a-ba94c1728c3a for generation 1 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,619] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Successfully joined group with generation 1 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:57:01,622] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Notifying assignor about the new Assignment(partitions=[outputTopic_2-0]) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:57:01,622] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Adding newly assigned partitions: outputTopic_2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:57:01,624] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Found no committed offset for partition outputTopic_2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:01,628] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Resetting offset for partition outputTopic_2-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:01,730] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Revoke previously assigned partitions outputTopic_2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:307)
[2020-05-27 17:57:01,731] INFO [Consumer clientId=consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2, groupId=428be263-5641-411b-9a0a-ba94c1728c3a] Member consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2-deb3d8f5-b931-4693-bab1-3160e156c9e6 sending LeaveGroup request to coordinator localhost:56486 (id: 2147483647 rack: null) due to the consumer is being closed (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:1005)
[2020-05-27 17:57:01,731] INFO [GroupCoordinator 0]: Member[group.instance.id None, member.id consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2-deb3d8f5-b931-4693-bab1-3160e156c9e6] in group 428be263-5641-411b-9a0a-ba94c1728c3a has left, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,732] INFO [GroupCoordinator 0]: Preparing to rebalance group 428be263-5641-411b-9a0a-ba94c1728c3a in state PreparingRebalance with old generation 1 (__consumer_offsets-4) (reason: removing member consumer-428be263-5641-411b-9a0a-ba94c1728c3a-2-deb3d8f5-b931-4693-bab1-3160e156c9e6 on LeaveGroup) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,732] INFO [GroupCoordinator 0]: Group 428be263-5641-411b-9a0a-ba94c1728c3a with generation 2 is now empty (__consumer_offsets-4) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:01,735] INFO stream-client [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:01,736] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:901)
[2020-05-27 17:57:01,736] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:01,834] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:915)
[2020-05-27 17:57:01,834] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [0_0] Prepared clean close (org.apache.kafka.streams.processor.internals.StreamTask:419)
[2020-05-27 17:57:01,835] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [1_0] Prepared clean close (org.apache.kafka.streams.processor.internals.StreamTask:419)
[2020-05-27 17:57:01,835] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [2_0] Prepared clean close (org.apache.kafka.streams.processor.internals.StreamTask:419)
[2020-05-27 17:57:01,851] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:01,851] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [1_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:428)
[2020-05-27 17:57:01,851] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:01,851] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [2_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:428)
[2020-05-27 17:57:01,852] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:01,852] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] task [0_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:428)
[2020-05-27 17:57:01,852] INFO [Producer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:01,854] INFO [Consumer clientId=regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:01,856] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:01,857] INFO stream-thread [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:940)
[2020-05-27 17:57:01,858] INFO stream-client [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:01,858] INFO stream-client [regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:986)
[2020-05-27 17:57:01,860] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:01,861] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,862] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,862] INFO Kafka startTimeMs: 1590627421861 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,870] INFO Creating topic outputTopic_3 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:57:01,876] INFO [Controller id=0] New topics: [HashSet(outputTopic_3)], deleted topics: [HashSet()], new partition replica assignment [Map(outputTopic_3-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:01,877] INFO [Controller id=0] New partition creation callback for outputTopic_3-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:01,877] INFO [Controller id=0 epoch=1] Changed partition outputTopic_3-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:57:01,877] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:01,878] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:01,884] INFO [Controller id=0 epoch=1] Changed partition outputTopic_3-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:57:01,885] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:57:01,885] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:01,885] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:01,886] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 29 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:01,886] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(outputTopic_3-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:01,886] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 29 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:01,890] INFO [Log partition=outputTopic_3-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:57:01,891] INFO [Log partition=outputTopic_3-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:57:01,892] INFO Created log for partition outputTopic_3-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/outputTopic_3-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:57:01,893] INFO [Partition outputTopic_3-0 broker=0] No checkpointed highwatermark is found for partition outputTopic_3-0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:01,893] INFO [Partition outputTopic_3-0 broker=0] Log loaded for partition outputTopic_3-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:01,893] INFO [Broker id=0] Leader outputTopic_3-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:57:01,896] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 30 (state.change.logger:66)
[2020-05-27 17:57:01,900] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:01,902] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,902] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,902] INFO Kafka startTimeMs: 1590627421902 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,914] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:01,916] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,917] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,917] INFO Kafka startTimeMs: 1590627421916 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,925] INFO Creating topic outputTopic_4 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:57:01,931] INFO [Controller id=0] New topics: [HashSet(outputTopic_4)], deleted topics: [HashSet()], new partition replica assignment [Map(outputTopic_4-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:01,931] INFO [Controller id=0] New partition creation callback for outputTopic_4-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:01,932] INFO [Controller id=0 epoch=1] Changed partition outputTopic_4-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:57:01,932] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:01,932] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:01,937] INFO [Controller id=0 epoch=1] Changed partition outputTopic_4-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:57:01,938] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:57:01,938] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:01,939] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:01,939] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 31 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:01,940] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(outputTopic_4-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:01,940] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 31 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:01,947] INFO [Log partition=outputTopic_4-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:57:01,949] INFO [Log partition=outputTopic_4-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:57:01,950] INFO Created log for partition outputTopic_4-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/outputTopic_4-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:57:01,951] INFO [Partition outputTopic_4-0 broker=0] No checkpointed highwatermark is found for partition outputTopic_4-0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:01,951] INFO [Partition outputTopic_4-0 broker=0] Log loaded for partition outputTopic_4-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:01,951] INFO [Broker id=0] Leader outputTopic_4-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:57:01,955] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 32 (state.change.logger:66)
[2020-05-27 17:57:01,962] INFO StreamsConfig values:
acceptable.recovery.lag = 10000
application.id = regex-source-integration-test
application.server =
bootstrap.servers = [localhost:56486]
buffered.records.per.partition = 1000
built.in.metrics.version = latest
cache.max.bytes.buffering = 0
client.id =
commit.interval.ms = 100
connections.max.idle.ms = 540000
default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler
default.key.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler
default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
default.value.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
max.task.idle.ms = 0
max.warmup.replicas = 2
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
num.standby.replicas = 0
num.stream.threads = 1
partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper
poll.ms = 100
probing.rebalance.interval.ms = 600000
processing.guarantee = at_least_once
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
replication.factor = 1
request.timeout.ms = 40000
retries = 0
retry.backoff.ms = 100
rocksdb.config.setter = null
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
state.cleanup.delay.ms = 600000
state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-5154128249393238509
topology.optimization = none
upgrade.from = null
windowstore.changelog.additional.retention.ms = 86400000
(org.apache.kafka.streams.StreamsConfig:347)
[2020-05-27 17:57:01,963] INFO stream-client [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:694)
[2020-05-27 17:57:01,963] INFO stream-client [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:695)
[2020-05-27 17:57:01,964] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id = regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-admin
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:01,966] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,966] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,967] INFO Kafka startTimeMs: 1590627421966 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,967] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:304)
[2020-05-27 17:57:01,968] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = none
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-restore-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = null
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:01,971] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,972] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,972] INFO Kafka startTimeMs: 1590627421971 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,972] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)
[2020-05-27 17:57:01,973] INFO ProducerConfig values:
acks = 1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-producer
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
linger.ms = 100
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 1000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 10000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:01,977] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,977] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,978] INFO Kafka startTimeMs: 1590627421976 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,978] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:353)
[2020-05-27 17:57:01,982] INFO [Producer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-producer] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:01,982] INFO ConsumerConfig values:
allow.auto.create.topics = false
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = regex-source-integration-test
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:01,985] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:216)
[2020-05-27 17:57:01,987] WARN The configuration 'admin.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:01,988] WARN The configuration 'admin.retries' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:01,988] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,988] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,989] INFO Kafka startTimeMs: 1590627421988 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,992] INFO StreamsConfig values:
acceptable.recovery.lag = 10000
application.id = regex-source-integration-test
application.server =
bootstrap.servers = [localhost:56486]
buffered.records.per.partition = 1000
built.in.metrics.version = latest
cache.max.bytes.buffering = 0
client.id =
commit.interval.ms = 100
connections.max.idle.ms = 540000
default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler
default.key.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler
default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
default.value.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
max.task.idle.ms = 0
max.warmup.replicas = 2
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
num.standby.replicas = 0
num.stream.threads = 1
partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper
poll.ms = 100
probing.rebalance.interval.ms = 600000
processing.guarantee = at_least_once
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
replication.factor = 1
request.timeout.ms = 40000
retries = 0
retry.backoff.ms = 100
rocksdb.config.setter = null
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
state.cleanup.delay.ms = 600000
state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-5154128249393238509
topology.optimization = none
upgrade.from = null
windowstore.changelog.additional.retention.ms = 86400000
(org.apache.kafka.streams.StreamsConfig:347)
[2020-05-27 17:57:01,993] INFO stream-client [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:694)
[2020-05-27 17:57:01,994] INFO stream-client [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:695)
[2020-05-27 17:57:01,995] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id = regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-admin
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:01,997] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:01,997] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:01,998] INFO Kafka startTimeMs: 1590627421997 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:01,998] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:304)
[2020-05-27 17:57:01,999] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = none
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-restore-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = null
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:02,002] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:02,003] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:02,003] INFO Kafka startTimeMs: 1590627422002 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:02,005] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)
[2020-05-27 17:57:02,006] INFO ProducerConfig values:
acks = 1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-producer
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
linger.ms = 100
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 1000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 10000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:02,010] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:02,010] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:02,011] INFO Kafka startTimeMs: 1590627422010 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:02,012] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:353)
[2020-05-27 17:57:02,014] INFO [Producer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-producer] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:02,014] INFO ConsumerConfig values:
allow.auto.create.topics = false
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = regex-source-integration-test
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:02,017] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:216)
[2020-05-27 17:57:02,019] WARN The configuration 'admin.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:02,019] WARN The configuration 'admin.retries' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:02,020] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:02,020] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:02,020] INFO Kafka startTimeMs: 1590627422020 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:02,022] INFO stream-client [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:02,023] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:502)
[2020-05-27 17:57:02,023] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:02,023] INFO stream-client [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:02,025] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:502)
[2020-05-27 17:57:02,025] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:02,026] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Subscribed to pattern: 'partitioned-\d' (org.apache.kafka.streams.integration.RegexSourceIntegrationTest$4$1:1025)
[2020-05-27 17:57:02,026] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Subscribed to pattern: 'partitioned-\d' (org.apache.kafka.streams.integration.RegexSourceIntegrationTest$3$1:1025)
[2020-05-27 17:57:02,036] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:02,036] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:02,036] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:57:02,037] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:57:02,044] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:02,045] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:02,048] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)
[2020-05-27 17:57:02,048] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:02,048] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)
[2020-05-27 17:57:02,049] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:02,050] INFO [GroupCoordinator 0]: Preparing to rebalance group regex-source-integration-test in state PreparingRebalance with old generation 4 (__consumer_offsets-3) (reason: Adding new member regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer-e5cdb53f-d806-40d5-8eac-8d15dbeefcf4 with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:11,837] INFO [GroupCoordinator 0]: Member regex-source-integration-test-d425dc29-c09e-42a0-aba8-9919a5968e38-StreamThread-1-consumer-8f66bcdb-f6a2-4dc0-8bc3-71ac95a81b42 in group regex-source-integration-test has failed, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:11,839] INFO [GroupCoordinator 0]: Stabilized group regex-source-integration-test generation 5 (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:11,841] INFO Decided on assignment: {e3f9ab94-803a-4c33-a0d3-eae78eb29b42=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1], 412a9784-f401-401b-b357-4eb85cf3be0e=[activeTasks: ([0_1]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)
[2020-05-27 17:57:11,842] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer] Assigned tasks to clients as
e3f9ab94-803a-4c33-a0d3-eae78eb29b42=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]
412a9784-f401-401b-b357-4eb85cf3be0e=[activeTasks: ([0_1]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:724)
[2020-05-27 17:57:11,842] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:938)
[2020-05-27 17:57:11,843] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Finished assignment for group at generation 5: {regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer-e5cdb53f-d806-40d5-8eac-8d15dbeefcf4=Assignment(partitions=[partitioned-1-1, partitioned-2-1], userDataSize=56), regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer-fc2de7d0-1c1a-4f80-9e97-dca359a450ca=Assignment(partitions=[partitioned-1-0, partitioned-2-0], userDataSize=56)} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:57:11,844] INFO [GroupCoordinator 0]: Assignment received from leader for group regex-source-integration-test for generation 5 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:11,846] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Successfully joined group with generation 5 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:57:11,846] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Successfully joined group with generation 5 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:57:11,850] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Updating assignment with
Assigned partitions: [partitioned-1-1, partitioned-2-1]
Current owned partitions: []
Added partitions (assigned - owned): [partitioned-1-1, partitioned-2-1]
Revoked partitions (owned - assigned): []
(org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)
[2020-05-27 17:57:11,851] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Notifying assignor about the new Assignment(partitions=[partitioned-1-1, partitioned-2-1], userDataSize=56) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:57:11,850] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Updating assignment with
Assigned partitions: [partitioned-1-0, partitioned-2-0]
Current owned partitions: []
Added partitions (assigned - owned): [partitioned-1-0, partitioned-2-0]
Revoked partitions (owned - assigned): []
(org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)
[2020-05-27 17:57:11,851] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1474)
[2020-05-27 17:57:11,851] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Notifying assignor about the new Assignment(partitions=[partitioned-1-0, partitioned-2-0], userDataSize=56) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:57:11,851] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] Handle new assignment with:
New active tasks: [0_1]
New standby tasks: []
Existing active tasks: []
Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:176)
[2020-05-27 17:57:11,851] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1474)
[2020-05-27 17:57:11,852] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] Handle new assignment with:
New active tasks: [0_0]
New standby tasks: []
Existing active tasks: []
Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:176)
[2020-05-27 17:57:11,856] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Adding newly assigned partitions: partitioned-1-0, partitioned-2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:57:11,856] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Adding newly assigned partitions: partitioned-1-1, partitioned-2-1 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:57:11,856] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:11,856] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:11,859] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition partitioned-1-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:11,859] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition partitioned-2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:11,859] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition partitioned-1-1 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:11,860] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition partitioned-2-1 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:11,861] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition partitioned-1-1 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:11,861] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition partitioned-2-1 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:11,861] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition partitioned-1-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:11,862] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition partitioned-2-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:11,877] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] task [0_1] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:212)
[2020-05-27 17:57:11,879] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition partitioned-1-1 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:11,879] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition partitioned-2-1 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:11,880] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] task [0_1] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:229)
[2020-05-27 17:57:11,880] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:11,880] INFO stream-client [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:11,930] INFO stream-client [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:11,931] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:901)
[2020-05-27 17:57:11,931] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:11,954] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] task [0_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:212)
[2020-05-27 17:57:11,955] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition partitioned-1-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:11,955] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition partitioned-2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:11,956] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] task [0_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:229)
[2020-05-27 17:57:11,956] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:11,956] INFO stream-client [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:11,990] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:915)
[2020-05-27 17:57:11,991] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] task [0_1] Prepared clean close (org.apache.kafka.streams.processor.internals.StreamTask:419)
[2020-05-27 17:57:11,991] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:11,992] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] task [0_1] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:428)
[2020-05-27 17:57:11,992] INFO [Producer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:11,994] INFO [Consumer clientId=regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:12,002] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:12,002] INFO stream-thread [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:940)
[2020-05-27 17:57:12,004] INFO stream-client [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:12,005] INFO stream-client [regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:986)
[2020-05-27 17:57:12,005] INFO stream-client [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:12,006] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:901)
[2020-05-27 17:57:12,006] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:12,056] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:915)
[2020-05-27 17:57:12,057] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] task [0_0] Prepared clean close (org.apache.kafka.streams.processor.internals.StreamTask:419)
[2020-05-27 17:57:12,057] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:12,058] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] task [0_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:428)
[2020-05-27 17:57:12,058] INFO [Producer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:12,060] INFO [Consumer clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:12,065] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:12,065] INFO stream-thread [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:940)
[2020-05-27 17:57:12,066] INFO [AdminClient clientId=regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-admin] Metadata update failed (org.apache.kafka.clients.admin.internals.AdminMetadataManager:235)
org.apache.kafka.common.errors.TimeoutException: Call(callName=fetchMetadata, deadlineMs=1590627462065, tries=1, nextAllowedTryMs=-9223372036854775709) timed out at 9223372036854775807 after 1 attempt(s)
Caused by: org.apache.kafka.common.errors.TimeoutException: The AdminClient thread has exited.
[2020-05-27 17:57:12,075] INFO stream-client [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:12,076] INFO stream-client [regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:986)
[2020-05-27 17:57:12,077] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:12,079] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:12,080] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:12,080] INFO Kafka startTimeMs: 1590627432079 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:12,089] INFO Creating topic outputTopic_5 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:57:12,095] INFO [Controller id=0] New topics: [HashSet(outputTopic_5)], deleted topics: [HashSet()], new partition replica assignment [Map(outputTopic_5-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:12,095] INFO [Controller id=0] New partition creation callback for outputTopic_5-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:12,095] INFO [Controller id=0 epoch=1] Changed partition outputTopic_5-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:57:12,096] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:12,096] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:12,100] INFO [Controller id=0 epoch=1] Changed partition outputTopic_5-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:57:12,101] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:57:12,101] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:12,102] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:12,102] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 33 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:12,103] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(outputTopic_5-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:12,103] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 33 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:12,108] INFO [Log partition=outputTopic_5-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:57:12,109] INFO [Log partition=outputTopic_5-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:57:12,110] INFO Created log for partition outputTopic_5-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/outputTopic_5-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:57:12,111] INFO [Partition outputTopic_5-0 broker=0] No checkpointed highwatermark is found for partition outputTopic_5-0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:12,111] INFO [Partition outputTopic_5-0 broker=0] Log loaded for partition outputTopic_5-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:12,112] INFO [Broker id=0] Leader outputTopic_5-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:57:12,115] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 34 (state.change.logger:66)
[2020-05-27 17:57:12,119] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:12,121] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:12,121] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:12,121] INFO Kafka startTimeMs: 1590627432121 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:12,129] INFO Creating topic TEST-TOPIC-A with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:57:12,135] INFO [Controller id=0] New topics: [HashSet(TEST-TOPIC-A)], deleted topics: [HashSet()], new partition replica assignment [Map(TEST-TOPIC-A-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:12,135] INFO [Controller id=0] New partition creation callback for TEST-TOPIC-A-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:12,136] INFO [Controller id=0 epoch=1] Changed partition TEST-TOPIC-A-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:57:12,136] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:12,137] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:12,143] INFO [Controller id=0 epoch=1] Changed partition TEST-TOPIC-A-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:57:12,144] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:57:12,144] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:12,145] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:12,145] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 35 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:12,146] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(TEST-TOPIC-A-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:12,146] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 35 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:12,151] INFO [Log partition=TEST-TOPIC-A-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:57:12,152] INFO [Log partition=TEST-TOPIC-A-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:57:12,153] INFO Created log for partition TEST-TOPIC-A-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/TEST-TOPIC-A-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:57:12,154] INFO [Partition TEST-TOPIC-A-0 broker=0] No checkpointed highwatermark is found for partition TEST-TOPIC-A-0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:12,154] INFO [Partition TEST-TOPIC-A-0 broker=0] Log loaded for partition TEST-TOPIC-A-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:12,155] INFO [Broker id=0] Leader TEST-TOPIC-A-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:57:12,158] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 36 (state.change.logger:66)
[2020-05-27 17:57:12,161] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:12,163] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:12,163] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:12,164] INFO Kafka startTimeMs: 1590627432163 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:12,170] INFO Creating topic TEST-TOPIC-B with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:57:12,176] INFO [Controller id=0] New topics: [HashSet(TEST-TOPIC-B)], deleted topics: [HashSet()], new partition replica assignment [Map(TEST-TOPIC-B-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:12,176] INFO [Controller id=0] New partition creation callback for TEST-TOPIC-B-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:12,176] INFO [Controller id=0 epoch=1] Changed partition TEST-TOPIC-B-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:57:12,176] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:12,177] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:12,182] INFO [Controller id=0 epoch=1] Changed partition TEST-TOPIC-B-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:57:12,182] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:57:12,182] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:12,183] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:12,183] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 37 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:12,184] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(TEST-TOPIC-B-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:12,184] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 37 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:12,189] INFO [Log partition=TEST-TOPIC-B-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:57:12,191] INFO [Log partition=TEST-TOPIC-B-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:57:12,191] INFO Created log for partition TEST-TOPIC-B-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/TEST-TOPIC-B-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:57:12,192] INFO [Partition TEST-TOPIC-B-0 broker=0] No checkpointed highwatermark is found for partition TEST-TOPIC-B-0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:12,192] INFO [Partition TEST-TOPIC-B-0 broker=0] Log loaded for partition TEST-TOPIC-B-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:12,193] INFO [Broker id=0] Leader TEST-TOPIC-B-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:57:12,196] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 38 (state.change.logger:66)
[2020-05-27 17:57:12,201] INFO StreamsConfig values:
acceptable.recovery.lag = 10000
application.id = regex-source-integration-test
application.server =
bootstrap.servers = [localhost:56486]
buffered.records.per.partition = 1000
built.in.metrics.version = latest
cache.max.bytes.buffering = 0
client.id =
commit.interval.ms = 100
connections.max.idle.ms = 540000
default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler
default.key.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler
default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
default.value.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
max.task.idle.ms = 0
max.warmup.replicas = 2
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
num.standby.replicas = 0
num.stream.threads = 1
partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper
poll.ms = 100
probing.rebalance.interval.ms = 600000
processing.guarantee = at_least_once
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
replication.factor = 1
request.timeout.ms = 40000
retries = 0
retry.backoff.ms = 100
rocksdb.config.setter = null
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
state.cleanup.delay.ms = 600000
state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-814280565460057519
topology.optimization = none
upgrade.from = null
windowstore.changelog.additional.retention.ms = 86400000
(org.apache.kafka.streams.StreamsConfig:347)
[2020-05-27 17:57:12,202] INFO stream-client [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:694)
[2020-05-27 17:57:12,202] INFO stream-client [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:695)
[2020-05-27 17:57:12,203] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id = regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-admin
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:12,205] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:12,205] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:12,205] INFO Kafka startTimeMs: 1590627432205 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:12,206] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:304)
[2020-05-27 17:57:12,206] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = none
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-restore-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = null
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:12,210] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:12,210] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:12,210] INFO Kafka startTimeMs: 1590627432210 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:12,211] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)
[2020-05-27 17:57:12,212] INFO ProducerConfig values:
acks = 1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-producer
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
linger.ms = 100
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 1000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 10000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:12,215] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:12,216] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:12,216] INFO Kafka startTimeMs: 1590627432215 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:12,217] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:353)
[2020-05-27 17:57:12,218] INFO [Producer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-producer] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:12,219] INFO ConsumerConfig values:
allow.auto.create.topics = false
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = regex-source-integration-test
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:12,222] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:216)
[2020-05-27 17:57:12,224] WARN The configuration 'admin.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:12,224] WARN The configuration 'admin.retries' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:12,225] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:12,225] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:12,225] INFO Kafka startTimeMs: 1590627432225 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:12,227] INFO stream-client [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:12,228] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:502)
[2020-05-27 17:57:12,228] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:12,228] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Subscribed to pattern: 'TEST-TOPIC-[A-Z]' (org.apache.kafka.streams.integration.RegexSourceIntegrationTest$2$1:1025)
[2020-05-27 17:57:12,233] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:12,234] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:57:12,239] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:12,241] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)
[2020-05-27 17:57:12,242] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:12,243] INFO [GroupCoordinator 0]: Preparing to rebalance group regex-source-integration-test in state PreparingRebalance with old generation 5 (__consumer_offsets-3) (reason: Adding new member regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer-cdde98d7-3502-47c9-baa2-2918d3050c81 with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:21,846] INFO [GroupCoordinator 0]: Member regex-source-integration-test-412a9784-f401-401b-b357-4eb85cf3be0e-StreamThread-1-consumer-e5cdb53f-d806-40d5-8eac-8d15dbeefcf4 in group regex-source-integration-test has failed, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:21,846] INFO [GroupCoordinator 0]: Member regex-source-integration-test-e3f9ab94-803a-4c33-a0d3-eae78eb29b42-StreamThread-1-consumer-fc2de7d0-1c1a-4f80-9e97-dca359a450ca in group regex-source-integration-test has failed, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:21,847] INFO [GroupCoordinator 0]: Stabilized group regex-source-integration-test generation 6 (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:21,848] INFO Decided on assignment: {3fc29d7b-c152-4989-87b8-96c5558904d1=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)
[2020-05-27 17:57:21,848] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer] Assigned tasks to clients as
3fc29d7b-c152-4989-87b8-96c5558904d1=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:724)
[2020-05-27 17:57:21,848] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:938)
[2020-05-27 17:57:21,848] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Finished assignment for group at generation 6: {regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer-cdde98d7-3502-47c9-baa2-2918d3050c81=Assignment(partitions=[TEST-TOPIC-A-0, TEST-TOPIC-B-0], userDataSize=56)} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:57:21,849] INFO [GroupCoordinator 0]: Assignment received from leader for group regex-source-integration-test for generation 6 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:21,850] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Successfully joined group with generation 6 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:57:21,852] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Updating assignment with
Assigned partitions: [TEST-TOPIC-B-0, TEST-TOPIC-A-0]
Current owned partitions: []
Added partitions (assigned - owned): [TEST-TOPIC-B-0, TEST-TOPIC-A-0]
Revoked partitions (owned - assigned): []
(org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)
[2020-05-27 17:57:21,852] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Notifying assignor about the new Assignment(partitions=[TEST-TOPIC-B-0, TEST-TOPIC-A-0], userDataSize=56) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:57:21,852] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1474)
[2020-05-27 17:57:21,852] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] Handle new assignment with:
New active tasks: [0_0]
New standby tasks: []
Existing active tasks: []
Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:176)
[2020-05-27 17:57:21,854] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Adding newly assigned partitions: TEST-TOPIC-B-0, TEST-TOPIC-A-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:57:21,854] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:21,855] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition TEST-TOPIC-B-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:21,855] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition TEST-TOPIC-A-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:21,856] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition TEST-TOPIC-B-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:21,856] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Resetting offset for partition TEST-TOPIC-A-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:21,860] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:21,861] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:21,861] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:21,862] INFO Kafka startTimeMs: 1590627441861 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:21,871] INFO [Controller id=0] Starting topic deletion for topics TEST-TOPIC-A (kafka.controller.KafkaController:66)
[2020-05-27 17:57:21,872] INFO [Topic Deletion Manager 0] Handling deletion for topics TEST-TOPIC-A (kafka.controller.TopicDeletionManager:66)
[2020-05-27 17:57:21,874] INFO [Topic Deletion Manager 0] Deletion of topic TEST-TOPIC-A (re)started (kafka.controller.TopicDeletionManager:66)
[2020-05-27 17:57:21,877] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:21,878] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:21,878] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:21,879] INFO [Broker id=0] Add 0 partitions and deleted 1 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 39 (state.change.logger:66)
[2020-05-27 17:57:21,882] INFO [GroupCoordinator 0]: Removed 0 offsets associated with deleted partitions: TEST-TOPIC-A-0. (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:21,891] INFO [Controller id=0 epoch=1] Partition TEST-TOPIC-A-0 state changed to (Leader:-1,ISR:0,LeaderEpoch:1,ControllerEpoch:1) after removing replica 0 from the ISR as part of transition to OfflineReplica (state.change.logger:66)
[2020-05-27 17:57:21,893] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:21,894] INFO [Controller id=0 epoch=1] Sending StopReplica request for 1 replicas to broker 0 (state.change.logger:66)
[2020-05-27 17:57:21,896] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:21,896] INFO [Controller id=0 epoch=1] Sending StopReplica request for 1 replicas to broker 0 (state.change.logger:66)
[2020-05-27 17:57:21,897] INFO [Broker id=0] Handling StopReplica request correlationId 40 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:21,899] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(TEST-TOPIC-A-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:21,899] INFO [ReplicaAlterLogDirsManager on broker 0] Removed fetcher for partitions Set(TEST-TOPIC-A-0) (kafka.server.ReplicaAlterLogDirsManager:66)
[2020-05-27 17:57:21,903] INFO [Broker id=0] Handling StopReplica request correlationId 41 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:21,903] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(TEST-TOPIC-A-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:21,904] INFO [ReplicaAlterLogDirsManager on broker 0] Removed fetcher for partitions Set(TEST-TOPIC-A-0) (kafka.server.ReplicaAlterLogDirsManager:66)
[2020-05-27 17:57:21,906] INFO The cleaning for partition TEST-TOPIC-A-0 is aborted and paused (kafka.log.LogCleaner:66)
[2020-05-27 17:57:21,906] INFO The cleaning for partition TEST-TOPIC-A-0 is aborted (kafka.log.LogCleaner:66)
[2020-05-27 17:57:21,912] INFO Log for partition TEST-TOPIC-A-0 is renamed to /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/TEST-TOPIC-A-0.1c14a4c1dd3d44c0a8c6dfd773649bc2-delete and is scheduled for deletion (kafka.log.LogManager:66)
[2020-05-27 17:57:21,916] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:21,916] INFO [Topic Deletion Manager 0] Handling deletion for topics TEST-TOPIC-A (kafka.controller.TopicDeletionManager:66)
[2020-05-27 17:57:21,918] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:21,929] INFO [Topic Deletion Manager 0] Deletion of topic TEST-TOPIC-A successfully completed (kafka.controller.TopicDeletionManager:66)
[2020-05-27 17:57:21,930] INFO [Controller id=0] New topics: [HashSet()], deleted topics: [HashSet()], new partition replica assignment [Map()] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:21,931] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] task [0_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:212)
[2020-05-27 17:57:21,932] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition TEST-TOPIC-B-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:21,932] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Found no committed offset for partition TEST-TOPIC-A-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:21,932] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] task [0_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:229)
[2020-05-27 17:57:21,933] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:21,933] INFO stream-client [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:22,435] WARN [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Received unknown topic or partition error in fetch for partition TEST-TOPIC-A-0 (org.apache.kafka.clients.consumer.internals.Fetcher:1254)
[2020-05-27 17:57:22,938] WARN [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Received unknown topic or partition error in fetch for partition TEST-TOPIC-A-0 (org.apache.kafka.clients.consumer.internals.Fetcher:1254)
[2020-05-27 17:57:22,938] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Revoke previously assigned partitions TEST-TOPIC-A-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:307)
[2020-05-27 17:57:22,938] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] State transition from RUNNING to PARTITIONS_REVOKED (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:22,938] INFO stream-client [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1] State transition from RUNNING to REBALANCING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:22,938] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] partition revocation took 0 ms. (org.apache.kafka.streams.processor.internals.StreamThread:78)
[2020-05-27 17:57:22,939] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:22,939] INFO [GroupCoordinator 0]: Preparing to rebalance group regex-source-integration-test in state PreparingRebalance with old generation 6 (__consumer_offsets-3) (reason: Updating metadata for member regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer-cdde98d7-3502-47c9-baa2-2918d3050c81) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:22,939] INFO [GroupCoordinator 0]: Stabilized group regex-source-integration-test generation 7 (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:22,940] INFO Decided on assignment: {3fc29d7b-c152-4989-87b8-96c5558904d1=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([0_0]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([TEST-TOPIC-B-0]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)
[2020-05-27 17:57:22,940] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer] Assigned tasks to clients as
3fc29d7b-c152-4989-87b8-96c5558904d1=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([0_0]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([TEST-TOPIC-B-0]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([]) capacity: 1 assigned: 1]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:724)
[2020-05-27 17:57:22,940] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:938)
[2020-05-27 17:57:22,941] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Finished assignment for group at generation 7: {regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer-cdde98d7-3502-47c9-baa2-2918d3050c81=Assignment(partitions=[TEST-TOPIC-B-0], userDataSize=48)} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:57:22,941] INFO [GroupCoordinator 0]: Assignment received from leader for group regex-source-integration-test for generation 7 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:22,942] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Successfully joined group with generation 7 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:57:22,942] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Updating assignment with
Assigned partitions: [TEST-TOPIC-B-0]
Current owned partitions: [TEST-TOPIC-B-0]
Added partitions (assigned - owned): []
Revoked partitions (owned - assigned): []
(org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)
[2020-05-27 17:57:22,942] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Notifying assignor about the new Assignment(partitions=[TEST-TOPIC-B-0], userDataSize=48) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:57:22,942] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1474)
[2020-05-27 17:57:22,943] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] Handle new assignment with:
New active tasks: [0_0]
New standby tasks: []
Existing active tasks: [0_0]
Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:176)
[2020-05-27 17:57:22,943] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer, groupId=regex-source-integration-test] Adding newly assigned partitions: (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:57:22,943] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] State transition from PARTITIONS_REVOKED to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:22,973] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:22,974] INFO stream-client [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:23,021] INFO stream-client [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:23,021] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:901)
[2020-05-27 17:57:23,021] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:23,079] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:915)
[2020-05-27 17:57:23,079] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] task [0_0] Prepared clean close (org.apache.kafka.streams.processor.internals.StreamTask:419)
[2020-05-27 17:57:23,079] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:23,080] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] task [0_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:428)
[2020-05-27 17:57:23,080] INFO [Producer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:23,081] INFO [Consumer clientId=regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:23,082] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:23,082] INFO stream-thread [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:940)
[2020-05-27 17:57:23,084] INFO stream-client [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:23,084] INFO stream-client [regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:986)
[2020-05-27 17:57:23,085] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:23,086] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:23,086] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:23,086] INFO Kafka startTimeMs: 1590627443086 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:23,090] INFO Creating topic outputTopic_6 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:57:23,094] INFO [Controller id=0] New topics: [HashSet(outputTopic_6)], deleted topics: [HashSet()], new partition replica assignment [Map(outputTopic_6-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:23,094] INFO [Controller id=0] New partition creation callback for outputTopic_6-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:23,094] INFO [Controller id=0 epoch=1] Changed partition outputTopic_6-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:57:23,095] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:23,095] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:23,098] INFO [Controller id=0 epoch=1] Changed partition outputTopic_6-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:57:23,098] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:57:23,099] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:23,099] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:23,099] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 42 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:23,100] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(outputTopic_6-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:23,100] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 42 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:23,103] INFO [Log partition=outputTopic_6-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:57:23,104] INFO [Log partition=outputTopic_6-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:57:23,104] INFO Created log for partition outputTopic_6-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/outputTopic_6-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:57:23,106] INFO [Partition outputTopic_6-0 broker=0] No checkpointed highwatermark is found for partition outputTopic_6-0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:23,106] INFO [Partition outputTopic_6-0 broker=0] Log loaded for partition outputTopic_6-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:23,106] INFO [Broker id=0] Leader outputTopic_6-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:57:23,109] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 43 (state.change.logger:66)
[2020-05-27 17:57:23,114] INFO StreamsConfig values:
acceptable.recovery.lag = 10000
application.id = regex-source-integration-test
application.server =
bootstrap.servers = [localhost:56486]
buffered.records.per.partition = 1000
built.in.metrics.version = latest
cache.max.bytes.buffering = 0
client.id =
commit.interval.ms = 100
connections.max.idle.ms = 540000
default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler
default.key.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler
default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
default.value.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
max.task.idle.ms = 0
max.warmup.replicas = 2
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
num.standby.replicas = 0
num.stream.threads = 1
partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper
poll.ms = 100
probing.rebalance.interval.ms = 600000
processing.guarantee = at_least_once
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
replication.factor = 1
request.timeout.ms = 40000
retries = 0
retry.backoff.ms = 100
rocksdb.config.setter = null
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
state.cleanup.delay.ms = 600000
state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-6234927904251092418
topology.optimization = none
upgrade.from = null
windowstore.changelog.additional.retention.ms = 86400000
(org.apache.kafka.streams.StreamsConfig:347)
[2020-05-27 17:57:23,115] INFO stream-client [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:694)
[2020-05-27 17:57:23,115] INFO stream-client [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:695)
[2020-05-27 17:57:23,116] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id = regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-admin
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:23,117] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:23,117] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:23,117] INFO Kafka startTimeMs: 1590627443117 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:23,118] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:304)
[2020-05-27 17:57:23,118] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = none
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-restore-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = null
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:23,120] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:23,121] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:23,121] INFO Kafka startTimeMs: 1590627443120 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:23,121] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)
[2020-05-27 17:57:23,122] INFO ProducerConfig values:
acks = 1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-producer
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
linger.ms = 100
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 1000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 10000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:23,124] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:23,125] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:23,125] INFO Kafka startTimeMs: 1590627443124 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:23,125] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:353)
[2020-05-27 17:57:23,126] INFO ConsumerConfig values:
allow.auto.create.topics = false
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = regex-source-integration-test
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:23,127] INFO [Producer clientId=regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-producer] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:23,128] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:216)
[2020-05-27 17:57:23,128] WARN The configuration 'admin.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:23,133] WARN The configuration 'admin.retries' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:23,133] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:23,133] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:23,133] INFO Kafka startTimeMs: 1590627443133 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:23,135] INFO stream-client [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:23,135] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:502)
[2020-05-27 17:57:23,135] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:23,135] INFO [Consumer clientId=regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-consumer, groupId=regex-source-integration-test] Subscribed to pattern: 'foo.*|f.*' (org.apache.kafka.clients.consumer.KafkaConsumer:1025)
[2020-05-27 17:57:23,136] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-14
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:23,138] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:23,139] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:23,139] INFO [Consumer clientId=regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-consumer, groupId=regex-source-integration-test] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:23,139] INFO Kafka startTimeMs: 1590627443138 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:23,139] INFO [Consumer clientId=regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-consumer, groupId=regex-source-integration-test] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:57:23,142] INFO [Producer clientId=producer-14] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:23,143] INFO [Consumer clientId=regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:23,144] INFO [Producer clientId=producer-14] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:23,145] ERROR stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] Encountered the following exception during processing and the thread is going to shut down: (org.apache.kafka.streams.processor.internals.StreamThread:529)
org.apache.kafka.streams.errors.TopologyException: Invalid topology: Topic foo is already matched for another regex pattern foo.* and hence cannot be matched to this regex pattern f.* any more.
at org.apache.kafka.streams.processor.internals.InternalTopologyBuilder$SourceNodeFactory.getTopics(InternalTopologyBuilder.java:261)
at org.apache.kafka.streams.processor.internals.InternalTopologyBuilder.setRegexMatchedTopicsToSourceNodes(InternalTopologyBuilder.java:1122)
at org.apache.kafka.streams.processor.internals.InternalTopologyBuilder.updateSubscribedTopics(InternalTopologyBuilder.java:1983)
at org.apache.kafka.streams.processor.internals.InternalTopologyBuilder.addSubscribedTopicsFromMetadata(InternalTopologyBuilder.java:1973)
at org.apache.kafka.streams.processor.internals.TaskManager.handleRebalanceStart(TaskManager.java:136)
at org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor.handleRebalanceStart(StreamsPartitionAssignor.java:1585)
at org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor.subscriptionUserData(StreamsPartitionAssignor.java:232)
at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.metadata(ConsumerCoordinator.java:222)
at org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest(AbstractCoordinator.java:561)
at org.apache.kafka.clients.consumer.internals.AbstractCoordinator.initiateJoinGroup(AbstractCoordinator.java:496)
at org.apache.kafka.clients.consumer.internals.AbstractCoordinator.joinGroupIfNeeded(AbstractCoordinator.java:418)
at org.apache.kafka.clients.consumer.internals.AbstractCoordinator.ensureActiveGroup(AbstractCoordinator.java:359)
at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.poll(ConsumerCoordinator.java:506)
at org.apache.kafka.clients.consumer.KafkaConsumer.updateAssignmentMetadataIfNeeded(KafkaConsumer.java:1265)
at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1231)
at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1206)
at org.apache.kafka.streams.processor.internals.StreamThread.pollRequests(StreamThread.java:770)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:630)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:550)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:509)
[2020-05-27 17:57:23,146] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] State transition from STARTING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:23,146] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:915)
[2020-05-27 17:57:23,146] INFO [Producer clientId=regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:23,148] INFO [Consumer clientId=regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:23,152] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:23,153] INFO stream-client [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a] State transition from REBALANCING to ERROR (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:23,153] ERROR stream-client [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a] All stream threads have died. The instance will be in error state and should be closed. (org.apache.kafka.streams.KafkaStreams:439)
[2020-05-27 17:57:23,153] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:940)
[2020-05-27 17:57:23,153] ERROR Thread StreamsThread threadId: regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1
TaskManager
MetadataState:
Tasks:
died (org.apache.zookeeper.server.NIOServerCnxnFactory:92)
org.apache.kafka.streams.errors.TopologyException: Invalid topology: Topic foo is already matched for another regex pattern foo.* and hence cannot be matched to this regex pattern f.* any more.
at org.apache.kafka.streams.processor.internals.InternalTopologyBuilder$SourceNodeFactory.getTopics(InternalTopologyBuilder.java:261)
at org.apache.kafka.streams.processor.internals.InternalTopologyBuilder.setRegexMatchedTopicsToSourceNodes(InternalTopologyBuilder.java:1122)
at org.apache.kafka.streams.processor.internals.InternalTopologyBuilder.updateSubscribedTopics(InternalTopologyBuilder.java:1983)
at org.apache.kafka.streams.processor.internals.InternalTopologyBuilder.addSubscribedTopicsFromMetadata(InternalTopologyBuilder.java:1973)
at org.apache.kafka.streams.processor.internals.TaskManager.handleRebalanceStart(TaskManager.java:136)
at org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor.handleRebalanceStart(StreamsPartitionAssignor.java:1585)
at org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor.subscriptionUserData(StreamsPartitionAssignor.java:232)
at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.metadata(ConsumerCoordinator.java:222)
at org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest(AbstractCoordinator.java:561)
at org.apache.kafka.clients.consumer.internals.AbstractCoordinator.initiateJoinGroup(AbstractCoordinator.java:496)
at org.apache.kafka.clients.consumer.internals.AbstractCoordinator.joinGroupIfNeeded(AbstractCoordinator.java:418)
at org.apache.kafka.clients.consumer.internals.AbstractCoordinator.ensureActiveGroup(AbstractCoordinator.java:359)
at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.poll(ConsumerCoordinator.java:506)
at org.apache.kafka.clients.consumer.KafkaConsumer.updateAssignmentMetadataIfNeeded(KafkaConsumer.java:1265)
at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1231)
at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1206)
at org.apache.kafka.streams.processor.internals.StreamThread.pollRequests(StreamThread.java:770)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:630)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:550)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:509)
[2020-05-27 17:57:23,155] INFO ProducerConfig values:
acks = -1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-15
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.StringSerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:23,157] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:23,158] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:23,158] INFO Kafka startTimeMs: 1590627443157 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:23,162] INFO [Producer clientId=producer-15] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:23,162] INFO [Producer clientId=producer-15] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:23,172] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = true
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = 7c5f532d-c14e-4d65-91df-d44309ddaecf
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = true
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 500
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:23,176] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:23,176] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:23,176] INFO Kafka startTimeMs: 1590627443176 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:23,177] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Subscribed to topic(s): outputTopic_6 (org.apache.kafka.clients.consumer.KafkaConsumer:961)
[2020-05-27 17:57:23,180] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:23,180] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:57:23,181] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:23,183] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)
[2020-05-27 17:57:23,184] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:23,185] INFO [GroupCoordinator 0]: Preparing to rebalance group 7c5f532d-c14e-4d65-91df-d44309ddaecf in state PreparingRebalance with old generation 0 (__consumer_offsets-4) (reason: Adding new member consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3-2cea1754-6eac-4ee9-a4c2-4aeb041af27f with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:23,185] INFO [GroupCoordinator 0]: Stabilized group 7c5f532d-c14e-4d65-91df-d44309ddaecf generation 1 (__consumer_offsets-4) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:23,186] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Finished assignment for group at generation 1: {consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3-2cea1754-6eac-4ee9-a4c2-4aeb041af27f=Assignment(partitions=[outputTopic_6-0])} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:57:23,186] INFO [GroupCoordinator 0]: Assignment received from leader for group 7c5f532d-c14e-4d65-91df-d44309ddaecf for generation 1 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:23,188] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Successfully joined group with generation 1 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:57:23,190] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Notifying assignor about the new Assignment(partitions=[outputTopic_6-0]) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:57:23,190] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Adding newly assigned partitions: outputTopic_6-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:57:23,191] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Found no committed offset for partition outputTopic_6-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)
[2020-05-27 17:57:23,193] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Resetting offset for partition outputTopic_6-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:387)
[2020-05-27 17:57:28,366] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Revoke previously assigned partitions outputTopic_6-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:307)
[2020-05-27 17:57:28,366] INFO [Consumer clientId=consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3, groupId=7c5f532d-c14e-4d65-91df-d44309ddaecf] Member consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3-2cea1754-6eac-4ee9-a4c2-4aeb041af27f sending LeaveGroup request to coordinator localhost:56486 (id: 2147483647 rack: null) due to the consumer is being closed (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:1005)
[2020-05-27 17:57:28,367] INFO [GroupCoordinator 0]: Member[group.instance.id None, member.id consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3-2cea1754-6eac-4ee9-a4c2-4aeb041af27f] in group 7c5f532d-c14e-4d65-91df-d44309ddaecf has left, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:28,367] INFO [GroupCoordinator 0]: Preparing to rebalance group 7c5f532d-c14e-4d65-91df-d44309ddaecf in state PreparingRebalance with old generation 1 (__consumer_offsets-4) (reason: removing member consumer-7c5f532d-c14e-4d65-91df-d44309ddaecf-3-2cea1754-6eac-4ee9-a4c2-4aeb041af27f on LeaveGroup) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:28,367] INFO [GroupCoordinator 0]: Group 7c5f532d-c14e-4d65-91df-d44309ddaecf with generation 2 is now empty (__consumer_offsets-4) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:28,369] INFO stream-client [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a] State transition from ERROR to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:28,369] INFO stream-thread [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:901)
[2020-05-27 17:57:28,371] INFO stream-client [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:28,371] INFO stream-client [regex-source-integration-test-9e3e325f-3fe3-4741-9862-eff246a8fb2a] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:986)
[2020-05-27 17:57:28,372] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id =
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:28,373] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:28,373] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:28,373] INFO Kafka startTimeMs: 1590627448373 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:28,378] INFO Creating topic outputTopic_7 with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:57:28,381] INFO [Controller id=0] New topics: [HashSet(outputTopic_7)], deleted topics: [HashSet()], new partition replica assignment [Map(outputTopic_7-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:28,381] INFO [Controller id=0] New partition creation callback for outputTopic_7-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:28,382] INFO [Controller id=0 epoch=1] Changed partition outputTopic_7-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:57:28,382] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:28,382] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:28,385] INFO [Controller id=0 epoch=1] Changed partition outputTopic_7-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:57:28,385] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:57:28,386] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:28,386] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:28,386] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 44 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:28,387] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(outputTopic_7-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:28,387] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 44 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:28,391] INFO [Log partition=outputTopic_7-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:57:28,392] INFO [Log partition=outputTopic_7-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:57:28,392] INFO Created log for partition outputTopic_7-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/outputTopic_7-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:57:28,393] INFO [Partition outputTopic_7-0 broker=0] No checkpointed highwatermark is found for partition outputTopic_7-0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:28,393] INFO [Partition outputTopic_7-0 broker=0] Log loaded for partition outputTopic_7-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:28,393] INFO [Broker id=0] Leader outputTopic_7-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:57:28,395] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 45 (state.change.logger:66)
[2020-05-27 17:57:28,398] INFO StreamsConfig values:
acceptable.recovery.lag = 10000
application.id = regex-source-integration-test
application.server =
bootstrap.servers = [localhost:56486]
buffered.records.per.partition = 1000
built.in.metrics.version = latest
cache.max.bytes.buffering = 0
client.id =
commit.interval.ms = 100
connections.max.idle.ms = 540000
default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler
default.key.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler
default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
default.value.serde = class org.apache.kafka.common.serialization.Serdes$StringSerde
max.task.idle.ms = 0
max.warmup.replicas = 2
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
num.standby.replicas = 0
num.stream.threads = 1
partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper
poll.ms = 100
probing.rebalance.interval.ms = 600000
processing.guarantee = at_least_once
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
replication.factor = 1
request.timeout.ms = 40000
retries = 0
retry.backoff.ms = 100
rocksdb.config.setter = null
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
state.cleanup.delay.ms = 600000
state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-1629757011841220824
topology.optimization = none
upgrade.from = null
windowstore.changelog.additional.retention.ms = 86400000
(org.apache.kafka.streams.StreamsConfig:347)
[2020-05-27 17:57:28,399] INFO stream-client [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:694)
[2020-05-27 17:57:28,399] INFO stream-client [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:695)
[2020-05-27 17:57:28,400] INFO AdminClientConfig values:
bootstrap.servers = [localhost:56486]
client.dns.lookup = default
client.id = regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-admin
connections.max.idle.ms = 300000
default.api.timeout.ms = 60000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
(org.apache.kafka.clients.admin.AdminClientConfig:347)
[2020-05-27 17:57:28,401] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:28,402] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:28,402] INFO Kafka startTimeMs: 1590627448401 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:28,402] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:304)
[2020-05-27 17:57:28,402] INFO ConsumerConfig values:
allow.auto.create.topics = true
auto.commit.interval.ms = 5000
auto.offset.reset = none
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-restore-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = null
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:28,404] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:28,404] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:28,404] INFO Kafka startTimeMs: 1590627448404 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:28,405] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)
[2020-05-27 17:57:28,405] INFO ProducerConfig values:
acks = 1
batch.size = 16384
bootstrap.servers = [localhost:56486]
buffer.memory = 33554432
client.dns.lookup = default
client.id = regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-producer
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 120000
enable.idempotence = false
interceptor.classes = []
internal.auto.downgrade.txn.commit = false
key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
linger.ms = 100
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 1000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retries = 2147483647
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 10000
transactional.id = null
value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
(org.apache.kafka.clients.producer.ProducerConfig:347)
[2020-05-27 17:57:28,407] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:28,408] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:28,408] INFO Kafka startTimeMs: 1590627448407 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:28,408] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:353)
[2020-05-27 17:57:28,409] INFO ConsumerConfig values:
allow.auto.create.topics = false
auto.commit.interval.ms = 5000
auto.offset.reset = earliest
bootstrap.servers = [localhost:56486]
check.crcs = true
client.dns.lookup = default
client.id = regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer
client.rack =
connections.max.idle.ms = 540000
default.api.timeout.ms = 60000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = regex-source-integration-test
group.instance.id = null
heartbeat.interval.ms = 3000
interceptor.classes = []
internal.leave.group.on.close = false
internal.throw.on.fetch.stable.offset.unsupported = false
isolation.level = read_uncommitted
key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 1000
metadata.max.age.ms = 1000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
partition.assignment.strategy = [org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 30000
retry.backoff.ms = 100
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = https
ssl.engine.factory.class = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLSv1.2
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
(org.apache.kafka.clients.consumer.ConsumerConfig:347)
[2020-05-27 17:57:28,410] INFO [Producer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-producer] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:28,411] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:216)
[2020-05-27 17:57:28,412] WARN The configuration 'admin.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:28,412] WARN The configuration 'admin.retries' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:355)
[2020-05-27 17:57:28,412] INFO Kafka version: 2.6.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)
[2020-05-27 17:57:28,412] INFO Kafka commitId: 8471eec6ad04d07a (org.apache.kafka.common.utils.AppInfoParser:118)
[2020-05-27 17:57:28,412] INFO Kafka startTimeMs: 1590627448412 (org.apache.kafka.common.utils.AppInfoParser:119)
[2020-05-27 17:57:28,413] INFO stream-client [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:28,413] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:502)
[2020-05-27 17:57:28,413] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:28,414] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] Subscribed to pattern: 'topic-\d+' (org.apache.kafka.clients.consumer.KafkaConsumer:1025)
[2020-05-27 17:57:28,415] INFO StreamsConfig values:
acceptable.recovery.lag = 10000
application.id = 40960d6d-d168-47d9-88a8-4f13bda635a6
application.server =
bootstrap.servers = [localhost:9091]
buffered.records.per.partition = 1000
built.in.metrics.version = latest
cache.max.bytes.buffering = 10485760
client.id =
commit.interval.ms = 30000
connections.max.idle.ms = 540000
default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler
default.key.serde = class org.apache.kafka.common.serialization.Serdes$ByteArraySerde
default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler
default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
default.value.serde = class org.apache.kafka.common.serialization.Serdes$ByteArraySerde
max.task.idle.ms = 0
max.warmup.replicas = 2
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
num.standby.replicas = 0
num.stream.threads = 1
partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper
poll.ms = 100
probing.rebalance.interval.ms = 600000
processing.guarantee = at_least_once
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
replication.factor = 1
request.timeout.ms = 40000
retries = 0
retry.backoff.ms = 100
rocksdb.config.setter = null
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
state.cleanup.delay.ms = 600000
state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-116171513935852134
topology.optimization = none
upgrade.from = null
windowstore.changelog.additional.retention.ms = 86400000
(org.apache.kafka.streams.StreamsConfig:347)
[2020-05-27 17:57:28,416] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] Cluster ID: XNmsyx4kR5in9tlu1-LxwQ (org.apache.kafka.clients.Metadata:277)
[2020-05-27 17:57:28,416] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] Discovered group coordinator localhost:56486 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)
[2020-05-27 17:57:28,418] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:28,420] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)
[2020-05-27 17:57:28,420] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)
[2020-05-27 17:57:28,420] INFO [GroupCoordinator 0]: Preparing to rebalance group regex-source-integration-test in state PreparingRebalance with old generation 7 (__consumer_offsets-3) (reason: Adding new member regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer-fe72542d-1ce7-4715-a194-91c454362dcf with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:28,524] INFO StreamsConfig values:
acceptable.recovery.lag = 10000
application.id = 02340d6a-09c1-403d-822b-a3d7caecb9a5
application.server =
bootstrap.servers = [localhost:9091]
buffered.records.per.partition = 1000
built.in.metrics.version = latest
cache.max.bytes.buffering = 10485760
client.id =
commit.interval.ms = 30000
connections.max.idle.ms = 540000
default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler
default.key.serde = class org.apache.kafka.common.serialization.Serdes$ByteArraySerde
default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler
default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
default.value.serde = class org.apache.kafka.common.serialization.Serdes$ByteArraySerde
max.task.idle.ms = 0
max.warmup.replicas = 2
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = DEBUG
metrics.sample.window.ms = 30000
num.standby.replicas = 0
num.stream.threads = 1
partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper
poll.ms = 100
probing.rebalance.interval.ms = 600000
processing.guarantee = at_least_once
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
replication.factor = 1
request.timeout.ms = 40000
retries = 0
retry.backoff.ms = 100
rocksdb.config.setter = null
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
state.cleanup.delay.ms = 600000
state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-6960889208209194116
topology.optimization = none
upgrade.from = null
windowstore.changelog.additional.retention.ms = 86400000
(org.apache.kafka.streams.StreamsConfig:347)
[2020-05-27 17:57:28,524] INFO stream-client [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21] State transition from REBALANCING to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:28,525] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:901)
[2020-05-27 17:57:28,525] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] State transition from STARTING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:32,942] INFO [GroupCoordinator 0]: Member regex-source-integration-test-3fc29d7b-c152-4989-87b8-96c5558904d1-StreamThread-1-consumer-cdde98d7-3502-47c9-baa2-2918d3050c81 in group regex-source-integration-test has failed, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:32,943] INFO [GroupCoordinator 0]: Stabilized group regex-source-integration-test generation 8 (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:32,945] INFO Creating topic regex-source-integration-test-testStateStore-changelog with configuration {message.timestamp.type=CreateTime, cleanup.policy=compact} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)
[2020-05-27 17:57:32,948] INFO [Controller id=0] New topics: [HashSet(regex-source-integration-test-testStateStore-changelog)], deleted topics: [HashSet()], new partition replica assignment [Map(regex-source-integration-test-testStateStore-changelog-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)
[2020-05-27 17:57:32,948] INFO [Controller id=0] New partition creation callback for regex-source-integration-test-testStateStore-changelog-0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:32,948] INFO [Controller id=0 epoch=1] Changed partition regex-source-integration-test-testStateStore-changelog-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)
[2020-05-27 17:57:32,948] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:32,948] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:32,951] INFO [Controller id=0 epoch=1] Changed partition regex-source-integration-test-testStateStore-changelog-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)
[2020-05-27 17:57:32,951] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)
[2020-05-27 17:57:32,951] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:32,951] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:32,951] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 46 from controller 0 for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:32,952] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(regex-source-integration-test-testStateStore-changelog-0) (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:32,952] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 46 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)
[2020-05-27 17:57:32,954] INFO [Log partition=regex-source-integration-test-testStateStore-changelog-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)
[2020-05-27 17:57:32,955] INFO [Log partition=regex-source-integration-test-testStateStore-changelog-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)
[2020-05-27 17:57:32,956] INFO Created log for partition regex-source-integration-test-testStateStore-changelog-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit1998193039627945819/junit5739127380229618770/regex-source-integration-test-testStateStore-changelog-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager:66)
[2020-05-27 17:57:32,956] INFO [Partition regex-source-integration-test-testStateStore-changelog-0 broker=0] No checkpointed highwatermark is found for partition regex-source-integration-test-testStateStore-changelog-0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:32,956] INFO [Partition regex-source-integration-test-testStateStore-changelog-0 broker=0] Log loaded for partition regex-source-integration-test-testStateStore-changelog-0 with initial high watermark 0 (kafka.cluster.Partition:66)
[2020-05-27 17:57:32,956] INFO [Broker id=0] Leader regex-source-integration-test-testStateStore-changelog-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR 0 addingReplicas removingReplicas .Previous leader epoch was -1. (state.change.logger:66)
[2020-05-27 17:57:32,958] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 47 (state.change.logger:66)
[2020-05-27 17:57:32,960] INFO Decided on assignment: {32a234e4-5ad8-402f-9fcd-68f8a26eea21=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([0_0=0]) capacity: 1 assigned: 1]} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)
[2020-05-27 17:57:32,960] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer] Assigned tasks to clients as
32a234e4-5ad8-402f-9fcd-68f8a26eea21=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([0_0=0]) capacity: 1 assigned: 1]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:724)
[2020-05-27 17:57:32,960] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:938)
[2020-05-27 17:57:32,960] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] Finished assignment for group at generation 8: {regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer-fe72542d-1ce7-4715-a194-91c454362dcf=Assignment(partitions=[topic-1-0, topic-2-0], userDataSize=56)} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)
[2020-05-27 17:57:32,961] INFO [GroupCoordinator 0]: Assignment received from leader for group regex-source-integration-test for generation 8 (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:32,961] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] Successfully joined group with generation 8 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)
[2020-05-27 17:57:32,962] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] Updating assignment with
Assigned partitions: [topic-1-0, topic-2-0]
Current owned partitions: []
Added partitions (assigned - owned): [topic-1-0, topic-2-0]
Revoked partitions (owned - assigned): []
(org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)
[2020-05-27 17:57:32,963] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] Notifying assignor about the new Assignment(partitions=[topic-1-0, topic-2-0], userDataSize=56) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)
[2020-05-27 17:57:32,963] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1474)
[2020-05-27 17:57:32,963] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] Handle new assignment with:
New active tasks: [0_0]
New standby tasks: []
Existing active tasks: []
Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:176)
[2020-05-27 17:57:32,964] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-consumer, groupId=regex-source-integration-test] Adding newly assigned partitions: topic-1-0, topic-2-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)
[2020-05-27 17:57:32,964] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:915)
[2020-05-27 17:57:32,965] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] task [0_0] Prepared clean close (org.apache.kafka.streams.processor.internals.StreamTask:419)
[2020-05-27 17:57:32,965] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:32,965] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] task [0_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:428)
[2020-05-27 17:57:32,965] INFO [Producer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1183)
[2020-05-27 17:57:32,966] INFO [Consumer clientId=regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1070)
[2020-05-27 17:57:32,968] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:220)
[2020-05-27 17:57:32,968] INFO stream-thread [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:940)
[2020-05-27 17:57:32,969] INFO stream-client [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:279)
[2020-05-27 17:57:32,969] INFO stream-client [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:986)
[2020-05-27 17:57:32,969] INFO stream-client [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21] Already in the pending shutdown state, wait to complete shutdown (org.apache.kafka.streams.KafkaStreams:934)
[2020-05-27 17:57:32,969] INFO stream-client [regex-source-integration-test-32a234e4-5ad8-402f-9fcd-68f8a26eea21] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:986)
[2020-05-27 17:57:32,970] INFO [KafkaServer id=0] shutting down (kafka.server.KafkaServer:66)
[2020-05-27 17:57:32,970] INFO [KafkaServer id=0] Starting controlled shutdown (kafka.server.KafkaServer:66)
[2020-05-27 17:57:32,973] INFO [Controller id=0] Shutting down broker 0 (kafka.controller.KafkaController:66)
[2020-05-27 17:57:32,974] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)
[2020-05-27 17:57:32,974] INFO [KafkaServer id=0] Controlled shutdown succeeded (kafka.server.KafkaServer:66)
[2020-05-27 17:57:32,975] INFO [/config/changes-event-process-thread]: Shutting down (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread:66)
[2020-05-27 17:57:32,975] INFO [/config/changes-event-process-thread]: Stopped (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread:66)
[2020-05-27 17:57:32,975] INFO [/config/changes-event-process-thread]: Shutdown completed (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread:66)
[2020-05-27 17:57:32,975] INFO [SocketServer brokerId=0] Stopping socket server request processors (kafka.network.SocketServer:66)
[2020-05-27 17:57:32,976] INFO [SocketServer brokerId=0] Stopped socket server request processors (kafka.network.SocketServer:66)
[2020-05-27 17:57:32,977] INFO [data-plane Kafka Request Handler on Broker 0], shutting down (kafka.server.KafkaRequestHandlerPool:66)
[2020-05-27 17:57:32,977] INFO [data-plane Kafka Request Handler on Broker 0], shut down completely (kafka.server.KafkaRequestHandlerPool:66)
[2020-05-27 17:57:32,977] INFO [ExpirationReaper-0-AlterAcls]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,043] INFO [ExpirationReaper-0-AlterAcls]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,043] INFO [ExpirationReaper-0-AlterAcls]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,043] INFO [KafkaApi-0] Shutdown complete. (kafka.server.KafkaApis:66)
[2020-05-27 17:57:33,044] INFO [ExpirationReaper-0-topic]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,054] INFO [ExpirationReaper-0-topic]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,054] INFO [ExpirationReaper-0-topic]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,055] INFO [TransactionCoordinator id=0] Shutting down. (kafka.coordinator.transaction.TransactionCoordinator:66)
[2020-05-27 17:57:33,055] INFO [ProducerId Manager 0]: Shutdown complete: last producerId assigned 0 (kafka.coordinator.transaction.ProducerIdManager:66)
[2020-05-27 17:57:33,055] INFO [Transaction State Manager 0]: Shutdown complete (kafka.coordinator.transaction.TransactionStateManager:66)
[2020-05-27 17:57:33,055] INFO [Transaction Marker Channel Manager 0]: Shutting down (kafka.coordinator.transaction.TransactionMarkerChannelManager:66)
[2020-05-27 17:57:33,056] INFO [Transaction Marker Channel Manager 0]: Stopped (kafka.coordinator.transaction.TransactionMarkerChannelManager:66)
[2020-05-27 17:57:33,056] INFO [Transaction Marker Channel Manager 0]: Shutdown completed (kafka.coordinator.transaction.TransactionMarkerChannelManager:66)
[2020-05-27 17:57:33,056] INFO [TransactionCoordinator id=0] Shutdown complete. (kafka.coordinator.transaction.TransactionCoordinator:66)
[2020-05-27 17:57:33,056] INFO [GroupCoordinator 0]: Shutting down. (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:33,056] INFO [ExpirationReaper-0-Heartbeat]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,142] INFO [ExpirationReaper-0-Heartbeat]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,142] INFO [ExpirationReaper-0-Heartbeat]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,143] INFO [ExpirationReaper-0-Rebalance]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,178] INFO [ExpirationReaper-0-Rebalance]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,178] INFO [ExpirationReaper-0-Rebalance]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,179] INFO [GroupCoordinator 0]: Shutdown complete. (kafka.coordinator.group.GroupCoordinator:66)
[2020-05-27 17:57:33,179] INFO [ReplicaManager broker=0] Shutting down (kafka.server.ReplicaManager:66)
[2020-05-27 17:57:33,180] INFO [LogDirFailureHandler]: Shutting down (kafka.server.ReplicaManager$LogDirFailureHandler:66)
[2020-05-27 17:57:33,180] INFO [LogDirFailureHandler]: Stopped (kafka.server.ReplicaManager$LogDirFailureHandler:66)
[2020-05-27 17:57:33,180] INFO [LogDirFailureHandler]: Shutdown completed (kafka.server.ReplicaManager$LogDirFailureHandler:66)
[2020-05-27 17:57:33,180] INFO [ReplicaFetcherManager on broker 0] shutting down (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:33,180] INFO [ReplicaFetcherManager on broker 0] shutdown completed (kafka.server.ReplicaFetcherManager:66)
[2020-05-27 17:57:33,180] INFO [ReplicaAlterLogDirsManager on broker 0] shutting down (kafka.server.ReplicaAlterLogDirsManager:66)
[2020-05-27 17:57:33,180] INFO [ReplicaAlterLogDirsManager on broker 0] shutdown completed (kafka.server.ReplicaAlterLogDirsManager:66)
[2020-05-27 17:57:33,181] INFO [ExpirationReaper-0-Fetch]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,221] INFO [ExpirationReaper-0-Fetch]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,221] INFO [ExpirationReaper-0-Fetch]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,221] INFO [ExpirationReaper-0-Produce]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,360] INFO [ExpirationReaper-0-Produce]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,360] INFO [ExpirationReaper-0-Produce]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,361] INFO [ExpirationReaper-0-DeleteRecords]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,386] INFO [ExpirationReaper-0-DeleteRecords]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,386] INFO [ExpirationReaper-0-DeleteRecords]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,387] INFO [ExpirationReaper-0-ElectLeader]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,400] INFO [ExpirationReaper-0-ElectLeader]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,400] INFO [ExpirationReaper-0-ElectLeader]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper:66)
[2020-05-27 17:57:33,403] INFO [ReplicaManager broker=0] Shut down completely (kafka.server.ReplicaManager:66)
[2020-05-27 17:57:33,403] INFO Shutting down. (kafka.log.LogManager:66)
[2020-05-27 17:57:33,403] INFO Shutting down the log cleaner. (kafka.log.LogCleaner:66)
[2020-05-27 17:57:33,403] INFO [kafka-log-cleaner-thread-0]: Shutting down (kafka.log.LogCleaner:66)
[2020-05-27 17:57:33,404] INFO [kafka-log-cleaner-thread-0]: Stopped (kafka.log.LogCleaner:66)
[2020-05-27 17:57:33,404] INFO [kafka-log-cleaner-thread-0]: Shutdown completed (kafka.log.LogCleaner:66)
[2020-05-27 17:57:33,405] INFO [ProducerStateManager partition=TEST-TOPIC-2-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,406] INFO [ProducerStateManager partition=topic-2-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,409] INFO [ProducerStateManager partition=topic-1-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,411] INFO [ProducerStateManager partition=topic-C-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,412] INFO [ProducerStateManager partition=outputTopic_2-0] Writing producer snapshot at offset 6 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,413] INFO [ProducerStateManager partition=foo-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,415] INFO [ProducerStateManager partition=__consumer_offsets-3] Writing producer snapshot at offset 20 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,416] INFO [ProducerStateManager partition=topic-A-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,417] INFO [ProducerStateManager partition=topic-Y-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,418] INFO [ProducerStateManager partition=__consumer_offsets-4] Writing producer snapshot at offset 22 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,420] INFO [ProducerStateManager partition=TEST-TOPIC-1-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,421] INFO [ProducerStateManager partition=topic-Z-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,423] INFO [ProducerStateManager partition=fa-0] Writing producer snapshot at offset 1 (kafka.log.ProducerStateManager:66)
[2020-05-27 17:57:33,431] INFO Shutdown complete. (kafka.log.LogManager:66)
[2020-05-27 17:57:33,431] INFO [ControllerEventThread controllerId=0] Shutting down (kafka.controller.ControllerEventManager$ControllerEventThread:66)
[2020-05-27 17:57:33,431] INFO [ControllerEventThread controllerId=0] Stopped (kafka.controller.ControllerEventManager$ControllerEventThread:66)
[2020-05-27 17:57:33,431] INFO [ControllerEventThread controllerId=0] Shutdown completed (kafka.controller.ControllerEventManager$ControllerEventThread:66)
[2020-05-27 17:57:33,431] INFO [PartitionStateMachine controllerId=0] Stopped partition state machine (kafka.controller.ZkPartitionStateMachine:66)
[2020-05-27 17:57:33,432] INFO [ReplicaStateMachine controllerId=0] Stopped replica state machine (kafka.controller.ZkReplicaStateMachine:66)
[2020-05-27 17:57:33,432] INFO [RequestSendThread controllerId=0] Shutting down (kafka.controller.RequestSendThread:66)
[2020-05-27 17:57:33,432] INFO [RequestSendThread controllerId=0] Stopped (kafka.controller.RequestSendThread:66)
[2020-05-27 17:57:33,432] INFO [RequestSendThread controllerId=0] Shutdown completed (kafka.controller.RequestSendThread:66)
[2020-05-27 17:57:33,432] INFO [Controller id=0] Resigned (kafka.controller.KafkaController:66)
[2020-05-27 17:57:33,433] INFO [ZooKeeperClient Kafka server] Closing. (kafka.zookeeper.ZooKeeperClient:66)
[2020-05-27 17:57:33,540] INFO Session: 0x100592733730000 closed (org.apache.zookeeper.ZooKeeper:1422)
[2020-05-27 17:57:33,540] INFO EventThread shut down for session: 0x100592733730000 (org.apache.zookeeper.ClientCnxn:524)
[2020-05-27 17:57:33,541] INFO [ZooKeeperClient Kafka server] Closed. (kafka.zookeeper.ZooKeeperClient:66)
[2020-05-27 17:57:33,541] INFO [ThrottledChannelReaper-Fetch]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:57:33,621] INFO [ThrottledChannelReaper-Fetch]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:57:33,621] INFO [ThrottledChannelReaper-Fetch]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:57:33,621] INFO [ThrottledChannelReaper-Produce]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:57:34,621] INFO [ThrottledChannelReaper-Produce]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:57:34,621] INFO [ThrottledChannelReaper-Produce]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:57:34,621] INFO [ThrottledChannelReaper-Request]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:57:34,654] INFO [ThrottledChannelReaper-Request]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:57:34,654] INFO [ThrottledChannelReaper-Request]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper:66)
[2020-05-27 17:57:34,655] INFO [SocketServer brokerId=0] Shutting down socket server (kafka.network.SocketServer:66)
[2020-05-27 17:57:34,663] INFO [SocketServer brokerId=0] Shutdown completed (kafka.network.SocketServer:66)
[2020-05-27 17:57:34,664] INFO [KafkaServer id=0] shut down completed (kafka.server.KafkaServer:66)
[2020-05-27 17:57:34,682] INFO ConnnectionExpirerThread interrupted (org.apache.zookeeper.server.NIOServerCnxnFactory:583)
[2020-05-27 17:57:34,682] INFO accept thread exitted run method (org.apache.zookeeper.server.NIOServerCnxnFactory:219)
[2020-05-27 17:57:34,682] INFO selector thread exitted run method (org.apache.zookeeper.server.NIOServerCnxnFactory:420)
[2020-05-27 17:57:34,682] INFO selector thread exitted run method (org.apache.zookeeper.server.NIOServerCnxnFactory:420)
[2020-05-27 17:57:34,682] INFO shutting down (org.apache.zookeeper.server.ZooKeeperServer:558)
[2020-05-27 17:57:34,682] INFO Shutting down (org.apache.zookeeper.server.SessionTrackerImpl:237)
[2020-05-27 17:57:34,683] INFO Shutting down (org.apache.zookeeper.server.PrepRequestProcessor:1007)
[2020-05-27 17:57:34,683] INFO Shutting down (org.apache.zookeeper.server.SyncRequestProcessor:191)
[2020-05-27 17:57:34,683] INFO PrepRequestProcessor exited loop! (org.apache.zookeeper.server.PrepRequestProcessor:155)
[2020-05-27 17:57:34,683] INFO SyncRequestProcessor exited! (org.apache.zookeeper.server.SyncRequestProcessor:169)
[2020-05-27 17:57:34,683] INFO shutdown of request processor complete (org.apache.zookeeper.server.FinalRequestProcessor:514)