{\rtf1\ansi\ansicpg1252\cocoartf2512
\cocoatextscaling0\cocoaplatform0{\fonttbl\f0\fmodern\fcharset0 Courier;}
{\colortbl;\red255\green255\blue255;\red0\green0\blue0;}
{\*\expandedcolortbl;;\cssrgb\c0\c0\c0;}
\margl1440\margr1440\vieww10800\viewh8400\viewkind0
\deftab720
\pard\pardeftab720\partightenfactor0

\f0\fs24 \cf2 \expnd0\expndtw0\kerning0
\outl0\strokewidth0 \strokec2 [2020-06-10 20:55:20,549] INFO AdminClientConfig values: \
	bootstrap.servers = [localhost:59678]\
	client.dns.lookup = use_all_dns_ips\
	client.id = \
	connections.max.idle.ms = 300000\
	default.api.timeout.ms = 60000\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	receive.buffer.bytes = 65536\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
 (org.apache.kafka.clients.admin.AdminClientConfig:354)\
[2020-06-10 20:55:20,571] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:20,572] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:20,573] INFO Kafka startTimeMs: 1591847720570 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:20,610] INFO Creating topic input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi with configuration \{\} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)\
[2020-06-10 20:55:20,646] INFO [Controller id=0] New topics: [HashSet(input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi)], deleted topics: [HashSet()], new partition replica assignment [Map(input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)\
[2020-06-10 20:55:20,646] INFO [Controller id=0] New partition creation callback for input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 (kafka.controller.KafkaController:66)\
[2020-06-10 20:55:20,647] INFO [Controller id=0 epoch=1] Changed partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2020-06-10 20:55:20,652] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2020-06-10 20:55:20,668] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2020-06-10 20:55:20,811] INFO [Controller id=0 epoch=1] Changed partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2020-06-10 20:55:20,812] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)\
[2020-06-10 20:55:20,823] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)\
[2020-06-10 20:55:20,826] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2020-06-10 20:55:20,842] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 11 from controller 0 for 1 partitions (state.change.logger:66)\
[2020-06-10 20:55:20,872] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0) (kafka.server.ReplicaFetcherManager:66)\
[2020-06-10 20:55:20,873] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 11 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)\
[2020-06-10 20:55:20,914] INFO [Log partition=input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit4556314779568824159/junit7538242801421088820] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)\
[2020-06-10 20:55:20,922] INFO [Log partition=input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit4556314779568824159/junit7538242801421088820] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)\
[2020-06-10 20:55:20,932] INFO Created log for partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit4556314779568824159/junit7538242801421088820/input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 with properties \{compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760\}. (kafka.log.LogManager:66)\
[2020-06-10 20:55:20,934] INFO [Partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 broker=0] No checkpointed highwatermark is found for partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 (kafka.cluster.Partition:66)\
[2020-06-10 20:55:20,936] INFO [Partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 broker=0] Log loaded for partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2020-06-10 20:55:20,937] INFO [Broker id=0] Leader input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2020-06-10 20:55:20,946] INFO [Broker id=0] Finished LeaderAndIsr request in 0ms correlationId 11 from controller 0 for 1 partitions (state.change.logger:66)\
[2020-06-10 20:55:20,956] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 12 (state.change.logger:66)\
[2020-06-10 20:55:20,987] INFO StreamsConfig values: \
	acceptable.recovery.lag = 10000\
	application.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi\
	application.server = \
	bootstrap.servers = [localhost:59678]\
	buffered.records.per.partition = 1000\
	built.in.metrics.version = latest\
	cache.max.bytes.buffering = 0\
	client.id = \
	commit.interval.ms = 1000\
	connections.max.idle.ms = 540000\
	default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler\
	default.key.serde = class org.apache.kafka.common.serialization.Serdes$IntegerSerde\
	default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler\
	default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp\
	default.value.serde = class org.apache.kafka.common.serialization.Serdes$IntegerSerde\
	max.task.idle.ms = 0\
	max.warmup.replicas = 2\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	num.standby.replicas = 0\
	num.stream.threads = 1\
	partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper\
	poll.ms = 100\
	probing.rebalance.interval.ms = 600000\
	processing.guarantee = at_least_once\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	replication.factor = 1\
	request.timeout.ms = 40000\
	retries = 0\
	retry.backoff.ms = 100\
	rocksdb.config.setter = null\
	security.protocol = PLAINTEXT\
	send.buffer.bytes = 131072\
	state.cleanup.delay.ms = 600000\
	state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-5921107926961761805\
	topology.optimization = none\
	upgrade.from = null\
	windowstore.changelog.additional.retention.ms = 86400000\
 (org.apache.kafka.streams.StreamsConfig:354)\
[2020-06-10 20:55:20,993] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:699)\
[2020-06-10 20:55:20,993] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:700)\
[2020-06-10 20:55:20,995] INFO AdminClientConfig values: \
	bootstrap.servers = [localhost:59678]\
	client.dns.lookup = use_all_dns_ips\
	client.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-admin\
	connections.max.idle.ms = 300000\
	default.api.timeout.ms = 60000\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	receive.buffer.bytes = 65536\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
 (org.apache.kafka.clients.admin.AdminClientConfig:354)\
[2020-06-10 20:55:21,051] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:21,052] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:21,052] INFO Kafka startTimeMs: 1591847721051 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:21,054] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:304)\
[2020-06-10 20:55:21,057] INFO ConsumerConfig values: \
	allow.auto.create.topics = true\
	auto.commit.interval.ms = 5000\
	auto.offset.reset = none\
	bootstrap.servers = [localhost:59678]\
	check.crcs = true\
	client.dns.lookup = use_all_dns_ips\
	client.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-restore-consumer\
	client.rack = \
	connections.max.idle.ms = 540000\
	default.api.timeout.ms = 60000\
	enable.auto.commit = false\
	exclude.internal.topics = true\
	fetch.max.bytes = 52428800\
	fetch.max.wait.ms = 500\
	fetch.min.bytes = 1\
	group.id = null\
	group.instance.id = null\
	heartbeat.interval.ms = 3000\
	interceptor.classes = []\
	internal.leave.group.on.close = false\
	internal.throw.on.fetch.stable.offset.unsupported = false\
	isolation.level = read_uncommitted\
	key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer\
	max.partition.fetch.bytes = 1048576\
	max.poll.interval.ms = 300000\
	max.poll.records = 1000\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]\
	receive.buffer.bytes = 65536\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	session.timeout.ms = 10000\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer\
 (org.apache.kafka.clients.consumer.ConsumerConfig:354)\
[2020-06-10 20:55:21,067] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:21,068] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:21,068] INFO Kafka startTimeMs: 1591847721067 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:21,069] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)\
[2020-06-10 20:55:21,073] INFO ProducerConfig values: \
	acks = 1\
	batch.size = 16384\
	bootstrap.servers = [localhost:59678]\
	buffer.memory = 33554432\
	client.dns.lookup = use_all_dns_ips\
	client.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-producer\
	compression.type = none\
	connections.max.idle.ms = 540000\
	delivery.timeout.ms = 120000\
	enable.idempotence = false\
	interceptor.classes = []\
	internal.auto.downgrade.txn.commit = false\
	key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer\
	linger.ms = 100\
	max.block.ms = 60000\
	max.in.flight.requests.per.connection = 5\
	max.request.size = 1048576\
	metadata.max.age.ms = 300000\
	metadata.max.idle.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.timeout.ms = 10000\
	transactional.id = null\
	value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer\
 (org.apache.kafka.clients.producer.ProducerConfig:354)\
[2020-06-10 20:55:21,079] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:21,080] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:21,081] INFO Kafka startTimeMs: 1591847721079 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:21,081] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:353)\
[2020-06-10 20:55:21,082] INFO ConsumerConfig values: \
	allow.auto.create.topics = false\
	auto.commit.interval.ms = 5000\
	auto.offset.reset = earliest\
	bootstrap.servers = [localhost:59678]\
	check.crcs = true\
	client.dns.lookup = use_all_dns_ips\
	client.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer\
	client.rack = \
	connections.max.idle.ms = 540000\
	default.api.timeout.ms = 60000\
	enable.auto.commit = false\
	exclude.internal.topics = true\
	fetch.max.bytes = 52428800\
	fetch.max.wait.ms = 500\
	fetch.min.bytes = 1\
	group.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi\
	group.instance.id = null\
	heartbeat.interval.ms = 3000\
	interceptor.classes = []\
	internal.leave.group.on.close = false\
	internal.throw.on.fetch.stable.offset.unsupported = false\
	isolation.level = read_uncommitted\
	key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer\
	max.partition.fetch.bytes = 1048576\
	max.poll.interval.ms = 300000\
	max.poll.records = 1000\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partition.assignment.strategy = [org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor]\
	receive.buffer.bytes = 65536\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	session.timeout.ms = 10000\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer\
 (org.apache.kafka.clients.consumer.ConsumerConfig:354)\
[2020-06-10 20:55:21,084] INFO [Producer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-producer] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:21,087] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:215)\
[2020-06-10 20:55:21,090] WARN The configuration 'admin.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:362)\
[2020-06-10 20:55:21,091] WARN The configuration 'admin.retries' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:362)\
[2020-06-10 20:55:21,091] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:21,091] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:21,091] INFO Kafka startTimeMs: 1591847721091 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:21,330] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:282)\
[2020-06-10 20:55:21,335] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:502)\
[2020-06-10 20:55:21,336] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:55:21,342] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Subscribed to topic(s): input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi (org.apache.kafka.clients.consumer.KafkaConsumer:959)\
[2020-06-10 20:55:21,366] INFO ProducerConfig values: \
	acks = -1\
	batch.size = 16384\
	bootstrap.servers = [localhost:59678]\
	buffer.memory = 33554432\
	client.dns.lookup = use_all_dns_ips\
	client.id = producer-26\
	compression.type = none\
	connections.max.idle.ms = 540000\
	delivery.timeout.ms = 120000\
	enable.idempotence = false\
	interceptor.classes = []\
	internal.auto.downgrade.txn.commit = false\
	key.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
	linger.ms = 0\
	max.block.ms = 60000\
	max.in.flight.requests.per.connection = 5\
	max.request.size = 1048576\
	metadata.max.age.ms = 300000\
	metadata.max.idle.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.timeout.ms = 60000\
	transactional.id = null\
	value.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
 (org.apache.kafka.clients.producer.ProducerConfig:354)\
[2020-06-10 20:55:21,372] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:21,489] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Discovered group coordinator localhost:59678 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)\
[2020-06-10 20:55:21,491] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)\
[2020-06-10 20:55:21,555] INFO [Producer clientId=producer-26] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:21,662] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:21,662] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)\
[2020-06-10 20:55:21,663] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)\
[2020-06-10 20:55:21,662] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:21,663] INFO Kafka startTimeMs: 1591847721662 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:21,667] INFO [GroupCoordinator 0]: Preparing to rebalance group app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi in state PreparingRebalance with old generation 0 (__consumer_offsets-3) (reason: Adding new member app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer-c3e8d275-9f14-4cf8-bb95-de672c972bf8 with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:21,670] INFO [GroupCoordinator 0]: Stabilized group app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi generation 1 (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:21,677] INFO [Producer clientId=producer-26] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1186)\
[2020-06-10 20:55:21,687] INFO Creating topic app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog with configuration \{message.timestamp.type=CreateTime, cleanup.policy=compact\} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient:66)\
[2020-06-10 20:55:21,696] INFO [Controller id=0] New topics: [HashSet(app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog)], deleted topics: [HashSet()], new partition replica assignment [Map(app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))] (kafka.controller.KafkaController:66)\
[2020-06-10 20:55:21,697] INFO [Controller id=0] New partition creation callback for app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 (kafka.controller.KafkaController:66)\
[2020-06-10 20:55:21,697] INFO [Controller id=0 epoch=1] Changed partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger:66)\
[2020-06-10 20:55:21,697] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2020-06-10 20:55:21,698] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2020-06-10 20:55:21,708] INFO [Controller id=0 epoch=1] Changed partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isr=List(0), zkVersion=0) (state.change.logger:66)\
[2020-06-10 20:55:21,708] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger:66)\
[2020-06-10 20:55:21,708] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger:66)\
[2020-06-10 20:55:21,709] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger:66)\
[2020-06-10 20:55:21,709] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 13 from controller 0 for 1 partitions (state.change.logger:66)\
[2020-06-10 20:55:21,713] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0) (kafka.server.ReplicaFetcherManager:66)\
[2020-06-10 20:55:21,713] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 13 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger:66)\
[2020-06-10 20:55:21,722] INFO [Log partition=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit4556314779568824159/junit7538242801421088820] Loading producer state till offset 0 with message format version 2 (kafka.log.Log:66)\
[2020-06-10 20:55:21,730] INFO [Log partition=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0, dir=/var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit4556314779568824159/junit7538242801421088820] Completed load of log with 1 segments, log start offset 0 and log end offset 0 in 0 ms (kafka.log.Log:66)\
[2020-06-10 20:55:21,736] INFO Created log for partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 in /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/junit4556314779568824159/junit7538242801421088820/app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 with properties \{compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.6-IV0, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1000000, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760\}. (kafka.log.LogManager:66)\
[2020-06-10 20:55:21,738] INFO [Partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 broker=0] No checkpointed highwatermark is found for partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 (kafka.cluster.Partition:66)\
[2020-06-10 20:55:21,740] INFO [Partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 broker=0] Log loaded for partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 with initial high watermark 0 (kafka.cluster.Partition:66)\
[2020-06-10 20:55:21,740] INFO [Broker id=0] Leader app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 starts at leader epoch 0 from offset 0 with high watermark 0 ISR [0] addingReplicas [] removingReplicas []. Previous leader epoch was -1. (state.change.logger:66)\
[2020-06-10 20:55:21,747] INFO [Broker id=0] Finished LeaderAndIsr request in 0ms correlationId 13 from controller 0 for 1 partitions (state.change.logger:66)\
[2020-06-10 20:55:21,751] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 14 (state.change.logger:66)\
[2020-06-10 20:55:21,765] INFO Decided on assignment: \{db370b48-7d46-4cd9-8d22-5bb7ccaa8052=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([0_0=0]) capacity: 1 assigned: 1]\} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)\
[2020-06-10 20:55:21,770] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer] Assigned tasks to clients as \
db370b48-7d46-4cd9-8d22-5bb7ccaa8052=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([]) taskLagTotals: ([0_0=0]) capacity: 1 assigned: 1]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:714)\
[2020-06-10 20:55:21,777] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:926)\
[2020-06-10 20:55:21,777] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Finished assignment for group at generation 1: \{app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer-c3e8d275-9f14-4cf8-bb95-de672c972bf8=Assignment(partitions=[input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0], userDataSize=48)\} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)\
[2020-06-10 20:55:21,781] INFO [GroupCoordinator 0]: Assignment received from leader for group app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi for generation 1 (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:21,786] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Successfully joined group with generation 1 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)\
[2020-06-10 20:55:21,798] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Updating assignment with\
	Assigned partitions:                       [input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0]\
	Current owned partitions:                  []\
	Added partitions (assigned - owned):       [input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0]\
	Revoked partitions (owned - assigned):     []\
 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)\
[2020-06-10 20:55:21,799] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Notifying assignor about the new Assignment(partitions=[input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0], userDataSize=48) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)\
[2020-06-10 20:55:21,802] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1370)\
[2020-06-10 20:55:21,802] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] Handle new assignment with:\
	New active tasks: [0_0]\
	New standby tasks: []\
	Existing active tasks: []\
	Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:188)\
[2020-06-10 20:55:21,807] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Adding newly assigned partitions: input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)\
[2020-06-10 20:55:21,808] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:55:21,811] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Found no committed offset for partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)\
[2020-06-10 20:55:21,819] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Resetting offset for partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:397)\
[2020-06-10 20:55:22,098] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] task [0_0] State store store did not find checkpoint offset, hence would default to the starting offset at changelog app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 (org.apache.kafka.streams.processor.internals.ProcessorStateManager:244)\
[2020-06-10 20:55:22,100] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] task [0_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:214)\
[2020-06-10 20:55:22,172] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-restore-consumer, groupId=null] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:22,217] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-restore-consumer, groupId=null] Subscribed to partition(s): app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 (org.apache.kafka.clients.consumer.KafkaConsumer:1114)\
[2020-06-10 20:55:22,224] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-restore-consumer, groupId=null] Seeking to EARLIEST offset of partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 (org.apache.kafka.clients.consumer.internals.SubscriptionState:602)\
[2020-06-10 20:55:22,247] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-restore-consumer, groupId=null] Resetting offset for partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 to offset 0. (org.apache.kafka.clients.consumer.internals.SubscriptionState:397)\
[2020-06-10 20:55:22,362] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] Finished restoring changelog app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 to store store with a total number of 0 records (org.apache.kafka.streams.processor.internals.StoreChangelogReader:534)\
[2020-06-10 20:55:22,402] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Found no committed offset for partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:1342)\
[2020-06-10 20:55:22,466] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] task [0_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:235)\
[2020-06-10 20:55:22,466] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:55:22,467] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:282)\
[2020-06-10 20:55:22,877] INFO ProducerConfig values: \
	acks = -1\
	batch.size = 16384\
	bootstrap.servers = [localhost:59678]\
	buffer.memory = 33554432\
	client.dns.lookup = use_all_dns_ips\
	client.id = producer-27\
	compression.type = none\
	connections.max.idle.ms = 540000\
	delivery.timeout.ms = 120000\
	enable.idempotence = false\
	interceptor.classes = []\
	internal.auto.downgrade.txn.commit = false\
	key.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
	linger.ms = 0\
	max.block.ms = 60000\
	max.in.flight.requests.per.connection = 5\
	max.request.size = 1048576\
	metadata.max.age.ms = 300000\
	metadata.max.idle.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.timeout.ms = 60000\
	transactional.id = null\
	value.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
 (org.apache.kafka.clients.producer.ProducerConfig:354)\
[2020-06-10 20:55:22,882] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:22,883] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:22,884] INFO Kafka startTimeMs: 1591847722882 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:22,891] INFO [Producer clientId=producer-27] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:22,894] INFO [Producer clientId=producer-27] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1186)\
[2020-06-10 20:55:22,908] INFO ProducerConfig values: \
	acks = -1\
	batch.size = 16384\
	bootstrap.servers = [localhost:59678]\
	buffer.memory = 33554432\
	client.dns.lookup = use_all_dns_ips\
	client.id = producer-28\
	compression.type = none\
	connections.max.idle.ms = 540000\
	delivery.timeout.ms = 120000\
	enable.idempotence = false\
	interceptor.classes = []\
	internal.auto.downgrade.txn.commit = false\
	key.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
	linger.ms = 0\
	max.block.ms = 60000\
	max.in.flight.requests.per.connection = 5\
	max.request.size = 1048576\
	metadata.max.age.ms = 300000\
	metadata.max.idle.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.timeout.ms = 60000\
	transactional.id = null\
	value.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
 (org.apache.kafka.clients.producer.ProducerConfig:354)\
[2020-06-10 20:55:22,939] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:22,941] INFO [Producer clientId=producer-28] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:22,973] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:22,977] INFO Kafka startTimeMs: 1591847722939 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:22,997] INFO [Producer clientId=producer-28] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1186)\
[2020-06-10 20:55:23,119] INFO ProducerConfig values: \
	acks = -1\
	batch.size = 16384\
	bootstrap.servers = [localhost:59678]\
	buffer.memory = 33554432\
	client.dns.lookup = use_all_dns_ips\
	client.id = producer-29\
	compression.type = none\
	connections.max.idle.ms = 540000\
	delivery.timeout.ms = 120000\
	enable.idempotence = false\
	interceptor.classes = []\
	internal.auto.downgrade.txn.commit = false\
	key.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
	linger.ms = 0\
	max.block.ms = 60000\
	max.in.flight.requests.per.connection = 5\
	max.request.size = 1048576\
	metadata.max.age.ms = 300000\
	metadata.max.idle.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.timeout.ms = 60000\
	transactional.id = null\
	value.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
 (org.apache.kafka.clients.producer.ProducerConfig:354)\
[2020-06-10 20:55:23,125] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:23,128] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:23,129] INFO Kafka startTimeMs: 1591847723125 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:23,130] INFO [Producer clientId=producer-29] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:23,137] INFO [Producer clientId=producer-29] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1186)\
[2020-06-10 20:55:23,257] INFO ProducerConfig values: \
	acks = -1\
	batch.size = 16384\
	bootstrap.servers = [localhost:59678]\
	buffer.memory = 33554432\
	client.dns.lookup = use_all_dns_ips\
	client.id = producer-30\
	compression.type = none\
	connections.max.idle.ms = 540000\
	delivery.timeout.ms = 120000\
	enable.idempotence = false\
	interceptor.classes = []\
	internal.auto.downgrade.txn.commit = false\
	key.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
	linger.ms = 0\
	max.block.ms = 60000\
	max.in.flight.requests.per.connection = 5\
	max.request.size = 1048576\
	metadata.max.age.ms = 300000\
	metadata.max.idle.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.timeout.ms = 60000\
	transactional.id = null\
	value.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
 (org.apache.kafka.clients.producer.ProducerConfig:354)\
[2020-06-10 20:55:23,273] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:23,277] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:23,277] INFO Kafka startTimeMs: 1591847723273 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:23,279] INFO [Producer clientId=producer-30] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:23,287] INFO [Producer clientId=producer-30] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1186)\
[2020-06-10 20:55:23,388] INFO ProducerConfig values: \
	acks = -1\
	batch.size = 16384\
	bootstrap.servers = [localhost:59678]\
	buffer.memory = 33554432\
	client.dns.lookup = use_all_dns_ips\
	client.id = producer-31\
	compression.type = none\
	connections.max.idle.ms = 540000\
	delivery.timeout.ms = 120000\
	enable.idempotence = false\
	interceptor.classes = []\
	internal.auto.downgrade.txn.commit = false\
	key.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
	linger.ms = 0\
	max.block.ms = 60000\
	max.in.flight.requests.per.connection = 5\
	max.request.size = 1048576\
	metadata.max.age.ms = 300000\
	metadata.max.idle.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.timeout.ms = 60000\
	transactional.id = null\
	value.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
 (org.apache.kafka.clients.producer.ProducerConfig:354)\
[2020-06-10 20:55:23,409] INFO [Producer clientId=producer-31] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:23,411] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:23,429] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:23,437] INFO Kafka startTimeMs: 1591847723401 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:23,457] INFO [Producer clientId=producer-31] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1186)\
[2020-06-10 20:55:23,527] INFO ProducerConfig values: \
	acks = -1\
	batch.size = 16384\
	bootstrap.servers = [localhost:59678]\
	buffer.memory = 33554432\
	client.dns.lookup = use_all_dns_ips\
	client.id = producer-32\
	compression.type = none\
	connections.max.idle.ms = 540000\
	delivery.timeout.ms = 120000\
	enable.idempotence = false\
	interceptor.classes = []\
	internal.auto.downgrade.txn.commit = false\
	key.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
	linger.ms = 0\
	max.block.ms = 60000\
	max.in.flight.requests.per.connection = 5\
	max.request.size = 1048576\
	metadata.max.age.ms = 300000\
	metadata.max.idle.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.timeout.ms = 60000\
	transactional.id = null\
	value.serializer = class org.apache.kafka.common.serialization.IntegerSerializer\
 (org.apache.kafka.clients.producer.ProducerConfig:354)\
[2020-06-10 20:55:23,532] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:23,535] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:23,535] INFO Kafka startTimeMs: 1591847723532 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:23,536] INFO [Producer clientId=producer-32] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:23,573] INFO [Producer clientId=producer-32] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1186)\
[2020-06-10 20:55:23,601] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:282)\
[2020-06-10 20:55:23,602] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:875)\
[2020-06-10 20:55:23,603] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:55:23,696] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:889)\
[2020-06-10 20:55:23,697] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] task [0_0] Suspended running (org.apache.kafka.streams.processor.internals.StreamTask:270)\
[2020-06-10 20:55:23,704] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1068)\
[2020-06-10 20:55:23,712] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] task [0_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:460)\
[2020-06-10 20:55:23,712] INFO [Producer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1186)\
[2020-06-10 20:55:23,718] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1068)\
[2020-06-10 20:55:23,754] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:55:23,754] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:914)\
[2020-06-10 20:55:23,789] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:282)\
[2020-06-10 20:55:23,789] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:991)\
[2020-06-10 20:55:23,796] INFO StreamsConfig values: \
	acceptable.recovery.lag = 10000\
	application.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi\
	application.server = \
	bootstrap.servers = [localhost:59678]\
	buffered.records.per.partition = 1000\
	built.in.metrics.version = latest\
	cache.max.bytes.buffering = 0\
	client.id = \
	commit.interval.ms = 1000\
	connections.max.idle.ms = 540000\
	default.deserialization.exception.handler = class org.apache.kafka.streams.errors.LogAndFailExceptionHandler\
	default.key.serde = class org.apache.kafka.common.serialization.Serdes$IntegerSerde\
	default.production.exception.handler = class org.apache.kafka.streams.errors.DefaultProductionExceptionHandler\
	default.timestamp.extractor = class org.apache.kafka.streams.processor.FailOnInvalidTimestamp\
	default.value.serde = class org.apache.kafka.common.serialization.Serdes$IntegerSerde\
	max.task.idle.ms = 0\
	max.warmup.replicas = 2\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	num.standby.replicas = 0\
	num.stream.threads = 1\
	partition.grouper = class org.apache.kafka.streams.processor.DefaultPartitionGrouper\
	poll.ms = 100\
	probing.rebalance.interval.ms = 600000\
	processing.guarantee = at_least_once\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	replication.factor = 1\
	request.timeout.ms = 40000\
	retries = 0\
	retry.backoff.ms = 100\
	rocksdb.config.setter = null\
	security.protocol = PLAINTEXT\
	send.buffer.bytes = 131072\
	state.cleanup.delay.ms = 600000\
	state.dir = /var/folders/fl/5v9zwt251szdlb5gk0j3cc7c0000gn/T/kafka-5921107926961761805\
	topology.optimization = none\
	upgrade.from = null\
	windowstore.changelog.additional.retention.ms = 86400000\
 (org.apache.kafka.streams.StreamsConfig:354)\
[2020-06-10 20:55:23,801] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5] Kafka Streams version: test-version (org.apache.kafka.streams.KafkaStreams:699)\
[2020-06-10 20:55:23,801] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5] Kafka Streams commit ID: test-commit-ID (org.apache.kafka.streams.KafkaStreams:700)\
[2020-06-10 20:55:23,803] INFO AdminClientConfig values: \
	bootstrap.servers = [localhost:59678]\
	client.dns.lookup = use_all_dns_ips\
	client.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-admin\
	connections.max.idle.ms = 300000\
	default.api.timeout.ms = 60000\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	receive.buffer.bytes = 65536\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
 (org.apache.kafka.clients.admin.AdminClientConfig:354)\
[2020-06-10 20:55:23,808] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:23,809] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:23,809] INFO Kafka startTimeMs: 1591847723808 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:23,809] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] Creating restore consumer client (org.apache.kafka.streams.processor.internals.StreamThread:304)\
[2020-06-10 20:55:23,810] INFO ConsumerConfig values: \
	allow.auto.create.topics = true\
	auto.commit.interval.ms = 5000\
	auto.offset.reset = none\
	bootstrap.servers = [localhost:59678]\
	check.crcs = true\
	client.dns.lookup = use_all_dns_ips\
	client.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-restore-consumer\
	client.rack = \
	connections.max.idle.ms = 540000\
	default.api.timeout.ms = 60000\
	enable.auto.commit = false\
	exclude.internal.topics = true\
	fetch.max.bytes = 52428800\
	fetch.max.wait.ms = 500\
	fetch.min.bytes = 1\
	group.id = null\
	group.instance.id = null\
	heartbeat.interval.ms = 3000\
	interceptor.classes = []\
	internal.leave.group.on.close = false\
	internal.throw.on.fetch.stable.offset.unsupported = false\
	isolation.level = read_uncommitted\
	key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer\
	max.partition.fetch.bytes = 1048576\
	max.poll.interval.ms = 300000\
	max.poll.records = 1000\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]\
	receive.buffer.bytes = 65536\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	session.timeout.ms = 10000\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer\
 (org.apache.kafka.clients.consumer.ConsumerConfig:354)\
[2020-06-10 20:55:23,815] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:23,816] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:23,816] INFO Kafka startTimeMs: 1591847723815 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:23,816] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] Creating thread producer client (org.apache.kafka.streams.processor.internals.StreamThread:96)\
[2020-06-10 20:55:23,817] INFO ProducerConfig values: \
	acks = 1\
	batch.size = 16384\
	bootstrap.servers = [localhost:59678]\
	buffer.memory = 33554432\
	client.dns.lookup = use_all_dns_ips\
	client.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-producer\
	compression.type = none\
	connections.max.idle.ms = 540000\
	delivery.timeout.ms = 120000\
	enable.idempotence = false\
	interceptor.classes = []\
	internal.auto.downgrade.txn.commit = false\
	key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer\
	linger.ms = 100\
	max.block.ms = 60000\
	max.in.flight.requests.per.connection = 5\
	max.request.size = 1048576\
	metadata.max.age.ms = 300000\
	metadata.max.idle.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner\
	receive.buffer.bytes = 32768\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retries = 2147483647\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	transaction.timeout.ms = 10000\
	transactional.id = null\
	value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer\
 (org.apache.kafka.clients.producer.ProducerConfig:354)\
[2020-06-10 20:55:23,951] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:23,955] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:23,958] INFO Kafka startTimeMs: 1591847723949 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:23,963] INFO [Producer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-producer] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:23,966] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] Creating consumer client (org.apache.kafka.streams.processor.internals.StreamThread:353)\
[2020-06-10 20:55:23,980] INFO ConsumerConfig values: \
	allow.auto.create.topics = false\
	auto.commit.interval.ms = 5000\
	auto.offset.reset = earliest\
	bootstrap.servers = [localhost:59678]\
	check.crcs = true\
	client.dns.lookup = use_all_dns_ips\
	client.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer\
	client.rack = \
	connections.max.idle.ms = 540000\
	default.api.timeout.ms = 60000\
	enable.auto.commit = false\
	exclude.internal.topics = true\
	fetch.max.bytes = 52428800\
	fetch.max.wait.ms = 500\
	fetch.min.bytes = 1\
	group.id = app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi\
	group.instance.id = null\
	heartbeat.interval.ms = 3000\
	interceptor.classes = []\
	internal.leave.group.on.close = false\
	internal.throw.on.fetch.stable.offset.unsupported = false\
	isolation.level = read_uncommitted\
	key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer\
	max.partition.fetch.bytes = 1048576\
	max.poll.interval.ms = 300000\
	max.poll.records = 1000\
	metadata.max.age.ms = 300000\
	metric.reporters = []\
	metrics.num.samples = 2\
	metrics.recording.level = INFO\
	metrics.sample.window.ms = 30000\
	partition.assignment.strategy = [org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor]\
	receive.buffer.bytes = 65536\
	reconnect.backoff.max.ms = 1000\
	reconnect.backoff.ms = 50\
	request.timeout.ms = 30000\
	retry.backoff.ms = 100\
	sasl.client.callback.handler.class = null\
	sasl.jaas.config = null\
	sasl.kerberos.kinit.cmd = /usr/bin/kinit\
	sasl.kerberos.min.time.before.relogin = 60000\
	sasl.kerberos.service.name = null\
	sasl.kerberos.ticket.renew.jitter = 0.05\
	sasl.kerberos.ticket.renew.window.factor = 0.8\
	sasl.login.callback.handler.class = null\
	sasl.login.class = null\
	sasl.login.refresh.buffer.seconds = 300\
	sasl.login.refresh.min.period.seconds = 60\
	sasl.login.refresh.window.factor = 0.8\
	sasl.login.refresh.window.jitter = 0.05\
	sasl.mechanism = GSSAPI\
	security.protocol = PLAINTEXT\
	security.providers = null\
	send.buffer.bytes = 131072\
	session.timeout.ms = 10000\
	ssl.cipher.suites = null\
	ssl.enabled.protocols = [TLSv1.2]\
	ssl.endpoint.identification.algorithm = https\
	ssl.engine.factory.class = null\
	ssl.key.password = null\
	ssl.keymanager.algorithm = SunX509\
	ssl.keystore.location = null\
	ssl.keystore.password = null\
	ssl.keystore.type = JKS\
	ssl.protocol = TLSv1.2\
	ssl.provider = null\
	ssl.secure.random.implementation = null\
	ssl.trustmanager.algorithm = PKIX\
	ssl.truststore.location = null\
	ssl.truststore.password = null\
	ssl.truststore.type = JKS\
	value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer\
 (org.apache.kafka.clients.consumer.ConsumerConfig:354)\
[2020-06-10 20:55:24,274] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer] Cooperative rebalancing enabled now (org.apache.kafka.streams.processor.internals.assignment.AssignorConfiguration:215)\
[2020-06-10 20:55:24,356] WARN The configuration 'admin.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:362)\
[2020-06-10 20:55:24,356] WARN The configuration 'admin.retries' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig:362)\
[2020-06-10 20:55:24,357] INFO Kafka version: 2.7.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117)\
[2020-06-10 20:55:24,357] INFO Kafka commitId: 64b684476a756063 (org.apache.kafka.common.utils.AppInfoParser:118)\
[2020-06-10 20:55:24,357] INFO Kafka startTimeMs: 1591847724356 (org.apache.kafka.common.utils.AppInfoParser:119)\
[2020-06-10 20:55:24,364] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5] State transition from CREATED to REBALANCING (org.apache.kafka.streams.KafkaStreams:282)\
[2020-06-10 20:55:24,364] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] Starting (org.apache.kafka.streams.processor.internals.StreamThread:502)\
[2020-06-10 20:55:24,365] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] State transition from CREATED to STARTING (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:55:24,372] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Subscribed to topic(s): input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi (org.apache.kafka.clients.consumer.KafkaConsumer:959)\
[2020-06-10 20:55:24,402] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:24,415] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Discovered group coordinator localhost:59678 (id: 2147483647 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:815)\
[2020-06-10 20:55:24,422] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)\
[2020-06-10 20:55:24,440] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:456)\
[2020-06-10 20:55:24,441] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:553)\
[2020-06-10 20:55:24,446] INFO [GroupCoordinator 0]: Preparing to rebalance group app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi in state PreparingRebalance with old generation 1 (__consumer_offsets-3) (reason: Adding new member app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer-d5a33acf-eeb5-49cd-8b9b-77489db20940 with group instance id None) (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:30,489] INFO [GroupCoordinator 0]: Member app-StoreUpgradeIntegrationTestshouldProxyWindowStoreToTimestampedWindowStoreUsingPapi-c3cfaf64-38f7-47e0-9418-388c7df32ff7-StreamThread-1-consumer-ddfdb669-204e-4e37-831e-1dc39a4020f4 in group app-StoreUpgradeIntegrationTestshouldProxyWindowStoreToTimestampedWindowStoreUsingPapi has failed, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:30,509] INFO [GroupCoordinator 0]: Preparing to rebalance group app-StoreUpgradeIntegrationTestshouldProxyWindowStoreToTimestampedWindowStoreUsingPapi in state PreparingRebalance with old generation 2 (__consumer_offsets-3) (reason: removing member app-StoreUpgradeIntegrationTestshouldProxyWindowStoreToTimestampedWindowStoreUsingPapi-c3cfaf64-38f7-47e0-9418-388c7df32ff7-StreamThread-1-consumer-ddfdb669-204e-4e37-831e-1dc39a4020f4 on heartbeat expiration) (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:30,513] INFO [GroupCoordinator 0]: Group app-StoreUpgradeIntegrationTestshouldProxyWindowStoreToTimestampedWindowStoreUsingPapi with generation 3 is now empty (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:33,701] INFO [GroupCoordinator 0]: Member app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-db370b48-7d46-4cd9-8d22-5bb7ccaa8052-StreamThread-1-consumer-c3e8d275-9f14-4cf8-bb95-de672c972bf8 in group app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi has failed, removing it from the group (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:33,706] INFO [GroupCoordinator 0]: Stabilized group app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi generation 2 (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:33,930] INFO Decided on assignment: \{39daefdf-442a-4d68-83a7-9049d14eebf5=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([0_0]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([0_0=5]) taskLagTotals: ([0_0=2]) capacity: 1 assigned: 1]\} with no followup probing rebalance. (org.apache.kafka.streams.processor.internals.assignment.HighAvailabilityTaskAssignor:95)\
[2020-06-10 20:55:33,936] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer] Assigned tasks to clients as \
39daefdf-442a-4d68-83a7-9049d14eebf5=[activeTasks: ([0_0]) standbyTasks: ([]) prevActiveTasks: ([]) prevStandbyTasks: ([0_0]) prevOwnedPartitionsByConsumerId: ([]) changelogOffsetTotalsByTask: ([0_0=5]) taskLagTotals: ([0_0=2]) capacity: 1 assigned: 1]. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:714)\
[2020-06-10 20:55:33,941] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer] Finished stable assignment of tasks, no followup rebalances required. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:926)\
[2020-06-10 20:55:33,943] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Finished assignment for group at generation 2: \{app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer-d5a33acf-eeb5-49cd-8b9b-77489db20940=Assignment(partitions=[input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0], userDataSize=48)\} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:620)\
[2020-06-10 20:55:33,963] INFO [GroupCoordinator 0]: Assignment received from leader for group app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi for generation 2 (kafka.coordinator.group.GroupCoordinator:66)\
[2020-06-10 20:55:33,995] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Successfully joined group with generation 2 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:504)\
[2020-06-10 20:55:34,017] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Updating assignment with\
	Assigned partitions:                       [input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0]\
	Current owned partitions:                  []\
	Added partitions (assigned - owned):       [input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0]\
	Revoked partitions (owned - assigned):     []\
 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:393)\
[2020-06-10 20:55:34,024] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Notifying assignor about the new Assignment(partitions=[input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0], userDataSize=48) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:276)\
[2020-06-10 20:55:34,027] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer] No followup rebalance was requested, resetting the rebalance schedule. (org.apache.kafka.streams.processor.internals.StreamsPartitionAssignor:1370)\
[2020-06-10 20:55:34,030] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] Handle new assignment with:\
	New active tasks: [0_0]\
	New standby tasks: []\
	Existing active tasks: []\
	Existing standby tasks: [] (org.apache.kafka.streams.processor.internals.TaskManager:188)\
[2020-06-10 20:55:34,364] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Adding newly assigned partitions: input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:288)\
[2020-06-10 20:55:34,368] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] State transition from STARTING to PARTITIONS_ASSIGNED (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:55:34,760] INFO Opening store store in upgrade mode (org.apache.kafka.streams.state.internals.RocksDBTimestampedStore:97)\
[2020-06-10 20:55:34,930] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] task [0_0] Initialized (org.apache.kafka.streams.processor.internals.StreamTask:214)\
[2020-06-10 20:55:34,940] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-restore-consumer, groupId=null] Cluster ID: iBTYG5jLSriZwNccfa-rdw (org.apache.kafka.clients.Metadata:279)\
[2020-06-10 20:55:34,950] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-restore-consumer, groupId=null] Subscribed to partition(s): app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 (org.apache.kafka.clients.consumer.KafkaConsumer:1114)\
[2020-06-10 20:55:34,951] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-restore-consumer, groupId=null] Seeking to offset 6 for partition app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 (org.apache.kafka.clients.consumer.KafkaConsumer:1591)\
[2020-06-10 20:55:34,984] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] Finished restoring changelog app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-store-changelog-0 to store store with a total number of 1 records (org.apache.kafka.streams.processor.internals.StoreChangelogReader:534)\
[2020-06-10 20:55:34,986] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-consumer, groupId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi] Setting offset for partition input-stream-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-0 to the committed offset FetchPosition\{offset=7, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch\{leader=Optional[localhost:59678 (id: 0 rack: null)], epoch=0\}\} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:808)\
[2020-06-10 20:55:34,992] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] task [0_0] Restored and ready to run (org.apache.kafka.streams.processor.internals.StreamTask:235)\
[2020-06-10 20:55:34,992] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] State transition from PARTITIONS_ASSIGNED to RUNNING (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:55:34,992] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5] State transition from REBALANCING to RUNNING (org.apache.kafka.streams.KafkaStreams:282)\
[2020-06-10 20:56:35,101] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.KafkaStreams:282)\
[2020-06-10 20:56:35,108] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] Informed to shut down (org.apache.kafka.streams.processor.internals.StreamThread:875)\
[2020-06-10 20:56:35,110] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] State transition from RUNNING to PENDING_SHUTDOWN (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:56:35,166] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] Shutting down (org.apache.kafka.streams.processor.internals.StreamThread:889)\
[2020-06-10 20:56:35,170] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] task [0_0] Suspended running (org.apache.kafka.streams.processor.internals.StreamTask:270)\
[2020-06-10 20:56:35,179] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1068)\
[2020-06-10 20:56:35,191] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] task [0_0] Closed clean (org.apache.kafka.streams.processor.internals.StreamTask:460)\
[2020-06-10 20:56:35,192] INFO [Producer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-producer] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. (org.apache.kafka.clients.producer.KafkaProducer:1186)\
[2020-06-10 20:56:35,198] INFO [Consumer clientId=app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1-restore-consumer, groupId=null] Unsubscribed all topics or patterns and assigned partitions (org.apache.kafka.clients.consumer.KafkaConsumer:1068)\
[2020-06-10 20:56:35,218] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] State transition from PENDING_SHUTDOWN to DEAD (org.apache.kafka.streams.processor.internals.StreamThread:220)\
[2020-06-10 20:56:35,219] INFO stream-thread [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5-StreamThread-1] Shutdown complete (org.apache.kafka.streams.processor.internals.StreamThread:914)\
[2020-06-10 20:56:35,225] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5] State transition from PENDING_SHUTDOWN to NOT_RUNNING (org.apache.kafka.streams.KafkaStreams:282)\
[2020-06-10 20:56:35,226] INFO stream-client [app-StoreUpgradeIntegrationTestshouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi-39daefdf-442a-4d68-83a7-9049d14eebf5] Streams client stopped completely (org.apache.kafka.streams.KafkaStreams:991)\
[2020-06-10 20:56:35,229] INFO stream-thread [Test worker] Deleting obsolete state directory 0_0 for task 0_0 as 229ms has elapsed (cleanup delay is 0ms). (org.apache.kafka.streams.processor.internals.StateDirectory:324)\
}