[2019-07-04 00:25:13,164] INFO Created server with tickTime 800 minSessionTimeout 1600 maxSessionTimeout 16000 datadir /tmp/kafka-5133934689749621819/version-2 snapdir /tmp/kafka-3856914065937872899/version-2 (org.apache.zookeeper.server.ZooKeeperServer:174) [2019-07-04 00:25:13,165] INFO binding to port /127.0.0.1:0 (org.apache.zookeeper.server.NIOServerCnxnFactory:89) [2019-07-04 00:25:13,174] INFO Initiating client connection, connectString=127.0.0.1:36791 sessionTimeout=6000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@3c25b243 (org.apache.zookeeper.ZooKeeper:442) [2019-07-04 00:25:13,176] INFO Opening socket connection to server localhost/127.0.0.1:36791. Will not attempt to authenticate using SASL (unknown error) (org.apache.zookeeper.ClientCnxn:1025) [2019-07-04 00:25:13,177] INFO Accepted socket connection from /127.0.0.1:33796 (org.apache.zookeeper.server.NIOServerCnxnFactory:222) [2019-07-04 00:25:13,177] INFO Socket connection established to localhost/127.0.0.1:36791, initiating session (org.apache.zookeeper.ClientCnxn:879) [2019-07-04 00:25:13,178] INFO Client attempting to establish new session at /127.0.0.1:33796 (org.apache.zookeeper.server.ZooKeeperServer:949) [2019-07-04 00:25:13,179] INFO Creating new log file: log.1 (org.apache.zookeeper.server.persistence.FileTxnLog:216) [2019-07-04 00:25:13,180] INFO Established session 0x1051cbb6ce60000 with negotiated timeout 6000 for client /127.0.0.1:33796 (org.apache.zookeeper.server.ZooKeeperServer:694) [2019-07-04 00:25:13,180] INFO Session establishment complete on server localhost/127.0.0.1:36791, sessionid = 0x1051cbb6ce60000, negotiated timeout = 6000 (org.apache.zookeeper.ClientCnxn:1299) [2019-07-04 00:25:13,184] INFO Got user-level KeeperException when processing sessionid:0x1051cbb6ce60000 type:create cxid:0x2 zxid:0x3 txntype:-1 reqpath:n/a Error Path:/brokers Error:KeeperErrorCode = NoNode for /brokers (org.apache.zookeeper.server.PrepRequestProcessor:653) [2019-07-04 00:25:13,189] INFO Got user-level KeeperException when processing sessionid:0x1051cbb6ce60000 type:create cxid:0x6 zxid:0x7 txntype:-1 reqpath:n/a Error Path:/config Error:KeeperErrorCode = NoNode for /config (org.apache.zookeeper.server.PrepRequestProcessor:653) [2019-07-04 00:25:13,192] INFO Got user-level KeeperException when processing sessionid:0x1051cbb6ce60000 type:create cxid:0x9 zxid:0xa txntype:-1 reqpath:n/a Error Path:/admin Error:KeeperErrorCode = NoNode for /admin (org.apache.zookeeper.server.PrepRequestProcessor:653) [2019-07-04 00:25:13,202] INFO Got user-level KeeperException when processing sessionid:0x1051cbb6ce60000 type:create cxid:0x15 zxid:0x15 txntype:-1 reqpath:n/a Error Path:/cluster Error:KeeperErrorCode = NoNode for /cluster (org.apache.zookeeper.server.PrepRequestProcessor:653) [2019-07-04 00:25:13,204] WARN No meta.properties file under dir /tmp/junit8459899247457741708/junit5416730848281225955/meta.properties (kafka.server.BrokerMetadataCheckpoint:70) [2019-07-04 00:25:13,311] WARN No meta.properties file under dir /tmp/junit8459899247457741708/junit5416730848281225955/meta.properties (kafka.server.BrokerMetadataCheckpoint:70) [2019-07-04 00:25:13,370] INFO Kafka version: 2.4.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117) [2019-07-04 00:25:13,370] INFO Kafka commitId: 9573ac576c472324 (org.apache.kafka.common.utils.AppInfoParser:118) [2019-07-04 00:25:13,371] INFO Kafka startTimeMs: 1562199913163 (org.apache.kafka.common.utils.AppInfoParser:119) [2019-07-04 00:25:13,373] INFO ProducerConfig values: acks = 1 batch.size = 16384 bootstrap.servers = [localhost:38433] buffer.memory = 33554432 client.dns.lookup = default client.id = compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 120000 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer linger.ms = 0 max.block.ms = 60000 max.in.flight.requests.per.connection = 5 max.request.size = 1048576 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = https ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer (org.apache.kafka.clients.producer.ProducerConfig:347) [2019-07-04 00:25:13,373] INFO Got user-level KeeperException when processing sessionid:0x1051cbb6ce60000 type:multi cxid:0x38 zxid:0x1c txntype:-1 reqpath:n/a aborting remaining multi ops. Error Path:/admin/preferred_replica_election Error:KeeperErrorCode = NoNode for /admin/preferred_replica_election (org.apache.zookeeper.server.PrepRequestProcessor:596) [2019-07-04 00:25:13,377] INFO Kafka version: 2.4.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117) [2019-07-04 00:25:13,379] INFO Kafka commitId: 9573ac576c472324 (org.apache.kafka.common.utils.AppInfoParser:118) [2019-07-04 00:25:13,379] INFO Kafka startTimeMs: 1562199913377 (org.apache.kafka.common.utils.AppInfoParser:119) [2019-07-04 00:25:13,379] INFO Starting Connect cluster 'connect-cluster' with 3 workers (org.apache.kafka.connect.util.clusters.EmbeddedConnectCluster:208) [2019-07-04 00:25:13,380] INFO Scanning for plugin classes. This might take a moment ... (org.apache.kafka.connect.cli.ConnectDistributed:90) [2019-07-04 00:25:13,490] INFO [Producer clientId=producer-11] Cluster ID: yogXvb3iQ0CX_k2gWCO8gA (org.apache.kafka.clients.Metadata:266) [2019-07-04 00:25:13,928] INFO SessionTrackerImpl exited loop! (org.apache.zookeeper.server.SessionTrackerImpl:163) [2019-07-04 00:25:14,777] DEBUG Skipping class org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest$BogusSourceConnector as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,778] DEBUG Skipping class org.apache.kafka.connect.runtime.AbstractHerderTest$BogusSourceConnector as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,778] DEBUG Skipping class org.apache.kafka.connect.source.SourceConnector as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,779] DEBUG Skipping class org.apache.kafka.connect.runtime.WorkerConnectorTest$TestConnector as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,780] DEBUG Skipping class org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest$BogusSinkConnector as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,780] DEBUG Skipping class org.apache.kafka.connect.runtime.ConnectorConfigTest$TestConnector as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,780] DEBUG Skipping class org.apache.kafka.connect.sink.SinkConnector as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,781] DEBUG Skipping class org.apache.kafka.connect.runtime.distributed.DistributedHerderTest$BogusSourceConnector as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,783] DEBUG Skipping class org.apache.kafka.connect.converters.NumberConverter as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,784] DEBUG Skipping class org.apache.kafka.connect.converters.NumberConverter as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,786] DEBUG Skipping class org.apache.kafka.connect.transforms.Cast as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,786] DEBUG Skipping class org.apache.kafka.connect.transforms.HoistField as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,787] DEBUG Skipping class org.apache.kafka.connect.transforms.InsertField as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,788] DEBUG Skipping class org.apache.kafka.connect.transforms.SetSchemaMetadata as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,788] DEBUG Skipping class org.apache.kafka.connect.transforms.MaskField as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,789] DEBUG Skipping class org.apache.kafka.connect.transforms.Flatten as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,789] DEBUG Skipping class org.apache.kafka.connect.transforms.TimestampConverter as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,789] DEBUG Skipping class org.apache.kafka.connect.transforms.ExtractField as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,790] DEBUG Skipping class org.apache.kafka.connect.transforms.ReplaceField as it is not concrete implementation (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:333) [2019-07-04 00:25:14,794] INFO Registered loader: sun.misc.Launcher$AppClassLoader@73d16e93 (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:245) [2019-07-04 00:25:14,794] INFO Added plugin 'org.apache.kafka.connect.tools.MockSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,795] INFO Added plugin 'org.apache.kafka.connect.runtime.WorkerTest$WorkerTestConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,795] INFO Added plugin 'org.apache.kafka.connect.integration.MonitorableSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,795] INFO Added plugin 'org.apache.kafka.connect.tools.VerifiableSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,795] INFO Added plugin 'org.apache.kafka.connect.runtime.TestSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,796] INFO Added plugin 'org.apache.kafka.connect.tools.VerifiableSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,796] INFO Added plugin 'org.apache.kafka.connect.runtime.TestSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,796] INFO Added plugin 'org.apache.kafka.connect.tools.MockConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,796] INFO Added plugin 'org.apache.kafka.connect.runtime.rest.resources.ConnectorPluginsResourceTest$ConnectorPluginsResourceTestConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,797] INFO Added plugin 'org.apache.kafka.connect.integration.MonitorableSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,797] INFO Added plugin 'org.apache.kafka.connect.tools.SchemaSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,797] INFO Added plugin 'org.apache.kafka.connect.tools.MockSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,798] INFO Added plugin 'org.apache.kafka.connect.converters.DoubleConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,798] INFO Added plugin 'org.apache.kafka.connect.converters.LongConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,798] INFO Added plugin 'org.apache.kafka.connect.converters.FloatConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,798] INFO Added plugin 'org.apache.kafka.connect.json.JsonConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,799] INFO Added plugin 'org.apache.kafka.connect.runtime.WorkerTest$TestConfigurableConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,799] INFO Added plugin 'org.apache.kafka.connect.runtime.isolation.PluginsTest$TestConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,799] INFO Added plugin 'org.apache.kafka.connect.runtime.WorkerTest$TestConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,799] INFO Added plugin 'org.apache.kafka.connect.runtime.isolation.PluginsTest$TestInternalConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,800] INFO Added plugin 'org.apache.kafka.connect.runtime.ErrorHandlingTaskTest$FaultyConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,800] INFO Added plugin 'org.apache.kafka.connect.converters.ByteArrayConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,800] INFO Added plugin 'org.apache.kafka.connect.converters.ShortConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,800] INFO Added plugin 'org.apache.kafka.connect.converters.IntegerConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,801] INFO Added plugin 'org.apache.kafka.connect.storage.StringConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,801] INFO Added plugin 'org.apache.kafka.connect.runtime.isolation.PluginsTest$TestHeaderConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,801] INFO Added plugin 'org.apache.kafka.connect.storage.SimpleHeaderConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,802] INFO Added plugin 'org.apache.kafka.connect.transforms.SetSchemaMetadata$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,802] INFO Added plugin 'org.apache.kafka.connect.transforms.SetSchemaMetadata$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,802] INFO Added plugin 'org.apache.kafka.connect.transforms.HoistField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,802] INFO Added plugin 'org.apache.kafka.connect.runtime.ErrorHandlingTaskTest$FaultyPassthrough' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,803] INFO Added plugin 'org.apache.kafka.connect.transforms.Flatten$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,803] INFO Added plugin 'org.apache.kafka.connect.transforms.ExtractField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,803] INFO Added plugin 'org.apache.kafka.connect.transforms.ReplaceField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,803] INFO Added plugin 'org.apache.kafka.connect.transforms.HoistField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,804] INFO Added plugin 'org.apache.kafka.connect.transforms.ReplaceField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,804] INFO Added plugin 'org.apache.kafka.connect.runtime.AbstractHerderTest$SampleTransformation' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,804] INFO Added plugin 'org.apache.kafka.connect.integration.ErrorHandlingIntegrationTest$FaultyPassthrough' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,804] INFO Added plugin 'org.apache.kafka.connect.transforms.MaskField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,805] INFO Added plugin 'org.apache.kafka.connect.runtime.ConnectorConfigTest$SimpleTransformation' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,805] INFO Added plugin 'org.apache.kafka.connect.transforms.MaskField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,805] INFO Added plugin 'org.apache.kafka.connect.transforms.TimestampConverter$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,806] INFO Added plugin 'org.apache.kafka.connect.transforms.ExtractField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,806] INFO Added plugin 'org.apache.kafka.connect.transforms.Cast$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,806] INFO Added plugin 'org.apache.kafka.connect.transforms.InsertField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,806] INFO Added plugin 'org.apache.kafka.connect.transforms.Cast$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,807] INFO Added plugin 'org.apache.kafka.connect.transforms.Flatten$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,807] INFO Added plugin 'org.apache.kafka.connect.transforms.TimestampConverter$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,807] INFO Added plugin 'org.apache.kafka.connect.transforms.InsertField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,807] INFO Added plugin 'org.apache.kafka.connect.transforms.ValueToKey' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,808] INFO Added plugin 'org.apache.kafka.connect.transforms.RegexRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,808] INFO Added plugin 'org.apache.kafka.connect.transforms.TimestampRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,808] INFO Added plugin 'org.apache.kafka.common.config.provider.FileConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,808] INFO Added plugin 'org.apache.kafka.connect.runtime.isolation.PluginsTest$TestConnectRestExtension' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,809] INFO Added plugin 'org.apache.kafka.connect.connector.policy.AllConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,809] INFO Added plugin 'org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,809] INFO Added plugin 'org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:174) [2019-07-04 00:25:14,810] INFO Added aliases 'MonitorableSinkConnector' and 'MonitorableSink' to plugin 'org.apache.kafka.connect.integration.MonitorableSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,810] INFO Added aliases 'MonitorableSourceConnector' and 'MonitorableSource' to plugin 'org.apache.kafka.connect.integration.MonitorableSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,810] INFO Added aliases 'TestSinkConnector' and 'TestSink' to plugin 'org.apache.kafka.connect.runtime.TestSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,811] INFO Added aliases 'TestSourceConnector' and 'TestSource' to plugin 'org.apache.kafka.connect.runtime.TestSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,811] INFO Added aliases 'WorkerTestConnector' and 'WorkerTest' to plugin 'org.apache.kafka.connect.runtime.WorkerTest$WorkerTestConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,811] INFO Added aliases 'ConnectorPluginsResourceTestConnector' and 'ConnectorPluginsResourceTest' to plugin 'org.apache.kafka.connect.runtime.rest.resources.ConnectorPluginsResourceTest$ConnectorPluginsResourceTestConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,811] INFO Added aliases 'MockConnector' and 'Mock' to plugin 'org.apache.kafka.connect.tools.MockConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,812] INFO Added aliases 'MockSinkConnector' and 'MockSink' to plugin 'org.apache.kafka.connect.tools.MockSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,812] INFO Added aliases 'MockSourceConnector' and 'MockSource' to plugin 'org.apache.kafka.connect.tools.MockSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,812] INFO Added aliases 'SchemaSourceConnector' and 'SchemaSource' to plugin 'org.apache.kafka.connect.tools.SchemaSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,813] INFO Added aliases 'VerifiableSinkConnector' and 'VerifiableSink' to plugin 'org.apache.kafka.connect.tools.VerifiableSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,813] INFO Added aliases 'VerifiableSourceConnector' and 'VerifiableSource' to plugin 'org.apache.kafka.connect.tools.VerifiableSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,813] INFO Added aliases 'ByteArrayConverter' and 'ByteArray' to plugin 'org.apache.kafka.connect.converters.ByteArrayConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,813] INFO Added aliases 'DoubleConverter' and 'Double' to plugin 'org.apache.kafka.connect.converters.DoubleConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,814] INFO Added aliases 'FloatConverter' and 'Float' to plugin 'org.apache.kafka.connect.converters.FloatConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,814] INFO Added aliases 'IntegerConverter' and 'Integer' to plugin 'org.apache.kafka.connect.converters.IntegerConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,814] INFO Added aliases 'LongConverter' and 'Long' to plugin 'org.apache.kafka.connect.converters.LongConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,814] INFO Added aliases 'ShortConverter' and 'Short' to plugin 'org.apache.kafka.connect.converters.ShortConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,815] INFO Added aliases 'JsonConverter' and 'Json' to plugin 'org.apache.kafka.connect.json.JsonConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,815] INFO Added aliases 'FaultyConverter' and 'Faulty' to plugin 'org.apache.kafka.connect.runtime.ErrorHandlingTaskTest$FaultyConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,815] INFO Added aliases 'TestConfigurableConverter' and 'TestConfigurable' to plugin 'org.apache.kafka.connect.runtime.WorkerTest$TestConfigurableConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,816] INFO Added aliases 'TestInternalConverter' and 'TestInternal' to plugin 'org.apache.kafka.connect.runtime.isolation.PluginsTest$TestInternalConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,816] INFO Added aliases 'StringConverter' and 'String' to plugin 'org.apache.kafka.connect.storage.StringConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,816] INFO Added aliases 'ByteArrayConverter' and 'ByteArray' to plugin 'org.apache.kafka.connect.converters.ByteArrayConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,816] INFO Added aliases 'DoubleConverter' and 'Double' to plugin 'org.apache.kafka.connect.converters.DoubleConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,817] INFO Added aliases 'FloatConverter' and 'Float' to plugin 'org.apache.kafka.connect.converters.FloatConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,817] INFO Added aliases 'IntegerConverter' and 'Integer' to plugin 'org.apache.kafka.connect.converters.IntegerConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,817] INFO Added aliases 'LongConverter' and 'Long' to plugin 'org.apache.kafka.connect.converters.LongConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,818] INFO Added aliases 'ShortConverter' and 'Short' to plugin 'org.apache.kafka.connect.converters.ShortConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,818] INFO Added aliases 'JsonConverter' and 'Json' to plugin 'org.apache.kafka.connect.json.JsonConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,818] INFO Added aliases 'FaultyConverter' and 'Faulty' to plugin 'org.apache.kafka.connect.runtime.ErrorHandlingTaskTest$FaultyConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,818] INFO Added alias 'TestHeaderConverter' to plugin 'org.apache.kafka.connect.runtime.isolation.PluginsTest$TestHeaderConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:394) [2019-07-04 00:25:14,819] INFO Added aliases 'TestInternalConverter' and 'TestInternal' to plugin 'org.apache.kafka.connect.runtime.isolation.PluginsTest$TestInternalConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,819] INFO Added alias 'SimpleHeaderConverter' to plugin 'org.apache.kafka.connect.storage.SimpleHeaderConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:394) [2019-07-04 00:25:14,819] INFO Added aliases 'StringConverter' and 'String' to plugin 'org.apache.kafka.connect.storage.StringConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,820] INFO Added aliases 'SampleTransformation' and 'Sample' to plugin 'org.apache.kafka.connect.runtime.AbstractHerderTest$SampleTransformation' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,820] INFO Added aliases 'SimpleTransformation' and 'Simple' to plugin 'org.apache.kafka.connect.runtime.ConnectorConfigTest$SimpleTransformation' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,820] INFO Added alias 'RegexRouter' to plugin 'org.apache.kafka.connect.transforms.RegexRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:394) [2019-07-04 00:25:14,821] INFO Added alias 'TimestampRouter' to plugin 'org.apache.kafka.connect.transforms.TimestampRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:394) [2019-07-04 00:25:14,821] INFO Added alias 'ValueToKey' to plugin 'org.apache.kafka.connect.transforms.ValueToKey' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:394) [2019-07-04 00:25:14,821] INFO Added aliases 'TestConnectRestExtension' and 'Test' to plugin 'org.apache.kafka.connect.runtime.isolation.PluginsTest$TestConnectRestExtension' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,822] INFO Added aliases 'AllConnectorClientConfigOverridePolicy' and 'All' to plugin 'org.apache.kafka.connect.connector.policy.AllConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,822] INFO Added aliases 'NoneConnectorClientConfigOverridePolicy' and 'None' to plugin 'org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,822] INFO Added aliases 'PrincipalConnectorClientConfigOverridePolicy' and 'Principal' to plugin 'org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader:397) [2019-07-04 00:25:14,823] INFO DistributedConfig values: access.control.allow.methods = access.control.allow.origin = bootstrap.servers = [localhost:38433] client.dns.lookup = default client.id = config.providers = [] config.storage.replication.factor = 1 config.storage.topic = connect-config-topic-connect-cluster connect.protocol = compatible connections.max.idle.ms = 540000 connector.client.config.override.policy = None group.id = connect-integration-test-connect-cluster header.converter = class org.apache.kafka.connect.storage.SimpleHeaderConverter heartbeat.interval.ms = 3000 internal.key.converter = class org.apache.kafka.connect.json.JsonConverter internal.value.converter = class org.apache.kafka.connect.json.JsonConverter key.converter = class org.apache.kafka.connect.storage.StringConverter listeners = null metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 offset.flush.interval.ms = 5000 offset.flush.timeout.ms = 5000 offset.storage.partitions = 25 offset.storage.replication.factor = 1 offset.storage.topic = connect-offset-topic-connect-cluster plugin.path = null rebalance.timeout.ms = 60000 receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 40000 rest.advertised.host.name = null rest.advertised.listener = null rest.advertised.port = null rest.extension.classes = [] rest.host.name = localhost rest.port = 0 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.mechanism = GSSAPI scheduled.rebalance.max.delay.ms = 300000 security.protocol = PLAINTEXT send.buffer.bytes = 131072 session.timeout.ms = 10000 ssl.cipher.suites = null ssl.client.auth = none ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = https ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS status.storage.partitions = 5 status.storage.replication.factor = 1 status.storage.topic = connect-storage-topic-connect-cluster task.shutdown.graceful.timeout.ms = 5000 value.converter = class org.apache.kafka.connect.storage.StringConverter worker.sync.timeout.ms = 3000 worker.unsync.backoff.ms = 300000 (org.apache.kafka.connect.runtime.distributed.DistributedConfig:347) [2019-07-04 00:25:14,824] INFO Creating Kafka admin client (org.apache.kafka.connect.util.ConnectUtils:43) [2019-07-04 00:25:14,824] INFO AdminClientConfig values: bootstrap.servers = [localhost:38433] client.dns.lookup = default client.id = connections.max.idle.ms = 300000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 120000 retries = 5 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = https ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS (org.apache.kafka.clients.admin.AdminClientConfig:347) [2019-07-04 00:25:14,826] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,826] WARN The configuration 'group.id' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,827] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,827] WARN The configuration 'rest.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,827] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,827] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,827] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,828] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,828] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,828] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,828] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,829] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,829] INFO Kafka version: 2.4.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117) [2019-07-04 00:25:14,829] INFO Kafka commitId: 9573ac576c472324 (org.apache.kafka.common.utils.AppInfoParser:118) [2019-07-04 00:25:14,829] INFO Kafka startTimeMs: 1562199914829 (org.apache.kafka.common.utils.AppInfoParser:119) [2019-07-04 00:25:14,830] DEBUG Looking up Kafka cluster ID (org.apache.kafka.connect.util.ConnectUtils:50) [2019-07-04 00:25:14,831] DEBUG Fetching Kafka cluster ID (org.apache.kafka.connect.util.ConnectUtils:57) [2019-07-04 00:25:14,842] INFO Kafka cluster ID: yogXvb3iQ0CX_k2gWCO8gA (org.apache.kafka.connect.util.ConnectUtils:59) [2019-07-04 00:25:14,845] DEBUG Kafka cluster ID: yogXvb3iQ0CX_k2gWCO8gA (org.apache.kafka.connect.cli.ConnectDistributed:96) [2019-07-04 00:25:14,845] INFO Added connector for http://localhost:0 (org.apache.kafka.connect.runtime.rest.RestServer:124) [2019-07-04 00:25:14,846] INFO Initializing REST server (org.apache.kafka.connect.runtime.rest.RestServer:168) [2019-07-04 00:25:14,846] INFO jetty-9.4.18.v20190429; built: 2019-04-29T20:42:08.989Z; git: e1bc35120a6617ee3df052294e433f3a25ce7097; jvm 1.8.0_191-b12 (org.eclipse.jetty.server.Server:370) [2019-07-04 00:25:14,855] INFO Started http_localhost0@7fb27489{HTTP/1.1,[http/1.1]}{localhost:46068} (org.eclipse.jetty.server.AbstractConnector:292) [2019-07-04 00:25:14,855] INFO Started @35210ms (org.eclipse.jetty.server.Server:410) [2019-07-04 00:25:14,856] INFO Advertised URI: http://localhost:46068/ (org.apache.kafka.connect.runtime.rest.RestServer:285) [2019-07-04 00:25:14,856] INFO REST server listening at http://localhost:46068/, advertising URL http://localhost:46068/ (org.apache.kafka.connect.runtime.rest.RestServer:183) [2019-07-04 00:25:14,856] INFO Advertised URI: http://localhost:46068/ (org.apache.kafka.connect.runtime.rest.RestServer:285) [2019-07-04 00:25:14,857] INFO Setting up None Policy for ConnectorClientConfigOverride. This will disallow any client configuration to be overridden (org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicy:45) [2019-07-04 00:25:14,857] DEBUG Registering Connect metrics with JMX for worker 'localhost:46068' (org.apache.kafka.connect.runtime.ConnectMetrics:83) [2019-07-04 00:25:14,858] INFO Kafka version: 2.4.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117) [2019-07-04 00:25:14,858] INFO Kafka commitId: 9573ac576c472324 (org.apache.kafka.common.utils.AppInfoParser:118) [2019-07-04 00:25:14,858] INFO Kafka startTimeMs: 1562199914857 (org.apache.kafka.common.utils.AppInfoParser:119) [2019-07-04 00:25:14,859] DEBUG Configuring the key converter with configuration keys: [] (org.apache.kafka.connect.runtime.isolation.Plugins:254) [2019-07-04 00:25:14,860] INFO JsonConverterConfig values: converter.type = key schemas.cache.size = 1000 schemas.enable = false (org.apache.kafka.connect.json.JsonConverterConfig:347) [2019-07-04 00:25:14,860] DEBUG Configuring the value converter with configuration keys: [] (org.apache.kafka.connect.runtime.isolation.Plugins:254) [2019-07-04 00:25:14,860] INFO JsonConverterConfig values: converter.type = value schemas.cache.size = 1000 schemas.enable = false (org.apache.kafka.connect.json.JsonConverterConfig:347) [2019-07-04 00:25:14,863] INFO Kafka version: 2.4.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117) [2019-07-04 00:25:14,863] INFO Kafka commitId: 9573ac576c472324 (org.apache.kafka.common.utils.AppInfoParser:118) [2019-07-04 00:25:14,863] INFO Kafka startTimeMs: 1562199914863 (org.apache.kafka.common.utils.AppInfoParser:119) [2019-07-04 00:25:14,864] DEBUG [Worker clientId=connect-4, groupId=connect-integration-test-connect-cluster] Connect group member created (org.apache.kafka.connect.runtime.distributed.WorkerGroupMember:139) [2019-07-04 00:25:14,864] DEBUG Kafka Connect instance created (org.apache.kafka.connect.runtime.Connect:42) [2019-07-04 00:25:14,864] INFO Kafka Connect distributed worker initialization took 1485ms (org.apache.kafka.connect.cli.ConnectDistributed:128) [2019-07-04 00:25:14,865] INFO Kafka Connect starting (org.apache.kafka.connect.runtime.Connect:50) [2019-07-04 00:25:14,865] INFO Initializing REST resources (org.apache.kafka.connect.runtime.rest.RestServer:187) [2019-07-04 00:25:14,865] INFO [Worker clientId=connect-4, groupId=connect-integration-test-connect-cluster] Herder starting (org.apache.kafka.connect.runtime.distributed.DistributedHerder:238) [2019-07-04 00:25:14,866] INFO Worker starting (org.apache.kafka.connect.runtime.Worker:182) [2019-07-04 00:25:14,866] INFO Starting KafkaOffsetBackingStore (org.apache.kafka.connect.storage.KafkaOffsetBackingStore:108) [2019-07-04 00:25:14,866] INFO Starting KafkaBasedLog with topic connect-offset-topic-connect-cluster (org.apache.kafka.connect.util.KafkaBasedLog:125) [2019-07-04 00:25:14,866] DEBUG Creating admin client to manage Connect internal offset topic (org.apache.kafka.connect.storage.KafkaOffsetBackingStore:97) [2019-07-04 00:25:14,867] INFO AdminClientConfig values: bootstrap.servers = [localhost:38433] client.dns.lookup = default client.id = connections.max.idle.ms = 300000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 120000 retries = 5 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = https ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS (org.apache.kafka.clients.admin.AdminClientConfig:347) [2019-07-04 00:25:14,869] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,869] WARN The configuration 'group.id' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,869] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,869] WARN The configuration 'rest.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,869] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,869] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,870] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,870] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,870] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,870] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,870] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,870] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:355) [2019-07-04 00:25:14,870] INFO Kafka version: 2.4.0-SNAPSHOT (org.apache.kafka.common.utils.AppInfoParser:117) [2019-07-04 00:25:14,870] INFO Kafka commitId: 9573ac576c472324 (org.apache.kafka.common.utils.AppInfoParser:118) [2019-07-04 00:25:14,871] INFO Kafka startTimeMs: 1562199914870 (org.apache.kafka.common.utils.AppInfoParser:119) [2019-07-04 00:25:14,871] INFO DefaultSessionIdManager workerName=node0 (org.eclipse.jetty.server.session:365) [2019-07-04 00:25:14,872] INFO No SessionScavenger set, using defaults (org.eclipse.jetty.server.session:370) [2019-07-04 00:25:14,874] INFO node0 Scavenging every 660000ms (org.eclipse.jetty.server.session:149) [2019-07-04 00:25:14,884] INFO Got user-level KeeperException when processing sessionid:0x1051cbb6ce60000 type:setData cxid:0x3e zxid:0x1d txntype:-1 reqpath:n/a Error Path:/config/topics/connect-offset-topic-connect-cluster Error:KeeperErrorCode = NoNode for /config/topics/connect-offset-topic-connect-cluster (org.apache.zookeeper.server.PrepRequestProcessor:653) [2019-07-04 00:25:15,020] INFO Started o.e.j.s.ServletContextHandler@c53a30e{/,null,AVAILABLE} (org.eclipse.jetty.server.handler.ContextHandler:855) [2019-07-04 00:25:15,021] INFO REST resources initialized; server is started and ready to handle requests (org.apache.kafka.connect.runtime.rest.RestServer:233) [2019-07-04 00:25:15,021] INFO Kafka Connect started (org.apache.kafka.connect.runtime.Connect:56) [2019-07-04 00:25:15,021] INFO Started worker WorkerHandle{workerName='connect-worker-0'workerURL='http://localhost:46068/'} (org.apache.kafka.connect.util.clusters.EmbeddedConnectCluster:163) [2019-07-04 00:25:15,021] INFO Scanning for plugin classes. This might take a moment ... (org.apache.kafka.connect.cli.ConnectDistributed:90) [2019-07-04 00:25:16,117] INFO Created topic (name=connect-offset-topic-connect-cluster, numPartitions=25, replicationFactor=1, replicasAssignments=null, configs={cleanup.policy=compact}) on brokers at localhost:38433 (org.apache.kafka.connect.util.TopicAdmin:230) [2019-07-04 00:25:16,122] INFO ProducerConfig values: acks = all batch.size = 16384 bootstrap.servers = [localhost:38433] buffer.memory = 33554432 client.dns.lookup = default client.id = compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 2147483647 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer linger.ms = 0 max.block.ms = 60000 max.in.flight.requests.per.connection = 1 max.request.size = 1048576 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jit ...[truncated 78620325 chars]... 0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,451] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24770}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24770, keySchema=Schema{STRING}, value=value-simple-conn-0-24770, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,451] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,451] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24771}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24771, keySchema=Schema{STRING}, value=value-simple-conn-0-24771, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,451] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,451] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24773}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24773, keySchema=Schema{STRING}, value=value-simple-conn-0-24773, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,451] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,451] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24774}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24774, keySchema=Schema{STRING}, value=value-simple-conn-0-24774, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,451] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,452] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24776}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24776, keySchema=Schema{STRING}, value=value-simple-conn-0-24776, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,452] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,452] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24785}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24785, keySchema=Schema{STRING}, value=value-simple-conn-0-24785, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,452] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,452] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24787}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24787, keySchema=Schema{STRING}, value=value-simple-conn-0-24787, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,452] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,452] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24799}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24799, keySchema=Schema{STRING}, value=value-simple-conn-0-24799, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,452] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,452] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24801}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24801, keySchema=Schema{STRING}, value=value-simple-conn-0-24801, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,453] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,453] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24810}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24810, keySchema=Schema{STRING}, value=value-simple-conn-0-24810, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,453] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,453] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24811}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24811, keySchema=Schema{STRING}, value=value-simple-conn-0-24811, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,453] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,453] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24813}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24813, keySchema=Schema{STRING}, value=value-simple-conn-0-24813, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,453] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,453] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24817}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24817, keySchema=Schema{STRING}, value=value-simple-conn-0-24817, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,453] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,453] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24818}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24818, keySchema=Schema{STRING}, value=value-simple-conn-0-24818, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,454] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,454] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24823}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24823, keySchema=Schema{STRING}, value=value-simple-conn-0-24823, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,454] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,454] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24824}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24824, keySchema=Schema{STRING}, value=value-simple-conn-0-24824, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,454] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,454] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24826}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24826, keySchema=Schema{STRING}, value=value-simple-conn-0-24826, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,454] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,454] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24830}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24830, keySchema=Schema{STRING}, value=value-simple-conn-0-24830, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,454] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,455] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24833}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24833, keySchema=Schema{STRING}, value=value-simple-conn-0-24833, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,455] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,455] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24834}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24834, keySchema=Schema{STRING}, value=value-simple-conn-0-24834, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,455] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,455] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24835}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24835, keySchema=Schema{STRING}, value=value-simple-conn-0-24835, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,455] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,455] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24837}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24837, keySchema=Schema{STRING}, value=value-simple-conn-0-24837, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,455] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,455] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24838}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24838, keySchema=Schema{STRING}, value=value-simple-conn-0-24838, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,455] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,456] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24839}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24839, keySchema=Schema{STRING}, value=value-simple-conn-0-24839, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,456] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,456] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24843}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24843, keySchema=Schema{STRING}, value=value-simple-conn-0-24843, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,456] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,456] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24845}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24845, keySchema=Schema{STRING}, value=value-simple-conn-0-24845, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,456] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,456] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24852}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24852, keySchema=Schema{STRING}, value=value-simple-conn-0-24852, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,456] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,456] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24853}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24853, keySchema=Schema{STRING}, value=value-simple-conn-0-24853, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,456] INFO shutting down (org.apache.zookeeper.server.ZooKeeperServer:502) [2019-07-04 00:26:15,456] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,457] INFO Shutting down (org.apache.zookeeper.server.SessionTrackerImpl:226) [2019-07-04 00:26:15,457] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24854}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24854, keySchema=Schema{STRING}, value=value-simple-conn-0-24854, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,457] INFO Shutting down (org.apache.zookeeper.server.PrepRequestProcessor:769) [2019-07-04 00:26:15,457] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,457] INFO PrepRequestProcessor exited loop! (org.apache.zookeeper.server.PrepRequestProcessor:144) [2019-07-04 00:26:15,457] INFO Shutting down (org.apache.zookeeper.server.SyncRequestProcessor:208) [2019-07-04 00:26:15,458] INFO SyncRequestProcessor exited! (org.apache.zookeeper.server.SyncRequestProcessor:186) [2019-07-04 00:26:15,457] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24859}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24859, keySchema=Schema{STRING}, value=value-simple-conn-0-24859, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,458] INFO shutdown of request processor complete (org.apache.zookeeper.server.FinalRequestProcessor:430) [2019-07-04 00:26:15,458] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,458] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24861}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24861, keySchema=Schema{STRING}, value=value-simple-conn-0-24861, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,458] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,459] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24872}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24872, keySchema=Schema{STRING}, value=value-simple-conn-0-24872, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,459] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,459] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24873}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24873, keySchema=Schema{STRING}, value=value-simple-conn-0-24873, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,459] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,459] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24874}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24874, keySchema=Schema{STRING}, value=value-simple-conn-0-24874, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,459] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,459] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24878}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24878, keySchema=Schema{STRING}, value=value-simple-conn-0-24878, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,459] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,459] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24879}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24879, keySchema=Schema{STRING}, value=value-simple-conn-0-24879, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,459] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,460] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24884}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24884, keySchema=Schema{STRING}, value=value-simple-conn-0-24884, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,460] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,460] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24885}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24885, keySchema=Schema{STRING}, value=value-simple-conn-0-24885, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,460] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,460] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24886}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24886, keySchema=Schema{STRING}, value=value-simple-conn-0-24886, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,460] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,460] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24887}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24887, keySchema=Schema{STRING}, value=value-simple-conn-0-24887, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,460] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,460] INFO NIOServerCnxn factory exited run method (org.apache.zookeeper.server.NIOServerCnxnFactory:249) [2019-07-04 00:26:15,460] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24890}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24890, keySchema=Schema{STRING}, value=value-simple-conn-0-24890, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,461] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,461] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24894}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24894, keySchema=Schema{STRING}, value=value-simple-conn-0-24894, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,461] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,461] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24895}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24895, keySchema=Schema{STRING}, value=value-simple-conn-0-24895, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,461] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,461] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24896}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24896, keySchema=Schema{STRING}, value=value-simple-conn-0-24896, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,462] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,462] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24898}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24898, keySchema=Schema{STRING}, value=value-simple-conn-0-24898, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,462] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748) [2019-07-04 00:26:15,462] DEBUG [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} Failed record: SourceRecord{sourcePartition={task.id=simple-conn-0}, sourceOffset={saved=24901}} ConnectRecord{topic='test-topic', kafkaPartition=null, key=key-simple-conn-0-24901, keySchema=Schema{STRING}, value=value-simple-conn-0-24901, valueSchema=Schema{STRING}, timestamp=null, headers=ConnectHeaders(headers=)} (org.apache.kafka.connect.runtime.WorkerSourceTask:331) [2019-07-04 00:26:15,462] ERROR [simple-conn|task-0] WorkerSourceTask{id=simple-conn-0} failed to send record to test-topic: (org.apache.kafka.connect.runtime.WorkerSourceTask:330) org.apache.kafka.common.KafkaException: Producer is closed forcefully. at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortBatches(RecordAccumulator.java:729) at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortIncompleteBatches(RecordAccumulator.java:716) at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:278) at java.lang.Thread.run(Thread.java:748)