10:46:49 [INFO ] AbstractConfig.logAll - ConsumerConfig values: auto.commit.interval.ms = 1000 auto.offset.reset = latest bootstrap.servers = [localhost:9092] check.crcs = true client.id = connections.max.idle.ms = 540000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = testOutputTopic heartbeat.interval.ms = 3000 interceptor.classes = null key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor] receive.buffer.bytes = 65536 reconnect.backoff.ms = 50 request.timeout.ms = 305000 retry.backoff.ms = 100 sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 session.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer (AbstractConfig.java:180) 10:46:49 [DEBUG] KafkaConsumer. - Starting the Kafka consumer (KafkaConsumer.java:607) 10:46:49 [INFO ] AbstractConfig.logAll - ConsumerConfig values: auto.commit.interval.ms = 1000 auto.offset.reset = latest bootstrap.servers = [localhost:9092] check.crcs = true client.id = consumer-1 connections.max.idle.ms = 540000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = testOutputTopic heartbeat.interval.ms = 3000 interceptor.classes = null key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor] receive.buffer.bytes = 65536 reconnect.backoff.ms = 50 request.timeout.ms = 305000 retry.backoff.ms = 100 sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 session.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer (AbstractConfig.java:180) 10:46:49 [INFO ] Logging$class.info - New leader is 1 (Logging.scala:70) 10:46:49 [DEBUG] Metadata.update - Updated cluster metadata version 1 to Cluster(id = null, nodes = [localhost:9092 (id: -1 rack: null)], partitions = []) (Metadata.java:244) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name connections-closed: (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name connections-created: (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent-received: (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent: (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name bytes-received: (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name select-time: (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name io-time: (Metrics.java:335) 10:46:49 [INFO ] Logging$class.info - [BrokerChangeListener on Controller 1]: Broker change listener fired for path /brokers/ids with children 1 (Logging.scala:70) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name heartbeat-latency (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name join-latency (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name sync-latency (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name commit-latency (Metrics.java:335) 10:46:49 [INFO ] Logging$class.info - [BrokerChangeListener on Controller 1]: Newly added brokers: 1, deleted brokers: , all live brokers: 1 (Logging.scala:70) 10:46:49 [DEBUG] Logging$class.debug - [Channel manager on controller 1]: Controller 1 trying to connect to broker 1 (Logging.scala:54) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name bytes-fetched (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name records-fetched (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name fetch-latency (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name records-lag (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name fetch-throttle-time (Metrics.java:335) 10:46:49 [INFO ] AppInfoParser$AppInfo. - Kafka version : 0.10.2.0 (AppInfoParser.java:83) 10:46:49 [INFO ] AppInfoParser$AppInfo. - Kafka commitId : 576d93a8dc0cf421 (AppInfoParser.java:84) 10:46:49 [DEBUG] KafkaConsumer. - Kafka consumer created (KafkaConsumer.java:711) 10:46:49 [DEBUG] KafkaConsumer.subscribe - Subscribed to topic(s): testOutputTopic (KafkaConsumer.java:824) 10:46:49 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:49 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker localhost:9092 (id: -1 rack: null) (AbstractCoordinator.java:548) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name connections-closed:broker-id-1 (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name connections-created:broker-id-1 (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent-received:broker-id-1 (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent:broker-id-1 (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name bytes-received:broker-id-1 (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name select-time:broker-id-1 (Metrics.java:335) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name io-time:broker-id-1 (Metrics.java:335) 10:46:49 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node -1 at localhost:9092. (NetworkClient.java:627) 10:46:49 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:49 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name node--1.bytes-sent (Metrics.java:335) 10:46:49 [INFO ] Logging$class.info - [Controller-1-to-broker-1-send-thread], Starting (Logging.scala:70) 10:46:49 [DEBUG] Logging$class.debug - Accepted connection from /127.0.0.1:53380 on /127.0.0.1:9092 and assigned it to processor 0, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name node--1.bytes-received (Metrics.java:335) 10:46:49 [INFO ] Logging$class.info - [Controller 1]: New broker startup callback for 1 (Logging.scala:70) 10:46:49 [DEBUG] Metrics.sensor - Added sensor with name node--1.latency (Metrics.java:335) 10:46:49 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 65536, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node -1 (Selector.java:339) 10:46:49 [DEBUG] NetworkClient.handleConnections - Completed connection to node -1. Fetching API versions. (NetworkClient.java:590) 10:46:49 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node -1. (NetworkClient.java:603) 10:46:49 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node -1. Assuming version 0. (NetworkClient.java:292) 10:46:49 [DEBUG] Logging$class.debug - Processor 0 listening to new connection from /127.0.0.1:53380 (Logging.scala:54) 10:46:49 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node 1 at ISI050.utenze.BANKIT.IT:9092. (NetworkClient.java:627) 10:46:49 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 8192, SO_SNDBUF = 64512, SO_TIMEOUT = 0 to node 1 (Selector.java:339) 10:46:49 [DEBUG] NetworkClient.handleConnections - Completed connection to node 1. Ready. (NetworkClient.java:593) 10:46:49 [DEBUG] Logging$class.debug - Accepted connection from /10.36.240.33:53381 on /10.36.240.33:9092 and assigned it to processor 1, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:49 [DEBUG] Logging$class.debug - Processor 1 listening to new connection from /10.36.240.33:53381 (Logging.scala:54) 10:46:49 [TRACE] NetworkClient.doSend - Sending {} to node -1. (NetworkClient.java:316) 10:46:49 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:49 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:49 [INFO ] Logging$class.info - [Controller-1-to-broker-1-send-thread], Controller 1 connected to ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) for sending state change requests (Logging.scala:70) 10:46:49 [TRACE] NetworkClient.doSend - Sending {controller_id=1,controller_epoch=1,partition_states=[],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} to node 1. (NetworkClient.java:316) 10:46:49 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=0,api_version=2,correlation_id=0,client_id=} -- {acks=0,timeout=0,topic_data=[]} (Logging.scala:36) 10:46:49 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=18,api_version=0,correlation_id=1,client_id=consumer-1} -- {} (Logging.scala:36) 10:46:49 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(0,127.0.0.1:9092-127.0.0.1:53380,Session(User:ANONYMOUS,/127.0.0.1),null,1491468409204,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:49 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=6,api_version=3,correlation_id=0,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} (Logging.scala:36) 10:46:49 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468409209,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:49 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=6,api_version=3,correlation_id=0,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} from connection 10.36.240.33:9092-10.36.240.33:53381;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:49 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=1,client_id=consumer-1} -- {} from connection 127.0.0.1:9092-127.0.0.1:53380;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:49 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,127.0.0.1:9092-127.0.0.1:53380,Session(User:ANONYMOUS,/127.0.0.1),null,1491468409204,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2cd6e8cb,SendAction) (Logging.scala:36) 10:46:49 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468409209,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@3dac4c8e,SendAction) (Logging.scala:36) 10:46:49 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 6, received {error_code=0} (NetworkClient.java:534) 10:46:49 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:49 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node -1: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:49 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:49 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node -1 (NetworkClient.java:751) 10:46:49 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node -1. (NetworkClient.java:316) 10:46:49 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node -1. (NetworkClient.java:316) 10:46:49 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=1,client_id=consumer-1} -- {} from connection 127.0.0.1:9092-127.0.0.1:53380;totalTime:46,requestQueueTime:10,localTime:30,remoteTime:0,responseQueueTime:2,sendTime:4,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:49 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=3,api_version=2,correlation_id=2,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:49 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=6,api_version=3,correlation_id=0,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} from connection 10.36.240.33:9092-10.36.240.33:53381;totalTime:42,requestQueueTime:9,localTime:28,remoteTime:0,responseQueueTime:0,sendTime:5,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:49 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(0,127.0.0.1:9092-127.0.0.1:53380,Session(User:ANONYMOUS,/127.0.0.1),null,1491468409323,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:49 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=2,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 127.0.0.1:9092-127.0.0.1:53380;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:49 [INFO ] Logging$class.info - Topic creation {"version":1,"partitions":{"0":[1]}} (Logging.scala:70) 10:46:50 [DEBUG] Logging$class.debug - Updated path /brokers/topics/testOutputTopic with {"version":1,"partitions":{"0":[1]}} for replica assignment (Logging.scala:54) 10:46:50 [INFO ] Logging$class.info - [KafkaApi-1] Auto creation of topic testOutputTopic with 1 partitions and replication factor 1 is successful (Logging.scala:70) 10:46:50 [DEBUG] Logging$class.debug - [TopicChangeListener on Controller 1]: Topic change listener fired for path /brokers/topics with children testOutputTopic (Logging.scala:54) 10:46:50 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@1d323934 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 2 to client consumer-1 (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,127.0.0.1:9092-127.0.0.1:53380,Session(User:ANONYMOUS,/127.0.0.1),null,1491468409323,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@60a8182f,SendAction) (Logging.scala:36) 10:46:50 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=2,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 127.0.0.1:9092-127.0.0.1:53380;totalTime:1249,requestQueueTime:1,localTime:1247,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:50 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=5,topic=testOutputTopic,is_internal=false,partition_metadata=[]}]} (NetworkClient.java:534) 10:46:50 [WARN ] NetworkClient$DefaultMetadataUpdater.handleCompletedMetadataResponse - Error while fetching metadata with correlation id 2 : {testOutputTopic=LEADER_NOT_AVAILABLE} (NetworkClient.java:707) 10:46:50 [DEBUG] Metadata.update - Updated cluster metadata version 2 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = []) (Metadata.java:244) 10:46:50 [DEBUG] Logging$class.debug - Replicas assigned to topic [testOutputTopic], partition [0] are [List(1)] (Logging.scala:54) 10:46:50 [INFO ] Logging$class.info - [TopicChangeListener on Controller 1]: New topics: [Set(testOutputTopic)], deleted topics: [Set()], new partition replica assignment [Map([testOutputTopic,0] -> List(1))] (Logging.scala:70) 10:46:50 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=10,api_version=0,correlation_id=0,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(0,127.0.0.1:9092-127.0.0.1:53380,Session(User:ANONYMOUS,/127.0.0.1),null,1491468410576,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=0,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 127.0.0.1:9092-127.0.0.1:53380;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:50 [INFO ] Logging$class.info - [Controller 1]: New topic creation callback for [testOutputTopic,0] (Logging.scala:70) 10:46:50 [INFO ] Logging$class.info - [Controller 1]: New partition creation callback for [testOutputTopic,0] (Logging.scala:70) 10:46:50 [INFO ] Logging$class.info - [Partition state machine on Controller 1]: Invoking state change to NewPartition for partitions [testOutputTopic,0] (Logging.scala:70) 10:46:50 [INFO ] Logging$class.info - [Replica state machine on controller 1]: Invoking state change to NewReplica for replicas [Topic=testOutputTopic,Partition=0,Replica=1] (Logging.scala:70) 10:46:50 [INFO ] Logging$class.info - [Partition state machine on Controller 1]: Invoking state change to OnlinePartition for partitions [testOutputTopic,0] (Logging.scala:70) 10:46:50 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [testOutputTopic,0] are: [List(1)] (Logging.scala:54) 10:46:50 [INFO ] Logging$class.info - Topic creation {"version":1,"partitions":{"45":[1],"34":[1],"12":[1],"8":[1],"19":[1],"23":[1],"4":[1],"40":[1],"15":[1],"11":[1],"9":[1],"44":[1],"33":[1],"22":[1],"26":[1],"37":[1],"13":[1],"46":[1],"24":[1],"35":[1],"16":[1],"5":[1],"10":[1],"48":[1],"21":[1],"43":[1],"32":[1],"49":[1],"6":[1],"36":[1],"1":[1],"39":[1],"17":[1],"25":[1],"14":[1],"47":[1],"31":[1],"42":[1],"0":[1],"20":[1],"27":[1],"2":[1],"38":[1],"18":[1],"30":[1],"7":[1],"29":[1],"41":[1],"3":[1],"28":[1]}} (Logging.scala:70) 10:46:50 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [testOutputTopic,0] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:50 [DEBUG] Logging$class.debug - Updated path /brokers/topics/__consumer_offsets with {"version":1,"partitions":{"45":[1],"34":[1],"12":[1],"8":[1],"19":[1],"23":[1],"4":[1],"40":[1],"15":[1],"11":[1],"9":[1],"44":[1],"33":[1],"22":[1],"26":[1],"37":[1],"13":[1],"46":[1],"24":[1],"35":[1],"16":[1],"5":[1],"10":[1],"48":[1],"21":[1],"43":[1],"32":[1],"49":[1],"6":[1],"36":[1],"1":[1],"39":[1],"17":[1],"25":[1],"14":[1],"47":[1],"31":[1],"42":[1],"0":[1],"20":[1],"27":[1],"2":[1],"38":[1],"18":[1],"30":[1],"7":[1],"29":[1],"41":[1],"3":[1],"28":[1]}} for replica assignment (Logging.scala:54) 10:46:50 [INFO ] Logging$class.info - [KafkaApi-1] Auto creation of topic __consumer_offsets with 50 partitions and replication factor 1 is successful (Logging.scala:70) 10:46:50 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 0 to client consumer-1. (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,127.0.0.1:9092-127.0.0.1:53380,Session(User:ANONYMOUS,/127.0.0.1),null,1491468410576,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6d9e2791,SendAction) (Logging.scala:36) 10:46:50 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:50 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=0,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 127.0.0.1:9092-127.0.0.1:53380;totalTime:212,requestQueueTime:6,localTime:202,remoteTime:0,responseQueueTime:0,sendTime:4,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:50 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468410788, latencyMs=1650, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=0,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:50 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:50 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:50 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (NetworkClient.java:443) 10:46:50 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Initialize connection to node 1 for sending metadata request (NetworkClient.java:767) 10:46:50 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node 1 at ISI050.utenze.BANKIT.IT:9092. (NetworkClient.java:627) 10:46:50 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (NetworkClient.java:443) 10:46:50 [DEBUG] Metrics.sensor - Added sensor with name node-1.bytes-sent (Metrics.java:335) 10:46:50 [DEBUG] Logging$class.debug - Accepted connection from /10.36.240.33:53382 on /10.36.240.33:9092 and assigned it to processor 2, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:50 [DEBUG] Logging$class.debug - Processor 2 listening to new connection from /10.36.240.33:53382 (Logging.scala:54) 10:46:50 [DEBUG] Metrics.sensor - Added sensor with name node-1.bytes-received (Metrics.java:335) 10:46:50 [DEBUG] Metrics.sensor - Added sensor with name node-1.latency (Metrics.java:335) 10:46:50 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 65536, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node 1 (Selector.java:339) 10:46:50 [DEBUG] NetworkClient.handleConnections - Completed connection to node 1. Fetching API versions. (NetworkClient.java:590) 10:46:50 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node 1. (NetworkClient.java:603) 10:46:50 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node 1. Assuming version 0. (NetworkClient.java:292) 10:46:50 [TRACE] NetworkClient.doSend - Sending {} to node 1. (NetworkClient.java:316) 10:46:50 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (NetworkClient.java:443) 10:46:50 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (NetworkClient.java:443) 10:46:50 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=18,api_version=0,correlation_id=3,client_id=consumer-1} -- {} (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410801,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=3,client_id=consumer-1} -- {} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410801,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@5e51c0b3,SendAction) (Logging.scala:36) 10:46:50 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=3,client_id=consumer-1} -- {} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:7,requestQueueTime:2,localTime:4,remoteTime:0,responseQueueTime:1,sendTime:2,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:50 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:50 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node 1: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:50 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:50 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:50 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:50 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=4,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410815,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=4,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@17c1688e and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 4 to client consumer-1 (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410815,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1b53451d,SendAction) (Logging.scala:36) 10:46:50 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=4,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:70,requestQueueTime:2,localTime:67,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:50 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=5,topic=testOutputTopic,is_internal=false,partition_metadata=[]}]} (NetworkClient.java:534) 10:46:50 [WARN ] NetworkClient$DefaultMetadataUpdater.handleCompletedMetadataResponse - Error while fetching metadata with correlation id 4 : {testOutputTopic=LEADER_NOT_AVAILABLE} (NetworkClient.java:707) 10:46:50 [DEBUG] Metadata.update - Updated cluster metadata version 3 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = []) (Metadata.java:244) 10:46:50 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:50 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:50 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:50 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=5,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410889,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=5,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:50 [INFO ] Logging$class.info - [Replica state machine on controller 1]: Invoking state change to OnlineReplica for replicas [Topic=testOutputTopic,Partition=0,Replica=1] (Logging.scala:70) 10:46:50 [TRACE] NetworkClient.doSend - Sending {controller_id=1,controller_epoch=1,partition_states=[{topic=testOutputTopic,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_leaders=[{id=1,host=ISI050.utenze.BANKIT.IT,port=9092}]} to node 1. (NetworkClient.java:316) 10:46:50 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=4,api_version=0,correlation_id=1,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=testOutputTopic,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_leaders=[{id=1,host=ISI050.utenze.BANKIT.IT,port=9092}]} (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410916,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=4,api_version=0,correlation_id=1,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=testOutputTopic,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_leaders=[{id=1,host=ISI050.utenze.BANKIT.IT,port=9092}]} from connection 10.36.240.33:9092-10.36.240.33:53381;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:50 [DEBUG] Logging$class.debug - [TopicChangeListener on Controller 1]: Topic change listener fired for path /brokers/topics with children testOutputTopic,__consumer_offsets (Logging.scala:54) 10:46:50 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 5 to client consumer-1. (Logging.scala:36) 10:46:50 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410889,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@23d0207,SendAction) (Logging.scala:36) 10:46:50 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=5,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:68,requestQueueTime:1,localTime:67,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:50 [INFO ] Logging$class.info - [ReplicaFetcherManager on broker 1] Removed fetcher for partitions testOutputTopic-0 (Logging.scala:70) 10:46:50 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:50 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468410960, latencyMs=72, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=5,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:50 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:50 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:50 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:50 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:50 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:51 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=6,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [45] are [List(1)] (Logging.scala:54) 10:46:51 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410988,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=6,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [34] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [12] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [8] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [19] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [23] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [4] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [40] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [15] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [11] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [9] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [44] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [33] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [22] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [26] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [37] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [13] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [46] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [24] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [35] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [16] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [5] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [10] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [48] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [21] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [43] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [32] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [49] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [6] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [36] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [1] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [39] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [17] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [25] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [14] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [47] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [31] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [42] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [0] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [20] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [27] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [2] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [38] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [18] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [30] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [7] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [29] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [41] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [3] are [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - Replicas assigned to topic [__consumer_offsets], partition [28] are [List(1)] (Logging.scala:54) 10:46:51 [INFO ] Logging$class.info - [TopicChangeListener on Controller 1]: New topics: [Set(__consumer_offsets)], deleted topics: [Set()], new partition replica assignment [Map([__consumer_offsets,19] -> List(1), [__consumer_offsets,30] -> List(1), [__consumer_offsets,47] -> List(1), [__consumer_offsets,29] -> List(1), [__consumer_offsets,41] -> List(1), [__consumer_offsets,39] -> List(1), [__consumer_offsets,10] -> List(1), [__consumer_offsets,17] -> List(1), [__consumer_offsets,14] -> List(1), [__consumer_offsets,40] -> List(1), [__consumer_offsets,18] -> List(1), [__consumer_offsets,26] -> List(1), [__consumer_offsets,0] -> List(1), [__consumer_offsets,24] -> List(1), [__consumer_offsets,33] -> List(1), [__consumer_offsets,20] -> List(1), [__consumer_offsets,21] -> List(1), [__consumer_offsets,3] -> List(1), [__consumer_offsets,5] -> List(1), [__consumer_offsets,22] -> List(1), [__consumer_offsets,12] -> List(1), [__consumer_offsets,8] -> List(1), [__consumer_offsets,23] -> List(1), [__consumer_offsets,15] -> List(1), [__consumer_offsets,48] -> List(1), [__consumer_offsets,11] -> List(1), [__consumer_offsets,13] -> List(1), [__consumer_offsets,49] -> List(1), [__consumer_offsets,6] -> List(1), [__consumer_offsets,28] -> List(1), [__consumer_offsets,4] -> List(1), [__consumer_offsets,37] -> List(1), [__consumer_offsets,31] -> List(1), [__consumer_offsets,44] -> List(1), [__consumer_offsets,42] -> List(1), [__consumer_offsets,34] -> List(1), [__consumer_offsets,46] -> List(1), [__consumer_offsets,25] -> List(1), [__consumer_offsets,45] -> List(1), [__consumer_offsets,27] -> List(1), [__consumer_offsets,32] -> List(1), [__consumer_offsets,43] -> List(1), [__consumer_offsets,36] -> List(1), [__consumer_offsets,35] -> List(1), [__consumer_offsets,7] -> List(1), [__consumer_offsets,9] -> List(1), [__consumer_offsets,38] -> List(1), [__consumer_offsets,1] -> List(1), [__consumer_offsets,16] -> List(1), [__consumer_offsets,2] -> List(1))] (Logging.scala:70) 10:46:51 [INFO ] Logging$class.info - [Controller 1]: New topic creation callback for [__consumer_offsets,19],[__consumer_offsets,30],[__consumer_offsets,47],[__consumer_offsets,29],[__consumer_offsets,41],[__consumer_offsets,39],[__consumer_offsets,10],[__consumer_offsets,17],[__consumer_offsets,14],[__consumer_offsets,40],[__consumer_offsets,18],[__consumer_offsets,26],[__consumer_offsets,0],[__consumer_offsets,24],[__consumer_offsets,33],[__consumer_offsets,20],[__consumer_offsets,21],[__consumer_offsets,3],[__consumer_offsets,5],[__consumer_offsets,22],[__consumer_offsets,12],[__consumer_offsets,8],[__consumer_offsets,23],[__consumer_offsets,15],[__consumer_offsets,48],[__consumer_offsets,11],[__consumer_offsets,13],[__consumer_offsets,49],[__consumer_offsets,6],[__consumer_offsets,28],[__consumer_offsets,4],[__consumer_offsets,37],[__consumer_offsets,31],[__consumer_offsets,44],[__consumer_offsets,42],[__consumer_offsets,34],[__consumer_offsets,46],[__consumer_offsets,25],[__consumer_offsets,45],[__consumer_offsets,27],[__consumer_offsets,32],[__consumer_offsets,43],[__consumer_offsets,36],[__consumer_offsets,35],[__consumer_offsets,7],[__consumer_offsets,9],[__consumer_offsets,38],[__consumer_offsets,1],[__consumer_offsets,16],[__consumer_offsets,2] (Logging.scala:70) 10:46:51 [INFO ] Logging$class.info - [Controller 1]: New partition creation callback for [__consumer_offsets,19],[__consumer_offsets,30],[__consumer_offsets,47],[__consumer_offsets,29],[__consumer_offsets,41],[__consumer_offsets,39],[__consumer_offsets,10],[__consumer_offsets,17],[__consumer_offsets,14],[__consumer_offsets,40],[__consumer_offsets,18],[__consumer_offsets,26],[__consumer_offsets,0],[__consumer_offsets,24],[__consumer_offsets,33],[__consumer_offsets,20],[__consumer_offsets,21],[__consumer_offsets,3],[__consumer_offsets,5],[__consumer_offsets,22],[__consumer_offsets,12],[__consumer_offsets,8],[__consumer_offsets,23],[__consumer_offsets,15],[__consumer_offsets,48],[__consumer_offsets,11],[__consumer_offsets,13],[__consumer_offsets,49],[__consumer_offsets,6],[__consumer_offsets,28],[__consumer_offsets,4],[__consumer_offsets,37],[__consumer_offsets,31],[__consumer_offsets,44],[__consumer_offsets,42],[__consumer_offsets,34],[__consumer_offsets,46],[__consumer_offsets,25],[__consumer_offsets,45],[__consumer_offsets,27],[__consumer_offsets,32],[__consumer_offsets,43],[__consumer_offsets,36],[__consumer_offsets,35],[__consumer_offsets,7],[__consumer_offsets,9],[__consumer_offsets,38],[__consumer_offsets,1],[__consumer_offsets,16],[__consumer_offsets,2] (Logging.scala:70) 10:46:51 [INFO ] Logging$class.info - [Partition state machine on Controller 1]: Invoking state change to NewPartition for partitions [__consumer_offsets,19],[__consumer_offsets,30],[__consumer_offsets,47],[__consumer_offsets,29],[__consumer_offsets,41],[__consumer_offsets,39],[__consumer_offsets,10],[__consumer_offsets,17],[__consumer_offsets,14],[__consumer_offsets,40],[__consumer_offsets,18],[__consumer_offsets,26],[__consumer_offsets,0],[__consumer_offsets,24],[__consumer_offsets,33],[__consumer_offsets,20],[__consumer_offsets,21],[__consumer_offsets,3],[__consumer_offsets,5],[__consumer_offsets,22],[__consumer_offsets,12],[__consumer_offsets,8],[__consumer_offsets,23],[__consumer_offsets,15],[__consumer_offsets,48],[__consumer_offsets,11],[__consumer_offsets,13],[__consumer_offsets,49],[__consumer_offsets,6],[__consumer_offsets,28],[__consumer_offsets,4],[__consumer_offsets,37],[__consumer_offsets,31],[__consumer_offsets,44],[__consumer_offsets,42],[__consumer_offsets,34],[__consumer_offsets,46],[__consumer_offsets,25],[__consumer_offsets,45],[__consumer_offsets,27],[__consumer_offsets,32],[__consumer_offsets,43],[__consumer_offsets,36],[__consumer_offsets,35],[__consumer_offsets,7],[__consumer_offsets,9],[__consumer_offsets,38],[__consumer_offsets,1],[__consumer_offsets,16],[__consumer_offsets,2] (Logging.scala:70) 10:46:51 [INFO ] Logging$class.info - [Replica state machine on controller 1]: Invoking state change to NewReplica for replicas [Topic=__consumer_offsets,Partition=28,Replica=1],[Topic=__consumer_offsets,Partition=48,Replica=1],[Topic=__consumer_offsets,Partition=5,Replica=1],[Topic=__consumer_offsets,Partition=21,Replica=1],[Topic=__consumer_offsets,Partition=2,Replica=1],[Topic=__consumer_offsets,Partition=18,Replica=1],[Topic=__consumer_offsets,Partition=23,Replica=1],[Topic=__consumer_offsets,Partition=9,Replica=1],[Topic=__consumer_offsets,Partition=39,Replica=1],[Topic=__consumer_offsets,Partition=31,Replica=1],[Topic=__consumer_offsets,Partition=19,Replica=1],[Topic=__consumer_offsets,Partition=10,Replica=1],[Topic=__consumer_offsets,Partition=22,Replica=1],[Topic=__consumer_offsets,Partition=43,Replica=1],[Topic=__consumer_offsets,Partition=40,Replica=1],[Topic=__consumer_offsets,Partition=27,Replica=1],[Topic=__consumer_offsets,Partition=6,Replica=1],[Topic=__consumer_offsets,Partition=1,Replica=1],[Topic=__consumer_offsets,Partition=47,Replica=1],[Topic=__consumer_offsets,Partition=30,Replica=1],[Topic=__consumer_offsets,Partition=42,Replica=1],[Topic=__consumer_offsets,Partition=41,Replica=1],[Topic=__consumer_offsets,Partition=3,Replica=1],[Topic=__consumer_offsets,Partition=13,Replica=1],[Topic=__consumer_offsets,Partition=4,Replica=1],[Topic=__consumer_offsets,Partition=16,Replica=1],[Topic=__consumer_offsets,Partition=46,Replica=1],[Topic=__consumer_offsets,Partition=49,Replica=1],[Topic=__consumer_offsets,Partition=14,Replica=1],[Topic=__consumer_offsets,Partition=45,Replica=1],[Topic=__consumer_offsets,Partition=37,Replica=1],[Topic=__consumer_offsets,Partition=29,Replica=1],[Topic=__consumer_offsets,Partition=20,Replica=1],[Topic=__consumer_offsets,Partition=8,Replica=1],[Topic=__consumer_offsets,Partition=38,Replica=1],[Topic=__consumer_offsets,Partition=7,Replica=1],[Topic=__consumer_offsets,Partition=0,Replica=1],[Topic=__consumer_offsets,Partition=34,Replica=1],[Topic=__consumer_offsets,Partition=33,Replica=1],[Topic=__consumer_offsets,Partition=26,Replica=1],[Topic=__consumer_offsets,Partition=44,Replica=1],[Topic=__consumer_offsets,Partition=32,Replica=1],[Topic=__consumer_offsets,Partition=25,Replica=1],[Topic=__consumer_offsets,Partition=11,Replica=1],[Topic=__consumer_offsets,Partition=36,Replica=1],[Topic=__consumer_offsets,Partition=12,Replica=1],[Topic=__consumer_offsets,Partition=35,Replica=1],[Topic=__consumer_offsets,Partition=15,Replica=1],[Topic=__consumer_offsets,Partition=17,Replica=1],[Topic=__consumer_offsets,Partition=24,Replica=1] (Logging.scala:70) 10:46:51 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@6051042e and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 6 to client consumer-1 (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410988,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2a027a19,SendAction) (Logging.scala:36) 10:46:51 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=5,topic=testOutputTopic,is_internal=false,partition_metadata=[]}]} (NetworkClient.java:534) 10:46:51 [WARN ] NetworkClient$DefaultMetadataUpdater.handleCompletedMetadataResponse - Error while fetching metadata with correlation id 6 : {testOutputTopic=LEADER_NOT_AVAILABLE} (NetworkClient.java:707) 10:46:51 [DEBUG] Metadata.update - Updated cluster metadata version 4 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = []) (Metadata.java:244) 10:46:51 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:51 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:51 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=6,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:93,requestQueueTime:18,localTime:74,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:51 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:51 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=7,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468411083,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=7,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 7 to client consumer-1. (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468411083,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7a8663b2,SendAction) (Logging.scala:36) 10:46:51 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=7,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:14,requestQueueTime:1,localTime:12,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:51 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:51 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468411098, latencyMs=16, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=7,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:51 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:51 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:51 [INFO ] Logging$class.info - [Partition state machine on Controller 1]: Invoking state change to OnlinePartition for partitions [__consumer_offsets,19],[__consumer_offsets,30],[__consumer_offsets,47],[__consumer_offsets,29],[__consumer_offsets,41],[__consumer_offsets,39],[__consumer_offsets,10],[__consumer_offsets,17],[__consumer_offsets,14],[__consumer_offsets,40],[__consumer_offsets,18],[__consumer_offsets,26],[__consumer_offsets,0],[__consumer_offsets,24],[__consumer_offsets,33],[__consumer_offsets,20],[__consumer_offsets,21],[__consumer_offsets,3],[__consumer_offsets,5],[__consumer_offsets,22],[__consumer_offsets,12],[__consumer_offsets,8],[__consumer_offsets,23],[__consumer_offsets,15],[__consumer_offsets,48],[__consumer_offsets,11],[__consumer_offsets,13],[__consumer_offsets,49],[__consumer_offsets,6],[__consumer_offsets,28],[__consumer_offsets,4],[__consumer_offsets,37],[__consumer_offsets,31],[__consumer_offsets,44],[__consumer_offsets,42],[__consumer_offsets,34],[__consumer_offsets,46],[__consumer_offsets,25],[__consumer_offsets,45],[__consumer_offsets,27],[__consumer_offsets,32],[__consumer_offsets,43],[__consumer_offsets,36],[__consumer_offsets,35],[__consumer_offsets,7],[__consumer_offsets,9],[__consumer_offsets,38],[__consumer_offsets,1],[__consumer_offsets,16],[__consumer_offsets,2] (Logging.scala:70) 10:46:51 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,19] are: [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,19] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:51 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:46:51 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:51 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:51 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:51 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=8,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468411182,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=8,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:51 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,30] are: [List(1)] (Logging.scala:54) 10:46:51 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@19068406 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 8 to client consumer-1 (Logging.scala:36) 10:46:51 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,30] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:51 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468411182,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7b7cba52,SendAction) (Logging.scala:36) 10:46:51 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=8,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:104,requestQueueTime:1,localTime:102,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:51 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=5,topic=testOutputTopic,is_internal=false,partition_metadata=[]}]} (NetworkClient.java:534) 10:46:51 [WARN ] NetworkClient$DefaultMetadataUpdater.handleCompletedMetadataResponse - Error while fetching metadata with correlation id 8 : {testOutputTopic=LEADER_NOT_AVAILABLE} (NetworkClient.java:707) 10:46:51 [DEBUG] Metadata.update - Updated cluster metadata version 5 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = []) (Metadata.java:244) 10:46:51 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:51 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:51 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:51 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=9,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468411289,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=9,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:51 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\testOutputTopic-0\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,47] are: [List(1)] (Logging.scala:54) 10:46:51 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,47] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:51 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 9 to client consumer-1. (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468411289,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@113bb200,SendAction) (Logging.scala:36) 10:46:51 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=9,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:701,requestQueueTime:1,localTime:699,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:51 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:51 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468411991, latencyMs=703, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=9,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:51 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:51 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:51 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:51 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:51 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:51 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=10,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468411992,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:51 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=10,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [INFO ] Logging$class.info - Completed load of log testOutputTopic-0 with 1 log segments and log end offset 0 in 1022 ms (Logging.scala:70) 10:46:52 [INFO ] Logging$class.info - Created log for partition [testOutputTopic,0] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 1073741824, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:52 [INFO ] Logging$class.info - Partition [testOutputTopic,0] on broker 1: No checkpointed highwatermark is found for partition testOutputTopic-0 (Logging.scala:70) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,29] are: [List(1)] (Logging.scala:54) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,29] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@15d66c41 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 10 to client consumer-1 (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468411992,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2850c08c,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=5,topic=testOutputTopic,is_internal=false,partition_metadata=[]}]} (NetworkClient.java:534) 10:46:52 [WARN ] NetworkClient$DefaultMetadataUpdater.handleCompletedMetadataResponse - Error while fetching metadata with correlation id 10 : {testOutputTopic=LEADER_NOT_AVAILABLE} (NetworkClient.java:707) 10:46:52 [DEBUG] Metadata.update - Updated cluster metadata version 6 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = []) (Metadata.java:244) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:52 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=10,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:183,requestQueueTime:1,localTime:181,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=11,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412177,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=11,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,41] are: [List(1)] (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 11 to client consumer-1. (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,41] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412177,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6182adc7,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=11,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:85,requestQueueTime:1,localTime:84,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468412262, latencyMs=86, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=11,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:52 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:52 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=12,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412276,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=12,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log testOutputTopic-0 of length 0 bytes (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,39] are: [List(1)] (Logging.scala:54) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,39] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@6e96f654 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 12 to client consumer-1 (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412276,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7f32258f,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=5,topic=testOutputTopic,is_internal=false,partition_metadata=[]}]} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=12,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:88,requestQueueTime:0,localTime:87,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [WARN ] NetworkClient$DefaultMetadataUpdater.handleCompletedMetadataResponse - Error while fetching metadata with correlation id 12 : {testOutputTopic=LEADER_NOT_AVAILABLE} (NetworkClient.java:707) 10:46:52 [DEBUG] Metadata.update - Updated cluster metadata version 7 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = []) (Metadata.java:244) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:52 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=13,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412366,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=13,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 13 to client consumer-1. (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,10] are: [List(1)] (Logging.scala:54) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,10] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412366,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@33d533db,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=13,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:74,requestQueueTime:0,localTime:73,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468412440, latencyMs=75, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=13,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:52 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:52 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [testOutputTopic,0]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:52 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=14,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412466,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=14,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - Scheduling task highwatermark-checkpoint with initial delay 0 ms and period 5000 ms. (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,17] are: [List(1)] (Logging.scala:54) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,17] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@1b11908b and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 14 to client consumer-1 (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412466,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1d2fc18d,SendAction) (Logging.scala:36) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=14,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:101,requestQueueTime:2,localTime:98,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=5,topic=testOutputTopic,is_internal=false,partition_metadata=[]}]} (NetworkClient.java:534) 10:46:52 [WARN ] NetworkClient$DefaultMetadataUpdater.handleCompletedMetadataResponse - Error while fetching metadata with correlation id 14 : {testOutputTopic=LEADER_NOT_AVAILABLE} (NetworkClient.java:707) 10:46:52 [DEBUG] Metadata.update - Updated cluster metadata version 8 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = []) (Metadata.java:244) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:52 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=15,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412570,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=15,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468410916,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@72304487,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 4, received {error_code=0,partitions=[{topic=testOutputTopic,partition=0,error_code=0}]} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=4,api_version=0,correlation_id=1,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=testOutputTopic,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_leaders=[{id=1,host=ISI050.utenze.BANKIT.IT,port=9092}]} from connection 10.36.240.33:9092-10.36.240.33:53381;totalTime:1673,requestQueueTime:1,localTime:1671,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [TRACE] NetworkClient.doSend - Sending {controller_id=1,controller_epoch=1,partition_states=[{topic=testOutputTopic,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=6,api_version=3,correlation_id=2,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=testOutputTopic,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412590,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=6,api_version=3,correlation_id=2,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=testOutputTopic,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} from connection 10.36.240.33:9092-10.36.240.33:53381;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412590,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@5970a8cc,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 6, received {error_code=0} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=6,api_version=3,correlation_id=2,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=testOutputTopic,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} from connection 10.36.240.33:9092-10.36.240.33:53381;totalTime:7,requestQueueTime:1,localTime:5,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,14] are: [List(1)] (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 15 to client consumer-1. (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,14] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412570,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@69b7ab7a,SendAction) (Logging.scala:36) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=15,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:112,requestQueueTime:3,localTime:106,remoteTime:0,responseQueueTime:0,sendTime:3,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468412682, latencyMs=113, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=15,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:52 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:52 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=16,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412687,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=16,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@31a51b3 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 16 to client consumer-1 (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412687,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@747f59e,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=16,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:11,requestQueueTime:1,localTime:8,remoteTime:0,responseQueueTime:0,sendTime:2,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [DEBUG] Metadata.update - Updated cluster metadata version 9 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:52 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=17,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412702,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=17,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 17 to client consumer-1. (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,40] are: [List(1)] (Logging.scala:54) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,40] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412702,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1d965546,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=17,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:49,requestQueueTime:0,localTime:48,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468412751, latencyMs=50, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=17,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:52 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:52 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=18,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412800,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=18,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@1f2e0731 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 18 to client consumer-1 (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412800,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7e56ccd0,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=18,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:1,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [DEBUG] Metadata.update - Updated cluster metadata version 10 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:52 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=19,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412805,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=19,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,18] are: [List(1)] (Logging.scala:54) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,18] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 19 to client consumer-1. (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412805,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2aee541b,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=19,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:46,requestQueueTime:1,localTime:44,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468412851, latencyMs=47, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=19,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:52 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,26] are: [List(1)] (Logging.scala:54) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,26] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:52 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=20,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412904,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=20,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@211d2427 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 20 to client consumer-1 (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412904,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@123dbfdd,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=20,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [DEBUG] Metadata.update - Updated cluster metadata version 11 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:52 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:52 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:52 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:52 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=21,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412909,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=21,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,0] are: [List(1)] (Logging.scala:54) 10:46:52 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,0] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:52 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 21 to client consumer-1. (Logging.scala:36) 10:46:52 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468412909,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@3d2ee4b9,SendAction) (Logging.scala:36) 10:46:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:52 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=21,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:64,requestQueueTime:1,localTime:62,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468412973, latencyMs=65, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=21,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:52 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:52 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:53 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=22,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413008,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=22,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@5890b228 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 22 to client consumer-1 (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413008,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@75d69f80,SendAction) (Logging.scala:36) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=22,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:53 [DEBUG] Metadata.update - Updated cluster metadata version 12 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:53 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=23,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413012,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=23,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,24] are: [List(1)] (Logging.scala:54) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,24] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 23 to client consumer-1. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413012,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2dc043ad,SendAction) (Logging.scala:36) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468413041, latencyMs=30, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=23,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:53 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=23,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:31,requestQueueTime:0,localTime:27,remoteTime:0,responseQueueTime:1,sendTime:3,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,33] are: [List(1)] (Logging.scala:54) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,33] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:53 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=24,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413111,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=24,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@1afda803 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 24 to client consumer-1 (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413111,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@52fcc29f,SendAction) (Logging.scala:36) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=24,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:4,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:2,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [DEBUG] Metadata.update - Updated cluster metadata version 13 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:53 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=25,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413117,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=25,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,20] are: [List(1)] (Logging.scala:54) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,20] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 25 to client consumer-1. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413117,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@18bdef55,SendAction) (Logging.scala:36) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=25,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:62,requestQueueTime:0,localTime:55,remoteTime:0,responseQueueTime:6,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468413179, latencyMs=63, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=25,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:53 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:53 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=26,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,21] are: [List(1)] (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413216,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=26,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,21] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@6aa96f5e and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 26 to client consumer-1 (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413216,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@59d6595d,SendAction) (Logging.scala:36) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=26,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:53 [DEBUG] Metadata.update - Updated cluster metadata version 14 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:53 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=27,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413221,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=27,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,3] are: [List(1)] (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 27 to client consumer-1. (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,3] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413221,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@76a4a4e1,SendAction) (Logging.scala:36) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=27,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:53,requestQueueTime:1,localTime:51,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468413274, latencyMs=54, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=27,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:53 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:53 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=28,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413320,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=28,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@5a2c9ab0 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 28 to client consumer-1 (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413320,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@42f86890,SendAction) (Logging.scala:36) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=28,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:53 [DEBUG] Metadata.update - Updated cluster metadata version 15 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:53 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=29,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413327,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=29,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,5] are: [List(1)] (Logging.scala:54) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,5] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 29 to client consumer-1. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413327,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@35c8e0dd,SendAction) (Logging.scala:36) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=29,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:35,requestQueueTime:1,localTime:33,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468413362, latencyMs=37, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=29,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:53 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:53 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=30,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413425,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=30,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@7027195e and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 30 to client consumer-1 (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413425,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@5e3f5b88,SendAction) (Logging.scala:36) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=30,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [DEBUG] Metadata.update - Updated cluster metadata version 16 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:53 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=31,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413430,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=31,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,22] are: [List(1)] (Logging.scala:54) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,22] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'partition-rebalance-thread'. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 31 to client consumer-1. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413430,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@d6f7f02,SendAction) (Logging.scala:36) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=31,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:286,requestQueueTime:0,localTime:285,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468413716, latencyMs=287, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=31,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:53 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:53 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=32,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413719,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=32,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@31a04c44 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 32 to client consumer-1 (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413719,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@c523ad7,SendAction) (Logging.scala:36) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=32,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:53 [DEBUG] Metadata.update - Updated cluster metadata version 17 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:53 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=33,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413724,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=33,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,12] are: [List(1)] (Logging.scala:54) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,12] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 33 to client consumer-1. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413724,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7e6efd88,SendAction) (Logging.scala:36) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=33,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:109,requestQueueTime:1,localTime:107,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468413833, latencyMs=110, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=33,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:53 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:53 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=34,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413835,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=34,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@1d3903fe and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 34 to client consumer-1 (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413835,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@5a291ad6,SendAction) (Logging.scala:36) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=34,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:4,requestQueueTime:1,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:53 [DEBUG] Metadata.update - Updated cluster metadata version 18 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:53 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=35,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413848,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=35,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,8] are: [List(1)] (Logging.scala:54) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,8] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 35 to client consumer-1. (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413848,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2c48e9fe,SendAction) (Logging.scala:36) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=35,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:47,requestQueueTime:1,localTime:46,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468413896, latencyMs=55, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=35,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:53 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,23] are: [List(1)] (Logging.scala:54) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,23] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:53 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=36,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413941,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=36,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@1716d5b and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 36 to client consumer-1 (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413941,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@3151fa22,SendAction) (Logging.scala:36) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=36,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [DEBUG] Metadata.update - Updated cluster metadata version 19 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:53 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:53 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:53 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:53 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=37,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413945,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=37,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:53 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 37 to client consumer-1. (Logging.scala:36) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,15] are: [List(1)] (Logging.scala:54) 10:46:53 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,15] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:53 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468413945,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@385ac33d,SendAction) (Logging.scala:36) 10:46:53 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468413995, latencyMs=51, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=37,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:53 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:53 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=37,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:50,requestQueueTime:0,localTime:50,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:53 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=38,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414044,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=38,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@2fabfc6f and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 38 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414044,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2ab7814a,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=38,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 20 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=39,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414049,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=39,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,48] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,48] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 39 to client consumer-1. (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414049,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@415b8afb,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468414089, latencyMs=41, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=39,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=39,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:39,requestQueueTime:0,localTime:39,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,11] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,11] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=40,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414147,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=40,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@4f71764a and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 40 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414147,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7f65e818,SendAction) (Logging.scala:36) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=40,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 21 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=41,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414152,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=41,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,13] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,13] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 41 to client consumer-1. (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414152,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@11489fe2,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=41,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:54,requestQueueTime:0,localTime:54,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468414206, latencyMs=55, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=41,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:54 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=42,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414250,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=42,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@2fdade74 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 42 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414250,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7e3e150,SendAction) (Logging.scala:36) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=42,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:5,requestQueueTime:1,localTime:3,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 22 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,49] are: [List(1)] (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=43,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,49] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414262,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=43,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 43 to client consumer-1. (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,6] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,6] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414262,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1a72af05,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468414329, latencyMs=68, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=43,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:54 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=43,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:67,requestQueueTime:1,localTime:65,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=44,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414358,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=44,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@583b8e3d and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 44 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414358,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4a87a17b,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=44,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 23 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=45,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414362,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=45,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,28] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,28] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 45 to client consumer-1. (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414362,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1fe58046,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=45,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:55,requestQueueTime:1,localTime:54,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468414417, latencyMs=55, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=45,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:54 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,4] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,4] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=46,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414461,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=46,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@475f8bcd and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 46 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414461,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7940b8b,SendAction) (Logging.scala:36) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=46,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:4,requestQueueTime:1,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 24 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=47,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414470,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=47,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,37] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,37] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 47 to client consumer-1. (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414470,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@81196fb,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=47,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:58,requestQueueTime:1,localTime:57,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468414528, latencyMs=60, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=47,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:54 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,31] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,31] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=48,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414566,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=48,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@5c154ecf and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 48 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414566,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7f6a62fb,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=48,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 25 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=49,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414570,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=49,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,44] are: [List(1)] (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 49 to client consumer-1. (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,44] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414570,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4cbe2ae0,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=49,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:36,requestQueueTime:0,localTime:36,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468414607, latencyMs=38, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=49,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:54 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,42] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,42] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=50,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414669,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=50,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@462bb1de and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 50 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414669,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6b51a8af,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=50,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:5,requestQueueTime:1,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:2,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 26 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=51,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414677,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=51,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,34] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,34] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 51 to client consumer-1. (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414677,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6f735d70,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=51,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:51,requestQueueTime:0,localTime:51,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468414728, latencyMs=53, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=51,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:54 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,46] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,46] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=52,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414775,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=52,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@7c3e51d0 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 52 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414775,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@43a297df,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=52,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 27 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=53,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414779,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=53,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,25] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,25] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 53 to client consumer-1. (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414779,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7e764c61,SendAction) (Logging.scala:36) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=53,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:50,requestQueueTime:0,localTime:49,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468414829, latencyMs=51, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=53,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:54 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,45] are: [List(1)] (Logging.scala:54) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,45] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=54,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414878,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=54,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@4733edc0 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 54 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414878,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@180497d0,SendAction) (Logging.scala:36) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=54,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:5,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:2,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 28 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=55,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414887,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=55,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,27] are: [List(1)] (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 55 to client consumer-1. (Logging.scala:36) 10:46:54 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,27] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414887,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6d69c17d,SendAction) (Logging.scala:36) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=55,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:57,requestQueueTime:2,localTime:54,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468414944, latencyMs=58, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=55,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:54 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:54 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:54 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=56,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414987,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=56,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@68f51a8 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 56 to client consumer-1 (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414987,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@3aa2bf60,SendAction) (Logging.scala:36) 10:46:54 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=56,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:54 [DEBUG] Metadata.update - Updated cluster metadata version 29 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:54 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:54 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=57,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414991,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=57,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,32] are: [List(1)] (Logging.scala:54) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,32] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 57 to client consumer-1. (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468414991,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@76fbf911,SendAction) (Logging.scala:36) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=57,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:38,requestQueueTime:1,localTime:36,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468415029, latencyMs=39, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=57,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:55 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,43] are: [List(1)] (Logging.scala:54) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,43] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:55 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=58,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415091,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=58,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@7f55f701 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 58 to client consumer-1 (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415091,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4afd4a09,SendAction) (Logging.scala:36) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=58,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:1,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:55 [DEBUG] Metadata.update - Updated cluster metadata version 30 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:55 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=59,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415099,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=59,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,36] are: [List(1)] (Logging.scala:54) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,36] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 59 to client consumer-1. (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415099,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@69bc0802,SendAction) (Logging.scala:36) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=59,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:41,requestQueueTime:1,localTime:39,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468415140, latencyMs=43, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=59,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:55 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,35] are: [List(1)] (Logging.scala:54) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,35] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:55 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=60,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415196,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=60,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@442218ac and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 60 to client consumer-1 (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415196,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@464ae3ea,SendAction) (Logging.scala:36) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=60,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:55 [DEBUG] Metadata.update - Updated cluster metadata version 31 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:55 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=61,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415201,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=61,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,7] are: [List(1)] (Logging.scala:54) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,7] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 61 to client consumer-1. (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415201,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6f50b943,SendAction) (Logging.scala:36) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=61,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:50,requestQueueTime:1,localTime:48,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468415251, latencyMs=51, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=61,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:55 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,9] are: [List(1)] (Logging.scala:54) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,9] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:55 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=62,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415301,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=62,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@6bbc0314 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 62 to client consumer-1 (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415301,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4f5d0ba2,SendAction) (Logging.scala:36) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=62,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:17,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:15,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [DEBUG] Metadata.update - Updated cluster metadata version 32 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:55 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=63,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415320,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=63,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,38] are: [List(1)] (Logging.scala:54) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,38] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 63 to client consumer-1. (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415320,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4628e653,SendAction) (Logging.scala:36) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=63,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:42,requestQueueTime:1,localTime:40,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468415362, latencyMs=43, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=63,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:55 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:55 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=64,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415419,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=64,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@2be54f57 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 64 to client consumer-1 (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415419,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@57ec18cf,SendAction) (Logging.scala:36) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=64,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:4,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:55 [DEBUG] Metadata.update - Updated cluster metadata version 33 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:55 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=65,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415425,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=65,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,1] are: [List(1)] (Logging.scala:54) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,1] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 65 to client consumer-1. (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415425,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@3d1c29d0,SendAction) (Logging.scala:36) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468415736, latencyMs=312, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=65,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:55 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:55 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=65,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:311,requestQueueTime:1,localTime:308,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=66,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415738,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=66,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@1e4e9724 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 66 to client consumer-1 (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415738,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1abb35ca,SendAction) (Logging.scala:36) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=66,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:55 [DEBUG] Metadata.update - Updated cluster metadata version 34 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:55 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=67,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415743,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=67,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,16] are: [List(1)] (Logging.scala:54) 10:46:55 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,16] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 67 to client consumer-1. (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415743,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1b2042b0,SendAction) (Logging.scala:36) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=67,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:141,requestQueueTime:1,localTime:140,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468415884, latencyMs=142, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=67,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:55 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:55 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:55 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=68,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415886,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=68,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@1e556bbe and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 68 to client consumer-1 (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415886,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2c3d0b7b,SendAction) (Logging.scala:36) 10:46:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:55 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=68,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:55 [DEBUG] Metadata.update - Updated cluster metadata version 35 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:55 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:55 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:55 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:55 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=69,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415890,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:55 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=69,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Live assigned replicas for partition [__consumer_offsets,2] are: [List(1)] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - [Partition state machine on Controller 1]: Initializing leader and isr for partition [__consumer_offsets,2] to (Leader:1,ISR:1,LeaderEpoch:0,ControllerEpoch:1) (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 69 to client consumer-1. (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468415890,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@70b3f585,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=69,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:195,requestQueueTime:0,localTime:194,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468416085, latencyMs=196, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=69,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:56 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:56 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=70,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416087,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=70,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@6afd2db0 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 70 to client consumer-1 (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416087,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@33e59b3c,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=70,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:56 [DEBUG] Metadata.update - Updated cluster metadata version 36 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:56 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=71,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416092,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=71,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] NetworkClient.doSend - Sending {controller_id=1,controller_epoch=1,partition_states=[{topic=__consumer_offsets,partition=49,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=38,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=27,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=16,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=8,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=19,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=2,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=13,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=24,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=46,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=35,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=5,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=43,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=32,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=21,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=10,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=37,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=48,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=18,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=40,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=29,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=7,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=45,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=34,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=23,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=26,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=15,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=4,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=42,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=20,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=31,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=9,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=12,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=1,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=17,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=28,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=6,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=39,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=44,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=47,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=36,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=3,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=14,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=25,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=30,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=41,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=22,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=33,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=11,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_leaders=[{id=1,host=ISI050.utenze.BANKIT.IT,port=9092}]} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=4,api_version=0,correlation_id=3,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=__consumer_offsets,partition=49,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=38,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=27,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=16,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=8,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=19,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=2,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=13,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=24,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=46,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=35,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=5,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=43,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=32,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=21,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=10,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=37,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=48,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=18,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=40,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=29,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=7,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=45,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=34,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=23,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=26,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=15,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=4,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=42,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=20,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=31,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=9,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=12,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=1,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=17,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=28,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=6,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=39,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=44,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=47,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=36,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=3,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=14,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=25,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=30,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=41,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=22,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=33,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=11,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_leaders=[{id=1,host=ISI050.utenze.BANKIT.IT,port=9092}]} (Logging.scala:36) 10:46:56 [INFO ] Logging$class.info - [Replica state machine on controller 1]: Invoking state change to OnlineReplica for replicas [Topic=__consumer_offsets,Partition=28,Replica=1],[Topic=__consumer_offsets,Partition=48,Replica=1],[Topic=__consumer_offsets,Partition=5,Replica=1],[Topic=__consumer_offsets,Partition=21,Replica=1],[Topic=__consumer_offsets,Partition=2,Replica=1],[Topic=__consumer_offsets,Partition=18,Replica=1],[Topic=__consumer_offsets,Partition=23,Replica=1],[Topic=__consumer_offsets,Partition=9,Replica=1],[Topic=__consumer_offsets,Partition=39,Replica=1],[Topic=__consumer_offsets,Partition=31,Replica=1],[Topic=__consumer_offsets,Partition=19,Replica=1],[Topic=__consumer_offsets,Partition=10,Replica=1],[Topic=__consumer_offsets,Partition=22,Replica=1],[Topic=__consumer_offsets,Partition=43,Replica=1],[Topic=__consumer_offsets,Partition=40,Replica=1],[Topic=__consumer_offsets,Partition=27,Replica=1],[Topic=__consumer_offsets,Partition=6,Replica=1],[Topic=__consumer_offsets,Partition=1,Replica=1],[Topic=__consumer_offsets,Partition=47,Replica=1],[Topic=__consumer_offsets,Partition=30,Replica=1],[Topic=__consumer_offsets,Partition=42,Replica=1],[Topic=__consumer_offsets,Partition=41,Replica=1],[Topic=__consumer_offsets,Partition=3,Replica=1],[Topic=__consumer_offsets,Partition=13,Replica=1],[Topic=__consumer_offsets,Partition=4,Replica=1],[Topic=__consumer_offsets,Partition=16,Replica=1],[Topic=__consumer_offsets,Partition=46,Replica=1],[Topic=__consumer_offsets,Partition=49,Replica=1],[Topic=__consumer_offsets,Partition=14,Replica=1],[Topic=__consumer_offsets,Partition=45,Replica=1],[Topic=__consumer_offsets,Partition=37,Replica=1],[Topic=__consumer_offsets,Partition=29,Replica=1],[Topic=__consumer_offsets,Partition=20,Replica=1],[Topic=__consumer_offsets,Partition=8,Replica=1],[Topic=__consumer_offsets,Partition=38,Replica=1],[Topic=__consumer_offsets,Partition=7,Replica=1],[Topic=__consumer_offsets,Partition=0,Replica=1],[Topic=__consumer_offsets,Partition=34,Replica=1],[Topic=__consumer_offsets,Partition=33,Replica=1],[Topic=__consumer_offsets,Partition=26,Replica=1],[Topic=__consumer_offsets,Partition=44,Replica=1],[Topic=__consumer_offsets,Partition=32,Replica=1],[Topic=__consumer_offsets,Partition=25,Replica=1],[Topic=__consumer_offsets,Partition=11,Replica=1],[Topic=__consumer_offsets,Partition=36,Replica=1],[Topic=__consumer_offsets,Partition=12,Replica=1],[Topic=__consumer_offsets,Partition=35,Replica=1],[Topic=__consumer_offsets,Partition=15,Replica=1],[Topic=__consumer_offsets,Partition=17,Replica=1],[Topic=__consumer_offsets,Partition=24,Replica=1] (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416132,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 71 to client consumer-1. (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=4,api_version=0,correlation_id=3,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=__consumer_offsets,partition=49,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=38,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=27,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=16,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=8,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=19,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=2,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=13,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=24,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=46,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=35,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=5,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=43,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=32,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=21,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=10,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=37,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=48,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=18,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=40,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=29,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=7,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=45,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=34,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=23,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=26,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=15,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=4,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=42,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=20,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=31,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=9,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=12,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=1,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=17,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=28,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=6,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=39,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=44,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=47,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=36,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=3,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=14,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=25,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=30,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=41,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=22,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=33,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=11,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_leaders=[{id=1,host=ISI050.utenze.BANKIT.IT,port=9092}]} from connection 10.36.240.33:9092-10.36.240.33:53381;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416092,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2ac97a32,SendAction) (Logging.scala:36) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468416148, latencyMs=57, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=71,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:56 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=71,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:56,requestQueueTime:0,localTime:54,remoteTime:0,responseQueueTime:0,sendTime:2,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] Logging$class.trace - [Controller 1]: checking need to trigger partition rebalance (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - [Controller 1]: preferred replicas by broker Map(1 -> Map([__consumer_offsets,19] -> List(1), [__consumer_offsets,30] -> List(1), [__consumer_offsets,47] -> List(1), [__consumer_offsets,29] -> List(1), [__consumer_offsets,41] -> List(1), [__consumer_offsets,39] -> List(1), [__consumer_offsets,10] -> List(1), [__consumer_offsets,17] -> List(1), [__consumer_offsets,14] -> List(1), [__consumer_offsets,40] -> List(1), [__consumer_offsets,18] -> List(1), [__consumer_offsets,0] -> List(1), [__consumer_offsets,26] -> List(1), [__consumer_offsets,24] -> List(1), [__consumer_offsets,33] -> List(1), [__consumer_offsets,20] -> List(1), [__consumer_offsets,21] -> List(1), [__consumer_offsets,3] -> List(1), [__consumer_offsets,5] -> List(1), [__consumer_offsets,22] -> List(1), [__consumer_offsets,12] -> List(1), [__consumer_offsets,8] -> List(1), [__consumer_offsets,23] -> List(1), [__consumer_offsets,15] -> List(1), [__consumer_offsets,48] -> List(1), [__consumer_offsets,11] -> List(1), [__consumer_offsets,13] -> List(1), [__consumer_offsets,49] -> List(1), [__consumer_offsets,6] -> List(1), [__consumer_offsets,28] -> List(1), [__consumer_offsets,4] -> List(1), [__consumer_offsets,37] -> List(1), [__consumer_offsets,31] -> List(1), [__consumer_offsets,44] -> List(1), [__consumer_offsets,42] -> List(1), [__consumer_offsets,34] -> List(1), [__consumer_offsets,46] -> List(1), [__consumer_offsets,25] -> List(1), [__consumer_offsets,45] -> List(1), [__consumer_offsets,27] -> List(1), [testOutputTopic,0] -> List(1), [__consumer_offsets,32] -> List(1), [__consumer_offsets,43] -> List(1), [__consumer_offsets,36] -> List(1), [__consumer_offsets,35] -> List(1), [__consumer_offsets,7] -> List(1), [__consumer_offsets,9] -> List(1), [__consumer_offsets,38] -> List(1), [__consumer_offsets,1] -> List(1), [__consumer_offsets,16] -> List(1), [__consumer_offsets,2] -> List(1))) (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - [Controller 1]: topics not in preferred replica Map() (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:56 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - [Controller 1]: leader imbalance ratio for broker 1 is 0,000000 (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Completed execution of scheduled task 'partition-rebalance-thread'. (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=72,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416194,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=72,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@58a8e957 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 72 to client consumer-1 (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416194,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6e6bdac3,SendAction) (Logging.scala:36) 10:46:56 [INFO ] Logging$class.info - [ReplicaFetcherManager on broker 1] Removed fetcher for partitions __consumer_offsets-22,__consumer_offsets-30,__consumer_offsets-8,__consumer_offsets-21,__consumer_offsets-4,__consumer_offsets-27,__consumer_offsets-7,__consumer_offsets-9,__consumer_offsets-46,__consumer_offsets-25,__consumer_offsets-35,__consumer_offsets-41,__consumer_offsets-33,__consumer_offsets-23,__consumer_offsets-49,__consumer_offsets-47,__consumer_offsets-16,__consumer_offsets-28,__consumer_offsets-31,__consumer_offsets-36,__consumer_offsets-42,__consumer_offsets-3,__consumer_offsets-18,__consumer_offsets-37,__consumer_offsets-15,__consumer_offsets-24,__consumer_offsets-38,__consumer_offsets-17,__consumer_offsets-48,__consumer_offsets-19,__consumer_offsets-11,__consumer_offsets-13,__consumer_offsets-2,__consumer_offsets-43,__consumer_offsets-6,__consumer_offsets-14,__consumer_offsets-20,__consumer_offsets-0,__consumer_offsets-44,__consumer_offsets-39,__consumer_offsets-12,__consumer_offsets-45,__consumer_offsets-1,__consumer_offsets-5,__consumer_offsets-26,__consumer_offsets-29,__consumer_offsets-34,__consumer_offsets-10,__consumer_offsets-32,__consumer_offsets-40 (Logging.scala:70) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=72,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:56 [DEBUG] Metadata.update - Updated cluster metadata version 37 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:56 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=73,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416200,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=73,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 73 to client consumer-1. (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416200,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@11b1556,SendAction) (Logging.scala:36) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468416212, latencyMs=13, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=73,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:56 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=73,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:12,requestQueueTime:1,localTime:10,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-0\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-0 with 1 log segments and log end offset 0 in 16 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,0] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,0] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-0 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-0 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,0] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,0]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-29\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-29 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,29] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,29] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-29 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-29 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,29] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,29]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-48\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-48 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,48] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,48] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-48 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-48 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,48] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,48]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-10\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-10 with 1 log segments and log end offset 0 in 10 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,10] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,10] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-10 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-10 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,10] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,10]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-45\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-45 with 1 log segments and log end offset 0 in 7 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,45] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,45] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-45 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-45 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,45] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,45]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:56 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=74,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416300,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=74,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@2ae939a2 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 74 to client consumer-1 (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416300,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1eb85232,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=74,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:56 [DEBUG] Metadata.update - Updated cluster metadata version 38 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:56 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-26\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=75,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416309,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=75,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-26 with 1 log segments and log end offset 0 in 11 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,26] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,26] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-26 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-26 of length 0 bytes (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 75 to client consumer-1. (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,26] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,26]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416309,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7f41b33d,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=75,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:14,requestQueueTime:2,localTime:10,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468416324, latencyMs=17, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=75,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:56 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-7\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-7 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,7] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,7] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-7 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-7 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,7] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,7]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-42\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-42 with 1 log segments and log end offset 0 in 6 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,42] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,42] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-42 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-42 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,42] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,42]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-4\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-4 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,4] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,4] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-4 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-4 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,4] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,4]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-23\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-23 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,23] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,23] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-23 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-23 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,23] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,23]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-1\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-1 with 1 log segments and log end offset 0 in 7 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,1] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,1] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-1 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-1 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,1] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,1]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:56 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=76,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416407,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=76,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@6929929a and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 76 to client consumer-1 (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416407,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4fccc3d4,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=76,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:56 [DEBUG] Metadata.update - Updated cluster metadata version 39 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-20\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=77,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416414,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=77,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-20 with 1 log segments and log end offset 0 in 7 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,20] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,20] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-20 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-20 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,20] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,20]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-39\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 77 to client consumer-1. (Logging.scala:36) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-39 with 1 log segments and log end offset 0 in 8 ms (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416414,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@16a4dede,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=77,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:24,requestQueueTime:1,localTime:22,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468416439, latencyMs=27, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=77,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:56 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,39] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,39] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-39 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-39 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,39] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,39]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-17\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-17 with 1 log segments and log end offset 0 in 7 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,17] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,17] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-17 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-17 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,17] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,17]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-36\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-36 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,36] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,36] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-36 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-36 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,36] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,36]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-14\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-14 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,14] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,14] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-14 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-14 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,14] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,14]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-33\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-33 with 1 log segments and log end offset 0 in 6 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,33] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,33] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-33 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-33 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,33] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,33]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:56 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=78,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416515,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=78,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@62cc3848 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 78 to client consumer-1 (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416515,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4ad127fa,SendAction) (Logging.scala:36) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=78,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [DEBUG] Metadata.update - Updated cluster metadata version 40 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:56 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=79,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416521,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-49\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=79,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-49 with 1 log segments and log end offset 0 in 6 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,49] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,49] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-49 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-49 of length 0 bytes (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 79 to client consumer-1. (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,49] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,49]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416521,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@387bbdc6,SendAction) (Logging.scala:36) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468416533, latencyMs=13, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=79,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:56 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=79,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:12,requestQueueTime:1,localTime:10,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-11\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-11 with 1 log segments and log end offset 0 in 7 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,11] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,11] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-11 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-11 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,11] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,11]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-30\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-30 with 1 log segments and log end offset 0 in 9 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,30] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,30] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-30 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-30 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,30] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,30]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-46\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-46 with 1 log segments and log end offset 0 in 28 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,46] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,46] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-46 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-46 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,46] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,46]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-27\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-27 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,27] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,27] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-27 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-27 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,27] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,27]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:56 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=80,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416619,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=80,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@44496dd and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 80 to client consumer-1 (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-8\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416619,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@11801b7e,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=80,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:5,requestQueueTime:1,localTime:3,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:56 [DEBUG] Metadata.update - Updated cluster metadata version 41 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:56 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-8 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=81,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,8] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,8] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-8 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-8 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,8] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,8]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416627,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=81,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-24\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 81 to client consumer-1. (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416627,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@216275be,SendAction) (Logging.scala:36) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=81,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:14,requestQueueTime:3,localTime:10,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468416641, latencyMs=15, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=81,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:56 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-24 with 1 log segments and log end offset 0 in 8 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,24] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,24] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-24 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-24 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,24] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,24]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-43\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-43 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,43] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,43] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-43 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-43 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,43] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,43]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-5\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-5 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,5] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,5] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-5 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-5 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,5] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,5]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-21\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-21 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,21] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,21] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-21 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-21 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,21] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,21]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-2\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-2 with 1 log segments and log end offset 0 in 8 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,2] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,2] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-2 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-2 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,2] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,2]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:56 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=82,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416788,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=82,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@6ee7d44f and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 82 to client consumer-1 (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416788,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@62cf689f,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=82,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:56 [DEBUG] Metadata.update - Updated cluster metadata version 42 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:56 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=83,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416793,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=83,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-40\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-40 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,40] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,40] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-40 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-40 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,40] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,40]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 83 to client consumer-1. (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416793,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@73f7c8d,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=83,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:17,requestQueueTime:0,localTime:17,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468416811, latencyMs=20, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=83,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:56 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-37\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-37 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,37] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,37] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-37 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-37 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,37] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,37]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-18\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-18 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,18] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,18] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-18 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-18 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,18] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,18]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-34\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-34 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,34] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,34] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-34 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-34 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,34] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,34]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-15\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-15 with 1 log segments and log end offset 0 in 3 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,15] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,15] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-15 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-15 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,15] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,15]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-12\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-12 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,12] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,12] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-12 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-12 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,12] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,12]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-31\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-31 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,31] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,31] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-31 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-31 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,31] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,31]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-9\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-9 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,9] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,9] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-9 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-9 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,9] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,9]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-47\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-47 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,47] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,47] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-47 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-47 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,47] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,47]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:56 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=84,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416892,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=84,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@3a0365ff and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 84 to client consumer-1 (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416892,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@16f5ccff,SendAction) (Logging.scala:36) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=84,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:6,requestQueueTime:1,localTime:4,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:56 [DEBUG] Metadata.update - Updated cluster metadata version 43 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-19\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:56 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=85,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416901,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=85,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 85 to client consumer-1. (Logging.scala:36) 10:46:56 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416901,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2408941a,SendAction) (Logging.scala:36) 10:46:56 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:56 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=85,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:8,requestQueueTime:0,localTime:7,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468416909, latencyMs=9, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=85,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:56 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:56 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-19 with 1 log segments and log end offset 0 in 19 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,19] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,19] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-19 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-19 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,19] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,19]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-28\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-28 with 1 log segments and log end offset 0 in 12 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,28] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,28] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-28 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-28 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,28] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,28]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-38\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-38 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,38] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,38] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-38 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-38 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,38] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,38]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-35\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-35 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,35] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,35] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-35 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-35 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,35] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,35]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-44\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-44 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,44] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,44] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-44 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-44 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,44] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,44]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-6\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-6 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,6] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,6] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-6 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-6 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,6] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,6]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-25\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-25 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,25] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,25] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-25 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-25 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,25] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,25]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-16\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:56 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-16 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,16] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:56 [INFO ] Logging$class.info - Partition [__consumer_offsets,16] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-16 (Logging.scala:70) 10:46:56 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-16 of length 0 bytes (Logging.scala:36) 10:46:56 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,16] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,16]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:56 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:56 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:56 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=86,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417000,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=86,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@2f62eff1 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 86 to client consumer-1 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417000,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@3d9a9b94,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=86,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [DEBUG] Metadata.update - Updated cluster metadata version 44 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:57 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=87,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417004,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=87,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 87 to client consumer-1. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417004,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@41bd43be,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=87,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:5,requestQueueTime:0,localTime:5,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468417009, latencyMs=6, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=87,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:57 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:57 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:57 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-22\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:57 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-22 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,22] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Partition [__consumer_offsets,22] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-22 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-22 of length 0 bytes (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,22] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,22]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-41\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:57 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-41 with 1 log segments and log end offset 0 in 6 ms (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,41] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Partition [__consumer_offsets,41] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-41 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-41 of length 0 bytes (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,41] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,41]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-32\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:57 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-32 with 1 log segments and log end offset 0 in 4 ms (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,32] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Partition [__consumer_offsets,32] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-32 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-32 of length 0 bytes (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,32] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,32]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-3\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:57 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-3 with 1 log segments and log end offset 0 in 5 ms (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,3] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Partition [__consumer_offsets,3] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-3 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-3 of length 0 bytes (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,3] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,3]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Loaded index file D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka\__consumer_offsets-13\00000000000000000000.index with maxEntries = 1310720, maxIndexSize = 10485760, entries = 0, lastOffset = 0, file position = 0 (Logging.scala:54) 10:46:57 [INFO ] Logging$class.info - Completed load of log __consumer_offsets-13 with 1 log segments and log end offset 0 in 6 ms (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Created log for partition [__consumer_offsets,13] in D:\Apps\NetBeans\workspace\SettlementEngine\.\target\tmp\kafka with properties {compression.type -> producer, message.format.version -> 0.10.2-IV0, file.delete.delay.ms -> 60000, max.message.bytes -> 1000012, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, min.insync.replicas -> 1, segment.jitter.ms -> 0, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, cleanup.policy -> compact, flush.ms -> 9223372036854775807, segment.ms -> 604800000, segment.bytes -> 104857600, retention.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segment.index.bytes -> 10485760, flush.messages -> 9223372036854775807}. (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - Partition [__consumer_offsets,13] on broker 1: No checkpointed highwatermark is found for partition __consumer_offsets-13 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Reading 1 bytes from offset 0 in log __consumer_offsets-13 of length 0 bytes (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,13] on broker 1: Skipping update high watermark since Old hw 0 [0 : 0] is larger than new hw 0 [0 : 0] for partition [__consumer_offsets,13]. All leo's are 0 [0 : 0] (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-22 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-25 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-22'. (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-28 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-31 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-34 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-37 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-40 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-43 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-46 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-49 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-41 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-44 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-47 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-1 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-4 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-7 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-10 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-13 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-16 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-19 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-2 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-5 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-8 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-11 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-14 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-17 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-20 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-23 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-26 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-29 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-32 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-35 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-38 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-0 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-3 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-6 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-9 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-12 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-15 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-18 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-21 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-24 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-27 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-30 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-33 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-36 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-39 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-42 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-45 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - Scheduling task __consumer_offsets-48 with initial delay 0 ms and period -1 ms. (Logging.scala:54) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-22 (Logging.scala:70) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:57 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:57 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=88,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417105,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=88,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@513d8a39 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 88 to client consumer-1 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417105,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7917ed42,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:57 [DEBUG] Metadata.update - Updated cluster metadata version 45 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:57 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=88,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=89,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468416132,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6675750,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 4, received {error_code=0,partitions=[{topic=__consumer_offsets,partition=49,error_code=0},{topic=__consumer_offsets,partition=38,error_code=0},{topic=__consumer_offsets,partition=27,error_code=0},{topic=__consumer_offsets,partition=16,error_code=0},{topic=__consumer_offsets,partition=8,error_code=0},{topic=__consumer_offsets,partition=19,error_code=0},{topic=__consumer_offsets,partition=13,error_code=0},{topic=__consumer_offsets,partition=2,error_code=0},{topic=__consumer_offsets,partition=24,error_code=0},{topic=__consumer_offsets,partition=46,error_code=0},{topic=__consumer_offsets,partition=35,error_code=0},{topic=__consumer_offsets,partition=5,error_code=0},{topic=__consumer_offsets,partition=43,error_code=0},{topic=__consumer_offsets,partition=21,error_code=0},{topic=__consumer_offsets,partition=32,error_code=0},{topic=__consumer_offsets,partition=10,error_code=0},{topic=__consumer_offsets,partition=37,error_code=0},{topic=__consumer_offsets,partition=48,error_code=0},{topic=__consumer_offsets,partition=29,error_code=0},{topic=__consumer_offsets,partition=40,error_code=0},{topic=__consumer_offsets,partition=18,error_code=0},{topic=__consumer_offsets,partition=7,error_code=0},{topic=__consumer_offsets,partition=45,error_code=0},{topic=__consumer_offsets,partition=23,error_code=0},{topic=__consumer_offsets,partition=34,error_code=0},{topic=__consumer_offsets,partition=26,error_code=0},{topic=__consumer_offsets,partition=4,error_code=0},{topic=__consumer_offsets,partition=15,error_code=0},{topic=__consumer_offsets,partition=42,error_code=0},{topic=__consumer_offsets,partition=20,error_code=0},{topic=__consumer_offsets,partition=31,error_code=0},{topic=__consumer_offsets,partition=9,error_code=0},{topic=__consumer_offsets,partition=1,error_code=0},{topic=__consumer_offsets,partition=12,error_code=0},{topic=__consumer_offsets,partition=17,error_code=0},{topic=__consumer_offsets,partition=28,error_code=0},{topic=__consumer_offsets,partition=6,error_code=0},{topic=__consumer_offsets,partition=39,error_code=0},{topic=__consumer_offsets,partition=44,error_code=0},{topic=__consumer_offsets,partition=36,error_code=0},{topic=__consumer_offsets,partition=47,error_code=0},{topic=__consumer_offsets,partition=14,error_code=0},{topic=__consumer_offsets,partition=25,error_code=0},{topic=__consumer_offsets,partition=3,error_code=0},{topic=__consumer_offsets,partition=30,error_code=0},{topic=__consumer_offsets,partition=41,error_code=0},{topic=__consumer_offsets,partition=11,error_code=0},{topic=__consumer_offsets,partition=33,error_code=0},{topic=__consumer_offsets,partition=22,error_code=0},{topic=__consumer_offsets,partition=0,error_code=0}]} (NetworkClient.java:534) 10:46:57 [TRACE] NetworkClient.doSend - Sending {controller_id=1,controller_epoch=1,partition_states=[{topic=__consumer_offsets,partition=49,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=38,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=27,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=16,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=8,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=19,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=2,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=13,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=24,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=46,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=35,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=5,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=43,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=32,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=21,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=10,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=37,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=48,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=18,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=40,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=29,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=7,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=45,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=34,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=23,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=26,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=15,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=4,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=42,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=20,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=31,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=9,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=12,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=1,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=17,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=28,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=6,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=39,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=44,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=47,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=36,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=3,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=14,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=25,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=30,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=41,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=22,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=33,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=11,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=4,api_version=0,correlation_id=3,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=__consumer_offsets,partition=49,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=38,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=27,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=16,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=8,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=19,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=2,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=13,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=24,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=46,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=35,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=5,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=43,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=32,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=21,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=10,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=37,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=48,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=18,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=40,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=29,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=7,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=45,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=34,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=23,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=26,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=15,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=4,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=42,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=20,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=31,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=9,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=12,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=1,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=17,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=28,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=6,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=39,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=44,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=47,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=36,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=3,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=14,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=25,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=30,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=41,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=22,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=33,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=11,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_leaders=[{id=1,host=ISI050.utenze.BANKIT.IT,port=9092}]} from connection 10.36.240.33:9092-10.36.240.33:53381;totalTime:980,requestQueueTime:10,localTime:956,remoteTime:0,responseQueueTime:13,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=6,api_version=3,correlation_id=4,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=__consumer_offsets,partition=49,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=38,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=27,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=16,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=8,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=19,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=2,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=13,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=24,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=46,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=35,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=5,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=43,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=32,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=21,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=10,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=37,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=48,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=18,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=40,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=29,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=7,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=45,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=34,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=23,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=26,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=15,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=4,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=42,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=20,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=31,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=9,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=12,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=1,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=17,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=28,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=6,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=39,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=44,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=47,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=36,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=3,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=14,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=25,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=30,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=41,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=22,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=33,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=11,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417111,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=89,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417136,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=6,api_version=3,correlation_id=4,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=__consumer_offsets,partition=49,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=38,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=27,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=16,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=8,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=19,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=2,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=13,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=24,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=46,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=35,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=5,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=43,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=32,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=21,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=10,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=37,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=48,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=18,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=40,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=29,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=7,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=45,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=34,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=23,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=26,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=15,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=4,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=42,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=20,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=31,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=9,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=12,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=1,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=17,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=28,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=6,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=39,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=44,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=47,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=36,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=3,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=14,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=25,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=30,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=41,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=22,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=33,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=11,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} from connection 10.36.240.33:9092-10.36.240.33:53381;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,10.36.240.33:9092-10.36.240.33:53381,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417136,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4ff1d5ec,SendAction) (Logging.scala:36) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=6,api_version=3,correlation_id=4,client_id=1} -- {controller_id=1,controller_epoch=1,partition_states=[{topic=__consumer_offsets,partition=49,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=38,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=27,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=16,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=8,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=19,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=2,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=13,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=24,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=46,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=35,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=5,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=43,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=32,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=21,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=10,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=37,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=48,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=18,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=40,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=29,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=7,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=45,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=34,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=23,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=26,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=15,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=4,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=42,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=20,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=31,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=9,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=12,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=1,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=17,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=28,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=6,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=39,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=44,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=47,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=36,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=3,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=14,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=25,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=30,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=41,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=22,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=33,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=11,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]},{topic=__consumer_offsets,partition=0,controller_epoch=1,leader=1,leader_epoch=0,isr=[1],zk_version=0,replicas=[1]}],live_brokers=[{id=1,end_points=[{port=9092,host=ISI050.utenze.BANKIT.IT,listener_name=PLAINTEXT,security_protocol_type=0}],rack=null}]} from connection 10.36.240.33:9092-10.36.240.33:53381;totalTime:15,requestQueueTime:4,localTime:9,remoteTime:0,responseQueueTime:0,sendTime:2,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-22 in 56 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-22'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-25'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-25 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-25 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-25'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-28'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-28 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-28 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-28'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-31'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-31 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-31 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-31'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-34'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-34 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-34 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-34'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-37'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-37 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-37 in 2 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 6, received {error_code=0} (NetworkClient.java:534) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-37'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-40'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-40 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-40 in 2 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-40'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-43'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-43 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-43 in 2 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-43'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-46'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-46 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-46 in 5 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-46'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-49'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-49 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-49 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-49'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-41'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-41 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-41 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-41'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-44'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-44 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-44 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-44'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-47'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-47 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-47 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-47'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-1'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-1 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-1 in 2 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-1'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-4'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-4 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-4 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-4'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-7'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-7 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-7 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-7'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-10'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-10 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-10 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-10'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-13'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-13 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-13 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-13'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-16'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-16 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-16 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-16'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-19'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-19 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-19 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-19'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-2'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-2 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-2 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-2'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-5'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-5 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-5 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-5'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-8'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-8 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=15,coordinator={node_id=-1,host=,port=-1}} for correlation id 89 to client consumer-1. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-8 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417111,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@195f0e3e,SendAction) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-8'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-11'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-11 (Logging.scala:70) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=89,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:103,requestQueueTime:29,localTime:73,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-11 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-11'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-14'. (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=15,coordinator={node_id=-1,host=,port=-1}} (NetworkClient.java:534) 10:46:57 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468417216, latencyMs=107, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=89,client_id=consumer-1}, responseBody={error_code=15,coordinator={node_id=-1,host=,port=-1}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:57 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Group coordinator lookup for group testOutputTopic failed: The group coordinator is not available. (AbstractCoordinator.java:581) 10:46:57 [DEBUG] AbstractCoordinator.ensureCoordinatorReady - Coordinator discovery failed for group testOutputTopic, refreshing metadata (AbstractCoordinator.java:215) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:57 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node 1 (NetworkClient.java:751) 10:46:57 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node 1. (NetworkClient.java:316) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-14 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=3,api_version=2,correlation_id=90,client_id=consumer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-14 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-14'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-17'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-17 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417219,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=90,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@257344f3 and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 90 to client consumer-1 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417219,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@707060ba,SendAction) (Logging.scala:36) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=90,client_id=consumer-1} -- {topics=[testOutputTopic]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:8,requestQueueTime:4,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:2,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:57 [DEBUG] Metadata.update - Updated cluster metadata version 46 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:57 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (AbstractCoordinator.java:548) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=10,api_version=0,correlation_id=91,client_id=consumer-1} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417230,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=91,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-17 in 7 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=0,coordinator={node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092}} for correlation id 91 to client consumer-1. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417230,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@5b7b6db9,SendAction) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-17'. (Logging.scala:36) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=91,client_id=consumer-1} -- {group_id=testOutputTopic} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:5,requestQueueTime:1,localTime:4,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 10, received {error_code=0,coordinator={node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092}} (NetworkClient.java:534) 10:46:57 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468417236, latencyMs=7, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=91,client_id=consumer-1}, responseBody={error_code=0,coordinator={node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:57 [INFO ] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Discovered coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) for group testOutputTopic. (AbstractCoordinator.java:573) 10:46:57 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node 2147483646 at ISI050.utenze.BANKIT.IT:9092. (NetworkClient.java:627) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-20'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-20 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-20 in 3 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-20'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-23'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-23 (Logging.scala:70) 10:46:57 [DEBUG] Logging$class.debug - Accepted connection from /10.36.240.33:53383 on /10.36.240.33:9092 and assigned it to processor 0, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-23 in 4 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-23'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-26'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-26 (Logging.scala:70) 10:46:57 [DEBUG] Logging$class.debug - Processor 0 listening to new connection from /10.36.240.33:53383 (Logging.scala:54) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-26 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-26'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-29'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-29 (Logging.scala:70) 10:46:57 [INFO ] ConsumerCoordinator.onJoinPrepare - Revoking previously assigned partitions [] for group testOutputTopic (ConsumerCoordinator.java:393) 10:46:57 [TRACE] AbstractCoordinator$HeartbeatThread.disable - Disabling heartbeat thread for group testOutputTopic (AbstractCoordinator.java:846) 10:46:57 [INFO ] AbstractCoordinator.sendJoinGroupRequest - (Re-)joining group testOutputTopic (AbstractCoordinator.java:407) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-29 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-29'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-32'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-32 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-32 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-32'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-35'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-35 (Logging.scala:70) 10:46:57 [DEBUG] AbstractCoordinator$HeartbeatThread.run - Heartbeat thread for group testOutputTopic started (AbstractCoordinator.java:869) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-35 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-35'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-38'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-38 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-38 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-38'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-0'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-0 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-0 in 0 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-0'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-3'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-3 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-3 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-3'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-6'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-6 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-6 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-6'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-9'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-9 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-9 in 2 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-9'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-12'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-12 (Logging.scala:70) 10:46:57 [DEBUG] AbstractCoordinator.sendJoinGroupRequest - Sending JoinGroup ((type: JoinGroupRequest, groupId=testOutputTopic, sessionTimeout=10000, rebalanceTimeout=300000, memberId=, protocolType=consumer, groupProtocols=org.apache.kafka.common.requests.JoinGroupRequest$ProtocolMetadata@22b53226)) to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:415) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-12 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-12'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-15'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-15 (Logging.scala:70) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node-2147483646.bytes-sent (Metrics.java:335) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-15 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-15'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-18'. (Logging.scala:36) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node-2147483646.bytes-received (Metrics.java:335) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-18 (Logging.scala:70) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node-2147483646.latency (Metrics.java:335) 10:46:57 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 65536, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node 2147483646 (Selector.java:339) 10:46:57 [DEBUG] NetworkClient.handleConnections - Completed connection to node 2147483646. Fetching API versions. (NetworkClient.java:590) 10:46:57 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node 2147483646. (NetworkClient.java:603) 10:46:57 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node 2147483646. Assuming version 0. (NetworkClient.java:292) 10:46:57 [TRACE] NetworkClient.doSend - Sending {} to node 2147483646. (NetworkClient.java:316) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-18 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-18'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-21'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-21 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=18,api_version=0,correlation_id=93,client_id=consumer-1} -- {} (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-21 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-21'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-24'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-24 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-24 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-24'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-27'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-27 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-27 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-27'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-30'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-30 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417272,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-30 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-30'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-33'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-33 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=93,client_id=consumer-1} -- {} from connection 10.36.240.33:9092-10.36.240.33:53383;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-33 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-33'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-36'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417272,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4fc24d65,SendAction) (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-36 (Logging.scala:70) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=93,client_id=consumer-1} -- {} from connection 10.36.240.33:9092-10.36.240.33:53383;totalTime:9,requestQueueTime:4,localTime:4,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-36 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-36'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-39'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-39 (Logging.scala:70) 10:46:57 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node 2147483646: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} to node 2147483646. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=11,api_version=1,correlation_id=92,client_id=consumer-1} -- {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-39 in 3 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417286,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-39'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=11,api_version=1,correlation_id=92,client_id=consumer-1} -- {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} from connection 10.36.240.33:9092-10.36.240.33:53383;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-42'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-42 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-42 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-42'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-45'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-45 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-45 in 20 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-45'. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task '__consumer_offsets-48'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Loading offsets and group metadata from __consumer_offsets-48 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [Group Metadata Manager on Broker 1]: Finished loading offsets from __consumer_offsets-48 in 1 milliseconds. (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task '__consumer_offsets-48'. (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [GroupCoordinator 1]: Preparing to restabilize group testOutputTopic with old generation 0 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [GroupCoordinator 1]: Stabilized group testOutputTopic generation 1 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending join group response {error_code=0,generation_id=1,group_protocol=range,leader_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,members=[{member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} for correlation id 92 to client consumer-1. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417286,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@53fc49a3,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 11, received {error_code=0,generation_id=1,group_protocol=range,leader_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,members=[{member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=11,api_version=1,correlation_id=92,client_id=consumer-1} -- {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} from connection 10.36.240.33:9092-10.36.240.33:53383;totalTime:66,requestQueueTime:1,localTime:64,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [DEBUG] AbstractCoordinator$JoinGroupResponseHandler.handle - Received successful JoinGroup response for group testOutputTopic: {error_code=0,generation_id=1,group_protocol=range,leader_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,members=[{member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} (AbstractCoordinator.java:425) 10:46:57 [DEBUG] ConsumerCoordinator.performAssignment - Performing assignment for group testOutputTopic using strategy range with subscriptions {consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226=Subscription(topics=[testOutputTopic])} (ConsumerCoordinator.java:336) 10:46:57 [DEBUG] ConsumerCoordinator.performAssignment - Finished assignment for group testOutputTopic: {consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226=Assignment(partitions=[testOutputTopic-0])} (ConsumerCoordinator.java:375) 10:46:57 [DEBUG] AbstractCoordinator.onJoinLeader - Sending leader SyncGroup for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null): (type=SyncGroupRequest, groupId=testOutputTopic, generationId=1, memberId=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226, groupAssignment=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226) (AbstractCoordinator.java:493) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,generation_id=1,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,group_assignment=[{member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]}]} to node 2147483646. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=14,api_version=0,correlation_id=94,client_id=consumer-1} -- {group_id=testOutputTopic,generation_id=1,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,group_assignment=[{member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417356,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=14,api_version=0,correlation_id=94,client_id=consumer-1} -- {group_id=testOutputTopic,generation_id=1,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,group_assignment=[{member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]}]} from connection 10.36.240.33:9092-10.36.240.33:53383;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [GroupCoordinator 1]: Assignment received from leader for group testOutputTopic for generation 1 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(__consumer_offsets-40 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 204240202, CreateTime = 1491468417369, key = 19 bytes, value = 230 bytes))])] to local log (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Inserting 283 bytes at offset 0 at position 0 with largest timestamp 1491468417369 at shallow offset 0 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Appended 283 to .\target\tmp\kafka\__consumer_offsets-40\00000000000000000000.log at offset 0 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Appended message set to log __consumer_offsets-40 with first offset: 0, next offset: 1, and messages: [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 204240202, CreateTime = 1491468417369, key = 19 bytes, value = 230 bytes))] (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 fetch requests. (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition __consumer_offsets-40 to [1 [0 : 283]] (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,40] on broker 1: High watermark for partition [__consumer_offsets,40] updated to 1 [0 : 283] (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 fetch requests. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 producer requests. (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 283 bytes written to log __consumer_offsets-40 beginning at offset 0 and ending at offset 0 (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 54 ms (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - Initial partition status for __consumer_offsets-40 is [acksPending: true, error: 7, startOffset: 0, requiredOffset: 1] (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Checking produce satisfaction for __consumer_offsets-40, current status [acksPending: true, error: 7, startOffset: 0, requiredOffset: 1] (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Partition [__consumer_offsets,40] on broker 1: 1 acks satisfied for __consumer_offsets-40 with acks = -1 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417356,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4105627c,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 14, received {error_code=0,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=14,api_version=0,correlation_id=94,client_id=consumer-1} -- {group_id=testOutputTopic,generation_id=1,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,group_assignment=[{member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]}]} from connection 10.36.240.33:9092-10.36.240.33:53383;totalTime:104,requestQueueTime:0,localTime:103,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [INFO ] AbstractCoordinator$1.onSuccess - Successfully joined group testOutputTopic with generation 1 (AbstractCoordinator.java:375) 10:46:57 [TRACE] AbstractCoordinator$HeartbeatThread.enable - Enabling heartbeat thread for group testOutputTopic (AbstractCoordinator.java:837) 10:46:57 [INFO ] ConsumerCoordinator.onJoinComplete - Setting newly assigned partitions [testOutputTopic-0] for group testOutputTopic (ConsumerCoordinator.java:252) 10:46:57 [DEBUG] ConsumerCoordinator.sendOffsetFetchRequest - Group testOutputTopic fetching committed offsets for partitions: [testOutputTopic-0] (ConsumerCoordinator.java:800) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,topics=[{topic=testOutputTopic,partitions=[{partition=0}]}]} to node 2147483646. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=9,api_version=2,correlation_id=95,client_id=consumer-1} -- {group_id=testOutputTopic,topics=[{topic=testOutputTopic,partitions=[{partition=0}]}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417465,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=9,api_version=2,correlation_id=95,client_id=consumer-1} -- {group_id=testOutputTopic,topics=[{topic=testOutputTopic,partitions=[{partition=0}]}]} from connection 10.36.240.33:9092-10.36.240.33:53383;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Group Metadata Manager on Broker 1]: Getting offsets of ArrayBuffer(testOutputTopic-0) for group testOutputTopic. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending offset fetch response {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,offset=-1,metadata=,error_code=0}]}],error_code=0} for correlation id 95 to client consumer-1. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417465,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@74f53e16,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 9, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,offset=-1,metadata=,error_code=0}]}],error_code=0} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=9,api_version=2,correlation_id=95,client_id=consumer-1} -- {group_id=testOutputTopic,topics=[{topic=testOutputTopic,partitions=[{partition=0}]}]} from connection 10.36.240.33:9092-10.36.240.33:53383;totalTime:6,requestQueueTime:1,localTime:5,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [DEBUG] ConsumerCoordinator$OffsetFetchResponseHandler.handle - Group testOutputTopic has no committed offset for partition testOutputTopic-0 (ConsumerCoordinator.java:851) 10:46:57 [DEBUG] Fetcher.resetOffset - Resetting offset for partition testOutputTopic-0 to latest offset. (Fetcher.java:362) 10:46:57 [TRACE] Fetcher.sendListOffsetRequest - Sending ListOffsetRequest (type=ListOffsetRequest, replicaId=-1, partitionTimestamps={testOutputTopic-0=-1}, minVersion=0) to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:614) 10:46:57 [TRACE] NetworkClient.doSend - Sending {replica_id=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,timestamp=-1}]}]} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=2,api_version=1,correlation_id=96,client_id=consumer-1} -- {replica_id=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,timestamp=-1}]}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417476,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=2,api_version=1,correlation_id=96,client_id=consumer-1} -- {replica_id=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,timestamp=-1}]}]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417476,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1dc0e669,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 2, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,timestamp=-1,offset=0}]}]} (NetworkClient.java:534) 10:46:57 [TRACE] Fetcher$3.onSuccess - Received ListOffsetResponse {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,timestamp=-1,offset=0}]}]} from broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:620) 10:46:57 [DEBUG] Fetcher.handleListOffsetResponse - Handling ListOffsetResponse response for testOutputTopic-0. Fetched offset 0, timestamp -1 (Fetcher.java:662) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=2,api_version=1,correlation_id=96,client_id=consumer-1} -- {replica_id=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,timestamp=-1}]}]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:5,requestQueueTime:0,localTime:4,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [TRACE] Fetcher.createFetchRequests - Added fetch request for partition testOutputTopic-0 at offset 0 to node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:726) 10:46:57 [DEBUG] Fetcher.sendFetches - Sending fetch for partitions [testOutputTopic-0] to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:180) 10:46:57 [TRACE] NetworkClient.doSend - Sending {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] KafkaConsumer.close - Closing the Kafka consumer. (KafkaConsumer.java:1553) 10:46:57 [DEBUG] AbstractCoordinator.maybeLeaveGroup - Sending LeaveGroup request to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) for group testOutputTopic (AbstractCoordinator.java:677) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226} to node 2147483646. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=13,api_version=0,correlation_id=98,client_id=consumer-1} -- {group_id=testOutputTopic,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417491,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=13,api_version=0,correlation_id=98,client_id=consumer-1} -- {group_id=testOutputTopic,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226} from connection 10.36.240.33:9092-10.36.240.33:53383;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=1,api_version=3,correlation_id=97,client_id=consumer-1} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417492,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=1,api_version=3,correlation_id=97,client_id=consumer-1} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53382;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [GroupCoordinator 1]: Member consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226 in group testOutputTopic has failed (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [GroupCoordinator 1]: Preparing to restabilize group testOutputTopic with old generation 1 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [GroupCoordinator 1]: Group testOutputTopic with generation 2 is now empty (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(__consumer_offsets-40 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1768798425, CreateTime = 1491468417506, key = 19 bytes, value = 24 bytes))])] to local log (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Inserting 77 bytes at offset 1 at position 283 with largest timestamp 1491468417506 at shallow offset 1 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Appended 77 to .\target\tmp\kafka\__consumer_offsets-40\00000000000000000000.log at offset 1 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Appended message set to log __consumer_offsets-40 with first offset: 1, next offset: 2, and messages: [(offset=1,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1768798425, CreateTime = 1491468417506, key = 19 bytes, value = 24 bytes))] (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 fetch requests. (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition __consumer_offsets-40 to [2 [0 : 360]] (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,40] on broker 1: High watermark for partition [__consumer_offsets,40] updated to 2 [0 : 360] (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 fetch requests. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 producer requests. (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 77 bytes written to log __consumer_offsets-40 beginning at offset 1 and ending at offset 1 (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 7 ms (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - Initial partition status for __consumer_offsets-40 is [acksPending: true, error: 7, startOffset: 1, requiredOffset: 2] (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Checking produce satisfaction for __consumer_offsets-40, current status [acksPending: true, error: 7, startOffset: 1, requiredOffset: 2] (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Partition [__consumer_offsets,40] on broker 1: 1 acks satisfied for __consumer_offsets-40 with acks = -1 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending leave group response {error_code=0} for correlation id 98 to client consumer-1. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Fetching log segment for partition testOutputTopic-0, offset 0, partition fetch size 1048576, remaining response limit 52428800, ignoring response/partition size limits (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Reading 1048576 bytes from offset 0 in log testOutputTopic-0 of length 0 bytes (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53383,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417491,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2901fc88,SendAction) (Logging.scala:36) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=13,api_version=0,correlation_id=98,client_id=consumer-1} -- {group_id=testOutputTopic,member_id=consumer-1-5e1269bc-12dd-42fa-a34d-4df43bdcd226} from connection 10.36.240.33:9092-10.36.240.33:53383;totalTime:30,requestQueueTime:1,localTime:26,remoteTime:0,responseQueueTime:3,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 13, received {error_code=0} (NetworkClient.java:534) 10:46:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:46:57 [DEBUG] AbstractCoordinator$LeaveGroupResponseHandler.handle - LeaveGroup request for group testOutputTopic returned successfully (AbstractCoordinator.java:693) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name connections-closed: (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name connections-created: (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name bytes-sent-received: (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name bytes-sent: (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name bytes-received: (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name select-time: (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name io-time: (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name node--1.bytes-sent (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name node--1.bytes-received (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name node--1.latency (Metrics.java:368) 10:46:57 [DEBUG] Selector.pollSelectionKeys - Connection with /127.0.0.1 disconnected (Selector.java:375) java.io.EOFException at org.apache.kafka.common.network.NetworkReceive.readFromReadableChannel(NetworkReceive.java:83) at org.apache.kafka.common.network.NetworkReceive.readFrom(NetworkReceive.java:71) at org.apache.kafka.common.network.KafkaChannel.receive(KafkaChannel.java:169) at org.apache.kafka.common.network.KafkaChannel.read(KafkaChannel.java:150) at org.apache.kafka.common.network.Selector.pollSelectionKeys(Selector.java:355) at org.apache.kafka.common.network.Selector.poll(Selector.java:303) at kafka.network.Processor.poll(SocketServer.scala:494) at kafka.network.Processor.run(SocketServer.scala:432) at java.lang.Thread.run(Thread.java:745) 10:46:57 [DEBUG] AbstractCoordinator$HeartbeatThread.run - Heartbeat thread for group testOutputTopic has closed (AbstractCoordinator.java:947) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name node-1.bytes-sent (Metrics.java:368) 10:46:57 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name node-1.bytes-received (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name node-1.latency (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name node-2147483646.bytes-sent (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name node-2147483646.bytes-received (Metrics.java:368) 10:46:57 [DEBUG] Metrics.removeSensor - Removed sensor with name node-2147483646.latency (Metrics.java:368) 10:46:57 [DEBUG] KafkaConsumer.close - The Kafka consumer has closed. (KafkaConsumer.java:1569) 10:46:57 [INFO ] AbstractConfig.logAll - ConsumerConfig values: auto.commit.interval.ms = 1000 auto.offset.reset = latest bootstrap.servers = [localhost:9092] check.crcs = true client.id = connections.max.idle.ms = 540000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = testOutputTopic heartbeat.interval.ms = 3000 interceptor.classes = null key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor] receive.buffer.bytes = 65536 reconnect.backoff.ms = 50 request.timeout.ms = 305000 retry.backoff.ms = 100 sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 session.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer (AbstractConfig.java:180) 10:46:57 [DEBUG] KafkaConsumer. - Starting the Kafka consumer (KafkaConsumer.java:607) 10:46:57 [INFO ] AbstractConfig.logAll - ConsumerConfig values: auto.commit.interval.ms = 1000 auto.offset.reset = latest bootstrap.servers = [localhost:9092] check.crcs = true client.id = consumer-2 connections.max.idle.ms = 540000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = testOutputTopic heartbeat.interval.ms = 3000 interceptor.classes = null key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor] receive.buffer.bytes = 65536 reconnect.backoff.ms = 50 request.timeout.ms = 305000 retry.backoff.ms = 100 sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 session.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer (AbstractConfig.java:180) 10:46:57 [DEBUG] Metadata.update - Updated cluster metadata version 1 to Cluster(id = null, nodes = [localhost:9092 (id: -1 rack: null)], partitions = []) (Metadata.java:244) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name connections-closed: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name connections-created: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent-received: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name bytes-received: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name select-time: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name io-time: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name heartbeat-latency (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name join-latency (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name sync-latency (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name commit-latency (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name bytes-fetched (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name records-fetched (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name fetch-latency (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name records-lag (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name fetch-throttle-time (Metrics.java:335) 10:46:57 [INFO ] AppInfoParser$AppInfo. - Kafka version : 0.10.2.0 (AppInfoParser.java:83) 10:46:57 [INFO ] AppInfoParser$AppInfo. - Kafka commitId : 576d93a8dc0cf421 (AppInfoParser.java:84) 10:46:57 [DEBUG] KafkaConsumer. - Kafka consumer created (KafkaConsumer.java:711) 10:46:57 [DEBUG] KafkaConsumer.subscribe - Subscribed to topic(s): testOutputTopic (KafkaConsumer.java:824) 10:46:57 [INFO ] AbstractConfig.logAll - ConsumerConfig values: auto.commit.interval.ms = 1000 auto.offset.reset = latest bootstrap.servers = [localhost:9092] check.crcs = true client.id = connections.max.idle.ms = 540000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = testOutputTopic heartbeat.interval.ms = 3000 interceptor.classes = null key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor] receive.buffer.bytes = 65536 reconnect.backoff.ms = 50 request.timeout.ms = 305000 retry.backoff.ms = 100 sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 session.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer (AbstractConfig.java:180) 10:46:57 [DEBUG] KafkaConsumer. - Starting the Kafka consumer (KafkaConsumer.java:607) 10:46:57 [INFO ] AbstractConfig.logAll - ConsumerConfig values: auto.commit.interval.ms = 1000 auto.offset.reset = latest bootstrap.servers = [localhost:9092] check.crcs = true client.id = consumer-3 connections.max.idle.ms = 540000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = testOutputTopic heartbeat.interval.ms = 3000 interceptor.classes = null key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor] receive.buffer.bytes = 65536 reconnect.backoff.ms = 50 request.timeout.ms = 305000 retry.backoff.ms = 100 sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 session.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer (AbstractConfig.java:180) 10:46:57 [DEBUG] Metadata.update - Updated cluster metadata version 1 to Cluster(id = null, nodes = [localhost:9092 (id: -1 rack: null)], partitions = []) (Metadata.java:244) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name connections-closed: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name connections-created: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent-received: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name bytes-received: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name select-time: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name io-time: (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name heartbeat-latency (Metrics.java:335) 10:46:57 [DEBUG] Selector.pollSelectionKeys - Connection with /10.36.240.33 disconnected (Selector.java:375) java.io.EOFException at org.apache.kafka.common.network.NetworkReceive.readFromReadableChannel(NetworkReceive.java:83) at org.apache.kafka.common.network.NetworkReceive.readFrom(NetworkReceive.java:71) at org.apache.kafka.common.network.KafkaChannel.receive(KafkaChannel.java:169) at org.apache.kafka.common.network.KafkaChannel.read(KafkaChannel.java:150) at org.apache.kafka.common.network.Selector.pollSelectionKeys(Selector.java:355) at org.apache.kafka.common.network.Selector.poll(Selector.java:303) at kafka.network.Processor.poll(SocketServer.scala:494) at kafka.network.Processor.run(SocketServer.scala:432) at java.lang.Thread.run(Thread.java:745) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name join-latency (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name sync-latency (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name commit-latency (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name bytes-fetched (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name records-fetched (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name fetch-latency (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name records-lag (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name fetch-throttle-time (Metrics.java:335) 10:46:57 [INFO ] AppInfoParser$AppInfo. - Kafka version : 0.10.2.0 (AppInfoParser.java:83) 10:46:57 [INFO ] AppInfoParser$AppInfo. - Kafka commitId : 576d93a8dc0cf421 (AppInfoParser.java:84) 10:46:57 [DEBUG] KafkaConsumer. - Kafka consumer created (KafkaConsumer.java:711) 10:46:57 [DEBUG] KafkaConsumer.subscribe - Subscribed to topic(s): testOutputTopic (KafkaConsumer.java:824) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:57 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker localhost:9092 (id: -1 rack: null) (AbstractCoordinator.java:548) 10:46:57 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node -1 at localhost:9092. (NetworkClient.java:627) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node--1.bytes-sent (Metrics.java:335) 10:46:57 [DEBUG] Logging$class.debug - Accepted connection from /127.0.0.1:53389 on /127.0.0.1:9092 and assigned it to processor 1, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node--1.bytes-received (Metrics.java:335) 10:46:57 [DEBUG] Logging$class.debug - Processor 1 listening to new connection from /127.0.0.1:53389 (Logging.scala:54) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node--1.latency (Metrics.java:335) 10:46:57 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 65536, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node -1 (Selector.java:339) 10:46:57 [DEBUG] NetworkClient.handleConnections - Completed connection to node -1. Fetching API versions. (NetworkClient.java:590) 10:46:57 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node -1. (NetworkClient.java:603) 10:46:57 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node -1. Assuming version 0. (NetworkClient.java:292) 10:46:57 [TRACE] NetworkClient.doSend - Sending {} to node -1. (NetworkClient.java:316) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:57 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=18,api_version=0,correlation_id=1,client_id=consumer-2} -- {} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(1,127.0.0.1:9092-127.0.0.1:53389,Session(User:ANONYMOUS,/127.0.0.1),null,1491468417653,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=1,client_id=consumer-2} -- {} from connection 127.0.0.1:9092-127.0.0.1:53389;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,127.0.0.1:9092-127.0.0.1:53389,Session(User:ANONYMOUS,/127.0.0.1),null,1491468417653,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@40edc06d,SendAction) (Logging.scala:36) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=1,client_id=consumer-2} -- {} from connection 127.0.0.1:9092-127.0.0.1:53389;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:57 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node -1: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:57 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:57 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node -1 (NetworkClient.java:751) 10:46:57 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node -1. (NetworkClient.java:316) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node -1. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=3,api_version=2,correlation_id=2,client_id=consumer-2} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(1,127.0.0.1:9092-127.0.0.1:53389,Session(User:ANONYMOUS,/127.0.0.1),null,1491468417656,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=2,client_id=consumer-2} -- {topics=[testOutputTopic]} from connection 127.0.0.1:9092-127.0.0.1:53389;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@343d03be and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 2 to client consumer-2 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,127.0.0.1:9092-127.0.0.1:53389,Session(User:ANONYMOUS,/127.0.0.1),null,1491468417656,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1f23329,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=2,client_id=consumer-2} -- {topics=[testOutputTopic]} from connection 127.0.0.1:9092-127.0.0.1:53389;totalTime:2,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [DEBUG] Metadata.update - Updated cluster metadata version 2 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:57 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=10,api_version=0,correlation_id=0,client_id=consumer-2} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(1,127.0.0.1:9092-127.0.0.1:53389,Session(User:ANONYMOUS,/127.0.0.1),null,1491468417659,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=0,client_id=consumer-2} -- {group_id=testOutputTopic} from connection 127.0.0.1:9092-127.0.0.1:53389;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=0,coordinator={node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092}} for correlation id 0 to client consumer-2. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,127.0.0.1:9092-127.0.0.1:53389,Session(User:ANONYMOUS,/127.0.0.1),null,1491468417659,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@4bb32a83,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 10, received {error_code=0,coordinator={node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092}} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=0,client_id=consumer-2} -- {group_id=testOutputTopic} from connection 127.0.0.1:9092-127.0.0.1:53389;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468417661, latencyMs=13, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=0,client_id=consumer-2}, responseBody={error_code=0,coordinator={node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:57 [INFO ] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Discovered coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) for group testOutputTopic. (AbstractCoordinator.java:573) 10:46:57 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node 2147483646 at ISI050.utenze.BANKIT.IT:9092. (NetworkClient.java:627) 10:46:57 [INFO ] ConsumerCoordinator.onJoinPrepare - Revoking previously assigned partitions [] for group testOutputTopic (ConsumerCoordinator.java:393) 10:46:57 [DEBUG] AbstractCoordinator$HeartbeatThread.run - Heartbeat thread for group testOutputTopic started (AbstractCoordinator.java:869) 10:46:57 [TRACE] AbstractCoordinator$HeartbeatThread.disable - Disabling heartbeat thread for group testOutputTopic (AbstractCoordinator.java:846) 10:46:57 [INFO ] AbstractCoordinator.sendJoinGroupRequest - (Re-)joining group testOutputTopic (AbstractCoordinator.java:407) 10:46:57 [DEBUG] AbstractCoordinator.sendJoinGroupRequest - Sending JoinGroup ((type: JoinGroupRequest, groupId=testOutputTopic, sessionTimeout=10000, rebalanceTimeout=300000, memberId=, protocolType=consumer, groupProtocols=org.apache.kafka.common.requests.JoinGroupRequest$ProtocolMetadata@749c877b)) to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:415) 10:46:57 [DEBUG] Logging$class.debug - Accepted connection from /10.36.240.33:53390 on /10.36.240.33:9092 and assigned it to processor 2, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node-2147483646.bytes-sent (Metrics.java:335) 10:46:57 [DEBUG] Logging$class.debug - Processor 2 listening to new connection from /10.36.240.33:53390 (Logging.scala:54) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node-2147483646.bytes-received (Metrics.java:335) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node-2147483646.latency (Metrics.java:335) 10:46:57 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 65536, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node 2147483646 (Selector.java:339) 10:46:57 [DEBUG] NetworkClient.handleConnections - Completed connection to node 2147483646. Fetching API versions. (NetworkClient.java:590) 10:46:57 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node 2147483646. (NetworkClient.java:603) 10:46:57 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node 2147483646. Assuming version 0. (NetworkClient.java:292) 10:46:57 [TRACE] NetworkClient.doSend - Sending {} to node 2147483646. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=18,api_version=0,correlation_id=4,client_id=consumer-2} -- {} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417667,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=4,client_id=consumer-2} -- {} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417667,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@f1f5d67,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=4,client_id=consumer-2} -- {} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:1,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node 2147483646: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} to node 2147483646. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=11,api_version=1,correlation_id=3,client_id=consumer-2} -- {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417670,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=11,api_version=1,correlation_id=3,client_id=consumer-2} -- {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [GroupCoordinator 1]: Preparing to restabilize group testOutputTopic with old generation 2 (Logging.scala:70) 10:46:57 [INFO ] Logging$class.info - [GroupCoordinator 1]: Stabilized group testOutputTopic generation 3 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending join group response {error_code=0,generation_id=3,group_protocol=range,leader_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,members=[{member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} for correlation id 3 to client consumer-2. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417670,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@15760143,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 11, received {error_code=0,generation_id=3,group_protocol=range,leader_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,members=[{member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=11,api_version=1,correlation_id=3,client_id=consumer-2} -- {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [DEBUG] AbstractCoordinator$JoinGroupResponseHandler.handle - Received successful JoinGroup response for group testOutputTopic: {error_code=0,generation_id=3,group_protocol=range,leader_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,members=[{member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} (AbstractCoordinator.java:425) 10:46:57 [DEBUG] ConsumerCoordinator.performAssignment - Performing assignment for group testOutputTopic using strategy range with subscriptions {consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56=Subscription(topics=[testOutputTopic])} (ConsumerCoordinator.java:336) 10:46:57 [DEBUG] ConsumerCoordinator.performAssignment - Finished assignment for group testOutputTopic: {consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56=Assignment(partitions=[testOutputTopic-0])} (ConsumerCoordinator.java:375) 10:46:57 [DEBUG] AbstractCoordinator.onJoinLeader - Sending leader SyncGroup for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null): (type=SyncGroupRequest, groupId=testOutputTopic, generationId=3, memberId=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56, groupAssignment=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56) (AbstractCoordinator.java:493) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,group_assignment=[{member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]}]} to node 2147483646. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=14,api_version=0,correlation_id=5,client_id=consumer-2} -- {group_id=testOutputTopic,generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,group_assignment=[{member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417674,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=14,api_version=0,correlation_id=5,client_id=consumer-2} -- {group_id=testOutputTopic,generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,group_assignment=[{member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]}]} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [INFO ] Logging$class.info - [GroupCoordinator 1]: Assignment received from leader for group testOutputTopic for generation 3 (Logging.scala:70) 10:46:57 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(__consumer_offsets-40 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2584810975, CreateTime = 1491468417676, key = 19 bytes, value = 230 bytes))])] to local log (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Inserting 283 bytes at offset 2 at position 360 with largest timestamp 1491468417676 at shallow offset 2 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Appended 283 to .\target\tmp\kafka\__consumer_offsets-40\00000000000000000000.log at offset 2 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Appended message set to log __consumer_offsets-40 with first offset: 2, next offset: 3, and messages: [(offset=2,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2584810975, CreateTime = 1491468417676, key = 19 bytes, value = 230 bytes))] (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 fetch requests. (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition __consumer_offsets-40 to [3 [0 : 643]] (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,40] on broker 1: High watermark for partition [__consumer_offsets,40] updated to 3 [0 : 643] (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 fetch requests. (Logging.scala:54) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 producer requests. (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 283 bytes written to log __consumer_offsets-40 beginning at offset 2 and ending at offset 2 (Logging.scala:36) 10:46:57 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 3 ms (Logging.scala:54) 10:46:57 [TRACE] Logging$class.trace - Initial partition status for __consumer_offsets-40 is [acksPending: true, error: 7, startOffset: 2, requiredOffset: 3] (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Checking produce satisfaction for __consumer_offsets-40, current status [acksPending: true, error: 7, startOffset: 2, requiredOffset: 3] (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Partition [__consumer_offsets,40] on broker 1: 1 acks satisfied for __consumer_offsets-40 with acks = -1 (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417674,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@10541951,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 14, received {error_code=0,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=14,api_version=0,correlation_id=5,client_id=consumer-2} -- {group_id=testOutputTopic,generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,group_assignment=[{member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,member_assignment=java.nio.HeapByteBuffer[pos=0 lim=35 cap=35]}]} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:7,requestQueueTime:1,localTime:5,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [INFO ] AbstractCoordinator$1.onSuccess - Successfully joined group testOutputTopic with generation 3 (AbstractCoordinator.java:375) 10:46:57 [TRACE] AbstractCoordinator$HeartbeatThread.enable - Enabling heartbeat thread for group testOutputTopic (AbstractCoordinator.java:837) 10:46:57 [INFO ] ConsumerCoordinator.onJoinComplete - Setting newly assigned partitions [testOutputTopic-0] for group testOutputTopic (ConsumerCoordinator.java:252) 10:46:57 [DEBUG] ConsumerCoordinator.sendOffsetFetchRequest - Group testOutputTopic fetching committed offsets for partitions: [testOutputTopic-0] (ConsumerCoordinator.java:800) 10:46:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,topics=[{topic=testOutputTopic,partitions=[{partition=0}]}]} to node 2147483646. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=9,api_version=2,correlation_id=6,client_id=consumer-2} -- {group_id=testOutputTopic,topics=[{topic=testOutputTopic,partitions=[{partition=0}]}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417684,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=9,api_version=2,correlation_id=6,client_id=consumer-2} -- {group_id=testOutputTopic,topics=[{topic=testOutputTopic,partitions=[{partition=0}]}]} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Group Metadata Manager on Broker 1]: Getting offsets of ArrayBuffer(testOutputTopic-0) for group testOutputTopic. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending offset fetch response {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,offset=-1,metadata=,error_code=0}]}],error_code=0} for correlation id 6 to client consumer-2. (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417684,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@50e1c17e,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 9, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,offset=-1,metadata=,error_code=0}]}],error_code=0} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=9,api_version=2,correlation_id=6,client_id=consumer-2} -- {group_id=testOutputTopic,topics=[{topic=testOutputTopic,partitions=[{partition=0}]}]} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [DEBUG] ConsumerCoordinator$OffsetFetchResponseHandler.handle - Group testOutputTopic has no committed offset for partition testOutputTopic-0 (ConsumerCoordinator.java:851) 10:46:57 [DEBUG] Fetcher.resetOffset - Resetting offset for partition testOutputTopic-0 to latest offset. (Fetcher.java:362) 10:46:57 [TRACE] Fetcher.sendListOffsetRequest - Sending ListOffsetRequest (type=ListOffsetRequest, replicaId=-1, partitionTimestamps={testOutputTopic-0=-1}, minVersion=0) to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:614) 10:46:57 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node 1 at ISI050.utenze.BANKIT.IT:9092. (NetworkClient.java:627) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node-1.bytes-sent (Metrics.java:335) 10:46:57 [DEBUG] Logging$class.debug - Accepted connection from /10.36.240.33:53391 on /10.36.240.33:9092 and assigned it to processor 0, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node-1.bytes-received (Metrics.java:335) 10:46:57 [DEBUG] Logging$class.debug - Processor 0 listening to new connection from /10.36.240.33:53391 (Logging.scala:54) 10:46:57 [DEBUG] Metrics.sensor - Added sensor with name node-1.latency (Metrics.java:335) 10:46:57 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 65536, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node 1 (Selector.java:339) 10:46:57 [DEBUG] NetworkClient.handleConnections - Completed connection to node 1. Fetching API versions. (NetworkClient.java:590) 10:46:57 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node 1. (NetworkClient.java:603) 10:46:57 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node 1. Assuming version 0. (NetworkClient.java:292) 10:46:57 [TRACE] NetworkClient.doSend - Sending {} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=18,api_version=0,correlation_id=8,client_id=consumer-2} -- {} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417692,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=8,client_id=consumer-2} -- {} from connection 10.36.240.33:9092-10.36.240.33:53391;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417692,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@ae143ef,SendAction) (Logging.scala:36) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=8,client_id=consumer-2} -- {} from connection 10.36.240.33:9092-10.36.240.33:53391;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:57 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node 1: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:57 [TRACE] NetworkClient.doSend - Sending {replica_id=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,timestamp=-1}]}]} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=2,api_version=1,correlation_id=7,client_id=consumer-2} -- {replica_id=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,timestamp=-1}]}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417697,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=2,api_version=1,correlation_id=7,client_id=consumer-2} -- {replica_id=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,timestamp=-1}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417697,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@44cb2af4,SendAction) (Logging.scala:36) 10:46:57 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 2, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,timestamp=-1,offset=0}]}]} (NetworkClient.java:534) 10:46:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=2,api_version=1,correlation_id=7,client_id=consumer-2} -- {replica_id=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,timestamp=-1}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:57 [TRACE] Fetcher$3.onSuccess - Received ListOffsetResponse {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,timestamp=-1,offset=0}]}]} from broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:620) 10:46:57 [DEBUG] Fetcher.handleListOffsetResponse - Handling ListOffsetResponse response for testOutputTopic-0. Fetched offset 0, timestamp -1 (Fetcher.java:662) 10:46:57 [TRACE] Fetcher.createFetchRequests - Added fetch request for partition testOutputTopic-0 at offset 0 to node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:726) 10:46:57 [DEBUG] Fetcher.sendFetches - Sending fetch for partitions [testOutputTopic-0] to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:180) 10:46:57 [TRACE] NetworkClient.doSend - Sending {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} to node 1. (NetworkClient.java:316) 10:46:57 [TRACE] Fetcher.createFetchRequests - Skipping fetch for partition testOutputTopic-0 because there is an in-flight request to ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:728) 10:46:57 [TRACE] Fetcher.createFetchRequests - Skipping fetch for partition testOutputTopic-0 because there is an in-flight request to ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:728) 10:46:57 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=1,api_version=3,correlation_id=9,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417702,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=1,api_version=3,correlation_id=9,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Fetching log segment for partition testOutputTopic-0, offset 0, partition fetch size 1048576, remaining response limit 52428800, ignoring response/partition size limits (Logging.scala:36) 10:46:57 [TRACE] Logging$class.trace - Reading 1048576 bytes from offset 0 in log testOutputTopic-0 of length 0 bytes (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Fetching log segment for partition testOutputTopic-0, offset 0, partition fetch size 1048576, remaining response limit 52428800, ignoring response/partition size limits (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Reading 1048576 bytes from offset 0 in log testOutputTopic-0 of length 0 bytes (Logging.scala:36) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name Fetch-:consumer-1 (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name FetchThrottleTime-:consumer-1 (Metrics.java:335) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Sending fetch response to client consumer-1 of 0 bytes (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53382,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417492,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.MultiSend@25e04a60,SendAction) (Logging.scala:36) 10:46:58 [DEBUG] Selector.pollSelectionKeys - Connection with /10.36.240.33 disconnected (Selector.java:375) java.io.IOException: Connessione interrotta dal software del computer host at sun.nio.ch.SocketDispatcher.writev0(Native Method) at sun.nio.ch.SocketDispatcher.writev(SocketDispatcher.java:55) at sun.nio.ch.IOUtil.write(IOUtil.java:148) at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:504) at java.nio.channels.SocketChannel.write(SocketChannel.java:502) at org.apache.kafka.common.network.PlaintextTransportLayer.write(PlaintextTransportLayer.java:159) at org.apache.kafka.common.network.ByteBufferSend.writeTo(ByteBufferSend.java:56) at org.apache.kafka.common.network.MultiSend.writeTo(MultiSend.java:76) at org.apache.kafka.common.network.KafkaChannel.send(KafkaChannel.java:173) at org.apache.kafka.common.network.KafkaChannel.write(KafkaChannel.java:161) at org.apache.kafka.common.network.Selector.pollSelectionKeys(Selector.java:361) at org.apache.kafka.common.network.Selector.poll(Selector.java:303) at kafka.network.Processor.poll(SocketServer.scala:494) at kafka.network.Processor.run(SocketServer.scala:432) at java.lang.Thread.run(Thread.java:745) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=1,api_version=3,correlation_id=97,client_id=consumer-1} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53382;totalTime:686,requestQueueTime:3,localTime:147,remoteTime:520,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Fetching log segment for partition testOutputTopic-0, offset 0, partition fetch size 1048576, remaining response limit 52428800, ignoring response/partition size limits (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Reading 1048576 bytes from offset 0 in log testOutputTopic-0 of length 0 bytes (Logging.scala:36) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name Fetch-:consumer-2 (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name FetchThrottleTime-:consumer-2 (Metrics.java:335) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Sending fetch response to client consumer-2 of 0 bytes (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468417702,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.MultiSend@42ad7b81,SendAction) (Logging.scala:36) 10:46:58 [TRACE] MultiSend.writeTo - Bytes written as part of multi-send call: 55, total bytes written so far: 55, expected bytes to write: 55 (MultiSend.java:88) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 1, received {throttle_time_ms=0,responses=[{topic=testOutputTopic,partition_responses=[{partition_header={partition=0,error_code=0,high_watermark=0},record_set=[]}]}]} (NetworkClient.java:534) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=1,api_version=3,correlation_id=9,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;totalTime:506,requestQueueTime:0,localTime:3,remoteTime:502,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] Fetcher.parseCompletedFetch - Adding fetched record for partition testOutputTopic-0 with offset 0 to buffered record list (Fetcher.java:783) 10:46:58 [TRACE] Fetcher.parseCompletedFetch - Received 0 records in fetch response for partition testOutputTopic-0 with offset 0 (Fetcher.java:805) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name topic.testOutputTopic.bytes-fetched (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name topic.testOutputTopic.records-fetched (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name testOutputTopic-0.records-lag (Metrics.java:335) 10:46:58 [TRACE] Fetcher.createFetchRequests - Added fetch request for partition testOutputTopic-0 at offset 0 to node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:726) 10:46:58 [DEBUG] Fetcher.sendFetches - Sending fetch for partitions [testOutputTopic-0] to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:180) 10:46:58 [TRACE] NetworkClient.doSend - Sending {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Fetcher.createFetchRequests - Skipping fetch for partition testOutputTopic-0 because there is an in-flight request to ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:728) 10:46:58 [TRACE] Fetcher.createFetchRequests - Skipping fetch for partition testOutputTopic-0 because there is an in-flight request to ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:728) 10:46:58 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=1,api_version=3,correlation_id=10,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418214,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=1,api_version=3,correlation_id=10,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Fetching log segment for partition testOutputTopic-0, offset 0, partition fetch size 1048576, remaining response limit 52428800, ignoring response/partition size limits (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Reading 1048576 bytes from offset 0 in log testOutputTopic-0 of length 0 bytes (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:46:58 [INFO ] AbstractConfig.logAll - ProducerConfig values: acks = 1 batch.size = 16384 block.on.buffer.full = false bootstrap.servers = [localhost:9092] buffer.memory = 33554432 client.id = compression.type = none connections.max.idle.ms = 540000 interceptor.classes = null key.serializer = class org.apache.kafka.common.serialization.StringSerializer linger.ms = 0 max.block.ms = 60000 max.in.flight.requests.per.connection = 5 max.request.size = 1048576 metadata.fetch.timeout.ms = 60000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner receive.buffer.bytes = 32768 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 0 retry.backoff.ms = 100 sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT send.buffer.bytes = 131072 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS timeout.ms = 30000 value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer (AbstractConfig.java:180) 10:46:58 [TRACE] KafkaProducer. - Starting the Kafka producer (KafkaProducer.java:211) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name bufferpool-wait-time (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name buffer-exhausted-records (Metrics.java:335) 10:46:58 [DEBUG] Metadata.update - Updated cluster metadata version 1 to Cluster(id = null, nodes = [localhost:9092 (id: -1 rack: null)], partitions = []) (Metadata.java:244) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name connections-closed: (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name connections-created: (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent-received: (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name bytes-sent: (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name bytes-received: (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name select-time: (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name io-time: (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name batch-size (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name compression-rate (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name queue-time (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name request-time (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name produce-throttle-time (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name records-per-request (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name record-retries (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name errors (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name record-size-max (Metrics.java:335) 10:46:58 [INFO ] AppInfoParser$AppInfo. - Kafka version : 0.10.2.0 (AppInfoParser.java:83) 10:46:58 [DEBUG] Sender.run - Starting Kafka producer I/O thread. (Sender.java:121) 10:46:58 [INFO ] AppInfoParser$AppInfo. - Kafka commitId : 576d93a8dc0cf421 (AppInfoParser.java:84) 10:46:58 [DEBUG] KafkaProducer. - Kafka producer started (KafkaProducer.java:336) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Fetching log segment for partition testOutputTopic-0, offset 0, partition fetch size 1048576, remaining response limit 52428800, ignoring response/partition size limits (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Reading 1048576 bytes from offset 0 in log testOutputTopic-0 of length 0 bytes (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Sending fetch response to client consumer-2 of 0 bytes (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418214,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.MultiSend@54703a14,SendAction) (Logging.scala:36) 10:46:58 [TRACE] MultiSend.writeTo - Bytes written as part of multi-send call: 55, total bytes written so far: 55, expected bytes to write: 55 (MultiSend.java:88) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=1,api_version=3,correlation_id=10,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;totalTime:505,requestQueueTime:0,localTime:1,remoteTime:502,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] KafkaProducer.waitOnMetadata - Requesting metadata update for topic testOutputTopic. (KafkaProducer.java:544) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:58 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Initialize connection to node -1 for sending metadata request (NetworkClient.java:767) 10:46:58 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node -1 at localhost:9092. (NetworkClient.java:627) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node--1.bytes-sent (Metrics.java:335) 10:46:58 [DEBUG] Logging$class.debug - Accepted connection from /127.0.0.1:53394 on /127.0.0.1:9092 and assigned it to processor 1, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node--1.bytes-received (Metrics.java:335) 10:46:58 [DEBUG] Logging$class.debug - Processor 1 listening to new connection from /127.0.0.1:53394 (Logging.scala:54) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node--1.latency (Metrics.java:335) 10:46:58 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 32768, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node -1 (Selector.java:339) 10:46:58 [DEBUG] NetworkClient.handleConnections - Completed connection to node -1. Fetching API versions. (NetworkClient.java:590) 10:46:58 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node -1. (NetworkClient.java:603) 10:46:58 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node -1. Assuming version 0. (NetworkClient.java:292) 10:46:58 [TRACE] NetworkClient.doSend - Sending {} to node -1. (NetworkClient.java:316) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:58 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=18,api_version=0,correlation_id=0,client_id=producer-1} -- {} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(1,127.0.0.1:9092-127.0.0.1:53394,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418734,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=0,client_id=producer-1} -- {} from connection 127.0.0.1:9092-127.0.0.1:53394;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,127.0.0.1:9092-127.0.0.1:53394,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418734,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@b5b67f2,SendAction) (Logging.scala:36) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=0,client_id=producer-1} -- {} from connection 127.0.0.1:9092-127.0.0.1:53394;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:58 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node -1: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:58 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node -1 (NetworkClient.java:751) 10:46:58 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node -1. (NetworkClient.java:316) 10:46:58 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=3,api_version=2,correlation_id=1,client_id=producer-1} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(1,127.0.0.1:9092-127.0.0.1:53394,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418739,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=1,client_id=producer-1} -- {topics=[testOutputTopic]} from connection 127.0.0.1:9092-127.0.0.1:53394;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@36a0122e and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 1 to client producer-1 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,127.0.0.1:9092-127.0.0.1:53394,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418739,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@596ea328,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=1,client_id=producer-1} -- {topics=[testOutputTopic]} from connection 127.0.0.1:9092-127.0.0.1:53394;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [DEBUG] Metadata.update - Updated cluster metadata version 2 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference0, value=[B@553f1d75, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node 1 at ISI050.utenze.BANKIT.IT:9092. (NetworkClient.java:627) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 1, received {throttle_time_ms=0,responses=[{topic=testOutputTopic,partition_responses=[{partition_header={partition=0,error_code=0,high_watermark=0},record_set=[]}]}]} (NetworkClient.java:534) 10:46:58 [DEBUG] Logging$class.debug - Accepted connection from /10.36.240.33:53395 on /10.36.240.33:9092 and assigned it to processor 2, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - Processor 2 listening to new connection from /10.36.240.33:53395 (Logging.scala:54) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node-1.bytes-sent (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node-1.bytes-received (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node-1.latency (Metrics.java:335) 10:46:58 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 32768, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node 1 (Selector.java:339) 10:46:58 [DEBUG] NetworkClient.handleConnections - Completed connection to node 1. Fetching API versions. (NetworkClient.java:590) 10:46:58 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node 1. (NetworkClient.java:603) 10:46:58 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node 1. Assuming version 0. (NetworkClient.java:292) 10:46:58 [TRACE] NetworkClient.doSend - Sending {} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=18,api_version=0,correlation_id=2,client_id=producer-1} -- {} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418759,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=2,client_id=producer-1} -- {} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418759,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@49371df2,SendAction) (Logging.scala:36) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=2,client_id=producer-1} -- {} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:58 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node 1: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name topic.testOutputTopic.records-per-batch (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name topic.testOutputTopic.bytes (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name topic.testOutputTopic.compression-rate (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name topic.testOutputTopic.record-retries (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name topic.testOutputTopic.record-errors (Metrics.java:335) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 882282754, CreateTime = 1491468418743, key = 10 bytes, value = 74 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 882282754, CreateTime = 1491468418743, key = 10 bytes, value = 74 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=3,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 882282754, CreateTime = 1491468418743, key = 10 bytes, value = 74 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418765,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=3,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 882282754, CreateTime = 1491468418743, key = 10 bytes, value = 74 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 882282754, CreateTime = 1491468418743, key = 10 bytes, value = 74 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 118 bytes at offset 0 at position 0 with largest timestamp 1491468418743 at shallow offset 0 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 118 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 0 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 0, next offset: 1, and messages: [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 882282754, CreateTime = 1491468418743, key = 10 bytes, value = 74 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [1 [0 : 118]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 1 [0 : 118] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 118 bytes written to log testOutputTopic-0 beginning at offset 0 and ending at offset 0 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 2 ms (Logging.scala:54) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name Produce-:producer-1 (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name ProduceThrottleTime-:producer-1 (Metrics.java:335) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418765,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1bea49,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=0,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=3,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:10,requestQueueTime:1,localTime:8,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 3 (Sender.java:262) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 0 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference1, value=[B@6cbcf243, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1705513017, CreateTime = 1491468418776, key = 10 bytes, value = 94 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1705513017, CreateTime = 1491468418776, key = 10 bytes, value = 94 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=4,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1705513017, CreateTime = 1491468418776, key = 10 bytes, value = 94 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418778,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=4,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1705513017, CreateTime = 1491468418776, key = 10 bytes, value = 94 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1705513017, CreateTime = 1491468418776, key = 10 bytes, value = 94 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 138 bytes at offset 1 at position 118 with largest timestamp 1491468418776 at shallow offset 1 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 138 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 1 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 1, next offset: 2, and messages: [(offset=1,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1705513017, CreateTime = 1491468418776, key = 10 bytes, value = 94 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [2 [0 : 256]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 2 [0 : 256] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 138 bytes written to log testOutputTopic-0 beginning at offset 1 and ending at offset 1 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 2 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418778,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@9a7e3b7,SendAction) (Logging.scala:36) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=4,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:4,requestQueueTime:1,localTime:3,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=1,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 4 (Sender.java:262) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 1 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference2, value=[B@62435e70, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3975510442, CreateTime = 1491468418784, key = 10 bytes, value = 104 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3975510442, CreateTime = 1491468418784, key = 10 bytes, value = 104 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=5,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3975510442, CreateTime = 1491468418784, key = 10 bytes, value = 104 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418787,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=5,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3975510442, CreateTime = 1491468418784, key = 10 bytes, value = 104 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3975510442, CreateTime = 1491468418784, key = 10 bytes, value = 104 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 148 bytes at offset 2 at position 256 with largest timestamp 1491468418784 at shallow offset 2 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 148 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 2 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 2, next offset: 3, and messages: [(offset=2,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3975510442, CreateTime = 1491468418784, key = 10 bytes, value = 104 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [3 [0 : 404]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 3 [0 : 404] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 148 bytes written to log testOutputTopic-0 beginning at offset 2 and ending at offset 2 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 2 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418787,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@783d9006,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=2,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=5,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:4,requestQueueTime:0,localTime:4,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 5 (Sender.java:262) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 2 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference3, value=[B@38be305c, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2726021079, CreateTime = 1491468418792, key = 10 bytes, value = 114 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2726021079, CreateTime = 1491468418792, key = 10 bytes, value = 114 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=6,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2726021079, CreateTime = 1491468418792, key = 10 bytes, value = 114 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418794,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=6,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2726021079, CreateTime = 1491468418792, key = 10 bytes, value = 114 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2726021079, CreateTime = 1491468418792, key = 10 bytes, value = 114 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 158 bytes at offset 3 at position 404 with largest timestamp 1491468418792 at shallow offset 3 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 158 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 3 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 3, next offset: 4, and messages: [(offset=3,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2726021079, CreateTime = 1491468418792, key = 10 bytes, value = 114 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [4 [0 : 562]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 4 [0 : 562] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 158 bytes written to log testOutputTopic-0 beginning at offset 3 and ending at offset 3 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 2 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418794,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@914da7e,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=3,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=6,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:4,requestQueueTime:1,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 6 (Sender.java:262) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 3 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference4, value=[B@5ed731d0, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 578730568, CreateTime = 1491468418799, key = 10 bytes, value = 124 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 578730568, CreateTime = 1491468418799, key = 10 bytes, value = 124 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=7,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 578730568, CreateTime = 1491468418799, key = 10 bytes, value = 124 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418800,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=7,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 578730568, CreateTime = 1491468418799, key = 10 bytes, value = 124 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 578730568, CreateTime = 1491468418799, key = 10 bytes, value = 124 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 168 bytes at offset 4 at position 562 with largest timestamp 1491468418799 at shallow offset 4 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 168 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 4 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 4, next offset: 5, and messages: [(offset=4,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 578730568, CreateTime = 1491468418799, key = 10 bytes, value = 124 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [5 [0 : 730]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 5 [0 : 730] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 168 bytes written to log testOutputTopic-0 beginning at offset 4 and ending at offset 4 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 2 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418800,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@560230b7,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=4,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=7,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:4,requestQueueTime:1,localTime:3,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 7 (Sender.java:262) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 4 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference5, value=[B@7bc10d84, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3838122683, CreateTime = 1491468418805, key = 10 bytes, value = 134 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3838122683, CreateTime = 1491468418805, key = 10 bytes, value = 134 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=8,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3838122683, CreateTime = 1491468418805, key = 10 bytes, value = 134 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418807,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=8,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3838122683, CreateTime = 1491468418805, key = 10 bytes, value = 134 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3838122683, CreateTime = 1491468418805, key = 10 bytes, value = 134 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 178 bytes at offset 5 at position 730 with largest timestamp 1491468418805 at shallow offset 5 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 178 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 5 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 5, next offset: 6, and messages: [(offset=5,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3838122683, CreateTime = 1491468418805, key = 10 bytes, value = 134 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [6 [0 : 908]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 6 [0 : 908] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 178 bytes written to log testOutputTopic-0 beginning at offset 5 and ending at offset 5 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 2 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418807,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@600edb9b,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=5,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 8 (Sender.java:262) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=8,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:3,requestQueueTime:0,localTime:3,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 5 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference6, value=[B@40e10ff8, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2305736210, CreateTime = 1491468418811, key = 10 bytes, value = 144 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2305736210, CreateTime = 1491468418811, key = 10 bytes, value = 144 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=9,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2305736210, CreateTime = 1491468418811, key = 10 bytes, value = 144 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418813,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=9,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2305736210, CreateTime = 1491468418811, key = 10 bytes, value = 144 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2305736210, CreateTime = 1491468418811, key = 10 bytes, value = 144 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 188 bytes at offset 6 at position 908 with largest timestamp 1491468418811 at shallow offset 6 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 188 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 6 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 6, next offset: 7, and messages: [(offset=6,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2305736210, CreateTime = 1491468418811, key = 10 bytes, value = 144 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [7 [0 : 1096]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 7 [0 : 1096] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 188 bytes written to log testOutputTopic-0 beginning at offset 6 and ending at offset 6 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 2 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418813,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@74a7da3c,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=6,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 9 (Sender.java:262) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 6 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=9,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:4,requestQueueTime:1,localTime:3,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference7, value=[B@26a4842b, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3555532154, CreateTime = 1491468418818, key = 10 bytes, value = 154 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3555532154, CreateTime = 1491468418818, key = 10 bytes, value = 154 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=10,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3555532154, CreateTime = 1491468418818, key = 10 bytes, value = 154 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418821,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=10,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3555532154, CreateTime = 1491468418818, key = 10 bytes, value = 154 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3555532154, CreateTime = 1491468418818, key = 10 bytes, value = 154 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 198 bytes at offset 7 at position 1096 with largest timestamp 1491468418818 at shallow offset 7 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 198 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 7 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 7, next offset: 8, and messages: [(offset=7,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3555532154, CreateTime = 1491468418818, key = 10 bytes, value = 154 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [8 [0 : 1294]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 8 [0 : 1294] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 198 bytes written to log testOutputTopic-0 beginning at offset 7 and ending at offset 7 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 3 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418821,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@5bcf29fc,SendAction) (Logging.scala:36) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=10,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:5,requestQueueTime:0,localTime:4,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=7,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 10 (Sender.java:262) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 7 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference8, value=[B@366ef90e, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2952636745, CreateTime = 1491468418828, key = 10 bytes, value = 164 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2952636745, CreateTime = 1491468418828, key = 10 bytes, value = 164 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=11,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2952636745, CreateTime = 1491468418828, key = 10 bytes, value = 164 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418831,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=11,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2952636745, CreateTime = 1491468418828, key = 10 bytes, value = 164 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2952636745, CreateTime = 1491468418828, key = 10 bytes, value = 164 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 208 bytes at offset 8 at position 1294 with largest timestamp 1491468418828 at shallow offset 8 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 208 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 8 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 8, next offset: 9, and messages: [(offset=8,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2952636745, CreateTime = 1491468418828, key = 10 bytes, value = 164 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [9 [0 : 1502]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 9 [0 : 1502] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 208 bytes written to log testOutputTopic-0 beginning at offset 8 and ending at offset 8 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 3 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418831,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1c0f2dba,SendAction) (Logging.scala:36) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=11,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:4,requestQueueTime:0,localTime:3,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=8,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 11 (Sender.java:262) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 8 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] KafkaProducer.doSend - Sending record ProducerRecord(topic=testOutputTopic, partition=null, key=Reference9, value=[B@31e75d13, timestamp=null) with callback null to topic testOutputTopic partition 0 (KafkaProducer.java:475) 10:46:58 [TRACE] RecordAccumulator.append - Allocating a new 16384 byte message buffer for topic testOutputTopic partition 0 (RecordAccumulator.java:182) 10:46:58 [TRACE] KafkaProducer.doSend - Waking up the sender since topic testOutputTopic partition 0 is either full or getting a new batch (KafkaProducer.java:480) 10:46:58 [TRACE] Sender.run - Nodes with data ready to send: [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)] (Sender.java:216) 10:46:58 [TRACE] NetworkClient.doSend - Sending {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 508306320, CreateTime = 1491468418836, key = 10 bytes, value = 174 bytes))]}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Sender.sendProduceRequest - Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=({testOutputTopic-0=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 508306320, CreateTime = 1491468418836, key = 10 bytes, value = 174 bytes))]})) (Sender.java:365) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=0,api_version=2,correlation_id=12,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 508306320, CreateTime = 1491468418836, key = 10 bytes, value = 174 bytes))]}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418838,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=0,api_version=2,correlation_id=12,client_id=producer-1} -- {acks=1,timeout=30000,topic_data=[{topic=testOutputTopic,data=[{partition=0,record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 508306320, CreateTime = 1491468418836, key = 10 bytes, value = 174 bytes))]}]}]} from connection 10.36.240.33:9092-10.36.240.33:53395;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(testOutputTopic-0 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 508306320, CreateTime = 1491468418836, key = 10 bytes, value = 174 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 218 bytes at offset 9 at position 1502 with largest timestamp 1491468418836 at shallow offset 9 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 218 to .\target\tmp\kafka\testOutputTopic-0\00000000000000000000.log at offset 9 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log testOutputTopic-0 with first offset: 9, next offset: 10, and messages: [(offset=9,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 508306320, CreateTime = 1491468418836, key = 10 bytes, value = 174 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition testOutputTopic-0 to [10 [0 : 1720]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [testOutputTopic,0] on broker 1: High watermark for partition [testOutputTopic,0] updated to 10 [0 : 1720] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key testOutputTopic-0 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 218 bytes written to log testOutputTopic-0 beginning at offset 9 and ending at offset 9 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 2 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53395,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418838,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@a4bffa3,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 0, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0,base_offset=9,log_append_time=-1}]}],throttle_time_ms=0} (NetworkClient.java:534) 10:46:58 [TRACE] Sender.handleProduceResponse - Received produce response from node 1 with correlation id 12 (Sender.java:262) 10:46:58 [TRACE] RecordBatch.done - Produced messages to topic-partition testOutputTopic-0 with base offset offset 9 and error: null. (RecordBatch.java:97) 10:46:58 [TRACE] KafkaProducer.flush - Flushing accumulated records in producer. (KafkaProducer.java:617) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=0,api_version=2,correlation_id=12,client_id=producer-1} -- {acks=null,timeout=null,topic_data=null} from connection 10.36.240.33:9092-10.36.240.33:53395;totalTime:5,requestQueueTime:1,localTime:3,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] Fetcher.parseCompletedFetch - Adding fetched record for partition testOutputTopic-0 with offset 0 to buffered record list (Fetcher.java:783) 10:46:58 [TRACE] Fetcher.parseCompletedFetch - Received 0 records in fetch response for partition testOutputTopic-0 with offset 0 (Fetcher.java:805) 10:46:58 [TRACE] Fetcher.createFetchRequests - Added fetch request for partition testOutputTopic-0 at offset 0 to node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:726) 10:46:58 [DEBUG] Fetcher.sendFetches - Sending fetch for partitions [testOutputTopic-0] to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:180) 10:46:58 [TRACE] NetworkClient.doSend - Sending {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] Fetcher.createFetchRequests - Skipping fetch for partition testOutputTopic-0 because there is an in-flight request to ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:728) 10:46:58 [TRACE] Fetcher.createFetchRequests - Skipping fetch for partition testOutputTopic-0 because there is an in-flight request to ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:728) 10:46:58 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=1,api_version=3,correlation_id=11,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418846,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=1,api_version=3,correlation_id=11,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Fetching log segment for partition testOutputTopic-0, offset 0, partition fetch size 1048576, remaining response limit 52428800, ignoring response/partition size limits (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Reading 1048576 bytes from offset 0 in log testOutputTopic-0 of length 1720 bytes (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Sending fetch response to client consumer-2 of 1720 bytes (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418846,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.MultiSend@540ef54e,SendAction) (Logging.scala:36) 10:46:58 [TRACE] Fetcher.createFetchRequests - Skipping fetch for partition testOutputTopic-0 because there is an in-flight request to ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:728) 10:46:58 [TRACE] Fetcher.createFetchRequests - Skipping fetch for partition testOutputTopic-0 because there is an in-flight request to ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:728) 10:46:58 [TRACE] MultiSend.writeTo - Bytes written as part of multi-send call: 1775, total bytes written so far: 1775, expected bytes to write: 1775 (MultiSend.java:88) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=1,api_version=3,correlation_id=11,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=0,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;totalTime:17,requestQueueTime:1,localTime:3,remoteTime:0,responseQueueTime:0,sendTime:13,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 1, received {throttle_time_ms=0,responses=[{topic=testOutputTopic,partition_responses=[{partition_header={partition=0,error_code=0,high_watermark=10},record_set=[(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 882282754, CreateTime = 1491468418743, key = 10 bytes, value = 74 bytes)), (offset=1,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1705513017, CreateTime = 1491468418776, key = 10 bytes, value = 94 bytes)), (offset=2,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3975510442, CreateTime = 1491468418784, key = 10 bytes, value = 104 bytes)), (offset=3,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2726021079, CreateTime = 1491468418792, key = 10 bytes, value = 114 bytes)), (offset=4,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 578730568, CreateTime = 1491468418799, key = 10 bytes, value = 124 bytes)), (offset=5,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3838122683, CreateTime = 1491468418805, key = 10 bytes, value = 134 bytes)), (offset=6,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2305736210, CreateTime = 1491468418811, key = 10 bytes, value = 144 bytes)), (offset=7,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 3555532154, CreateTime = 1491468418818, key = 10 bytes, value = 154 bytes)), (offset=8,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 2952636745, CreateTime = 1491468418828, key = 10 bytes, value = 164 bytes)), (offset=9,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 508306320, CreateTime = 1491468418836, key = 10 bytes, value = 174 bytes))]}]}]} (NetworkClient.java:534) 10:46:58 [TRACE] Fetcher.parseCompletedFetch - Adding fetched record for partition testOutputTopic-0 with offset 0 to buffered record list (Fetcher.java:783) 10:46:58 [TRACE] Fetcher.parseCompletedFetch - Received 10 records in fetch response for partition testOutputTopic-0 with offset 0 (Fetcher.java:805) 10:46:58 [TRACE] Fetcher.drainRecords - Returning fetched records at offset 0 for assigned partition testOutputTopic-0 and update position to 10 (Fetcher.java:512) 10:46:58 [DEBUG] Fetcher.drainRecords - Ignoring fetched records for testOutputTopic-0 at offset 0 since the current position is 10 (Fetcher.java:526) 10:46:58 [TRACE] Fetcher.createFetchRequests - Added fetch request for partition testOutputTopic-0 at offset 10 to node ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:726) 10:46:58 [DEBUG] Fetcher.sendFetches - Sending fetch for partitions [testOutputTopic-0] to broker ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null) (Fetcher.java:180) 10:46:58 [TRACE] NetworkClient.doSend - Sending {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=10,max_bytes=1048576}]}]} to node 1. (NetworkClient.java:316) 10:46:58 [TRACE] ConsumerCoordinator.sendOffsetCommitRequest - Sending OffsetCommit request with {testOutputTopic-0=OffsetAndMetadata{offset=10, metadata=''}} to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) for group testOutputTopic (ConsumerCoordinator.java:706) 10:46:58 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,retention_time=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,offset=10,metadata=}]}]} to node 2147483646. (NetworkClient.java:316) 10:46:58 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=1,api_version=3,correlation_id=12,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=10,max_bytes=1048576}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=8,api_version=2,correlation_id=13,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,retention_time=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,offset=10,metadata=}]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418871,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 5 on Broker 1], Kafka request handler 5 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418871,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=1,api_version=3,correlation_id=12,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=10,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=8,api_version=2,correlation_id=13,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,retention_time=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,offset=10,metadata=}]}]} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Fetching log segment for partition testOutputTopic-0, offset 10, partition fetch size 1048576, remaining response limit 52428800, ignoring response/partition size limits (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Reading 1048576 bytes from offset 10 in log testOutputTopic-0 of length 1720 bytes (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Append [Map(__consumer_offsets-40 -> [(offset=0,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1781641225, CreateTime = 1491468418876, key = 40 bytes, value = 28 bytes))])] to local log (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Inserting 102 bytes at offset 3 at position 643 with largest timestamp 1491468418876 at shallow offset 3 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended 102 to .\target\tmp\kafka\__consumer_offsets-40\00000000000000000000.log at offset 3 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Appended message set to log __consumer_offsets-40 with first offset: 3, next offset: 4, and messages: [(offset=3,record=Record(magic = 1, attributes = 0, compression = NONE, crc = 1781641225, CreateTime = 1491468418876, key = 40 bytes, value = 28 bytes))] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Setting high watermark for replica 1 partition __consumer_offsets-40 to [4 [0 : 745]] (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - Partition [__consumer_offsets,40] on broker 1: High watermark for partition [__consumer_offsets,40] updated to 4 [0 : 745] (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 fetch requests. (Logging.scala:54) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Request key __consumer_offsets-40 unblocked 0 producer requests. (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: 102 bytes written to log __consumer_offsets-40 beginning at offset 3 and ending at offset 3 (Logging.scala:36) 10:46:58 [DEBUG] Logging$class.debug - [Replica Manager on Broker 1]: Produce to local log in 2 ms (Logging.scala:54) 10:46:58 [TRACE] Logging$class.trace - Initial partition status for __consumer_offsets-40 is [acksPending: true, error: 7, startOffset: 3, requiredOffset: 4] (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Checking produce satisfaction for __consumer_offsets-40, current status [acksPending: true, error: 7, startOffset: 3, requiredOffset: 4] (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Partition [__consumer_offsets,40] on broker 1: 1 acks satisfied for __consumer_offsets-40 with acks = -1 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418871,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@40f7a2ec,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 8, received {responses=[{topic=testOutputTopic,partition_responses=[{partition=0,error_code=0}]}]} (NetworkClient.java:534) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=8,api_version=2,correlation_id=13,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56,retention_time=-1,topics=[{topic=testOutputTopic,partitions=[{partition=0,offset=10,metadata=}]}]} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:14,requestQueueTime:1,localTime:12,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [DEBUG] ConsumerCoordinator$OffsetCommitResponseHandler.handle - Group testOutputTopic committed offset 10 for partition testOutputTopic-0 (ConsumerCoordinator.java:732) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:58 [DEBUG] AbstractCoordinator.sendGroupCoordinatorRequest - Sending GroupCoordinator request for group testOutputTopic to broker localhost:9092 (id: -1 rack: null) (AbstractCoordinator.java:548) 10:46:58 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node -1 at localhost:9092. (NetworkClient.java:627) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node--1.bytes-sent (Metrics.java:335) 10:46:58 [DEBUG] Logging$class.debug - Accepted connection from /127.0.0.1:53396 on /127.0.0.1:9092 and assigned it to processor 0, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node--1.bytes-received (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node--1.latency (Metrics.java:335) 10:46:58 [DEBUG] Logging$class.debug - Processor 0 listening to new connection from /127.0.0.1:53396 (Logging.scala:54) 10:46:58 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 65536, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node -1 (Selector.java:339) 10:46:58 [DEBUG] NetworkClient.handleConnections - Completed connection to node -1. Fetching API versions. (NetworkClient.java:590) 10:46:58 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node -1. (NetworkClient.java:603) 10:46:58 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node -1. Assuming version 0. (NetworkClient.java:292) 10:46:58 [TRACE] NetworkClient.doSend - Sending {} to node -1. (NetworkClient.java:316) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) (NetworkClient.java:443) 10:46:58 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=18,api_version=0,correlation_id=1,client_id=consumer-3} -- {} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 1 on Broker 1], Kafka request handler 1 on broker 1 handling request Request(0,127.0.0.1:9092-127.0.0.1:53396,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418891,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=1,client_id=consumer-3} -- {} from connection 127.0.0.1:9092-127.0.0.1:53396;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,127.0.0.1:9092-127.0.0.1:53396,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418891,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@3522c4e6,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=1,client_id=consumer-3} -- {} from connection 127.0.0.1:9092-127.0.0.1:53396;totalTime:1,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node -1: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:58 [TRACE] NetworkClient.leastLoadedNode - Found least loaded node localhost:9092 (id: -1 rack: null) connected with no in-flight requests (NetworkClient.java:430) 10:46:58 [DEBUG] NetworkClient$DefaultMetadataUpdater.maybeUpdate - Sending metadata request (type=MetadataRequest, topics=testOutputTopic) to node -1 (NetworkClient.java:751) 10:46:58 [TRACE] NetworkClient.doSend - Sending {topics=[testOutputTopic]} to node -1. (NetworkClient.java:316) 10:46:58 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic} to node -1. (NetworkClient.java:316) 10:46:58 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=3,api_version=2,correlation_id=2,client_id=consumer-3} -- {topics=[testOutputTopic]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 3 on Broker 1], Kafka request handler 3 on broker 1 handling request Request(0,127.0.0.1:9092-127.0.0.1:53396,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418894,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=3,api_version=2,correlation_id=2,client_id=consumer-3} -- {topics=[testOutputTopic]} from connection 127.0.0.1:9092-127.0.0.1:53396;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Sending topic metadata org.apache.kafka.common.requests.MetadataResponse$TopicMetadata@7f499dec and brokers 1 : (EndPoint(ISI050.utenze.BANKIT.IT,9092,ListenerName(PLAINTEXT),PLAINTEXT)) : null for correlation id 2 to client consumer-3 (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,127.0.0.1:9092-127.0.0.1:53396,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418894,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@1625f389,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 3, received {brokers=[{node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092,rack=null}],cluster_id=Ticpy3hDShevs90Ty0g1rg,controller_id=1,topic_metadata=[{topic_error_code=0,topic=testOutputTopic,is_internal=false,partition_metadata=[{partition_error_code=0,partition_id=0,leader=1,replicas=[1],isr=[1]}]}]} (NetworkClient.java:534) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=3,api_version=2,correlation_id=2,client_id=consumer-3} -- {topics=[testOutputTopic]} from connection 127.0.0.1:9092-127.0.0.1:53396;totalTime:2,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [DEBUG] Metadata.update - Updated cluster metadata version 2 to Cluster(id = Ticpy3hDShevs90Ty0g1rg, nodes = [ISI050.utenze.BANKIT.IT:9092 (id: 1 rack: null)], partitions = [Partition(topic = testOutputTopic, partition = 0, leader = 1, replicas = [1], isr = [1])]) (Metadata.java:244) 10:46:58 [TRACE] Logging$class.trace - Processor 0 received request : {api_key=10,api_version=0,correlation_id=0,client_id=consumer-3} -- {group_id=testOutputTopic} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 6 on Broker 1], Kafka request handler 6 on broker 1 handling request Request(0,127.0.0.1:9092-127.0.0.1:53396,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418897,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=10,api_version=0,correlation_id=0,client_id=consumer-3} -- {group_id=testOutputTopic} from connection 127.0.0.1:9092-127.0.0.1:53396;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Sending consumer metadata {error_code=0,coordinator={node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092}} for correlation id 0 to client consumer-3. (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,127.0.0.1:9092-127.0.0.1:53396,Session(User:ANONYMOUS,/127.0.0.1),null,1491468418897,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@20d1428d,SendAction) (Logging.scala:36) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node -1, for key 10, received {error_code=0,coordinator={node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092}} (NetworkClient.java:534) 10:46:58 [DEBUG] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Received GroupCoordinator response ClientResponse(receivedTimeMs=1491468418900, latencyMs=14, disconnected=false, requestHeader={api_key=10,api_version=0,correlation_id=0,client_id=consumer-3}, responseBody={error_code=0,coordinator={node_id=1,host=ISI050.utenze.BANKIT.IT,port=9092}}) for group testOutputTopic (AbstractCoordinator.java:559) 10:46:58 [INFO ] AbstractCoordinator$GroupCoordinatorResponseHandler.onSuccess - Discovered coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) for group testOutputTopic. (AbstractCoordinator.java:573) 10:46:58 [DEBUG] NetworkClient.initiateConnect - Initiating connection to node 2147483646 at ISI050.utenze.BANKIT.IT:9092. (NetworkClient.java:627) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=10,api_version=0,correlation_id=0,client_id=consumer-3} -- {group_id=testOutputTopic} from connection 127.0.0.1:9092-127.0.0.1:53396;totalTime:3,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [INFO ] ConsumerCoordinator.onJoinPrepare - Revoking previously assigned partitions [] for group testOutputTopic (ConsumerCoordinator.java:393) 10:46:58 [TRACE] AbstractCoordinator$HeartbeatThread.disable - Disabling heartbeat thread for group testOutputTopic (AbstractCoordinator.java:846) 10:46:58 [INFO ] AbstractCoordinator.sendJoinGroupRequest - (Re-)joining group testOutputTopic (AbstractCoordinator.java:407) 10:46:58 [DEBUG] AbstractCoordinator.sendJoinGroupRequest - Sending JoinGroup ((type: JoinGroupRequest, groupId=testOutputTopic, sessionTimeout=10000, rebalanceTimeout=300000, memberId=, protocolType=consumer, groupProtocols=org.apache.kafka.common.requests.JoinGroupRequest$ProtocolMetadata@4763c727)) to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:415) 10:46:58 [DEBUG] AbstractCoordinator$HeartbeatThread.run - Heartbeat thread for group testOutputTopic started (AbstractCoordinator.java:869) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node-2147483646.bytes-sent (Metrics.java:335) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node-2147483646.bytes-received (Metrics.java:335) 10:46:58 [DEBUG] Logging$class.debug - Accepted connection from /10.36.240.33:53397 on /10.36.240.33:9092 and assigned it to processor 1, sendBufferSize [actual|requested]: [102400|102400] recvBufferSize [actual|requested]: [102400|102400] (Logging.scala:54) 10:46:58 [DEBUG] Metrics.sensor - Added sensor with name node-2147483646.latency (Metrics.java:335) 10:46:58 [DEBUG] Logging$class.debug - Processor 1 listening to new connection from /10.36.240.33:53397 (Logging.scala:54) 10:46:58 [DEBUG] Selector.pollSelectionKeys - Created socket with SO_RCVBUF = 65536, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node 2147483646 (Selector.java:339) 10:46:58 [DEBUG] NetworkClient.handleConnections - Completed connection to node 2147483646. Fetching API versions. (NetworkClient.java:590) 10:46:58 [DEBUG] NetworkClient.handleInitiateApiVersionRequests - Initiating API versions fetch from node 2147483646. (NetworkClient.java:603) 10:46:58 [TRACE] NetworkClient.doSend - No version information found when sending message of type API_VERSIONS to node 2147483646. Assuming version 0. (NetworkClient.java:292) 10:46:58 [TRACE] NetworkClient.doSend - Sending {} to node 2147483646. (NetworkClient.java:316) 10:46:58 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=18,api_version=0,correlation_id=4,client_id=consumer-3} -- {} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 0 on Broker 1], Kafka request handler 0 on broker 1 handling request Request(1,10.36.240.33:9092-10.36.240.33:53397,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418910,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=18,api_version=0,correlation_id=4,client_id=consumer-3} -- {} from connection 10.36.240.33:9092-10.36.240.33:53397;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(1,Request(1,10.36.240.33:9092-10.36.240.33:53397,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418910,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@25b56372,SendAction) (Logging.scala:36) 10:46:58 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=18,api_version=0,correlation_id=4,client_id=consumer-3} -- {} from connection 10.36.240.33:9092-10.36.240.33:53397;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 18, received {error_code=0,api_versions=[{api_key=0,min_version=0,max_version=2},{api_key=1,min_version=0,max_version=3},{api_key=2,min_version=0,max_version=1},{api_key=3,min_version=0,max_version=2},{api_key=4,min_version=0,max_version=0},{api_key=5,min_version=0,max_version=0},{api_key=6,min_version=0,max_version=3},{api_key=7,min_version=1,max_version=1},{api_key=8,min_version=0,max_version=2},{api_key=9,min_version=0,max_version=2},{api_key=10,min_version=0,max_version=0},{api_key=11,min_version=0,max_version=1},{api_key=12,min_version=0,max_version=0},{api_key=13,min_version=0,max_version=0},{api_key=14,min_version=0,max_version=0},{api_key=15,min_version=0,max_version=0},{api_key=16,min_version=0,max_version=0},{api_key=17,min_version=0,max_version=0},{api_key=18,min_version=0,max_version=0},{api_key=19,min_version=0,max_version=1},{api_key=20,min_version=0,max_version=0}]} (NetworkClient.java:534) 10:46:58 [DEBUG] NetworkClient.handleApiVersionsResponse - Recorded API versions for node 2147483646: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 3 [usable: 3], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 2 [usable: 2], GroupCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 to 1 [usable: 1], DeleteTopics(20): 0 [usable: 0]) (NetworkClient.java:558) 10:46:58 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} to node 2147483646. (NetworkClient.java:316) 10:46:58 [TRACE] Logging$class.trace - Processor 1 received request : {api_key=11,api_version=1,correlation_id=3,client_id=consumer-3} -- {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [Kafka Request Handler 4 on Broker 1], Kafka request handler 4 on broker 1 handling request Request(1,10.36.240.33:9092-10.36.240.33:53397,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418915,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:46:58 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=11,api_version=1,correlation_id=3,client_id=consumer-3} -- {group_id=testOutputTopic,session_timeout=10000,rebalance_timeout=300000,member_id=,protocol_type=consumer,group_protocols=[{protocol_name=range,protocol_metadata=java.nio.HeapByteBuffer[pos=0 lim=27 cap=27]}]} from connection 10.36.240.33:9092-10.36.240.33:53397;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:46:58 [INFO ] Logging$class.info - [GroupCoordinator 1]: Preparing to restabilize group testOutputTopic with old generation 3 (Logging.scala:70) FROM NOW ON THE SERVER HANGS WHILE "Preparing to restabilize group testOutputTopic", NO "Stabilized group testOutputTopic" CAN BE FOUND ! ANY HEARTBEAT FOR GROUP testOutputTopic FROM THE CLIENT FAILS "since it is rebalancing", AND THE CONSUMER HANGS AS WELL 10:46:59 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Fetching log segment for partition testOutputTopic-0, offset 10, partition fetch size 1048576, remaining response limit 52428800, ignoring response/partition size limits (Logging.scala:36) 10:46:59 [TRACE] Logging$class.trace - Reading 1048576 bytes from offset 10 in log testOutputTopic-0 of length 1720 bytes (Logging.scala:36) 10:46:59 [TRACE] Logging$class.trace - [KafkaApi-1] Sending fetch response to client consumer-2 of 0 bytes (Logging.scala:36) 10:46:59 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(0,Request(0,10.36.240.33:9092-10.36.240.33:53391,Session(User:ANONYMOUS,/10.36.240.33),null,1491468418871,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.MultiSend@7fca6017,SendAction) (Logging.scala:36) 10:46:59 [TRACE] MultiSend.writeTo - Bytes written as part of multi-send call: 55, total bytes written so far: 55, expected bytes to write: 55 (MultiSend.java:88) 10:46:59 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=1,api_version=3,correlation_id=12,client_id=consumer-2} -- {replica_id=-1,max_wait_time=500,min_bytes=1,max_bytes=52428800,topics=[{topic=testOutputTopic,partitions=[{partition=0,fetch_offset=10,max_bytes=1048576}]}]} from connection 10.36.240.33:9092-10.36.240.33:53391;totalTime:504,requestQueueTime:1,localTime:1,remoteTime:501,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:46:59 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 1, for key 1, received {throttle_time_ms=0,responses=[{topic=testOutputTopic,partition_responses=[{partition_header={partition=0,error_code=0,high_watermark=10},record_set=[]}]}]} (NetworkClient.java:534) 10:47:00 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:00 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:00 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=14,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:00 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468420854,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:00 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=14,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:00 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 14 to client consumer-2. (Logging.scala:36) 10:47:00 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468420854,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6255e3c7,SendAction) (Logging.scala:36) 10:47:00 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=14,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:3,requestQueueTime:1,localTime:2,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:00 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:00 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:01 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:01 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:02 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:02 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:03 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:03 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:03 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:03 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:03 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:03 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:03 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:03 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=15,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:03 [TRACE] Logging$class.trace - [Kafka Request Handler 7 on Broker 1], Kafka request handler 7 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468423862,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:03 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=15,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:03 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 15 to client consumer-2. (Logging.scala:36) 10:47:03 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468423862,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@75270ca,SendAction) (Logging.scala:36) 10:47:03 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=15,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:03 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:03 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:06 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:06 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:06 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:06 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:06 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=16,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:06 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468426864,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:06 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=16,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:06 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 16 to client consumer-2. (Logging.scala:36) 10:47:06 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468426864,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@52312164,SendAction) (Logging.scala:36) 10:47:06 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=16,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:0,localTime:2,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:06 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:06 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:07 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:07 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:08 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'kafka-log-retention'. (Logging.scala:36) 10:47:08 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'kafka-log-flusher'. (Logging.scala:36) 10:47:08 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'kafka-recovery-point-checkpoint'. (Logging.scala:36) 10:47:08 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'kafka-delete-logs'. (Logging.scala:36) 10:47:08 [TRACE] Logging$class.trace - Completed execution of scheduled task 'kafka-delete-logs'. (Logging.scala:36) 10:47:08 [DEBUG] Logging$class.debug - Checking for dirty logs to flush... (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Beginning log cleanup... (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468417058 time since last flush: 10980 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416571 time since last flush: 11468 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416872 time since last flush: 11167 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416343 time since last flush: 11696 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416671 time since last flush: 11368 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on testOutputTopic flush interval 9223372036854775807 last flushed 1491468411115 time since last flush: 16924 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416450 time since last flush: 11589 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416553 time since last flush: 11487 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416304 time since last flush: 11736 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416660 time since last flush: 11380 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416942 time since last flush: 11098 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416390 time since last flush: 11650 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416831 time since last flush: 11209 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416992 time since last flush: 11048 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416283 time since last flush: 11757 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416848 time since last flush: 11192 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468417021 time since last flush: 11019 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416636 time since last flush: 11405 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416409 time since last flush: 11632 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416519 time since last flush: 11522 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416204 time since last flush: 11837 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416228 time since last flush: 11813 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416983 time since last flush: 11058 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416621 time since last flush: 11421 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416811 time since last flush: 11231 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416359 time since last flush: 11683 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416501 time since last flush: 11541 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416840 time since last flush: 11202 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416239 time since last flush: 11803 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416537 time since last flush: 11505 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416962 time since last flush: 11080 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416373 time since last flush: 11669 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416897 time since last flush: 11145 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468417036 time since last flush: 11006 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416923 time since last flush: 11119 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416328 time since last flush: 11714 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416791 time since last flush: 11251 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468417046 time since last flush: 10997 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416471 time since last flush: 11572 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416881 time since last flush: 11162 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416486 time since last flush: 11557 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416649 time since last flush: 11394 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416251 time since last flush: 11792 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468417011 time since last flush: 11032 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416822 time since last flush: 11221 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416859 time since last flush: 11184 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416608 time since last flush: 11436 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416429 time since last flush: 11615 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416972 time since last flush: 11072 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416952 time since last flush: 11092 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Checking if flush is needed on __consumer_offsets flush interval 9223372036854775807 last flushed 1491468416686 time since last flush: 11358 (Logging.scala:54) 10:47:08 [TRACE] Logging$class.trace - Completed execution of scheduled task 'kafka-log-flusher'. (Logging.scala:36) 10:47:08 [DEBUG] Logging$class.debug - Garbage collecting 'testOutputTopic-0' (Logging.scala:54) 10:47:08 [TRACE] Logging$class.trace - Completed execution of scheduled task 'kafka-recovery-point-checkpoint'. (Logging.scala:36) 10:47:08 [DEBUG] Logging$class.debug - Log cleanup completed. 0 files deleted in 0 seconds (Logging.scala:54) 10:47:08 [TRACE] Logging$class.trace - Completed execution of scheduled task 'kafka-log-retention'. (Logging.scala:36) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-13 topicPartition=__consumer_offsets-13. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-46 topicPartition=__consumer_offsets-46. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-9 topicPartition=__consumer_offsets-9. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-42 topicPartition=__consumer_offsets-42. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-21 topicPartition=__consumer_offsets-21. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-17 topicPartition=__consumer_offsets-17. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-30 topicPartition=__consumer_offsets-30. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-26 topicPartition=__consumer_offsets-26. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-5 topicPartition=__consumer_offsets-5. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-38 topicPartition=__consumer_offsets-38. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-1 topicPartition=__consumer_offsets-1. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-34 topicPartition=__consumer_offsets-34. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-16 topicPartition=__consumer_offsets-16. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-45 topicPartition=__consumer_offsets-45. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-12 topicPartition=__consumer_offsets-12. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-41 topicPartition=__consumer_offsets-41. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-24 topicPartition=__consumer_offsets-24. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-20 topicPartition=__consumer_offsets-20. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-49 topicPartition=__consumer_offsets-49. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-0 topicPartition=__consumer_offsets-0. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-29 topicPartition=__consumer_offsets-29. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-25 topicPartition=__consumer_offsets-25. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-8 topicPartition=__consumer_offsets-8. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-37 topicPartition=__consumer_offsets-37. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-4 topicPartition=__consumer_offsets-4. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-33 topicPartition=__consumer_offsets-33. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-15 topicPartition=__consumer_offsets-15. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-48 topicPartition=__consumer_offsets-48. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-11 topicPartition=__consumer_offsets-11. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-44 topicPartition=__consumer_offsets-44. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-23 topicPartition=__consumer_offsets-23. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-19 topicPartition=__consumer_offsets-19. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-32 topicPartition=__consumer_offsets-32. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-28 topicPartition=__consumer_offsets-28. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-7 topicPartition=__consumer_offsets-7. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-40 topicPartition=__consumer_offsets-40. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-3 topicPartition=__consumer_offsets-3. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-36 topicPartition=__consumer_offsets-36. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-47 topicPartition=__consumer_offsets-47. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-14 topicPartition=__consumer_offsets-14. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-43 topicPartition=__consumer_offsets-43. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-10 topicPartition=__consumer_offsets-10. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-22 topicPartition=__consumer_offsets-22. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-18 topicPartition=__consumer_offsets-18. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-31 topicPartition=__consumer_offsets-31. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-27 topicPartition=__consumer_offsets-27. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-39 topicPartition=__consumer_offsets-39. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-6 topicPartition=__consumer_offsets-6. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-35 topicPartition=__consumer_offsets-35. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-2 topicPartition=__consumer_offsets-2. Last clean offset=None now=1491468428069 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:08 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:08 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:08 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:08 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:08 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:09 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:09 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:09 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=17,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:09 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468429866,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:09 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=17,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:09 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 17 to client consumer-2. (Logging.scala:36) 10:47:09 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468429866,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2f2c190c,SendAction) (Logging.scala:36) 10:47:09 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=17,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:09 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:09 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:11 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:11 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:12 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:12 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:12 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:12 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:12 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=18,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:12 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468432868,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:12 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=18,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:12 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 18 to client consumer-2. (Logging.scala:36) 10:47:12 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468432868,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@62aa8cf5,SendAction) (Logging.scala:36) 10:47:12 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=18,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:12 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:12 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:13 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:13 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:13 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:13 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:13 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:15 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:15 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:15 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=19,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:15 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468435871,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:15 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=19,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:15 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 19 to client consumer-2. (Logging.scala:36) 10:47:15 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468435871,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@76ea644a,SendAction) (Logging.scala:36) 10:47:15 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=19,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:15 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:15 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:16 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:16 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:17 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:17 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:18 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:18 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:18 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:18 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:18 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:18 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:18 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:18 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=20,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:18 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468438873,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:18 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=20,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:18 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 20 to client consumer-2. (Logging.scala:36) 10:47:18 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468438873,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@2d8b51f2,SendAction) (Logging.scala:36) 10:47:18 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=20,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:18 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:18 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:21 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:21 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:21 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:21 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:21 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=21,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:21 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468441877,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:21 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=21,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:21 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 21 to client consumer-2. (Logging.scala:36) 10:47:21 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468441877,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6220c936,SendAction) (Logging.scala:36) 10:47:21 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=21,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:21 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:21 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:22 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:22 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-13 topicPartition=__consumer_offsets-13. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-46 topicPartition=__consumer_offsets-46. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-9 topicPartition=__consumer_offsets-9. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-42 topicPartition=__consumer_offsets-42. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-21 topicPartition=__consumer_offsets-21. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-17 topicPartition=__consumer_offsets-17. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-30 topicPartition=__consumer_offsets-30. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-26 topicPartition=__consumer_offsets-26. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-5 topicPartition=__consumer_offsets-5. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-38 topicPartition=__consumer_offsets-38. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-1 topicPartition=__consumer_offsets-1. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-34 topicPartition=__consumer_offsets-34. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-16 topicPartition=__consumer_offsets-16. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-45 topicPartition=__consumer_offsets-45. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-12 topicPartition=__consumer_offsets-12. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-41 topicPartition=__consumer_offsets-41. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-24 topicPartition=__consumer_offsets-24. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-20 topicPartition=__consumer_offsets-20. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-49 topicPartition=__consumer_offsets-49. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-0 topicPartition=__consumer_offsets-0. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-29 topicPartition=__consumer_offsets-29. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-25 topicPartition=__consumer_offsets-25. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-8 topicPartition=__consumer_offsets-8. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-37 topicPartition=__consumer_offsets-37. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-4 topicPartition=__consumer_offsets-4. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-33 topicPartition=__consumer_offsets-33. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-15 topicPartition=__consumer_offsets-15. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-48 topicPartition=__consumer_offsets-48. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-11 topicPartition=__consumer_offsets-11. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-44 topicPartition=__consumer_offsets-44. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-23 topicPartition=__consumer_offsets-23. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-19 topicPartition=__consumer_offsets-19. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-32 topicPartition=__consumer_offsets-32. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-28 topicPartition=__consumer_offsets-28. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-7 topicPartition=__consumer_offsets-7. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-40 topicPartition=__consumer_offsets-40. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-3 topicPartition=__consumer_offsets-3. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-36 topicPartition=__consumer_offsets-36. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-47 topicPartition=__consumer_offsets-47. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-14 topicPartition=__consumer_offsets-14. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-43 topicPartition=__consumer_offsets-43. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-10 topicPartition=__consumer_offsets-10. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-22 topicPartition=__consumer_offsets-22. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-18 topicPartition=__consumer_offsets-18. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-31 topicPartition=__consumer_offsets-31. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-27 topicPartition=__consumer_offsets-27. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-39 topicPartition=__consumer_offsets-39. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-6 topicPartition=__consumer_offsets-6. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-35 topicPartition=__consumer_offsets-35. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-2 topicPartition=__consumer_offsets-2. Last clean offset=None now=1491468443128 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:23 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:23 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:23 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:23 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:23 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:24 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:24 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:24 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=22,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:24 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468444880,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:24 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=22,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:24 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 22 to client consumer-2. (Logging.scala:36) 10:47:24 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468444880,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@545b2ca3,SendAction) (Logging.scala:36) 10:47:24 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=22,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:24 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:24 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:26 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:26 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:27 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:27 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:27 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:27 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:27 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=23,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:27 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468447883,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:27 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=23,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:27 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 23 to client consumer-2. (Logging.scala:36) 10:47:27 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468447883,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@71b6ca56,SendAction) (Logging.scala:36) 10:47:27 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=23,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:27 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:27 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:28 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:28 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:28 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:28 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:28 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:30 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:30 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:30 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=24,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:30 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468450885,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:30 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=24,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:30 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 24 to client consumer-2. (Logging.scala:36) 10:47:30 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468450885,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6c51153,SendAction) (Logging.scala:36) 10:47:30 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=24,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:30 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:30 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:31 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:31 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:32 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:32 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:33 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:33 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:33 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:33 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:33 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:33 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:33 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:33 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=25,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:33 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468453888,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:33 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=25,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:33 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 25 to client consumer-2. (Logging.scala:36) 10:47:33 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468453888,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@62c811c4,SendAction) (Logging.scala:36) 10:47:33 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=25,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:33 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:33 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:36 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:36 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:36 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:36 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:36 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=26,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:36 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468456890,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:36 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=26,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:36 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 26 to client consumer-2. (Logging.scala:36) 10:47:36 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468456890,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6ffb070a,SendAction) (Logging.scala:36) 10:47:36 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=26,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:36 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:36 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:37 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:37 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-13 topicPartition=__consumer_offsets-13. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-46 topicPartition=__consumer_offsets-46. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-9 topicPartition=__consumer_offsets-9. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-42 topicPartition=__consumer_offsets-42. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-21 topicPartition=__consumer_offsets-21. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-17 topicPartition=__consumer_offsets-17. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-30 topicPartition=__consumer_offsets-30. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-26 topicPartition=__consumer_offsets-26. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-5 topicPartition=__consumer_offsets-5. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-38 topicPartition=__consumer_offsets-38. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-1 topicPartition=__consumer_offsets-1. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-34 topicPartition=__consumer_offsets-34. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-16 topicPartition=__consumer_offsets-16. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-45 topicPartition=__consumer_offsets-45. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-12 topicPartition=__consumer_offsets-12. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-41 topicPartition=__consumer_offsets-41. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-24 topicPartition=__consumer_offsets-24. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-20 topicPartition=__consumer_offsets-20. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-49 topicPartition=__consumer_offsets-49. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-0 topicPartition=__consumer_offsets-0. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-29 topicPartition=__consumer_offsets-29. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-25 topicPartition=__consumer_offsets-25. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-8 topicPartition=__consumer_offsets-8. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-37 topicPartition=__consumer_offsets-37. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-4 topicPartition=__consumer_offsets-4. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-33 topicPartition=__consumer_offsets-33. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-15 topicPartition=__consumer_offsets-15. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-48 topicPartition=__consumer_offsets-48. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-11 topicPartition=__consumer_offsets-11. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-44 topicPartition=__consumer_offsets-44. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-23 topicPartition=__consumer_offsets-23. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-19 topicPartition=__consumer_offsets-19. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-32 topicPartition=__consumer_offsets-32. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-28 topicPartition=__consumer_offsets-28. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-7 topicPartition=__consumer_offsets-7. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-40 topicPartition=__consumer_offsets-40. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-3 topicPartition=__consumer_offsets-3. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-36 topicPartition=__consumer_offsets-36. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-47 topicPartition=__consumer_offsets-47. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-14 topicPartition=__consumer_offsets-14. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-43 topicPartition=__consumer_offsets-43. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-10 topicPartition=__consumer_offsets-10. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-22 topicPartition=__consumer_offsets-22. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-18 topicPartition=__consumer_offsets-18. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-31 topicPartition=__consumer_offsets-31. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-27 topicPartition=__consumer_offsets-27. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-39 topicPartition=__consumer_offsets-39. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-6 topicPartition=__consumer_offsets-6. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-35 topicPartition=__consumer_offsets-35. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-2 topicPartition=__consumer_offsets-2. Last clean offset=None now=1491468458138 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:38 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:38 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:38 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:38 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:38 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:39 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:39 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:39 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=27,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:39 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468459891,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:39 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=27,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:39 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 27 to client consumer-2. (Logging.scala:36) 10:47:39 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468459891,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@285884d9,SendAction) (Logging.scala:36) 10:47:39 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=27,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:39 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:39 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:41 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:41 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:42 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:42 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:42 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:42 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:42 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=28,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:42 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468462893,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:42 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=28,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:42 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 28 to client consumer-2. (Logging.scala:36) 10:47:42 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468462893,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@6b95708,SendAction) (Logging.scala:36) 10:47:42 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=28,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:42 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:42 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:43 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:43 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:43 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:43 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:43 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:45 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:45 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:45 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=29,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:45 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468465901,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:45 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=29,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:45 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 29 to client consumer-2. (Logging.scala:36) 10:47:45 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468465901,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@245d202a,SendAction) (Logging.scala:36) 10:47:45 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=29,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:46 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:46 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:46 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:46 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:47 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:47 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:48 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:48 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:48 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:48 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:48 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:48 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:48 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:48 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=30,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:48 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468468903,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:48 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=30,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:48 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 30 to client consumer-2. (Logging.scala:36) 10:47:48 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468468903,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@13a9a0fc,SendAction) (Logging.scala:36) 10:47:48 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=30,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:3,requestQueueTime:1,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:49 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:49 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:51 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:51 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:51 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:51 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:51 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=31,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:51 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468471906,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:51 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=31,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:51 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 31 to client consumer-2. (Logging.scala:36) 10:47:51 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468471906,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7fbb801c,SendAction) (Logging.scala:36) 10:47:51 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=31,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:0,localTime:1,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:52 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:52 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:52 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:52 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-13 topicPartition=__consumer_offsets-13. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-46 topicPartition=__consumer_offsets-46. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-9 topicPartition=__consumer_offsets-9. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-42 topicPartition=__consumer_offsets-42. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-21 topicPartition=__consumer_offsets-21. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-17 topicPartition=__consumer_offsets-17. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-30 topicPartition=__consumer_offsets-30. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-26 topicPartition=__consumer_offsets-26. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-5 topicPartition=__consumer_offsets-5. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-38 topicPartition=__consumer_offsets-38. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-1 topicPartition=__consumer_offsets-1. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-34 topicPartition=__consumer_offsets-34. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-16 topicPartition=__consumer_offsets-16. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-45 topicPartition=__consumer_offsets-45. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-12 topicPartition=__consumer_offsets-12. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-41 topicPartition=__consumer_offsets-41. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-24 topicPartition=__consumer_offsets-24. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-20 topicPartition=__consumer_offsets-20. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-49 topicPartition=__consumer_offsets-49. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-0 topicPartition=__consumer_offsets-0. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-29 topicPartition=__consumer_offsets-29. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-25 topicPartition=__consumer_offsets-25. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-8 topicPartition=__consumer_offsets-8. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-37 topicPartition=__consumer_offsets-37. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-4 topicPartition=__consumer_offsets-4. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-33 topicPartition=__consumer_offsets-33. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-15 topicPartition=__consumer_offsets-15. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-48 topicPartition=__consumer_offsets-48. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-11 topicPartition=__consumer_offsets-11. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-44 topicPartition=__consumer_offsets-44. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-23 topicPartition=__consumer_offsets-23. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-19 topicPartition=__consumer_offsets-19. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-32 topicPartition=__consumer_offsets-32. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-28 topicPartition=__consumer_offsets-28. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-7 topicPartition=__consumer_offsets-7. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-40 topicPartition=__consumer_offsets-40. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-3 topicPartition=__consumer_offsets-3. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-36 topicPartition=__consumer_offsets-36. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-47 topicPartition=__consumer_offsets-47. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-14 topicPartition=__consumer_offsets-14. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-43 topicPartition=__consumer_offsets-43. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-10 topicPartition=__consumer_offsets-10. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-22 topicPartition=__consumer_offsets-22. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-18 topicPartition=__consumer_offsets-18. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-31 topicPartition=__consumer_offsets-31. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-27 topicPartition=__consumer_offsets-27. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-39 topicPartition=__consumer_offsets-39. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-6 topicPartition=__consumer_offsets-6. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-35 topicPartition=__consumer_offsets-35. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [DEBUG] Logging$class.debug - Finding range of cleanable offsets for log=__consumer_offsets-2 topicPartition=__consumer_offsets-2. Last clean offset=None now=1491468473151 => firstDirtyOffset=0 firstUncleanableOffset=0 activeSegment.baseOffset=0 (Logging.scala:54) 10:47:53 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:53 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:53 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:53 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:53 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:54 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:54 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:54 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=32,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:54 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468474908,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:54 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=32,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:54 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 32 to client consumer-2. (Logging.scala:36) 10:47:54 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468474908,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@337689f4,SendAction) (Logging.scala:36) 10:47:54 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=32,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:1,sendTime:0,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:55 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:55 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:56 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:56 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:57 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:57 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:47:57 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:47:57 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:47:57 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=33,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:47:57 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468477911,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:47:57 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=33,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:47:57 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 33 to client consumer-2. (Logging.scala:36) 10:47:57 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468477911,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@7e67a3c0,SendAction) (Logging.scala:36) 10:47:57 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=33,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:2,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:0,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:47:58 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:47:58 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:47:58 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:58 [TRACE] Logging$class.trace - [Replica Manager on Broker 1]: Evaluating ISR list of partitions to see which replicas can be removed from the ISR (Logging.scala:36) 10:47:58 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-expiration'. (Logging.scala:36) 10:47:58 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:47:58 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:48:00 [DEBUG] AbstractCoordinator.sendHeartbeatRequest - Sending Heartbeat request for group testOutputTopic to coordinator ISI050.utenze.BANKIT.IT:9092 (id: 2147483646 rack: null) (AbstractCoordinator.java:704) 10:48:00 [TRACE] NetworkClient.doSend - Sending {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} to node 2147483646. (NetworkClient.java:316) 10:48:00 [TRACE] Logging$class.trace - Processor 2 received request : {api_key=12,api_version=0,correlation_id=34,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} (Logging.scala:36) 10:48:00 [TRACE] Logging$class.trace - [Kafka Request Handler 2 on Broker 1], Kafka request handler 2 on broker 1 handling request Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468480913,ListenerName(PLAINTEXT),PLAINTEXT) (Logging.scala:36) 10:48:00 [TRACE] Logging$class.trace - [KafkaApi-1] Handling request:{api_key=12,api_version=0,correlation_id=34,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;securityProtocol:PLAINTEXT,principal:User:ANONYMOUS (Logging.scala:36) 10:48:00 [TRACE] Logging$class.trace - [KafkaApi-1] Sending heartbeat response {error_code=27} for correlation id 34 to client consumer-2. (Logging.scala:36) 10:48:00 [TRACE] Logging$class.trace - Socket server received response to send, registering for write and sending data: Response(2,Request(2,10.36.240.33:9092-10.36.240.33:53390,Session(User:ANONYMOUS,/10.36.240.33),null,1491468480913,ListenerName(PLAINTEXT),PLAINTEXT),org.apache.kafka.common.network.NetworkSend@201742b0,SendAction) (Logging.scala:36) 10:48:00 [TRACE] RequestChannel$Request.updateRequestMetrics - Completed request:{api_key=12,api_version=0,correlation_id=34,client_id=consumer-2} -- {group_id=testOutputTopic,group_generation_id=3,member_id=consumer-2-51f84513-1279-4b8b-b3a0-6626277b7f56} from connection 10.36.240.33:9092-10.36.240.33:53390;totalTime:3,requestQueueTime:1,localTime:0,remoteTime:0,responseQueueTime:1,sendTime:1,securityProtocol:PLAINTEXT,principal:User:ANONYMOUS,listener:PLAINTEXT (RequestChannel.scala:157) 10:48:01 [TRACE] NetworkClient.handleCompletedReceives - Completed receive from node 2147483646, for key 12, received {error_code=27} (NetworkClient.java:534) 10:48:01 [DEBUG] AbstractCoordinator$HeartbeatResponseHandler.handle - Attempt to heartbeat failed for group testOutputTopic since it is rebalancing. (AbstractCoordinator.java:726) 10:48:01 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:48:01 [TRACE] Logging$class.trace - Completed execution of scheduled task 'isr-change-propagation'. (Logging.scala:36) 10:48:02 [TRACE] Logging$class.trace - Beginning execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36) 10:48:02 [TRACE] Logging$class.trace - Completed execution of scheduled task 'highwatermark-checkpoint'. (Logging.scala:36)