Uploaded image for project: 'Kafka'
  1. Kafka
  2. KAFKA-1718

"Message Size Too Large" error when only small messages produced with Snappy


    • Type: Bug
    • Status: Open
    • Priority: Critical
    • Resolution: Unresolved
    • Affects Version/s:
    • Fix Version/s: None
    • Component/s: core
    • Labels:


      I'm the primary author of the Go bindings, and while I originally received this as a bug against my bindings, I'm coming to the conclusion that it's a bug in the broker somehow.

      Specifically, take a look at the last two kafka packets in the following packet capture: https://dl.dropboxusercontent.com/u/171647/kafka.pcapng (you will need a trunk build of Wireshark to fully decode the kafka part of the packets).

      The produce request contains two partitions on one topic. Each partition has one message set (sizes 977205 bytes and 967362 bytes respectively). Each message set is a sequential collection of snappy-compressed messages, each message of size 46899. When uncompressed, each message contains a message set of 999600 bytes, containing a sequence of uncompressed 1024-byte messages.

      However, the broker responds to this with a MessageSizeTooLarge error, full stacktrace from the broker logs being:
      kafka.common.MessageSizeTooLargeException: Message size is 1070127 bytes which exceeds the maximum configured message size of 1000012.
      at kafka.log.Log$$anonfun$append$1.apply(Log.scala:267)
      at kafka.log.Log$$anonfun$append$1.apply(Log.scala:265)
      at scala.collection.Iterator$class.foreach(Iterator.scala:727)
      at kafka.utils.IteratorTemplate.foreach(IteratorTemplate.scala:32)
      at kafka.log.Log.append(Log.scala:265)
      at kafka.cluster.Partition.appendMessagesToLeader(Partition.scala:354)
      at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:376)
      at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:366)
      at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
      at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
      at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98)
      at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98)
      at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226)
      at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39)
      at scala.collection.mutable.HashMap.foreach(HashMap.scala:98)
      at scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
      at scala.collection.AbstractTraversable.map(Traversable.scala:105)
      at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:366)
      at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:292)
      at kafka.server.KafkaApis.handle(KafkaApis.scala:185)
      at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:42)
      at java.lang.Thread.run(Thread.java:695)

      Since as far as I can tell none of the sizes in the actual produced packet exceed the defined maximum, I can only assume that the broker is miscalculating something somewhere and throwing the exception improperly.

      This issue can be reliably reproduced using an out-of-the-box binary download of and the following gist: https://gist.github.com/eapache/ce0f15311c605a165ce7 (you will need to use the `producer-ng` branch of the Sarama library).

      I am happy to provide any more information you might need, or to do relevant experiments etc.




            • Assignee:
              eapache Evan Huus
            • Votes:
              3 Vote for this issue
              13 Start watching this issue


              • Created: