I was able to make this error disappear by upgrading my client library from 0.10.0.0 to 0.10.0.1
On Wed, Jan 18, 2017 at 10:40 AM, Ryan Thompson <ryan.thomp...@uptake.com> wrote: > Hello, > > I'm attempting to upgrade an application from 0.8 to 0.10 broker / client > libs, and integrate kafka streams. I am currently using the following > producer / consumer configs: > > Producer: > > Properties props = new Properties(); > props.put("bootstrap.servers", brokerList); > props = ProducerConfig.addSerializerToConfig(props, > Serdes.String().serializer(), Serdes.String().serializer()); > return props; > > > Consumer (kafka streams Processor) > > Properties streamsSettings = new Properties(); > streamsSettings.put("bootstrap.servers", brokerList); > streamsSettings.put("application.id", consumerGroupId); > streamsSettings.put("key.serde", Serdes.StringSerde.class. > getName()); > streamsSettings.put("value.serde", Serdes.StringSerde.class. > getName()); > StreamsConfig config = new StreamsConfig(streamsSettings); > > I'm running a 0.10 broker. However, when I publish a message, I see the > following error on the consumer side: > > org.apache.kafka.common.KafkaException: Error deserializing key/value for > partition <topic name>-0 at offset 0 > at org.apache.kafka.clients.consumer.internals.Fetcher. > parseRecord(Fetcher.java:665) > at org.apache.kafka.clients.consumer.internals.Fetcher. > handleFetchResponse(Fetcher.java:593) > at org.apache.kafka.clients.consumer.internals.Fetcher. > access$000(Fetcher.java:71) > at org.apache.kafka.clients.consumer.internals.Fetcher$1. > onSuccess(Fetcher.java:142) > at org.apache.kafka.clients.consumer.internals.Fetcher$1. > onSuccess(Fetcher.java:139) > at org.apache.kafka.clients.consumer.internals.RequestFuture.fireSuccess( > RequestFuture.java:133) > at org.apache.kafka.clients.consumer.internals.RequestFuture.complete( > RequestFuture.java:107) > at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient$ > RequestFutureCompletionHandler.onComplete(ConsumerNetworkClient.java:426) > at org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:278) > at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient. > clientPoll(ConsumerNetworkClient.java:360) > at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll( > ConsumerNetworkClient.java:224) > at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll( > ConsumerNetworkClient.java:201) > at org.apache.kafka.clients.consumer.KafkaConsumer. > pollOnce(KafkaConsumer.java:998) > at org.apache.kafka.clients.consumer.KafkaConsumer.poll( > KafkaConsumer.java:937) > at org.apache.kafka.streams.processor.internals.StreamThread.runLoop( > StreamThread.java:295) > at org.apache.kafka.streams.processor.internals. > StreamThread.run(StreamThread.java:218) > Caused by: java.lang.IllegalArgumentException: null > at java.nio.Buffer.limit(Buffer.java:275) > at org.apache.kafka.common.record.Record.sliceDelimited(Record.java:392) > at org.apache.kafka.common.record.Record.key(Record.java:376) > at org.apache.kafka.clients.consumer.internals.Fetcher. > parseRecord(Fetcher.java:650) > ... 15 common frames omitted > > > What could be causing this error? I see the same error if I attempt to > use the kafka bin tools to consume from the topic. > > Thanks, > Ryan >