[ 
https://issues.apache.org/jira/browse/KAFKA-1196?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jun Rao updated KAFKA-1196:
---------------------------

    Labels: newbie  (was: )

To really support fetching more than 2GB of data in a single fetch response 
requires wire protocol change. One simple thing that we can do immediately is 
to sanity check the response size. If it's more than 2GB, just throw an error 
to the client.

> java.lang.IllegalArgumentException Buffer.limit on FetchResponse.scala + 33
> ---------------------------------------------------------------------------
>
>                 Key: KAFKA-1196
>                 URL: https://issues.apache.org/jira/browse/KAFKA-1196
>             Project: Kafka
>          Issue Type: Bug
>          Components: consumer
>    Affects Versions: 0.8.0
>         Environment: running java 1.7, linux and kafka compiled against scala 
> 2.9.2
>            Reporter: Gerrit Jansen van Vuuren
>            Assignee: Neha Narkhede
>            Priority: Blocker
>              Labels: newbie
>             Fix For: 0.9.0
>
>
> I have 6 topics each with 8 partitions spread over 4 kafka servers.
> the servers are 24 core 72 gig ram.
> While consuming from the topics I get an IlegalArgumentException and all 
> consumption stops, the error keeps on throwing.
> I've tracked it down to FectchResponse.scala line 33
> The error happens when the FetchResponsePartitionData object's readFrom 
> method calls:
> messageSetBuffer.limit(messageSetSize)
> I put in some debug code the the messageSetSize is 671758648, while the 
> buffer.capacity() gives 155733313, for some reason the buffer is smaller than 
> the required message size.
> I don't know the consumer code enough to debug this. It doesn't matter if 
> compression is used or not.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)

Reply via email to