Hi Prashant,

I tried a local test using a very short topic.metadata.refresh.interval.ms
on the producer. The server had two partitions and both of them appended
data. Could you check if you have set the
topic.metadata.refresh.interval.ms on your producer to a very high value?

Swapnil

On 9/13/13 8:46 PM, "Jun Rao" <jun...@gmail.com> wrote:

>Without fixing KAFKA-1017, the issue is that the producer will maintain a
>socket connection per min(#partitions, #brokers). If you have lots of
>producers, the open file handlers on the broker could be an issue.
>
>So, what KAFKA-1017 fixes is to pick a random partition and stick to it
>for
>a configurable amount of time, and then switch to another random
>partition.
>This is the behavior in 0.7 when a load balancer is used and reduces #
>socket connections significantly.
>
>The issue you are reporting seems like a bug though. Which revision in 0.8
>are you using?
>
>Thanks,
>
>Jun
>
>
>On Fri, Sep 13, 2013 at 8:28 PM, prashant amar <amasin...@gmail.com>
>wrote:
>
>> Hi Guozhang, Joe, Drew
>>
>> In our case we have been running for the past 3 weeks and it has been
>> consistently writing only to to the first partition. The rest of the
>> partitions have empty index files.
>>
>> Not sure if I am hitting any issue here.
>>
>> I am using  offset checker as my barometer. Also introspect r&d the
>>folder
>> and it indicates the same.
>>
>> On Friday, September 13, 2013, Guozhang Wang wrote:
>>
>> > Hello Joe,
>> >
>> > The reason we make the producers to produce to a fixed partition for
>>each
>> > metadata-refresh interval are the following:
>> >
>> > https://issues.apache.org/jira/browse/KAFKA-1017
>> >
>> > https://issues.apache.org/jira/browse/KAFKA-959
>> >
>> > So in a word the randomness is still preserved but within one
>> > metadata-refresh interval the assignment is fixed.
>> >
>> > I agree that the document should be updated accordingly.
>> >
>> > Guozhang
>> >
>> >
>> > On Fri, Sep 13, 2013 at 1:48 PM, Joe Stein <crypt...@gmail.com> wrote:
>> >
>> > > Isn't this a bug?
>> > >
>> > > I don't see why we would want users to have to code and generate
>>random
>> > > partition keys to randomly distributed the data to partitions, that
>>is
>> > > Kafka's job isn't it?
>> > >
>> > > Or if supplying a null value tell the user this is not supported
>>(throw
>> > > exception) in KeyedMessage like we do for topic and not treat null
>>as a
>> > key
>> > > to hash?
>> > >
>> > > My preference is to put those three lines back in and let key be
>>null
>> and
>> > > give folks randomness unless its not a bug and there is a good
>>reason
>> for
>> > > it?
>> > >
>> > > Is there something about
>> > > https://issues.apache.org/jira/browse/KAFKA-691that requires the
>>lines
>> > > taken out? I haven't had a chance to look through
>> > > it yet
>> > >
>> > > My thought is a new person coming in they would expect to see the
>> > > partitions filling up in a round robin fashion as our docs says and
>> > unless
>> > > we force them in the API to know they have to-do this or give them
>>the
>> > > ability for this to happen when passing nothing in
>> > >
>> > > /*******************************************
>> > >  Joe Stein
>> > >  Founder, Principal Consultant
>> > >  Big Data Open Source Security LLC
>> > >  http://www.stealth.ly
>> > >  Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop>
>> > > ********************************************/
>> > >
>> > >
>> > > On Fri, Sep 13, 2013 at 4:17 PM, Drew Goya <d...@gradientx.com>
>>wrote:
>> > >
>> > > > I ran into this problem as well Prashant.  The default partition
>>key
>> > was
>> > > > recently changed:
>> > > >
>> > > >
>> > > >
>> > >
>> >
>> 
>>https://github.com/apache/kafka/commit/b71e6dc352770f22daec0c9a3682138666
>>f032be
>> > > >
>> > > > It no longer assigns a random partition to data with a null
>>partition
>> > > key.
>> > > >  I had to change my code to generate random partition keys to get
>>the
>> > > > randomly distributed behavior the producer used to have.
>> > > >
>> > > >
>> > > > On Fri, Sep 13, 2013 at 11:42 AM, prashant amar
>><amasin...@gmail.com
>> >
>> > > > wrote:
>> > > >
>> > > > > Thanks Neha
>> > > > >
>> > > > > I will try applying this property and circle back.
>> > > > >
>> > > > > Also, I have been attempting to execute
>>kafka-producer-perf-test.sh
>> > > and I
>> > > > > receive the following error
>> > > > >
>> > > > >        Error: Could not find or load main class
>> > > > > kafka.perf.ProducerPerformance
>> > > > >
>> > > > > I am running against 0.8.0-beta1
>> > > > >
>> > > > > Seems like perf is a separate project in the workspace.
>> > > > >
>> > > > > Does sbt package-assembly bundle the perf jar as well?
>> > > > >
>> > > > > Neither producer-perf-test not consumer-test are working with
>>this
>> > > build
>> > > > >
>> > > > >
>> > > > >
>> > > > > On Fri, Sep 13, 2013 at 9:56 AM, Neha Narkhede <
>> > > neha.narkh...@gmail.com
>> > > > > >wrote:
>> > > > >
>> > > > > > As Jun suggested, one reason could be that the
>> > > > > > topic.metadata.refresh.interval.ms is too high. Did you
>>observe
>> if
>> > > the
>> > > > > > distribution improves after
>>topic.metadata.refresh.interval.mshas
>> > > > > passed
>> > > > > > ?
>> > > > > >
>> > > > > > Thanks
>> > > > > > Neha
>> > > > > >
>> > > > > >
>> > > > > > On Fri, Sep 13, 2013 at 4:47 AM, prashant amar <
>> > amasin...@gmail.com>
>> > > > > > wrote:
>> > > > > >
>> > > > > --
>> > -- Guozhang
>> >
>>

Reply via email to