Any error in state-change.log? Also, are you using the latest code in the
0.8 branch?

Thanks,

Jun


On Wed, Jun 12, 2013 at 9:27 AM, Alexandre Rodrigues <
alexan...@blismedia.com> wrote:

> Hi Jun,
>
> Thanks for your prompt answer. The producer yields those errors in the
> beginning, so I think the topic metadata refresh has nothing to do with it.
>
> The problem is one of the brokers isn't leader on any partition assigned to
> it and because topics were created with a replication factor of 1, the
> producer will never connect to that broker at all. What I don't understand
> is why doesn't the broker assume the lead of those partitions.
>
> I deleted all the topics and tried now with a replication factor of two
>
> topic: A  partition: 0    leader: 1       replicas: 1,0   isr: 1
> topic: A  partition: 1    leader: 0       replicas: 0,1   isr: 0,1
> topic: B partition: 0    leader: 0       replicas: 0,1   isr: 0,1
> topic: B partition: 1    leader: 1       replicas: 1,0   isr: 1
> topic: C      partition: 0    leader: 1       replicas: 1,0   isr: 1
> topic: C      partition: 1    leader: 0       replicas: 0,1   isr: 0,1
>
>
> Now producer doesn't yield errors. However, one of the brokers ( broker 0 )
> generates lots of lines like this:
>
> [2013-06-12 16:19:41,805] WARN [KafkaApi-0] Produce request with
> correlation id 404999 from client  on partition [B,0] failed due to
> Partition [B,0] doesn't exist on 0 (kafka.server.KafkaApis)
>
> There should be a replica there, so I don't know why it complains about
> that message.
>
> Have you ever found anything like this?
>
>
>
> On 12 June 2013 16:27, Jun Rao <jun...@gmail.com> wrote:
>
> > If the leaders exist in both brokers, the producer should be able to
> > connect to both of them, assuming you don't provide any key when sending
> > the data. Could you try restarting the producer? If there has been broker
> > failures, it may take topic.metadata.refresh.interval.ms for the
> producer
> > to pick up the newly available partitions (see
> > http://kafka.apache.org/08/configuration.html for details).
> >
> > Thanks,
> >
> > Jun
> >
> >
> > On Wed, Jun 12, 2013 at 8:01 AM, Alexandre Rodrigues <
> > alexan...@blismedia.com> wrote:
> >
> > > Hi,
> > >
> > > I have a Kafka 0.8 cluster with two nodes connected to three ZKs, with
> > the
> > > same configuration but the brokerId (one is 0 and the other 1). I
> created
> > > three topics A, B and C with 4 partitions and a replication factor of
> 1.
> > My
> > > idea was to have 2 partitions per topic in each broker. However, when I
> > > connect a producer, I can't have both brokers to write at the same time
> > and
> > > I don't know what's going on.
> > >
> > > My server.config has the following entries:
> > >
> > > auto.create.topics.enable=true
> > > num.partitions=2
> > >
> > >
> > > When I run bin/kafka-list-topic.sh --zookeeper localhost:2181   I get
> the
> > > following partition leader assignments:
> > >
> > > topic: A  partition: 0    leader: 1       replicas: 1     isr: 1
> > > topic: A  partition: 1    leader: 0       replicas: 0     isr: 0
> > > topic: A  partition: 2    leader: 1       replicas: 1     isr: 1
> > > topic: A  partition: 3    leader: 0       replicas: 0     isr: 0
> > > topic: B partition: 0    leader: 0       replicas: 0     isr: 0
> > > topic: B partition: 1    leader: 1       replicas: 1     isr: 1
> > > topic: B partition: 2    leader: 0       replicas: 0     isr: 0
> > > topic: B partition: 3    leader: 1       replicas: 1     isr: 1
> > > topic: C      partition: 0    leader: 0       replicas: 0     isr: 0
> > > topic: C      partition: 1    leader: 1       replicas: 1     isr: 1
> > > topic: C      partition: 2    leader: 0       replicas: 0     isr: 0
> > > topic: C      partition: 3    leader: 1       replicas: 1     isr: 1
> > >
> > >
> > > I've forced reassignment using the kafka-reassign-partitions tool with
> > the
> > > following JSON:
> > >
> > > {"partitions":  [
> > >    {"topic": "A", "partition": 1, "replicas": [0] },
> > >    {"topic": "A", "partition": 3, "replicas": [0] },
> > >    {"topic": "A", "partition": 0, "replicas": [1] },
> > >    {"topic": "A", "partition": 2, "replicas": [1] },
> > >    {"topic": "B", "partition": 1, "replicas": [0] },
> > >    {"topic": "B", "partition": 3, "replicas": [0] },
> > >    {"topic": "B", "partition": 0, "replicas": [1] },
> > >    {"topic": "B", "partition": 2, "replicas": [1] },
> > >    {"topic": "C", "partition": 0, "replicas": [0] },
> > >    {"topic": "C", "partition": 1, "replicas": [1] },
> > >    {"topic": "C", "partition": 2, "replicas": [0] },
> > >    {"topic": "C", "partition": 3, "replicas": [1] }
> > > ]}
> > >
> > > After reassignment, I've restarted producer and nothing worked. I've
> > tried
> > > also to restart both brokers and producer and nothing.
> > >
> > > The producer contains this logs:
> > >
> > > 2013-06-12 14:48:46,467] WARN Error while fetching metadata
>  partition
> > 0
> > >     leader: none    replicas:       isr:    isUnderReplicated: false
> for
> > > topic partition [C,0]: [class kafka.common.LeaderNotAvailableException]
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2013-06-12 14:48:46,467] WARN Error while fetching metadata
> >  partition 0
> > >     leader: none    replicas:       isr:    isUnderReplicated: false
> for
> > > topic partition [C,0]: [class kafka.common.LeaderNotAvailableException]
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2013-06-12 14:48:46,468] WARN Error while fetching metadata
> >  partition 2
> > >     leader: none    replicas:       isr:    isUnderReplicated: false
> for
> > > topic partition [C,2]: [class kafka.common.LeaderNotAvailableException]
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2013-06-12 14:48:46,468] WARN Error while fetching metadata
> >  partition 2
> > >     leader: none    replicas:       isr:    isUnderReplicated: false
> for
> > > topic partition [C,2]: [class kafka.common.LeaderNotAvailableException]
> > > (kafka.producer.BrokerPartitionInfo)
> > >
> > >
> > > And sometimes lines like this:
> > >
> > > [2013-06-12 14:55:37,339] WARN Error while fetching metadata
> > > [{TopicMetadata for topic B ->
> > > No partition metadata for topic B due to
> > > kafka.common.LeaderNotAvailableException}] for topic [B]: class
> > > kafka.common.LeaderNotAvailableException
> > >  (kafka.producer.BrokerPartitionInfo)
> > >
> > >
> > > Do you guys have any idea what's going on?
> > >
> > > Thanks in advance,
> > > Alex
> > >
> > > --
> > >
> > > @BlisMedia <http://twitter.com/BlisMedia>
> > >
> > > www.blismedia.com <http://blismedia.com>
> > >
> > > This email and any attachments to it may be confidential and are
> intended
> > > solely
> > > for the use of the individual to whom it is addressed. Any views or
> > > opinions
> > > expressed are solely those of the author and do not necessarily
> represent
> > > those of BlisMedia Ltd, a company registered in England and Wales with
> > > registered number 06455773. Its registered office is 3rd Floor, 101 New
> > > Cavendish St, London, W1W 6XH, United Kingdom.
> > >
> > > If you are not the intended recipient of this email, you must neither
> > take
> > > any action based upon its contents, nor copy or show it to anyone.
> Please
> > > contact the sender if you believe you have received this email in
> error.
> > >
> >
>
> --
>
> @BlisMedia <http://twitter.com/BlisMedia>
>
> www.blismedia.com <http://blismedia.com>
>
> This email and any attachments to it may be confidential and are intended
> solely
> for the use of the individual to whom it is addressed. Any views or
> opinions
> expressed are solely those of the author and do not necessarily represent
> those of BlisMedia Ltd, a company registered in England and Wales with
> registered number 06455773. Its registered office is 3rd Floor, 101 New
> Cavendish St, London, W1W 6XH, United Kingdom.
>
> If you are not the intended recipient of this email, you must neither take
> any action based upon its contents, nor copy or show it to anyone. Please
> contact the sender if you believe you have received this email in error.
>

Reply via email to