The easiest way is to wipe out both ZK and kafka data and start from
scratch.

Thanks,

Jun

On Fri, Mar 22, 2013 at 6:51 AM, Jason Huang <jason.hu...@icare.com> wrote:

> Thanks Jun.
>
> I have built the new kafka version and start the services. You
> mentioned that ZK data structure has been changed - does that mean we
> can't reload the previous messages from current log files? I actually
> tried to copy the log files (.logs and .index) to the new kafka
> instance but get the same "topic doesn't exist" error after running
> the new kafka services.
>
> Any comments on how I might be able to recover previous messages?
>
> thanks
>
> Jason
>
> On Wed, Mar 20, 2013 at 12:15 PM, Jun Rao <jun...@gmail.com> wrote:
> > The latest version of 0.8 can be found in the 0.8 branch, not trunk.
> >
> > Thanks,
> >
> > Jun
> >
> > On Wed, Mar 20, 2013 at 7:47 AM, Jason Huang <jason.hu...@icare.com>
> wrote:
> >
> >> The 0.8 version I use was built from trunk last Dec. Since then, this
> >> error happened 3 times. Each time we had to remove all the ZK and
> >> Kafka log data and restart the services.
> >>
> >> I will try newer versions with more recent patches and keep monitoring
> it.
> >>
> >> thanks!
> >>
> >> Jason
> >>
> >> On Wed, Mar 20, 2013 at 10:39 AM, Jun Rao <jun...@gmail.com> wrote:
> >> > Ok, so you are using the same broker id. What the error is saying is
> that
> >> > broker 1 doesn't seem to be up.
> >> >
> >> > Not sure what revision of 0.8 you are using. Could you try the latest
> >> > revision in 0.8 and see if the problem still exists? You may have to
> wipe
> >> > out all ZK and Kafka data first since some ZK data structures have
> been
> >> > rename a few weeks ago.
> >> >
> >> > Thanks,
> >> >
> >> > Jun
> >> >
> >> > On Wed, Mar 20, 2013 at 6:57 AM, Jason Huang <jason.hu...@icare.com>
> >> wrote:
> >> >
> >> >> I restarted the zookeeper server first, then broker. It's the same
> >> >> instance of kafka 0.8 and I am using the same config file. In
> >> >> server.properties I have: brokerid=1
> >> >>
> >> >> Is that sufficient to ensure the broker get restarted with the same
> >> >> broker id as before?
> >> >>
> >> >> thanks,
> >> >>
> >> >> Jason
> >> >>
> >> >> On Wed, Mar 20, 2013 at 12:30 AM, Jun Rao <jun...@gmail.com> wrote:
> >> >> > Did the broker get restarted with the same broker id?
> >> >> >
> >> >> > Thanks,
> >> >> >
> >> >> > Jun
> >> >> >
> >> >> > On Tue, Mar 19, 2013 at 1:34 PM, Jason Huang <
> jason.hu...@icare.com>
> >> >> wrote:
> >> >> >
> >> >> >> Hello,
> >> >> >>
> >> >> >> My kafka (0.8) server went down today for unknown reason and when
> I
> >> >> >> restarted both zookeeper and kafka server I got the following
> error
> >> at
> >> >> >> the kafka server log:
> >> >> >>
> >> >> >> [2013-03-19 13:39:16,131] INFO [Partition state machine on
> Controller
> >> >> >> 1]: Invoking state change to OnlinePartition for partitions
> >> >> >> (kafka.controller.PartitionStateMachine)
> >> >> >> [2013-03-19 13:39:16,262] INFO [Partition state machine on
> Controller
> >> >> >> 1]: Electing leader for partition
> >> >> >> [topic_a937ac27-1883-4ca0-95bc-c9a740d08947, 0]
> >> >> >> (kafka.controller.PartitionStateMachine)
> >> >> >> [2013-03-19 13:39:16,451] ERROR [Partition state machine on
> >> Controller
> >> >> >> 1]: State change for partition
> >> >> >> [topic_a937ac27-1883-4ca0-95bc-c9a740d08947, 0] from
> OfflinePartition
> >> >> >> to OnlinePartition failed (kafka.controller.PartitionStateMachine)
> >> >> >> kafka.common.PartitionOfflineException: All replicas for partition
> >> >> >> [topic_a937ac27-1883-4ca0-95bc-c9a740d08947, 0] are dead. Marking
> >> this
> >> >> >> partition offline
> >> >> >>         at
> >> >> >>
> >> >>
> >>
> kafka.controller.PartitionStateMachine.electLeaderForPartition(PartitionStateMachine.scala:300)
> >> >> >> .....
> >> >> >> Caused by: kafka.common.PartitionOfflineException: No replica for
> >> >> >> partition ([topic_a937ac27-1883-4ca0-95bc-c9a740d08947, 0]) is
> alive.
> >> >> >> Live brokers are: [Set()], Assigned replicas are: [List(1)]
> >> >> >> .......
> >> >> >>
> >> >> >> I am using one single server to host kafka and zookeeper.
> Replication
> >> >> >> factor is set to 1.
> >> >> >>
> >> >> >> This happened for all the existing topics. Not sure how this
> happened
> >> >> >> but it appeared to be a bug. I did some search and the only
> possible
> >> >> >> fix for this bug seems to be KAFKA-708.
> >> >> >>
> >> >> >> Any comments on this?  Thanks!
> >> >> >>
> >> >> >> Jason
> >> >> >>
> >> >>
> >>
>

Reply via email to