Makes sense. thanks!
Jason On Fri, Mar 22, 2013 at 11:17 AM, Neha Narkhede <neha.narkh...@gmail.com> wrote: > Jason, > > We will document the migration steps. The only reason you have to wipe out > data this time is that you were running an older version and we made some > zookeeper format changes. Such changes are expected until the final > release. Once it is released, we don't expect to make such big changes. > > Thanks, > Neha > > > On Fri, Mar 22, 2013 at 8:09 AM, Jason Huang <jason.hu...@icare.com> wrote: > >> I see. >> >> Since I am not running anything for production at this point - I will >> probably just do this. >> >> However, if someone runs 0.8.0 at production and they want to upgrade >> to the latest version, how should they migrate their message? Maybe >> there should be something documented in the wiki for this? >> >> thanks, >> >> Jason >> >> On Fri, Mar 22, 2013 at 11:07 AM, Jun Rao <jun...@gmail.com> wrote: >> > The easiest way is to wipe out both ZK and kafka data and start from >> > scratch. >> > >> > Thanks, >> > >> > Jun >> > >> > On Fri, Mar 22, 2013 at 6:51 AM, Jason Huang <jason.hu...@icare.com> >> wrote: >> > >> >> Thanks Jun. >> >> >> >> I have built the new kafka version and start the services. You >> >> mentioned that ZK data structure has been changed - does that mean we >> >> can't reload the previous messages from current log files? I actually >> >> tried to copy the log files (.logs and .index) to the new kafka >> >> instance but get the same "topic doesn't exist" error after running >> >> the new kafka services. >> >> >> >> Any comments on how I might be able to recover previous messages? >> >> >> >> thanks >> >> >> >> Jason >> >> >> >> On Wed, Mar 20, 2013 at 12:15 PM, Jun Rao <jun...@gmail.com> wrote: >> >> > The latest version of 0.8 can be found in the 0.8 branch, not trunk. >> >> > >> >> > Thanks, >> >> > >> >> > Jun >> >> > >> >> > On Wed, Mar 20, 2013 at 7:47 AM, Jason Huang <jason.hu...@icare.com> >> >> wrote: >> >> > >> >> >> The 0.8 version I use was built from trunk last Dec. Since then, this >> >> >> error happened 3 times. Each time we had to remove all the ZK and >> >> >> Kafka log data and restart the services. >> >> >> >> >> >> I will try newer versions with more recent patches and keep >> monitoring >> >> it. >> >> >> >> >> >> thanks! >> >> >> >> >> >> Jason >> >> >> >> >> >> On Wed, Mar 20, 2013 at 10:39 AM, Jun Rao <jun...@gmail.com> wrote: >> >> >> > Ok, so you are using the same broker id. What the error is saying >> is >> >> that >> >> >> > broker 1 doesn't seem to be up. >> >> >> > >> >> >> > Not sure what revision of 0.8 you are using. Could you try the >> latest >> >> >> > revision in 0.8 and see if the problem still exists? You may have >> to >> >> wipe >> >> >> > out all ZK and Kafka data first since some ZK data structures have >> >> been >> >> >> > rename a few weeks ago. >> >> >> > >> >> >> > Thanks, >> >> >> > >> >> >> > Jun >> >> >> > >> >> >> > On Wed, Mar 20, 2013 at 6:57 AM, Jason Huang < >> jason.hu...@icare.com> >> >> >> wrote: >> >> >> > >> >> >> >> I restarted the zookeeper server first, then broker. It's the same >> >> >> >> instance of kafka 0.8 and I am using the same config file. In >> >> >> >> server.properties I have: brokerid=1 >> >> >> >> >> >> >> >> Is that sufficient to ensure the broker get restarted with the >> same >> >> >> >> broker id as before? >> >> >> >> >> >> >> >> thanks, >> >> >> >> >> >> >> >> Jason >> >> >> >> >> >> >> >> On Wed, Mar 20, 2013 at 12:30 AM, Jun Rao <jun...@gmail.com> >> wrote: >> >> >> >> > Did the broker get restarted with the same broker id? >> >> >> >> > >> >> >> >> > Thanks, >> >> >> >> > >> >> >> >> > Jun >> >> >> >> > >> >> >> >> > On Tue, Mar 19, 2013 at 1:34 PM, Jason Huang < >> >> jason.hu...@icare.com> >> >> >> >> wrote: >> >> >> >> > >> >> >> >> >> Hello, >> >> >> >> >> >> >> >> >> >> My kafka (0.8) server went down today for unknown reason and >> when >> >> I >> >> >> >> >> restarted both zookeeper and kafka server I got the following >> >> error >> >> >> at >> >> >> >> >> the kafka server log: >> >> >> >> >> >> >> >> >> >> [2013-03-19 13:39:16,131] INFO [Partition state machine on >> >> Controller >> >> >> >> >> 1]: Invoking state change to OnlinePartition for partitions >> >> >> >> >> (kafka.controller.PartitionStateMachine) >> >> >> >> >> [2013-03-19 13:39:16,262] INFO [Partition state machine on >> >> Controller >> >> >> >> >> 1]: Electing leader for partition >> >> >> >> >> [topic_a937ac27-1883-4ca0-95bc-c9a740d08947, 0] >> >> >> >> >> (kafka.controller.PartitionStateMachine) >> >> >> >> >> [2013-03-19 13:39:16,451] ERROR [Partition state machine on >> >> >> Controller >> >> >> >> >> 1]: State change for partition >> >> >> >> >> [topic_a937ac27-1883-4ca0-95bc-c9a740d08947, 0] from >> >> OfflinePartition >> >> >> >> >> to OnlinePartition failed >> (kafka.controller.PartitionStateMachine) >> >> >> >> >> kafka.common.PartitionOfflineException: All replicas for >> partition >> >> >> >> >> [topic_a937ac27-1883-4ca0-95bc-c9a740d08947, 0] are dead. >> Marking >> >> >> this >> >> >> >> >> partition offline >> >> >> >> >> at >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> kafka.controller.PartitionStateMachine.electLeaderForPartition(PartitionStateMachine.scala:300) >> >> >> >> >> ..... >> >> >> >> >> Caused by: kafka.common.PartitionOfflineException: No replica >> for >> >> >> >> >> partition ([topic_a937ac27-1883-4ca0-95bc-c9a740d08947, 0]) is >> >> alive. >> >> >> >> >> Live brokers are: [Set()], Assigned replicas are: [List(1)] >> >> >> >> >> ....... >> >> >> >> >> >> >> >> >> >> I am using one single server to host kafka and zookeeper. >> >> Replication >> >> >> >> >> factor is set to 1. >> >> >> >> >> >> >> >> >> >> This happened for all the existing topics. Not sure how this >> >> happened >> >> >> >> >> but it appeared to be a bug. I did some search and the only >> >> possible >> >> >> >> >> fix for this bug seems to be KAFKA-708. >> >> >> >> >> >> >> >> >> >> Any comments on this? Thanks! >> >> >> >> >> >> >> >> >> >> Jason >> >> >> >> >> >> >> >> >> >> >> >> >> >> >>