We never tried manipulating ZK directly. So it may or may not work. Our plan is to have tools to change # partitions, replication factor, etc in the future.
Thanks, Jun On Wed, Jun 19, 2013 at 6:18 PM, Jason Rosenberg <j...@squareup.com> wrote: > Is there a way to manually do this? Or is there a way to manually change > the replication factor for a topic? Possibly by manipulating zookeeper > data directly? > > Jason > > > On Wed, Jun 19, 2013 at 8:17 AM, Jun Rao <jun...@gmail.com> wrote: > > > Yogesh, > > > > Delete topic is not supported in 0.8 beta yet. We decided to take this > > feature out in 0.8 beta to make the release available sooner. > > > > Thanks, > > > > Jun > > > > > > On Tue, Jun 18, 2013 at 6:25 AM, Yogesh Sangvikar < > > yogesh.sangvi...@gmail.com> wrote: > > > > > Hi Team, > > > > > > I am exploring kafka 0.8 beta release to understand data flow, > > replication > > > features. > > > While testing i found that, the partition data for data for deleted > topic > > > is preseved in kafka-logs, why this behavior? suppose below case, > > > > > > A topic (suppose test1) is created with partition 6 and replication 3 > > on a > > > system with 4 brokers, respective log and index files will be prepared > > per > > > partition in kafka-logs. If I delete the topic and recreate the same > > topic > > > ‘test1’ after some time with partition 2 and replication 2. The > > kafka-logs > > > directory seems to be confusing to understand why the partitions for > > > previous topic are present. *Please help to understand this scenario*. > > > > > > Also, while testing the replication and leader selection feature > observed > > > leader -1 status, > > > > > > Original status: > > > topic: test1 partition: 0 leader: 4 replicas: 4,2,3 isr: > > 4,2,3 > > > topic: test1 partition: 1 leader: 0 replicas: 0,3,4 isr: > > 0,3,4 > > > topic: test1 partition: 2 leader: 1 replicas: 1,4,0 isr: > > 1,4,0 > > > if leader 4 goes down: > > > topic: test1 partition: 0 leader: 2 replicas: 4,2,3 isr: > 2,3 > > > topic: test1 partition: 1 leader: 0 replicas: 0,3,4 isr: > > 0,3,4 > > > topic: test1 partition: 2 leader: 1 replicas: 1,4,0 isr: > > 1,0,4 > > > > > > if leader 2 goes down: > > > topic: test1 partition: 0 leader: 3 replicas: 4,2,3 isr: 3 > > > topic: test1 partition: 1 leader: 0 replicas: 0,3,4 isr: > > 0,3,4 > > > topic: test1 partition: 2 leader: 1 replicas: 1,4,0 isr: > > 1,0,4 > > > > > > if again leader 3 goes down: > > > topic: test1 partition: 0 leader: -1 replicas: 4,2,3 isr: > > > topic: test1 partition: 1 leader: 0 replicas: 0,3,4 isr: > 0,4 > > > topic: test1 partition: 2 leader: 1 replicas: 1,4,0 isr: > > 1,0,4 > > > > > > As per kafka protocol guide, *leader: -1 means If no leader exists > > because > > > we are in the middle of a leader election this id will be -1.* > > > > > > *Does it mean that, the data from partition 0 will be unavailable due > no > > > leader (leader selection in progress)?* > > > As per my understanding, can we have auto re-balancer facility to > > > re-balance the partition replications to available brokers if one of > the > > > broker is down, as in above case of (if leader 4 goes down), we can > > > replicate the partition 0 to broker 0/1 to re-balance the replication. > > > > > > Please correct me for any wrong understanding as those are my initial > > > observations. > > > > > > Thanks in advance. > > > > > > Thanks, > > > Yogesh Sangvikar > > > > > >