Sorry, i meant 5 nodes in previous question.

On Tue, Jun 24, 2014 at 12:36 PM, Kane Kane <kane.ist...@gmail.com> wrote:
> Hello Neha,
>
>>>ZK cluster of 3 nodes will tolerate the loss of 1 node, but if there is a
> subsequent leader election for any reason, there is a chance that the
> cluster does not reach a quorum. It is less likely but still risky to some
> extent.
>
> Does it mean if you have to tolerate 1 node loss without any issues,
> you need *at least* 4 nodes?
>
> On Tue, Jun 24, 2014 at 11:16 AM, Neha Narkhede <neha.narkh...@gmail.com> 
> wrote:
>> Can you elaborate your notion of "smooth"? I thought if you have
>> replication factor=3 in this case, you should be able to tolerate loss
>> of a node?
>>
>> Yes, you should be able to tolerate the loss of a node but if controlled
>> shutdown is not enabled, the delay between loss of the old leader and
>> election of the new leader will be longer.
>>
>> So, you mean ZK cluster of 3 nodes can't tolerate 1 node loss? I've
>> seen many recommendations to run 3-nodes cluster, does it mean in
>> cluster of 3 you won't be able to operate after loosing 1 node?
>>
>> ZK cluster of 3 nodes will tolerate the loss of 1 node, but if there is a
>> subsequent leader election for any reason, there is a chance that the
>> cluster does not reach a quorum. It is less likely but still risky to some
>> extent.
>>
>>
>> On Tue, Jun 24, 2014 at 2:44 AM, Hemath Kumar <hksrckmur...@gmail.com>
>> wrote:
>>
>>> Yes kane i have the replication factor configured as 3
>>>
>>>
>>> On Tue, Jun 24, 2014 at 2:42 AM, Kane Kane <kane.ist...@gmail.com> wrote:
>>>
>>> > Hello Neha, can you explain your statements:
>>> > >>Bringing one node down in a cluster will go smoothly only if your
>>> > replication factor is 1 and you enabled controlled shutdown on the
>>> brokers.
>>> >
>>> > Can you elaborate your notion of "smooth"? I thought if you have
>>> > replication factor=3 in this case, you should be able to tolerate loss
>>> > of a node?
>>> >
>>> > >>Also, bringing down 1 node our of a 3 node zookeeper cluster is risky,
>>> > since any subsequent leader election might not reach a quorum.
>>> >
>>> > So, you mean ZK cluster of 3 nodes can't tolerate 1 node loss? I've
>>> > seen many recommendations to run 3-nodes cluster, does it mean in
>>> > cluster of 3 you won't be able to operate after loosing 1 node?
>>> >
>>> > Thanks.
>>> >
>>> > On Mon, Jun 23, 2014 at 9:04 AM, Neha Narkhede <neha.narkh...@gmail.com>
>>> > wrote:
>>> > > Bringing one node down in a cluster will go smoothly only if your
>>> > > replication factor is 1 and you enabled controlled shutdown on the
>>> > brokers.
>>> > > Also, bringing down 1 node our of a 3 node zookeeper cluster is risky,
>>> > > since any subsequent leader election might not reach a quorum. Having
>>> > said
>>> > > that, a partition going offline shouldn't cause a consumer's offset to
>>> > > reset to an old value. How did you find out what the consumer's offset
>>> > was?
>>> > > Do you have your consumer's logs around?
>>> > >
>>> > > Thanks,
>>> > > Neha
>>> > >
>>> > >
>>> > > On Mon, Jun 23, 2014 at 12:28 AM, Hemath Kumar <hksrckmur...@gmail.com
>>> >
>>> > > wrote:
>>> > >
>>> > >> We have a 3 node cluster ( 3 kafka + 3 ZK nodes ) . Recently we came
>>> > across
>>> > >> a strange issue where we wanted to bring one of the node down from
>>> > cluster
>>> > >> ( 1 kafka + 1 zookeeper) for doing a maintenance . But the movement we
>>> > >> brought it to down on some of the topics ( only some partitions)
>>> > consumers
>>> > >> offset is reset some old value.
>>> > >>
>>> > >> Any reason why this is happened?. As of my knowledge when brought one
>>> > node
>>> > >> down its should work smoothly with out any impact.
>>> > >>
>>> > >> Thanks,
>>> > >> Murthy Chelankuri
>>> > >>
>>> >
>>>

Reply via email to