Nevermind the question. It was a firewall problem. Now the nodes between
different versions are able to see ach other! =)

Cheers,

Paulo


2013/10/2 Paulo Motta <pauloricard...@gmail.com>

> Hello,
>
> I just started the rolling upgrade procedure from 1.1.10 to 2.1.10. Our
> strategy is to simultaneously upgrade one server from each replication
> group. So, if we have a 6 nodes with RF=2, we will upgrade 3 nodes at a
> time (from distinct replication groups).
>
> My question is: do the newly upgraded nodes show as "Down" in the
> "nodetool ring" of the old cluster (1.1.10)? Because I thought that network
> compatibility meant nodes from a newer version would receive traffic (write
> + reads) from the previous version without problems.
>
> Cheers,
>
> Paulo
>
>
> 2013/9/26 Paulo Motta <pauloricard...@gmail.com>
>
>> Hello Charles,
>>
>> Thank you very much for your detailed upgrade report. It'll be very
>> helpful during our upgrade operation (even though we'll do a rolling
>> production upgrade).
>>
>> I'll also share our findings during the upgrade here.
>>
>> Cheers,
>>
>> Paulo
>>
>>
>> 2013/9/24 Charles Brophy <cbro...@zulily.com>
>>
>>> Hi Paulo,
>>>
>>> I just completed a migration from 1.1.10 to 1.2.10 and it was
>>> surprisingly painless.
>>>
>>> The course of action that I took:
>>> 1) describe cluster - make sure all nodes are on the same schema
>>> 2) shutoff all maintenance tasks; i.e. make sure no scheduled repair is
>>> going to kick off in the middle of what you're doing
>>> 3) snapshot - maybe not necessary but it's so quick it makes no sense to
>>> skip this step
>>> 4) drain the nodes - I shut down the entire cluster rather than chance
>>> any incompatible gossip concerns that might come from a rolling upgrade. I
>>> have the luxury of controlling both the providers and consumers of our
>>> data, so this wasn't so disruptive for us.
>>> 5) Upgrade the nodes, turn them on one-by-one, monitor the logs for
>>> funny business.
>>> 6) nodetool upgradesstables
>>> 7) Turn various maintenance tasks back on, etc.
>>>
>>> The worst part was managing the yaml/config changes between the
>>> versions. It wasn't horrible, but the diff was "noisier" than a more
>>> incremental upgrade typically is. A few things I recall that were special:
>>> 1) Since you have an existing cluster, you'll probably need to set the
>>> default partitioner back to RandomPartitioner in cassandra.yaml. I believe
>>> that is outlined in NEWS.
>>> 2) I set the initial tokens to be the same as what the nodes held
>>> previously.
>>> 3) The timeout is now divided into more atomic settings and you get to
>>> decided how (or if) to configure it from the default appropriately.
>>>
>>> tldr; I did a standard upgrade and payed careful attention to the
>>> NEWS.txt upgrade notices. I did a full cluster restart and NOT a rolling
>>> upgrade. It went without a hitch.
>>>
>>> Charles
>>>
>>>
>>>
>>>
>>>
>>>
>>> On Tue, Sep 24, 2013 at 2:33 PM, Paulo Motta 
>>> <pauloricard...@gmail.com>wrote:
>>>
>>>> Cool, sounds fair enough. Thanks for the help, Rob!
>>>>
>>>> If anyone has upgraded from 1.1.X to 1.2.X, please feel invited to
>>>> share any tips on issues you're encountered that are not yet documented.
>>>>
>>>> Cheers,
>>>>
>>>> Paulo
>>>>
>>>>
>>>> 2013/9/24 Robert Coli <rc...@eventbrite.com>
>>>>
>>>>> On Tue, Sep 24, 2013 at 1:41 PM, Paulo Motta <pauloricard...@gmail.com
>>>>> > wrote:
>>>>>
>>>>>> Doesn't the probability of something going wrong increases as the gap
>>>>>> between the versions increase? So, using this reasoning, upgrading from
>>>>>> 1.1.10 to 1.2.6 would have less chance of something going wrong then from
>>>>>> 1.1.10 to 1.2.9 or 1.2.10.
>>>>>>
>>>>>
>>>>> Sorta, but sorta not.
>>>>>
>>>>> https://github.com/apache/cassandra/blob/trunk/NEWS.txt
>>>>>
>>>>> Is the canonical source of concerns on upgrade. There are a few cases
>>>>> where upgrading to the "root" of X.Y.Z creates issues that do not exist if
>>>>> you upgrade to the "head" of that line. AFAIK there have been no cases
>>>>> where upgrading to the "head" of a line (where that line is mature, like
>>>>> 1.2.10) has created problems which would have been avoided by upgrading to
>>>>> the "root" first.
>>>>>
>>>>>
>>>>>> I'm hoping this reasoning is wrong and I can update directly from
>>>>>> 1.1.10 to 1.2.10. :-)
>>>>>>
>>>>>
>>>>> That's what I plan to do when we move to 1.2.X, FWIW.
>>>>>
>>>>> =Rob
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Paulo Ricardo
>>>>
>>>> --
>>>> European Master in Distributed Computing***
>>>> Royal Institute of Technology - KTH
>>>> *
>>>> *Instituto Superior Técnico - IST*
>>>> *http://paulormg.com*
>>>>
>>>
>>>
>>
>>
>> --
>> Paulo Ricardo
>>
>> --
>> European Master in Distributed Computing***
>> Royal Institute of Technology - KTH
>> *
>> *Instituto Superior Técnico - IST*
>> *http://paulormg.com*
>>
>
>
>
> --
> Paulo Ricardo
>
> --
> European Master in Distributed Computing***
> Royal Institute of Technology - KTH
> *
> *Instituto Superior Técnico - IST*
> *http://paulormg.com*
>



-- 
Paulo Ricardo

-- 
European Master in Distributed Computing***
Royal Institute of Technology - KTH
*
*Instituto Superior Técnico - IST*
*http://paulormg.com*

Reply via email to