No.  This is not going to work.  The vnodes feature requires the murmur3
partitioner which was introduced with Cassandra 1.2.

Since you are currently using 1.1, you must be using the random
partitioner, which is not compatible with vnodes.

Because the partitioner determines the physical layout of all of your data
on disk and across the cluster, it is not possible to change partitioner
without taking some downtime to rewrite all of your data.

You should probably plan on an upgrade to 1.2 but without also switching to
vnodes at this point.

-Tupshin
On Dec 30, 2013 9:46 AM, "Katriel Traum" <katr...@google.com> wrote:

> Hello list,
>
> I have a 2 DC set up with DC1:3, DC2:3 replication factor. DC1 has 6
> nodes, DC2 has 3. This whole setup runs on AWS, running cassandra 1.1.
> Here's my nodetool ring:
> 1.1.1.1  eu-west     1a          Up     Normal  55.07 GB        50.00%
>          0
> 2.2.2.1  us-east     1b          Up     Normal  107.82 GB       100.00%
>           1
> 1.1.1.2  eu-west     1b          Up     Normal  53.98 GB        50.00%
>          28356863910078205288614550619314017622
> 1.1.1.3  eu-west     1c          Up     Normal  54.85 GB        50.00%
>          56713727820156410577229101238628035242
> 2.2.2.2  us-east     1d          Up     Normal  107.25 GB       100.00%
>           56713727820156410577229101238628035243
> 1.1.1.4  eu-west     1a          Up     Normal  54.99 GB        50.00%
>          85070591730234615865843651857942052863
> 1.1.1.5  eu-west     1b          Up     Normal  55.1 GB         50.00%
>          113427455640312821154458202477256070484
> 2.2.2.3  us-east     1e          Up     Normal  106.78 GB       100.00%
>           113427455640312821154458202477256070485
> 1.1.1.6  eu-west     1c          Up     Normal  55.01 GB        50.00%
>          141784319550391026443072753096570088105
>
>
> I am going to upgrade my machine type, upgrade to 1.2 and change the
> 6-node to 3 nodes. I will have to do it on the live system.
> I'd appreciate any comments about my plan.
> 1. Decommission a 1.1 node.
> 2. Bootstrap a new one in-place, cassandra 1.2, vnodes enabled (I am
> trying to avoid a re-balance later on).
> 3. When done, decommission nodes 4-6 at DC1
>
> Issues i've spotted:
> 1. I'm guessing I will have an unbalanced cluster for the time period
> where I have 1.2+vnodes and 1.1 mixed.
> 2. Rollback is cumbersome, snapshots won't help here.
>
> Any feedback appreciated
>
> Katriel
>
>

Reply via email to