Thanks all - a little clarification: - The node has fully joined at this point with the duplicates - Cleanup has been run on older nodes - Currently using LCS
Rob - thanks for that, I was wondering whether either of those would successfully deduplicate the data. We were hypothesizing that a decommission would merely stream the duplicates out as well as though they were valid data - is this not the case? After some discussion just now a colleague is suggesting we force them to L0[1] - would you agree this should be equivalent to option 2, albeit with downtime? Cheers, Alain [1] https://issues.apache.org/jira/browse/CASSANDRA-5271 On Tue, Nov 18, 2014 at 12:13 AM, Robert Coli <rc...@eventbrite.com> wrote: > On Mon, Nov 17, 2014 at 12:04 PM, Alain Vandendorpe <al...@tapstream.com> > wrote: > >> With bootstrapping and initial compactions finished that node now has >> what seems to be duplicate data, with almost exactly 2x the expected disk >> usage. CQL returns correct results but we depend on the ability to directly >> read the SSTable files (hence also RF=1.) >> >> Would anyone have suggestions on a good way to resolve this? >> > > (If I understand correctly, the new node is now joined to the cluster; the > below assumes this.) > > ** The simplest, slightly inconvenient, way, which temporarily reduces > capacity : > > 1) nodetool cleanup # on the original node > > This removes obsolete data from the original node, which has moved to the > new node. I mention this in case you did not already do it as part of > joining the new node. There are some edge cases [1] where you will "wake > up" old data if you haven't done this before decommission of the new node. > > 2) nodetool decommission # on the new node > > This moves the data from the new node back onto the old node and removes > the new node from the cluster. > > 3) wipe data from new node # including the system keyspace > > 4) re-bootstrap new node > > ** The second simplest way, which requires using Size Tiered Compaction > Strategy (STS) but does not reduce capacity until step 2) : > > 1) nodetool compact > > This will merge all your duplicates into One Big SSTable. > > 2) if necessary, once RF>1, use sstablesplit [1] (with the node down) to > split up your One Big SSTable. > > If you're not using STS, you can temporarily switch to it, but 2) becomes > less straightforward. > > =Rob > http://twitter.com/rcolidba > [1] https://issues.apache.org/jira/browse/CASSANDRA-7764 > [2] > http://www.datastax.com/documentation/cassandra/2.0/cassandra/tools/toolsSSTableSplit.html > >