Did you run cleanup before you shrank the cluster? 

-- 
Jeff Jirsa


> On Dec 13, 2017, at 4:49 AM, Python_Max <python....@gmail.com> wrote:
> 
> Hello.
> 
> I have a situation similar to 
> https://issues.apache.org/jira/browse/CASSANDRA-13153 except mine cassandra 
> is 3.11.1 (that issue should be fixed according to jira).
> 
> Cluster consist of 40 nodes which I have to shrink to 25 more powerful nodes 
> moving less powerful out from the cluster.
> 
> SizeTieredCompactionStrategy is used for target table with default compaction 
> options and default gc_grace_seconds (10 days).
> 
> Full sequential repair executed every day using 
> https://github.com/thelastpickle/cassandra-reaper
> 
> Before reaper was involved the default (incremental) 'nodetool repair' was 
> used.
> 
> 
> There are couple of problems that I observe.
> 
> 1. Compaction.
> 
> There are rows in target table which deleted long time ago (gc_grace_seconds 
> passed) but they are not compacting. I tried 'nodetool compact' and 'nodetool 
> repair -full -seq' with same outcome: sstables recreated but that rows still 
> there (I used sstabledump to detect this state).
> 
> Some of that rows have tombstone and data in wrong order: the data located at 
> more recent sstable by index but relative tombstone is located in previous 
> sstable (shouldn't it be opposite?). CQL select does not return that rows 
> (correct behaviour) until node decommission.
> 
> sstables with non-compacted rows have very old repairedAt value (about 10 
> days before the first deleted row in sstable which should have been compacted 
> long time ago).
> 
> 
> 2. Streaming.
> 
> When moving the node out of cluster 'nodetool decommission' is used. After 
> streaming complete some old rows that was not compacted earlier is back to 
> life in shrinked cluster.
> 
> CQL select does return that rows as alive until running full sequential 
> repair using cassandra-reaper.
> 
> As workaround I tried to shut down a node and using 'nodetool removenode' in 
> case the node itself is streaming wrong data on decommission but that did not 
> work either (deleted data is back to life).
> 
> 
> Is this a known issue?
> 
> 
> PS: I have not tried 'nodetool scrub' yet nor dropping repairedAt for 
> affected sstables.
> 
> 
> -- 
> 
> Best Regards,
> Python_Max.
> 
> 
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
> For additional commands, e-mail: user-h...@cassandra.apache.org
> 

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org

Reply via email to