Did you get chance to look at tlp reaper tool i.e. http://cassandra-reaper.io/ It is pretty awesome – Thanks to TLP team.
From: Martin Xue [mailto:martin...@gmail.com] Sent: Wednesday, July 31, 2019 12:09 AM To: user@cassandra.apache.org Subject: Repair / compaction for 6 nodes, 2 DC cluster Hello, Good day. This is Martin. Can someone help me with the following query regarding Cassandra repair and compaction? Currently we have a large keyspace (keyspace_event) with 1TB of data (in /var/lib/cassandra/data/keyspace_event); There is a cluster with Datacenter 1 contains 3 nodes, Data center 2 containing 3 nodes; All together 6 nodes; As part of maintenance, I run the repair on this keyspace with the following command: nodetool repair -pr --full keyspace_event; now it has been run for 2 days. yes 2 days, when doing nodetool tpstats, it shows there is a compaction running: CompactionExecutor 1 1 5783732 0 0 nodetool compactionstats shows: pending tasks: 6 id compaction type keyspace table completed total unit progress 249ec5f1-b225-11e9-82bd-5b36ef02cadd Anticompaction after repair keyspace_event table_event 1916937740948 2048931045927 bytes 93.56% Now my questions are: 1. why running repair (with primary range option, -pr, as I want to limit the repair node by node), triggered the compaction running on other nodes? 2. when I run the repair on the second node with nodetool repair -pr --full keyspace_event; will the subsequent compaction run again on all the 6 nodes? I want to know what are the best option to run the repair (full repair) as we did not run it before, especially if it can take less time (in current speed it will take 2 weeks to finish all). I am running Cassandra 3.0.14 Any suggestions will be appreciated. Thanks Regards Martin