Just a side note: increase system_auth keyspace replication factor if you're using authentication.
Le Jeudi 12 janvier 2017 14h52, Alain RODRIGUEZ <arodr...@gmail.com> a écrit : Hi, Nodetool repair always list lots of data and never stays repaired. I think. This might be the reason: "incremental: true" Incremental repairs is the default in your version. It marks data as being repaired in order to only repair each data only once. It is a clever feature, but with some caveats. I would read about it as it is not trivial to understand impacts and in some cases it can create issues and not be such a good idea to use incremental repairs. Make sure to run a full repair instead when a node goes down for example. C*heers,-----------------------Alain Rodriguez - @arodream - alain@thelastpickle.comFrance The Last Pickle - Apache Cassandra Consultinghttp://www.thelastpickle.com 2017-01-11 15:21 GMT+01:00 Cogumelos Maravilha <cogumelosmaravi...@sapo.pt>: Nodetool repair always list lots of data and never stays repaired. I think. Cheers On 01/11/2017 02:15 PM, Hannu Kröger wrote: > Just to understand: > > What exactly is the problem? > > Cheers, > Hannu > >> On 11 Jan 2017, at 16.07, Cogumelos Maravilha <cogumelosmaravi...@sapo.pt> >> wrote: >> >> Cassandra 3.9. >> >> nodetool status >> Datacenter: dc1 >> =============== >> Status=Up/Down >> |/ State=Normal/Leaving/Joining/ Moving >> -- Address Load Tokens Owns (effective) Host >> ID Rack >> UN 10.0.120.145 1.21 MiB 256 49.5% >> da6683cd-c3cf-4c14-b3cc- e7af4080c24f rack1 >> UN 10.0.120.179 1020.51 KiB 256 48.1% >> fb695bea-d5e8-4bde-99db- 9f756456a035 rack1 >> UN 10.0.120.55 1.02 MiB 256 53.3% >> eb911989-3555-4aef-b11c- 4a684a89a8c4 rack1 >> UN 10.0.120.46 1.01 MiB 256 49.1% >> 8034c30a-c1bc-44d4-bf84- 36742e0ec21c rack1 >> >> nodetool repair >> [2017-01-11 13:58:27,274] Replication factor is 1. No repair is needed >> for keyspace 'system_auth' >> [2017-01-11 13:58:27,284] Starting repair command #4, repairing keyspace >> system_traces with repair options (parallelism: parallel, primary range: >> false, incremental: true, job threads: 1, ColumnFamilies: [], >> dataCenters: [], hosts: [], # of ranges: 515) >> [2017-01-11 14:01:55,628] Repair session >> 82a25960-d806-11e6-8ac4- 73b93fe4986d for range >> [(-1278992819359672027,- 1209509957304098060], >> (-2593749995021251600,- 2592266543457887959], >> (-6451044457481580778,- 6438233936014720969], >> (-1917989291840804877,- 1912580903456869648], >> (-3693090304802198257,- 3681923561719364766], >> (-380426998894740867,- 350094836653869552], >> (1890591246410309420, 1899294587910578387], >> (6561031217224224632, 6580230317350171440], >> ... 4 pages of data >> , (6033828815719998292, 6079920177089043443]] finished (progress: 1%) >> [2017-01-11 13:58:27,986] Repair completed successfully >> [2017-01-11 13:58:27,988] Repair command #4 finished in 0 seconds >> >> nodetool gcstats >> Interval (ms) Max GC Elapsed (ms)Total GC Elapsed (ms)Stdev GC Elapsed >> (ms) GC Reclaimed (MB) Collections Direct Memory Bytes >> 360134 23 >> 23 0 333975216 >> 1 -1 >> >> (wait) >> nodetool gcstats >> Interval (ms) Max GC Elapsed (ms)Total GC Elapsed (ms)Stdev GC Elapsed >> (ms) GC Reclaimed (MB) Collections Direct Memory Bytes >> 60016 0 0 >> NaN 0 0 -1 >> >> nodetool repair >> [2017-01-11 14:00:45,888] Replication factor is 1. No repair is needed >> for keyspace 'system_auth' >> [2017-01-11 14:00:45,896] Starting repair command #5, repairing keyspace >> system_traces with repair options (parallelism: parallel, primary range: >> false, incremental: true, job threads: 1, ColumnFamilies: [], >> dataCenters: [], hosts: [], # of ranges: 515) >> ... 4 pages of data >> , (94613607632078948, 219237792837906432], >> (6033828815719998292, 6079920177089043443]] finished (progress: 1%) >> [2017-01-11 14:00:46,567] Repair completed successfully >> [2017-01-11 14:00:46,576] Repair command #5 finished in 0 seconds >> >> nodetool gcstats >> Interval (ms) Max GC Elapsed (ms)Total GC Elapsed (ms)Stdev GC Elapsed >> (ms) GC Reclaimed (MB) Collections Direct Memory Bytes >> 9169 25 25 >> 0 330518688 1 -1 >> >> >> Always in loop, I think! >> >> Thanks in advance. >>