Are you using vnodes by any chance? If so, how many? How many nodes and
what's the replication factor? How was data inserted (at what consistency
level)?

Streaming might create a large number of sstables with vnodes (see
CASSANDRA-10495), so in case data is inconsistent between nodes (detected
during the validation phase) a large number of sstables might be created
during the streaming phase of the repair. This might happen with or without
incremental repairs.

There is also a migration step from non-incremental to incremental repair
that must be followed for LCS tables to avoid recompacting all sstables
again. You may find more information in the migration procedure on
http://docs.datastax.com/en/cassandra/2.1/cassandra/operations/opsRepairNodesMigration.html

2016-02-10 14:16 GMT-03:00 Jean Carlo <jean.jeancar...@gmail.com>:

> Hello Kai
>
> This is for *cf3*
>
> nodetool cfstats pns_nonreg_bench.cf3 -H
> Keyspace: pns_nonreg_bench
>     Read Count: 23594
>     Read Latency: 1.2980987539204882 ms.
>     Write Count: 148161
>     Write Latency: 0.04608940274431193 ms.
>     Pending Flushes: 0
>         Table: cf3
>         SSTable count: 11489
>         SSTables in each level: [11485/4, 4, 0, 0, 0, 0, 0, 0, 0]
>         Space used (live): 954.89 MB
>         Space used (total): 954.89 MB
>         Space used by snapshots (total): 0 bytes
>         Off heap memory used (total): 3.74 MB
>         SSTable Compression Ratio: 0.4245371280396339
>         Number of keys (estimate): 1051613
>         Memtable cell count: 0
>         Memtable data size: 0 bytes
>         Memtable off heap memory used: 0 bytes
>         Memtable switch count: 889
>         Local read count: 23594
>         Local read latency: 1.299 ms
>         Local write count: 148161
>         Local write latency: 0.047 ms
>         Pending flushes: 0
>         Bloom filter false positives: 16609
>         Bloom filter false ratio: 0.00000
>         Bloom filter space used: 1.64 MB
>         Bloom filter off heap memory used: 1.55 MB
>         Index summary off heap memory used: 1.88 MB
>         Compression metadata off heap memory used: 318.56 KB
>         Compacted partition minimum bytes: 643 bytes
>         Compacted partition maximum bytes: 924 bytes
>         Compacted partition mean bytes: 914 bytes
>         Average live cells per slice (last five minutes): 1.0
>         Maximum live cells per slice (last five minutes): 1.0
>         Average tombstones per slice (last five minutes): 0.0
>         Maximum tombstones per slice (last five minutes): 0.0
>
> ----------------
>
> This is for *cf1*
>
> nodetool cfstats pns_nonreg_bench.cf1 -H
> Keyspace: pns_nonreg_bench
>     Read Count: 24207
>     Read Latency: 0.9072712851654481 ms.
>     Write Count: 148380
>     Write Latency: 0.04561746866154468 ms.
>     Pending Flushes: 0
>         Table: cf1
>         SSTable count: 1942
>
> It seems that nodetool does not retrieve the rest of the information, I
> think it is due to compactions
>
> nodetool tpstats
> CompactionExecutor                2        83           1424
> 0                 0
>
>
>
>
>
> Saludos
>
> Jean Carlo
>
> "The best way to predict the future is to invent it" Alan Kay
>
> On Wed, Feb 10, 2016 at 6:00 PM, Kai Wang <dep...@gmail.com> wrote:
>
>> Jean,
>>
>> What does your cfstats look like? Especially "SSTables in each level"
>> line.
>>
>> On Wed, Feb 10, 2016 at 8:33 AM, Jean Carlo <jean.jeancar...@gmail.com>
>> wrote:
>>
>>> Hello guys!
>>>
>>> I am testing the repair inc in my custer cassandra. I am doing my test
>>> over these tables
>>>
>>> *CREATE TABLE pns_nonreg_bench.cf3* (
>>>     s text,
>>>     sp int,
>>>     d text,
>>>     dp int,
>>>     m map<text, text>,
>>>     t timestamp,
>>>     PRIMARY KEY (s, sp, d, dp)
>>> ) WITH CLUSTERING ORDER BY (sp ASC, d ASC, dp ASC)
>>>
>>> AND compaction = {'class': 'org.apache.cassandra.db.compaction.
>>> *LeveledCompactionStrategy*'}
>>>     AND compression = {'sstable_compression':
>>> 'org.apache.cassandra.io.compress.*SnappyCompressor*'}
>>>
>>> *CREATE TABLE pns_nonreg_bench.cf1* (
>>>     ise text PRIMARY KEY,
>>>     int_col int,
>>>     text_col text,
>>>     ts_col timestamp,
>>>     uuid_col uuid
>>> ) WITH bloom_filter_fp_chance = 0.01
>>>  AND compaction = {'class': 'org.apache.cassandra.db.compaction.
>>> *LeveledCompactionStrategy*'}
>>>     AND compression = {'sstable_compression':
>>> 'org.apache.cassandra.io.compress.*SnappyCompressor*'}
>>>
>>>
>>>
>>> *table cf1        Space used (live): 665.7 MB*
>>>
>>> *table cf2*
>>> *        Space used (live): 697.03 MB*
>>>
>>> It happens that when I do repair -inc -par on theses tables, *cf2 got a
>>> pick of 3k sstables*. When the repair finish, it takes 30 min or more
>>> to finish all the compactations and return to 6 sstable.
>>>
>>> I am a little concern about if this will happen on production. is it
>>> normal?
>>>
>>> Saludos
>>>
>>> Jean Carlo
>>>
>>> "The best way to predict the future is to invent it" Alan Kay
>>>
>>
>>
>

Reply via email to