Hi Jonathan, both I believe.

The window size is 1 day, full settings:
    AND compaction = {'timestamp_resolution': 'MILLISECONDS',
'unchecked_tombstone_compaction': 'true', 'compaction_window_size': '1',
'compaction_window_unit': 'DAYS', 'tombstone_compaction_interval': '86400',
'tombstone_threshold': '0.2', 'class':
'com.jeffjirsa.cassandra.db.compaction.TimeWindowCompactionStrategy'}


nodetool tpstats

Pool Name                    Active   Pending      Completed   Blocked  All
time blocked
MutationStage                     0         0    68582241832         0
           0
ReadStage                         0         0      209566303         0
           0
RequestResponseStage              0         0    44680860850         0
           0
ReadRepairStage                   0         0       24562722         0
           0
CounterMutationStage              0         0              0         0
           0
MiscStage                         0         0              0         0
           0
HintedHandoff                     1         1            203         0
           0
GossipStage                       0         0        8471784         0
           0
CacheCleanupExecutor              0         0            122         0
           0
InternalResponseStage             0         0         552125         0
           0
CommitLogArchiver                 0         0              0         0
           0
CompactionExecutor                8        42        1433715         0
           0
ValidationExecutor                0         0           2521         0
           0
MigrationStage                    0         0         527549         0
           0
AntiEntropyStage                  0         0           7697         0
           0
PendingRangeCalculator            0         0             17         0
           0
Sampler                           0         0              0         0
           0
MemtableFlushWriter               0         0         116966         0
           0
MemtablePostFlush                 0         0         209103         0
           0
MemtableReclaimMemory             0         0         116966         0
           0
Native-Transport-Requests         1         0     1715937778         0
      176262

Message type           Dropped
READ                         2
RANGE_SLICE                  0
_TRACE                       0
MUTATION                  4390
COUNTER_MUTATION             0
BINARY                       0
REQUEST_RESPONSE          1882
PAGED_RANGE                  0
READ_REPAIR                  0


On Tue, Aug 7, 2018 at 7:57 PM Jonathan Haddad <j...@jonhaddad.com> wrote:

> What's your window size?
>
> When you say backed up, how are you measuring that?  Are there pending
> tasks or do you just see more files than you expect?
>
> On Tue, Aug 7, 2018 at 4:38 PM Brian Spindler <brian.spind...@gmail.com>
> wrote:
>
>> Hey guys, quick question:
>>
>> I've got a v2.1 cassandra cluster, 12 nodes on aws i3.2xl, commit log on
>> one drive, data on nvme.  That was working very well, it's a ts db and has
>> been accumulating data for about 4weeks.
>>
>> The nodes have increased in load and compaction seems to be falling
>> behind.  I used to get about 1 file per day for this column family, about
>> ~30GB Data.db file per day.  I am now getting hundreds per day at  1mb -
>> 50mb.
>>
>> How to recover from this?
>>
>> I can scale out to give some breathing room but will it go back and
>> compact the old days into nicely packed files for the day?
>>
>> I tried setting compaction throughput to 1000 from 256 and it seemed to
>> make things worse for the CPU, it's configured on i3.2xl with 8 compaction
>> threads.
>>
>> -B
>>
>> Lastly, I have mixed TTLs in this CF and need to run a repair (I think)
>> to get rid of old tombstones, however running repairs in 2.1 on TWCS column
>> families causes a very large spike in sstable counts due to anti-compaction
>> which causes a lot of disruption, is there any other way?
>>
>>
>>
>
> --
> Jon Haddad
> http://www.rustyrazorblade.com
> twitter: rustyrazorblade
>

Reply via email to