Hi,

Thanks for the fast response!

We are not using any materialized views, but there are several indexes.  I
don't have a recent heap dump, and it will be about 24 before I can
generate an interesting one, but most of the memory was allocated to byte
buffers, so not entirely helpful.

nodetool cfstats is also below.

I also see a lot of flushing happening, but it seems like there are too
many small allocations to be effective.  Here are the messages I see,

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,459 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='gpsmessages') to
> free up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04,
> this: 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,459 ColumnFamilyStore.java:915
> - Enqueuing flush of gpsmessages: 0.000KiB (0%) on-heap, 0.014KiB (0%)
> off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,460 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='user_history') to
> free up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04,
> this: 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,461 ColumnFamilyStore.java:915
> - Enqueuing flush of user_history: 0.000KiB (0%) on-heap, 0.011KiB (0%)
> off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,465 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='tweets') to free
> up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04, this:
> 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,465 ColumnFamilyStore.java:915
> - Enqueuing flush of tweets: 0.000KiB (0%) on-heap, 0.188KiB (0%) off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='user_history') to
> free up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04,
> this: 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:915
> - Enqueuing flush of user_history: 0.000KiB (0%) on-heap, 0.024KiB (0%)
> off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='tweets') to free
> up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04, this:
> 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:915
> - Enqueuing flush of tweets: 0.000KiB (0%) on-heap, 0.188KiB (0%) off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,472 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='gpsmessages') to
> free up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04,
> this: 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,472 ColumnFamilyStore.java:915
> - Enqueuing flush of gpsmessages: 0.000KiB (0%) on-heap, 0.013KiB (0%)
> off-heap


>

Stack traces from errors are below.


> java.io.IOException: Broken pipe

        at sun.nio.ch.FileDispatcherImpl.write0(Native Method)
> ~[na:1.8.0_181]

        at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)
> ~[na:1.8.0_181]

        at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)
> ~[na:1.8.0_181]

        at sun.nio.ch.IOUtil.write(IOUtil.java:51) ~[na:1.8.0_181]

        at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:471)
> ~[na:1.8.0_181]

        at
> org.apache.cassandra.io.util.BufferedDataOutputStreamPlus.doFlush(BufferedDataOutputStreamPlus.java:323)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.io.util.BufferedDataOutputStreamPlus.flush(BufferedDataOutputStreamPlus.java:331)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.streaming.ConnectionHandler$OutgoingMessageHandler.sendMessage(ConnectionHandler.java:409)
> [apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.streaming.ConnectionHandler$OutgoingMessageHandler.run(ConnectionHandler.java:380)
> [apache-cassandra-3.11.1.jar:3.11.1]

        at java.lang.Thread.run(Thread.java:748) [na:1.8.0_181]

ERROR [MutationStage-226] 2018-08-06 07:16:08,236
> JVMStabilityInspector.java:142 - JVM state determined to be unstable.
> Exiting forcefully due to:

java.lang.OutOfMemoryError: Direct buffer memory

        at java.nio.Bits.reserveMemory(Bits.java:694) ~[na:1.8.0_181]

        at java.nio.DirectByteBuffer.<init>(DirectByteBuffer.java:123)
> ~[na:1.8.0_181]

        at java.nio.ByteBuffer.allocateDirect(ByteBuffer.java:311)
> ~[na:1.8.0_181]

        at
> org.apache.cassandra.utils.memory.SlabAllocator.getRegion(SlabAllocator.java:139)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.utils.memory.SlabAllocator.allocate(SlabAllocator.java:104)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.utils.memory.ContextAllocator.allocate(ContextAllocator.java:57)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.utils.memory.ContextAllocator.clone(ContextAllocator.java:47)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.utils.memory.MemtableBufferAllocator.clone(MemtableBufferAllocator.java:40)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at org.apache.cassandra.db.Memtable.put(Memtable.java:269)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.db.ColumnFamilyStore.apply(ColumnFamilyStore.java:1332)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.db.Keyspace.applyInternal(Keyspace.java:618)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at org.apache.cassandra.db.Keyspace.applyFuture(Keyspace.java:425)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at org.apache.cassandra.db.Mutation.applyFuture(Mutation.java:222)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.db.MutationVerbHandler.doVerb(MutationVerbHandler.java:68)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.net.MessageDeliveryTask.run(MessageDeliveryTask.java:66)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> ~[na:1.8.0_181]

        at
> org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$FutureTask.run(AbstractLocalAwareExecutorService.java:162)
> ~[apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$LocalSessionFutureTask.run(AbstractLocalAwareExecutorService.java:134)
> [apache-cassandra-3.11.1.jar:3.11.1]

        at
> org.apache.cassandra.concurrent.SEPWorker.run(SEPWorker.java:109)
> [apache-cassandra-3.11.1.jar:3.11.1]

        at java.lang.Thread.run(Thread.java:748) [na:1.8.0_181]

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,459 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='gpsmessages') to
> free up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04,
> this: 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,459 ColumnFamilyStore.java:915
> - Enqueuing flush of gpsmessages: 0.000KiB (0%) on-heap, 0.014KiB (0%)
> off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,460 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='user_history') to
> free up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04,
> this: 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,461 ColumnFamilyStore.java:915
> - Enqueuing flush of user_history: 0.000KiB (0%) on-heap, 0.011KiB (0%)
> off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,465 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='tweets') to free
> up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04, this:
> 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,465 ColumnFamilyStore.java:915
> - Enqueuing flush of tweets: 0.000KiB (0%) on-heap, 0.188KiB (0%) off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='user_history') to
> free up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04,
> this: 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:915
> - Enqueuing flush of user_history: 0.000KiB (0%) on-heap, 0.024KiB (0%)
> off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='tweets') to free
> up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04, this:
> 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:915
> - Enqueuing flush of tweets: 0.000KiB (0%) on-heap, 0.188KiB (0%) off-heap

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,472 ColumnFamilyStore.java:1305
> - Flushing largest CFS(Keyspace='userinfo', ColumnFamily='gpsmessages') to
> free up room. Used total: 0.54/0.05, live: 0.00/0.00, flushing: 0.40/0.04,
> this: 0.00/0.00

DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,472 ColumnFamilyStore.java:915
> - Enqueuing flush of gpsmessages: 0.000KiB (0%) on-heap, 0.013KiB (0%)
> off-heap


Total number of tables: 40
----------------
Keyspace : userinfo
Read Count: 143301
Read Latency: 14.945587623254548 ms.
Write Count: 2754603904
Write Latency: 0.020883145284324698 ms.
Pending Flushes: 0
Table (index):
gpsmessages.gpsmessages_addresscount_idxgpsmessages.gpsmessages_addresscount_idx
SSTable count: 9
Space used (live): 19043463189
Space used (total): 19043463189
Space used by snapshots (total): 0
Off heap memory used (total): 6259448
SSTable Compression Ratio: 0.3704785164266614
Number of partitions (estimate): 1025
Memtable cell count: 309066
Memtable data size: 13602774
Memtable off heap memory used: 0
Memtable switch count: 0
Local read count: 0
Local read latency: NaN ms
Local write count: 46025778
Local write latency: 0.034 ms
Pending flushes: 0
Percent repaired: 0.0
Bloom filter false positives: 0
Bloom filter false ratio: 0.00000
Bloom filter space used: 2504
Bloom filter off heap memory used: 2432
Index summary off heap memory used: 320
Compression metadata off heap memory used: 6256696
Compacted partition minimum bytes: 43
Compacted partition maximum bytes: 44285675122
Compacted partition mean bytes: 30405277
Average live cells per slice (last five minutes): NaN
Maximum live cells per slice (last five minutes): 0
Average tombstones per slice (last five minutes): NaN
Maximum tombstones per slice (last five minutes): 0
Dropped Mutations: 0

Table (index):
gpsmessages.gpsmessages_addresses_idxgpsmessages.gpsmessages_addresses_idx
SSTable count: 18
Space used (live): 409514565570
Space used (total): 409514565570
Space used by snapshots (total): 0
Off heap memory used (total): 153405673
SSTable Compression Ratio: 0.4447731157134059
Number of partitions (estimate): 6013125
Memtable cell count: 1110334
Memtable data size: 67480140
Memtable off heap memory used: 0
Memtable switch count: 0
Local read count: 0
Local read latency: NaN ms
Local write count: 147639252
Local write latency: 0.015 ms
Pending flushes: 0
Percent repaired: 0.0
Bloom filter false positives: 0
Bloom filter false ratio: 0.00000
Bloom filter space used: 34175400
Bloom filter off heap memory used: 34175256
Index summary off heap memory used: 7432177
Compression metadata off heap memory used: 111798240
Compacted partition minimum bytes: 61
Compacted partition maximum bytes: 322381140
Compacted partition mean bytes: 36692
Average live cells per slice (last five minutes): NaN
Maximum live cells per slice (last five minutes): 0
Average tombstones per slice (last five minutes): NaN
Maximum tombstones per slice (last five minutes): 0
Dropped Mutations: 0

Table (index):
gpsmessages.addressreceivedtime_idxgpsmessages.addressreceivedtime_idx
SSTable count: 10
Space used (live): 52738155477
Space used (total): 52738155477
Space used by snapshots (total): 0
Off heap memory used (total): 1909362628
SSTable Compression Ratio: 0.4106961621795128
Number of partitions (estimate): 1338730016
Memtable cell count: 308990
Memtable data size: 13410867
Memtable off heap memory used: 0
Memtable switch count: 0
Local read count: 0
Local read latency: NaN ms
Local write count: 46012614
Local write latency: 0.012 ms
Pending flushes: 0
Percent repaired: 0.0
Bloom filter false positives: 0
Bloom filter false ratio: 0.00000
Bloom filter space used: 1687550888
Bloom filter off heap memory used: 1687550808
Index summary off heap memory used: 213249180
Compression metadata off heap memory used: 8562640
Compacted partition minimum bytes: 36
Compacted partition maximum bytes: 2759
Compacted partition mean bytes: 54
Average live cells per slice (last five minutes): NaN
Maximum live cells per slice (last five minutes): 0
Average tombstones per slice (last five minutes): NaN
Maximum tombstones per slice (last five minutes): 0
Dropped Mutations: 0

Table: gpsmessages
SSTable count: 13
Space used (live): 337974446627
Space used (total): 337974446627
Space used by snapshots (total): 0
Off heap memory used (total): 77833540
SSTable Compression Ratio: 0.5300637241381126
Number of partitions (estimate): 22034
Memtable cell count: 308904
Memtable data size: 72074512
Memtable off heap memory used: 0
Memtable switch count: 110
Local read count: 0
Local read latency: NaN ms
Local write count: 45996652
Local write latency: 0.281 ms
Pending flushes: 0
Percent repaired: 0.0
Bloom filter false positives: 0
Bloom filter false ratio: 0.00000
Bloom filter space used: 67904
Bloom filter off heap memory used: 67800
Index summary off heap memory used: 11756
Compression metadata off heap memory used: 77753984
Compacted partition minimum bytes: 73
Compacted partition maximum bytes: 1155149911
Compacted partition mean bytes: 13158224
Average live cells per slice (last five minutes): NaN
Maximum live cells per slice (last five minutes): 0
Average tombstones per slice (last five minutes): NaN
Maximum tombstones per slice (last five minutes): 0
Dropped Mutations: 13699

Table: user_history
SSTable count: 17
Space used (live): 116361158882
Space used (total): 116361158882
Space used by snapshots (total): 0
Off heap memory used (total): 29562319
SSTable Compression Ratio: 0.5683114352331539
Number of partitions (estimate): 1337206
Memtable cell count: 773277
Memtable data size: 40623368
Memtable off heap memory used: 0
Memtable switch count: 57
Local read count: 209
Local read latency: NaN ms
Local write count: 145853733
Local write latency: 0.020 ms
Pending flushes: 0
Percent repaired: 0.0
Bloom filter false positives: 0
Bloom filter false ratio: 0.00000
Bloom filter space used: 3844416
Bloom filter off heap memory used: 3844280
Index summary off heap memory used: 800991
Compression metadata off heap memory used: 24917048
Compacted partition minimum bytes: 61
Compacted partition maximum bytes: 464228842
Compacted partition mean bytes: 72182
Average live cells per slice (last five minutes): NaN
Maximum live cells per slice (last five minutes): 0
Average tombstones per slice (last five minutes): NaN
Maximum tombstones per slice (last five minutes): 0
Dropped Mutations: 66702

Table: users
SSTable count: 3
Space used (live): 89945186
Space used (total): 89945186
Space used by snapshots (total): 0
Off heap memory used (total): 2092053
SSTable Compression Ratio: 0.5712127629253333
Number of partitions (estimate): 1365645
Memtable cell count: 3556
Memtable data size: 150903
Memtable off heap memory used: 0
Memtable switch count: 42
Local read count: 143087
Local read latency: 6.094 ms
Local write count: 250971
Local write latency: 0.024 ms
Pending flushes: 0
Percent repaired: 0.0
Bloom filter false positives: 0
Bloom filter false ratio: 0.00000
Bloom filter space used: 1709848
Bloom filter off heap memory used: 1709824
Index summary off heap memory used: 372125
Compression metadata off heap memory used: 10104
Compacted partition minimum bytes: 36
Compacted partition maximum bytes: 310
Compacted partition mean bytes: 66
Average live cells per slice (last five minutes): 1.0
Maximum live cells per slice (last five minutes): 1
Average tombstones per slice (last five minutes): 1.0
Maximum tombstones per slice (last five minutes): 1
Dropped Mutations: 114

Table: tweets
SSTable count: 18
Space used (live): 1809145656486
Space used (total): 1809145656486
Space used by snapshots (total): 0
Off heap memory used (total): 435915908
SSTable Compression Ratio: 0.5726200929451171
Number of partitions (estimate): 26217889
Memtable cell count: 710146
Memtable data size: 31793929
Memtable off heap memory used: 0
Memtable switch count: 399
Local read count: 5
Local read latency: NaN ms
Local write count: 2322829524
Local write latency: 0.019 ms
Pending flushes: 0
Percent repaired: 0.0
Bloom filter false positives: 0
Bloom filter false ratio: 0.00000
Bloom filter space used: 35019224
Bloom filter off heap memory used: 35019080
Index summary off heap memory used: 16454076
Compression metadata off heap memory used: 384442752
Compacted partition minimum bytes: 104
Compacted partition maximum bytes: 3379391
Compacted partition mean bytes: 124766
Average live cells per slice (last five minutes): NaN
Maximum live cells per slice (last five minutes): 0
Average tombstones per slice (last five minutes): NaN
Maximum tombstones per slice (last five minutes): 0
Dropped Mutations: 697696

----------------


On Mon, Aug 6, 2018 at 8:57 PM, Jeff Jirsa <jji...@gmail.com> wrote:

>
>
> Upgrading to 3.11.3 May fix it (there were some memory recycling bugs
> fixed recently), but analyzing the heap will be the best option
>
> If you can print out the heap histogram and stack trace or open a heap
> dump in your kit or visualvm or MAT and show us what’s at the top of the
> reclaimed objects, we may be able to figure out what’s going on
>
> --
> Jeff Jirsa
>
>
> On Aug 6, 2018, at 5:42 PM, Jeff Jirsa <jji...@gmail.com> wrote:
>
> Are you using materialized views or secondary indices?
>
> --
> Jeff Jirsa
>
>
> On Aug 6, 2018, at 3:49 PM, Laszlo Szabo <laszlo.viktor.sz...@gmail.com>
> wrote:
>
> Hello All,
>
> I'm having JVM unstable / OOM errors when attempting to auto bootstrap a
> 9th node to an existing 8 node cluster (256 tokens).  Each machine has 24
> cores 148GB RAM and 10TB (2TB used).  Under normal operation the 8 nodes
> have JVM memory configured with Xms35G and Xmx35G, and handle 2-4 billion
> inserts per day.  There are never updates, deletes, or sparsely populated
> rows.
>
> For the bootstrap node, I've tried memory values from 35GB to 135GB in
> 10GB increments. I've tried using both memtable_allocation_types
> (heap_buffers and offheap_buffers).  I've not tried modifying the
> memtable_cleanup_threshold but instead have tried memtable_flush_writers
> from 2 to 8.  I've tried memtable_(off)heap_space_in_mb from 20000 to
> 60000.  I've tried both CMS and G1 garbage collection with various
> settings.
>
> Typically, after streaming about ~2TB of data, CPU load will hit a
> maximum, and the "nodetool info" heap memory will, over the course of an
> hour, approach the maximum.  At that point, CPU load will drop to a single
> thread with minimal activity until the system becomes unstable and
> eventually the OOM error occurs.
>
> Excerpt of the system log is below, and what I consistently see is the
> MemtableFlushWriter and the MemtableReclaimMemory pending queues grow as
> the memory becomes depleted, but the number of completed seems to stop
> changing a few minutes after the CPU load spikes.
>
> One other data point is there seems to be a huge number of mutations that
> occur after most of the stream has occured.  Concurrent_writes is set at
> 256 with the queue getting as high as 200K before dropping down.
>
> Any suggestions for yaml changes or jvm changes?  JVM.options is currently
> the default with the memory set to the max, the current YAML file is below.
>
> Thanks!
>
>
> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,329 StatusLogger.java:51 -
>> MutationStage                     1         2      191498052         0
>>            0
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,331 StatusLogger.java:51 -
>> ViewMutationStage                 0         0              0         0
>>            0
>> INFO  [Service Thread] 2018-08-06 17:49:26,338 StatusLogger.java:51 -
>> PerDiskMemtableFlushWriter_0         0         0           5865         0
>>                0
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,343 StatusLogger.java:51 -
>> ReadStage                         0         0              0         0
>>            0
>> INFO  [Service Thread] 2018-08-06 17:49:26,347 StatusLogger.java:51 -
>> ValidationExecutor                0         0              0         0
>>            0
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,360 StatusLogger.java:51 -
>> RequestResponseStage              0         0              8         0
>>            0
>> INFO  [Service Thread] 2018-08-06 17:49:26,380 StatusLogger.java:51 -
>> Sampler                           0         0              0         0
>>            0
>> INFO  [Service Thread] 2018-08-06 17:49:26,382 StatusLogger.java:51 - 
>> *MemtableFlushWriter
>>              8     74293           4716         0  *               0
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,388 StatusLogger.java:51 -
>> ReadRepairStage                   0         0              0         0
>>            0
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,389 StatusLogger.java:51 -
>> CounterMutationStage              0         0              0         0
>>            0
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,404 StatusLogger.java:51 -
>> MiscStage                         0         0              0         0
>>            0
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,407 StatusLogger.java:51 -
>> CompactionExecutor                8        13            493         0
>>            0
>> INFO  [Service Thread] 2018-08-06 17:49:26,410 StatusLogger.java:51 -
>> InternalResponseStage             0         0             16         0
>>            0
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,413 StatusLogger.java:51 - 
>> *MemtableReclaimMemory
>>            1      6066            356         0     *            0
>> INFO  [Service Thread] 2018-08-06 17:49:26,421 StatusLogger.java:51 -
>> AntiEntropyStage                  0         0              0         0
>>            0
>> INFO  [Service Thread] 2018-08-06 17:49:26,430 StatusLogger.java:51 -
>> CacheCleanupExecutor              0         0              0         0
>>            0
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,431 StatusLogger.java:51 -
>> PendingRangeCalculator            0         0              9         0
>>            0
>> INFO  [Service Thread] 2018-08-06 17:49:26,436 StatusLogger.java:61 -
>> CompactionManager                 8        19
>
>
>
>
>  Current Yaml
>
> num_tokens: 256
>
> hinted_handoff_enabled: true
>
> hinted_handoff_throttle_in_kb: 10240
>
> max_hints_delivery_threads: 8
>
> hints_flush_period_in_ms: 10000
>
> max_hints_file_size_in_mb: 128
>
> batchlog_replay_throttle_in_kb: 10240
>
> authenticator: AllowAllAuthenticator
>
> authorizer: AllowAllAuthorizer
>
> role_manager: CassandraRoleManager
>
> roles_validity_in_ms: 2000
>
> permissions_validity_in_ms: 2000
>
> credentials_validity_in_ms: 2000
>
> partitioner: org.apache.cassandra.dht.Murmur3Partitioner
>
> data_file_directories:
>
>     - /data/cassandra/data
>
> commitlog_directory: /data/cassandra/commitlog
>
> cdc_enabled: false
>
> disk_failure_policy: stop
>
> commit_failure_policy: stop
>
> prepared_statements_cache_size_mb:
>
> thrift_prepared_statements_cache_size_mb:
>
> key_cache_size_in_mb:
>
> key_cache_save_period: 14400
>
> row_cache_size_in_mb: 0
>
> row_cache_save_period: 0
>
> counter_cache_size_in_mb:
>
> counter_cache_save_period: 7200
>
> saved_caches_directory: /data/cassandra/saved_caches
>
> commitlog_sync: periodic
>
> commitlog_sync_period_in_ms: 10000
>
> commitlog_segment_size_in_mb: 32
>
> seed_provider:
>
>     - class_name: org.apache.cassandra.locator.SimpleSeedProvider
>
>       parameters:
>
>           - seeds: "10.1.1.11,10.1.1.12,10.1.1.13"
>
> concurrent_reads: 128
>
> concurrent_writes: 256
>
> concurrent_counter_writes: 96
>
> concurrent_materialized_view_writes: 32
>
> disk_optimization_strategy: spinning
>
> memtable_heap_space_in_mb: 61440
>
> memtable_offheap_space_in_mb: 61440
>
> memtable_allocation_type: heap_buffers
>
> commitlog_total_space_in_mb: 81920
>
> memtable_flush_writers: 8
>
>
>

Reply via email to