Sorry, the command is actually nodetool tablestats and you should watch the
bloom filter size or similar metrics.

2016-03-14 17:35 GMT-03:00 Mohamed Lrhazi <mohamed.lrh...@georgetown.edu>:

> Hi Paulo,
>
> Which metric should I watch for this ?
>
> [root@avesterra-prod-1 ~]# rpm -qa| grep datastax
> datastax-ddc-3.2.1-1.noarch
> datastax-ddc-tools-3.2.1-1.noarch
> [root@avesterra-prod-1 ~]# cassandra -v
> 3.2.1
> [root@avesterra-prod-1 ~]#
>
> [root@avesterra-prod-1 ~]# nodetool -u cassandra -pw '########'  tpstats
>
>
> Pool Name                    Active   Pending      Completed   Blocked
>  All time blocked
> MutationStage                     0         0          13609         0
>             0
> ViewMutationStage                 0         0              0         0
>             0
> ReadStage                         0         0              0         0
>             0
> RequestResponseStage              0         0              8         0
>             0
> ReadRepairStage                   0         0              0         0
>             0
> CounterMutationStage              0         0              0         0
>             0
> MiscStage                         0         0              0         0
>             0
> CompactionExecutor                1         1          17556         0
>             0
> MemtableReclaimMemory             0         0             38         0
>             0
> PendingRangeCalculator            0         0              8         0
>             0
> GossipStage                       0         0         118094         0
>             0
> SecondaryIndexManagement          0         0              0         0
>             0
> HintsDispatcher                   0         0              0         0
>             0
> MigrationStage                    0         0              0         0
>             0
> MemtablePostFlush                 0         0             55         0
>             0
> PerDiskMemtableFlushWriter_0         0         0             38         0
>                 0
> ValidationExecutor                0         0              0         0
>             0
> Sampler                           0         0              0         0
>             0
> MemtableFlushWriter               0         0             38         0
>             0
> InternalResponseStage             0         0              0         0
>             0
> AntiEntropyStage                  0         0              0         0
>             0
> CacheCleanupExecutor              0         0              0         0
>             0
> Native-Transport-Requests         0         0              0         0
>             0
>
> Message type           Dropped
> READ                         0
> RANGE_SLICE                  0
> _TRACE                       0
> HINT                         0
> MUTATION                     0
> COUNTER_MUTATION             0
> BATCH_STORE                  0
> BATCH_REMOVE                 0
> REQUEST_RESPONSE             0
> PAGED_RANGE                  0
> READ_REPAIR                  0
> [root@avesterra-prod-1 ~]#
>
>
>
>
> Thanks a lot,
> Mohamed.
>
>
>
> On Mon, Mar 14, 2016 at 8:22 AM, Paulo Motta <pauloricard...@gmail.com>
> wrote:
>
>> Can you check with nodetool tpstats if bloom filter mem space utilization
>> is very large/ramping up before the node gets killed? You could be hitting
>> CASSANDRA-11344.
>>
>> 2016-03-12 19:43 GMT-03:00 Mohamed Lrhazi <mohamed.lrh...@georgetown.edu>
>> :
>>
>>> In my case, all nodes seem to be constantly logging messages like these:
>>>
>>> DEBUG [GossipStage:1] 2016-03-12 17:41:19,123 FailureDetector.java:456 -
>>> Ignoring interval time of 2000928319 for /10.212.18.170
>>>
>>> What does that mean?
>>>
>>> Thanks a lot,
>>> Mohamed.
>>>
>>>
>>> On Sat, Mar 12, 2016 at 5:39 PM, Mohamed Lrhazi <
>>> mohamed.lrh...@georgetown.edu> wrote:
>>>
>>>> Oh wow, similar behavior with different version all together!!
>>>>
>>>> On Sat, Mar 12, 2016 at 5:28 PM, ssiv...@gmail.com <ssiv...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi, I'll duplicate here my email with the same issue
>>>>>
>>>>> "
>>>>>
>>>>>
>>>>> *I have 7 nodes of C* v2.2.5 running on CentOS 7 and using jemalloc
>>>>> for dynamic storage allocation. Use only one keyspace and one table with
>>>>> Leveled compaction strategy. I've loaded ~500 GB of data into the cluster
>>>>> with replication factor equals to 3 and waiting until compaction is
>>>>> finished. But during compaction each of the C* nodes allocates all the
>>>>> available memory (~128GB) and just stops its process. This is a known bug 
>>>>> ?
>>>>> *"
>>>>>
>>>>>
>>>>> On 03/13/2016 12:56 AM, Mohamed Lrhazi wrote:
>>>>>
>>>>> Hello,
>>>>>
>>>>> We installed Datastax community edition, on 8 nodes, RHEL7. We
>>>>> inserted some 7 billion rows into a pretty simple table. the inserts seem
>>>>> to have completed without issues. but ever since, we find that the nodes
>>>>> reliably run out of RAM after few hours, without any user activity at all.
>>>>> No reads nor write are sent at all.  What should we look for to try and
>>>>> identify root cause?
>>>>>
>>>>>
>>>>> [root@avesterra-prod-1 ~]# cat /etc/redhat-release
>>>>> Red Hat Enterprise Linux Server release 7.2 (Maipo)
>>>>> [root@avesterra-prod-1 ~]# rpm -qa| grep datastax
>>>>> datastax-ddc-3.2.1-1.noarch
>>>>> datastax-ddc-tools-3.2.1-1.noarch
>>>>> [root@avesterra-prod-1 ~]#
>>>>>
>>>>> The nodes had 8 GB RAM, which we doubled twice and now are trying with
>>>>> 40GB... they still manage to consume it all and cause oom_killer to kick 
>>>>> in.
>>>>>
>>>>> Pretty much all the settings are the default ones the installation
>>>>> created.
>>>>>
>>>>> Thanks,
>>>>> Mohamed.
>>>>>
>>>>>
>>>>> --
>>>>> Thanks,
>>>>> Serj
>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to