Indeed, sorry. Subscribed to both so missed which one this was. 

Sent from my iPhone

> On 11 May 2017, at 19:56, Michael Kjellman <mkjell...@internalcircle.com> 
> wrote:
> 
> This discussion should be on the C* user mailing list. Thanks!
> 
> best,
> kjellman
> 
>> On May 11, 2017, at 10:53 AM, Oskar Kjellin <oskar.kjel...@gmail.com> wrote:
>> 
>> That seems way too low. Depending on what type of disk you have it should be 
>> closer to 1-200MB.
>> That's probably causing your problems. It would still take a while for you 
>> to compact all your data tho 
>> 
>> Sent from my iPhone
>> 
>>> On 11 May 2017, at 19:50, varun saluja <saluj...@gmail.com> wrote:
>>> 
>>> nodetool getcompactionthrougput
>>> 
>>> ./nodetool getcompactionthroughput
>>> Current compaction throughput: 16 MB/s
>>> 
>>> Regards,
>>> Varun Saluja
>>> 
>>>> On 11 May 2017 at 23:18, varun saluja <saluj...@gmail.com> wrote:
>>>> Hi,
>>>> 
>>>> PFB results for same. Numbers are scary here.
>>>> 
>>>> [root@WA-CASSDB2 bin]# ./nodetool compactionstats
>>>> pending tasks: 137
>>>>  compaction type         keyspace                 table    completed       
>>>>    total    unit   progress
>>>>       Compaction           system                 hints   5762711108   
>>>> 837522028005   bytes      0.69%
>>>>       Compaction   walletkeyspace   user_txn_history_v2    101477894     
>>>> 4722068388   bytes      2.15%
>>>>       Compaction   walletkeyspace   user_txn_history_v2   1511866634   
>>>> 753221762663   bytes      0.20%
>>>>       Compaction   walletkeyspace   user_txn_history_v2   3664734135    
>>>> 18605501268   bytes     19.70%
>>>> Active compaction remaining time :  26h32m28s
>>>> 
>>>> 
>>>> 
>>>>> On 11 May 2017 at 23:15, Oskar Kjellin <oskar.kjel...@gmail.com> wrote:
>>>>> What does nodetool compactionstats show?
>>>>> 
>>>>> I meant compaction throttling. nodetool getcompactionthrougput
>>>>> 
>>>>> 
>>>>>> On 11 May 2017, at 19:41, varun saluja <saluj...@gmail.com> wrote:
>>>>>> 
>>>>>> Hi Oskar,
>>>>>> 
>>>>>> Thanks for response.
>>>>>> 
>>>>>> Yes, could see lot of threads for compaction. Actually we are loading 
>>>>>> around 400GB data  per node on 3 node cassandra cluster.
>>>>>> Throttling was set to write around 7k TPS per node. Job ran fine for 2 
>>>>>> days and then, we start getting Mutation drops  , longer GC and very 
>>>>>> high load on system.
>>>>>> 
>>>>>> System log reports:
>>>>>> Enqueuing flush of compactions_in_progress: 1156 (0%) on-heap, 1132 (0%) 
>>>>>> off-heap
>>>>>> 
>>>>>> The job was stopped 12 hours back. But, still these failures can be 
>>>>>> seen. Can you Please let me know how shall i proceed further. If 
>>>>>> possible, Please suggest some parameters for high write intensive jobs.
>>>>>> 
>>>>>> 
>>>>>> Regards,
>>>>>> Varun Saluja
>>>>>> 
>>>>>> 
>>>>>>> On 11 May 2017 at 23:01, Oskar Kjellin <oskar.kjel...@gmail.com> wrote:
>>>>>>> Do you have a lot of compactions going on? It sounds like you might've 
>>>>>>> built up a huge backlog. Is your throttling configured properly?
>>>>>>> 
>>>>>>>> On 11 May 2017, at 18:50, varun saluja <saluj...@gmail.com> wrote:
>>>>>>>> 
>>>>>>>> Hi Experts,
>>>>>>>> 
>>>>>>>> Seeking your help on a production issue.  We were running high write 
>>>>>>>> intensive job on our 3 node cassandra cluster V 2.1.7.
>>>>>>>> 
>>>>>>>> TPS on nodes were high. Job ran for more than 2 days and thereafter, 
>>>>>>>> loadavg on 1 of the node increased to very high number like loadavg : 
>>>>>>>> 29.
>>>>>>>> 
>>>>>>>> System log reports:
>>>>>>>> 
>>>>>>>> INFO  [ScheduledTasks:1] 2017-05-11 22:11:04,466 
>>>>>>>> MessagingService.java:888 - 839 MUTATION messages dropped in last 
>>>>>>>> 5000ms
>>>>>>>> INFO  [ScheduledTasks:1] 2017-05-11 22:11:04,466 
>>>>>>>> MessagingService.java:888 - 2 READ messages dropped in last 5000ms
>>>>>>>> INFO  [ScheduledTasks:1] 2017-05-11 22:11:04,466 
>>>>>>>> MessagingService.java:888 - 1 REQUEST_RESPONSE messages dropped in 
>>>>>>>> last 5000ms
>>>>>>>> 
>>>>>>>> The job was stopped due to heavy load. But sill after 12 hours , we 
>>>>>>>> can see mutation drops messages and sudden increase on avgload
>>>>>>>> 
>>>>>>>> Are these hintedhandoff mutations? Can we stop these.
>>>>>>>> Strangely this behaviour is seen only on 2 nodes. Node 1 does not show 
>>>>>>>> any load or any such activity.
>>>>>>>> 
>>>>>>>> Due to heavy load and GC , there are intermittent gossip failures 
>>>>>>>> among node. Can you someone Please help.
>>>>>>>> 
>>>>>>>> PS: Load job was stopped on cluster. Everything ran fine for few hours 
>>>>>>>> and and Later issue started again like mutation messages drops.
>>>>>>>> 
>>>>>>>> Thanks and Regards,
>>>>>>>> Varun Saluja
>>>>>>>> 
>>>>>>>> ---------------------------------------------------------------------
>>>>>>>> To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org
>>>>>>>> For additional commands, e-mail: dev-h...@cassandra.apache.org
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
> 

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org
For additional commands, e-mail: dev-h...@cassandra.apache.org

Reply via email to