Frank,

The same advice for investigating holds: add the VM flags
-XX:+PrintSafepointStatistics -XX:PrintSafepointStatisticsCount=1
(you could put something above 1 there, to reduce the amount of
logging, since a pause of 52s will be pretty obvious even if
aggregated with lots of other safe points; the count is the number of
safepoints to aggregate into one log message)


52s is a very extreme pause, and I would be surprised if revoke bias
could cause this. I wonder if the VM is swapping out.



On 29 January 2014 19:02, Frank Ng <fnt...@gmail.com> wrote:

> Thanks for the update.  Our logs indicated that there were 0 pending for
> CompactionManager at that time.  Also, there were no nodetool repairs
> running at that time.  The log statements above state that the application
> had to stop to reach a safepoint.  Yet, it doesn't say what is requesting
> the safepoint.
>
>
> On Wed, Jan 29, 2014 at 1:20 PM, Shao-Chuan Wang <
> shaochuan.w...@bloomreach.com> wrote:
>
>> We had similar latency spikes when pending compactions can't keep it up
>> or repair/streaming taking too much cycles.
>>
>>
>> On Wed, Jan 29, 2014 at 10:07 AM, Frank Ng <fnt...@gmail.com> wrote:
>>
>>> All,
>>>
>>> We've been having intermittent long application pauses (version 1.2.8)
>>> and not sure if it's a cassandra bug.  During these pauses, there are
>>> dropped messages in the cassandra log file along with the node seeing other
>>> nodes as down.  We've turned on gc logging and the following is an example
>>> of a long "stopped" or pause event in the gc.log file.
>>>
>>> 2014-01-28T23:11:12.183-0500: 1337654.424: Total time for which
>>> application threads were stopped: 0.091450 seconds
>>> 2014-01-28T23:14:11.161-0500: 1337833.401: Total time for which
>>> application threads were stopped: 51.8190260 seconds
>>> 2014-01-28T23:14:19.870-0500: 1337842.111: Total time for which
>>> application threads were stopped: 0.005470 seconds
>>>
>>> As seen above, there was a 0.091450 secs pause, then a 51.8190260 secs
>>> pause.  There were no GC log events between those 2 log statements.  Since
>>> there's no GC logs in between, something else must be causing the long stop
>>> time to reach a safepoint.
>>>
>>> Could there be a Cassandra thread that is taking a long time to reach a
>>> safepoint and what is it trying to do? Along with the node seeing other
>>> nodes as down in the cassandra log file, the StatusLogger shows 1599
>>> Pending in ReadStage and 9 Pending in MutationStage.
>>>
>>> There is mention of cassandra batch revoke bias locks as a possible
>>> cause (not GC) via:
>>> http://www.mail-archive.com/user@cassandra.apache.org/msg34401.html
>>>
>>> We have JNA, no swap, and the cluster runs fine besides there
>>> intermittent long pause that can cause a node to appear down to other
>>> nodes.  Any ideas as the cause of the long pause above? It seems not
>>> related to GC.
>>>
>>> thanks.
>>>
>>>
>>
>

Reply via email to