Quick follow up : triggering a repair did the trick, sstables are starting
to get compacted.

Thank you

On 13 November 2017 at 15:53, Nicolas Guyomar <nicolas.guyo...@gmail.com>
wrote:

> Hi,
>
> I'm using default "nodetool repair" in 3.0.13 which I believe is full by
> default. I'm not using subrange repair
>
> Jeff you're right, "Nov 11 01:23 mc-6474-big-Data.db" is not yet marked as
> repaired, my repair routine is broken  (sorry Alexander I'm not using
> repear yet ;)  )
>
> I'm gonna fix my repair script and see if it unlock this situation
>
> Thank you
>
>
> On 13 November 2017 at 15:41, Alexander Dejanovski <a...@thelastpickle.com
> > wrote:
>
>> And actually, full repair with 3.0/3.x would have the same effect
>> (anticompation) unless you're using subrange repair.
>>
>> On Mon, Nov 13, 2017 at 3:28 PM Jeff Jirsa <jji...@gmail.com> wrote:
>>
>>> Running incremental repair puts sstables into a “repaired” set (and an
>>> unrepaired set), which results in something similar to what you’re
>>> describing.
>>>
>>> Were you running / did you run incremental repair ?
>>>
>>>
>>> --
>>> Jeff Jirsa
>>>
>>>
>>> On Nov 13, 2017, at 5:04 AM, Nicolas Guyomar <nicolas.guyo...@gmail.com>
>>> wrote:
>>>
>>> Hi everyone,
>>>
>>> I'm facing quite a strange behavior on STCS on 3.0.13, the strategy
>>> seems to have "forgotten" about old sstables, and started a completely new
>>> cycle from scratch, leaving the old sstables on disk untouched :
>>>
>>> Something happened on Nov 10 on every node, which resulted in all those
>>> sstables left behind :
>>>
>>> -rw-r--r--.  8 cassandra cassandra   15G Nov  9 22:22 mc-4828-big-Data.db
>>> -rw-r--r--.  8 cassandra cassandra  4.8G Nov 10 01:39 mc-4955-big-Data.db
>>> -rw-r--r--.  8 cassandra cassandra  2.4G Nov 10 01:45 mc-4957-big-Data.db
>>> -rw-r--r--.  8 cassandra cassandra  662M Nov 10 01:47 mc-4959-big-Data.db
>>> -rw-r--r--.  8 cassandra cassandra  2.8G Nov 10 03:46 mc-5099-big-Data.db
>>> -rw-r--r--.  8 cassandra cassandra  4.6G Nov 10 03:58 mc-5121-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra   53M Nov 10 08:45 mc-5447-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra  219M Nov 10 08:46 mc-5454-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra  650M Nov 10 08:46 mc-5452-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra  1.2G Nov 10 08:48 mc-5458-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra  1.5G Nov 10 08:50 mc-5465-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra  504M Nov 10 09:39 mc-5526-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra   57M Nov 10 09:40 mc-5527-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra  101M Nov 10 09:41 mc-5532-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra   86M Nov 10 09:41 mc-5533-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra  134M Nov 10 09:42 mc-5537-big-Data.db
>>> -rw-r--r--.  7 cassandra cassandra  3.9G Nov 10 09:54 mc-5538-big-Data.db
>>> *-rw-r--r--.  7 cassandra cassandra  1.3G Nov 10 09:57
>>> mc-5548-big-Data.db*
>>> -rw-r--r--.  6 cassandra cassandra   16G Nov 11 01:23 mc-6474-big-Data.db
>>> -rw-r--r--.  4 cassandra cassandra   17G Nov 12 06:44 mc-7898-big-Data.db
>>> -rw-r--r--.  3 cassandra cassandra  8.2G Nov 12 13:45 mc-8226-big-Data.db
>>> -rw-r--r--.  2 cassandra cassandra  6.8G Nov 12 22:38 mc-8581-big-Data.db
>>> -rw-r--r--.  2 cassandra cassandra  6.1G Nov 13 03:10 mc-8937-big-Data.db
>>> -rw-r--r--.  2 cassandra cassandra  3.1G Nov 13 04:12 mc-9019-big-Data.db
>>> -rw-r--r--.  2 cassandra cassandra  3.0G Nov 13 05:56 mc-9112-big-Data.db
>>> -rw-r--r--.  2 cassandra cassandra  1.2G Nov 13 06:14 mc-9138-big-Data.db
>>> -rw-r--r--.  2 cassandra cassandra  1.1G Nov 13 06:27 mc-9159-big-Data.db
>>> -rw-r--r--.  2 cassandra cassandra  1.2G Nov 13 06:46 mc-9182-big-Data.db
>>> -rw-r--r--.  1 cassandra cassandra  1.9G Nov 13 07:18 mc-9202-big-Data.db
>>> -rw-r--r--.  1 cassandra cassandra  353M Nov 13 07:22 mc-9207-big-Data.db
>>> -rw-r--r--.  1 cassandra cassandra  120M Nov 13 07:22 mc-9208-big-Data.db
>>> -rw-r--r--.  1 cassandra cassandra  100M Nov 13 07:23 mc-9209-big-Data.db
>>> -rw-r--r--.  1 cassandra cassandra   67M Nov 13 07:25 mc-9210-big-Data.db
>>> -rw-r--r--.  1 cassandra cassandra   51M Nov 13 07:25 mc-9211-big-Data.db
>>> -rw-r--r--.  1 cassandra cassandra   73M Nov 13 07:27 mc-9212-big-Data.db
>>>
>>>
>>> TRACE logs for the Compaction Manager shows that sstables before Nov 10
>>> are grouped in different buckets than the one after Nov 10.
>>>
>>> At first I thought off some coldness behavior that would filter those
>>> "old" sstables, but looking at the code https://github.com/apache/cass
>>> andra/blob/cassandra-3.0/src/java/org/apache/cassandra/db/
>>> compaction/SizeTieredCompactionStrategy.java#L237 I don't see any
>>> coldness or time pattern used to create bucket.
>>>
>>> I tried restarting the node but the buckets are still grouping in 2
>>> "groups" splitted around Nov 10
>>>
>>> I may have missed sthg from the logs, but they are clear from error/warn
>>> at that Nov 10 time
>>>
>>> For what it's worth, restarting the node fixed nodetool status from
>>> reporting a wrong Load (nearly 2TB per node instead à 300Gb) => we are
>>> loading some data for a week now, it seems that this can happen sometimes
>>>
>>> If anyone ever experienced that kind of behavior I'd be glad to know
>>> whether it is OK or not, I'd like to avoid manually triggering JMX
>>> UserDefinedCompaction ;)
>>>
>>> Thank you
>>>
>>> --
>> -----------------
>> Alexander Dejanovski
>> France
>> @alexanderdeja
>>
>> Consultant
>> Apache Cassandra Consulting
>> http://www.thelastpickle.com
>>
>
>

Reply via email to