As Jonathan said earlier, you are hitting
https://issues.apache.org/jira/browse/CASSANDRA-2765

This will be fixed in 0.8.1 that is currently under a vote and should be
released soon (let's say beginning of next week, maybe sooner).

--
Sylvain

2011/6/23 Héctor Izquierdo Seliva <izquie...@strands.com>:
> Hi Aaron. Reverted back to 4-32. Did the flush but it did not trigger
> any minor compaction. Ran compact by hand, and it picked only two
> sstables.
>
> Here's the ls before:
>
> http://pastebin.com/xDtvVZvA
>
> And this is the ls after:
>
> http://pastebin.com/DcpbGvK6
>
> Any suggestions?
>
>
>
> El jue, 23-06-2011 a las 10:55 +1200, aaron morton escribió:
>> Setting them to 2 and 2 means compaction can only ever compact 2 files at 
>> time, so it will be worse off.
>>
>> Lets the try following:
>>
>> - restore the compactions settings to the default 4 and 32
>> - run `ls -lah` in the data dir and grab the output
>> - run `nodetool flush` this will trigger minor compaction once the memtables 
>> have been flushed
>> - check the logs for messages from 'CompactionManager'
>> - when done grab the output from  `ls -lah` again.
>>
>> Hope that helps.
>>
>>
>> -----------------
>> Aaron Morton
>> Freelance Cassandra Developer
>> @aaronmorton
>> http://www.thelastpickle.com
>>
>> On 23 Jun 2011, at 02:04, Héctor Izquierdo Seliva wrote:
>>
>> > Hi All. I set the compaction threshold at minimum 2, maximum 2 and try
>> > to run compact, but it's not doing anything. There are over 69 sstables
>> > now, read performance is horrible, and it's taking an insane amount of
>> > space. Maybe I don't quite get how the new per bucket stuff works, but I
>> > think this is not normal behaviour.
>> >
>> > El lun, 13-06-2011 a las 10:32 -0500, Jonathan Ellis escribió:
>> >> As Terje already said in this thread, the threshold is per bucket
>> >> (group of similarly sized sstables) not per CF.
>> >>
>> >> 2011/6/13 Héctor Izquierdo Seliva <izquie...@strands.com>:
>> >>> I was already way over the minimum. There were 12 sstables. Also, is
>> >>> there any reason why scrub got stuck? I did not see anything in the
>> >>> logs. Via jmx I saw that the scrubbed bytes were equal to one of the
>> >>> sstables size, and it stuck there for a couple hours .
>> >>>
>> >>> El lun, 13-06-2011 a las 22:55 +0900, Terje Marthinussen escribió:
>> >>>> That most likely happened just because after scrub you had new files
>> >>>> and got over the "4" file minimum limit.
>> >>>>
>> >>>> https://issues.apache.org/jira/browse/CASSANDRA-2697
>> >>>>
>> >>>> Is the bug report.
>> >>>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>
>> >>
>> >>
>> >
>> >
>>
>
>
>

Reply via email to