Perhaps doing a sstable2json on some of the small tables would shed some
illumination.  I was going to suggest the anticompaction feature of C*2.1
(which I'm not familiar with), but you're on 2.0.

On Tue, Jun 9, 2015 at 11:11 AM, Anuj Wadehra <anujw_2...@yahoo.co.in>
wrote:

> We were facing dropped mutations earlier and we increased flush writers.
> Now there are no dropped mutations in tpstats. To repair the damaged vnodes
> / inconsistent data we executed repair -pr on all nodes. Still, we see the
> same problem.
>
> When we analyze repair logs we see 2 strange things:
>
> 1. "Out of sync" ranges for cf which are not being actively being
> written/updated while the repair is going on. When we repaired all data by
> repair -pr on all nodes, why out of sync data?
>
> 2. For some cf , repair logs shows that all ranges are consistent. Still
> we get so many sstables created during repair. When everything is in sync ,
> why repair creates tiny sstables to repair data?
>
> Thanks
> Anuj Wadehra
>
> Sent from Yahoo Mail on Android
> <https://overview.mail.yahoo.com/mobile/?.src=Android>
> ------------------------------
>   *From*:"Ken Hancock" <ken.hanc...@schange.com>
> *Date*:Tue, 9 Jun, 2015 at 8:24 pm
> *Subject*:Re: Hundreds of sstables after every Repair
>
> I think this came up recently in another thread.  If you're getting large
> numbers of SSTables after repairs, that means that your nodes are diverging
> from the keys that they're supposed to be having.  Likely you're dropping
> mutations.  Do a nodetool tpstats on each of your nodes and look at the
> mutation droppped counters.  If you're seeing dropped message, my money you
> have a non-zero FlushWriter "All time blocked" stat which is causing
> mutations to be dropped.
>
>
>
> On Tue, Jun 9, 2015 at 10:35 AM, Anuj Wadehra <anujw_2...@yahoo.co.in>
> wrote:
>
>> Any suggestions or comments on this one?
>>
>> Thanks
>> Anuj Wadehra
>>
>> Sent from Yahoo Mail on Android
>> <https://overview.mail.yahoo.com/mobile/?.src=Android>
>> ------------------------------
>>   *From*:"Anuj Wadehra" <anujw_2...@yahoo.co.in>
>> *Date*:Sun, 7 Jun, 2015 at 1:54 am
>> *Subject*:Hundreds of sstables after every Repair
>>
>> Hi,
>>
>> We are using 2.0.3 and vnodes. After every repair -pr operation  50+ tiny
>> sstables( <10K) get created. And these sstables never get compacted due to
>> coldness issue. I have raised
>> https://issues.apache.org/jira/browse/CASSANDRA-9146 for this issue but
>> I have been told to upgrade. Till we upgrade to latest 2.0.x , we are
>> stuck. Upgrade takes time, testing and planning in Production systems :(
>>
>> I have observed that even if vnodes are NOT damaged, hundreds of tiny
>> sstables are created during repair for a wide row CF. This is beyond my
>> understanding. If everything is consistent, and for the entire repair
>> process Cassandra is saying "Endpoints /x.x.x.x and /x.x.x.y are consistent
>> for <CF>". Whats the need of creating sstables?
>>
>> Is there any alternative to regular major compaction to deal with
>> situation?
>>
>>
>> Thanks
>> Anuj Wadehra
>>
>>
>
>
>
>
>
>

Reply via email to