Ja, Please upgrade to official 2.1.3 we've fixed many things related to
compaction.  Are you seeing the compactions % complete progress at all?

On Wed, Feb 18, 2015 at 11:58 AM, Roni Balthazar <ronibaltha...@gmail.com>
wrote:

> Try repair -pr on all nodes.
>
> If after that you still have issues, you can try to rebuild the SSTables
> using nodetool upgradesstables or scrub.
>
> Regards,
>
> Roni Balthazar
>
> Em 18/02/2015, às 14:13, Ja Sam <ptrstp...@gmail.com> escreveu:
>
> ad 3)  I did this already yesterday (setcompactionthrouput also). But
> still SSTables are increasing.
>
> ad 1) What do you think I should use -pr or try to use incremental?
>
>
>
> On Wed, Feb 18, 2015 at 4:54 PM, Roni Balthazar <ronibaltha...@gmail.com>
> wrote:
>
>> You are right... Repair makes the data consistent between nodes.
>>
>> I understand that you have 2 issues going on.
>>
>> You need to run repair periodically without errors and need to decrease
>> the numbers of compactions pending.
>>
>> So I suggest:
>>
>> 1) Run repair -pr on all nodes. If you upgrade to the new 2.1.3, you can
>> use incremental repairs. There were some bugs on 2.1.2.
>> 2) Run cleanup on all nodes
>> 3) Since you have too many cold SSTables, set cold_reads_to_omit to 0.0,
>> and increase setcompactionthroughput for some time and see if the number
>> of SSTables is going down.
>>
>> Let us know what errors are you getting when running repairs.
>>
>> Regards,
>>
>> Roni Balthazar
>>
>>
>> On Wed, Feb 18, 2015 at 1:31 PM, Ja Sam <ptrstp...@gmail.com> wrote:
>>
>>> Can you explain me what is the correlation between growing SSTables and
>>> repair?
>>> I was sure, until your  mail, that repair is only to make data
>>> consistent between nodes.
>>>
>>> Regards
>>>
>>>
>>> On Wed, Feb 18, 2015 at 4:20 PM, Roni Balthazar <ronibaltha...@gmail.com
>>> > wrote:
>>>
>>>> Which error are you getting when running repairs?
>>>> You need to run repair on your nodes within gc_grace_seconds (eg:
>>>> weekly). They have data that are not read frequently. You can run
>>>> "repair -pr" on all nodes. Since you do not have deletes, you will not
>>>> have trouble with that. If you have deletes, it's better to increase
>>>> gc_grace_seconds before the repair.
>>>>
>>>> http://www.datastax.com/documentation/cassandra/2.0/cassandra/operations/ops_repair_nodes_c.html
>>>> After repair, try to run a "nodetool cleanup".
>>>>
>>>> Check if the number of SSTables goes down after that... Pending
>>>> compactions must decrease as well...
>>>>
>>>> Cheers,
>>>>
>>>> Roni Balthazar
>>>>
>>>>
>>>>
>>>>
>>>> On Wed, Feb 18, 2015 at 12:39 PM, Ja Sam <ptrstp...@gmail.com> wrote:
>>>> > 1) we tried to run repairs but they usually does not succeed. But we
>>>> had
>>>> > Leveled compaction before. Last week we ALTER tables to STCS, because
>>>> guys
>>>> > from DataStax suggest us that we should not use Leveled and alter
>>>> tables in
>>>> > STCS, because we don't have SSD. After this change we did not run any
>>>> > repair. Anyway I don't think it will change anything in SSTable count
>>>> - if I
>>>> > am wrong please give me an information
>>>> >
>>>> > 2) I did this. My tables are 99% write only. It is audit system
>>>> >
>>>> > 3) Yes I am using default values
>>>> >
>>>> > 4) In both operations I am using LOCAL_QUORUM.
>>>> >
>>>> > I am almost sure that READ timeout happens because of too much
>>>> SSTables.
>>>> > Anyway firstly I would like to fix to many pending compactions. I
>>>> still
>>>> > don't know how to speed up them.
>>>> >
>>>> >
>>>> > On Wed, Feb 18, 2015 at 2:49 PM, Roni Balthazar <
>>>> ronibaltha...@gmail.com>
>>>> > wrote:
>>>> >>
>>>> >> Are you running repairs within gc_grace_seconds? (default is 10 days)
>>>> >>
>>>> >>
>>>> http://www.datastax.com/documentation/cassandra/2.0/cassandra/operations/ops_repair_nodes_c.html
>>>> >>
>>>> >> Double check if you set cold_reads_to_omit to 0.0 on tables with STCS
>>>> >> that you do not read often.
>>>> >>
>>>> >> Are you using default values for the properties
>>>> >> min_compaction_threshold(4) and max_compaction_threshold(32)?
>>>> >>
>>>> >> Which Consistency Level are you using for reading operations? Check
>>>> if
>>>> >> you are not reading from DC_B due to your Replication Factor and CL.
>>>> >>
>>>> >>
>>>> http://www.datastax.com/documentation/cassandra/2.0/cassandra/dml/dml_config_consistency_c.html
>>>> >>
>>>> >>
>>>> >> Cheers,
>>>> >>
>>>> >> Roni Balthazar
>>>> >>
>>>> >> On Wed, Feb 18, 2015 at 11:07 AM, Ja Sam <ptrstp...@gmail.com>
>>>> wrote:
>>>> >> > I don't have problems with DC_B (replica) only in DC_A(my system
>>>> write
>>>> >> > only
>>>> >> > to it) I have read timeouts.
>>>> >> >
>>>> >> > I checked in OpsCenter SSTable count  and I have:
>>>> >> > 1) in DC_A  same +-10% for last week, a small increase for last
>>>> 24h (it
>>>> >> > is
>>>> >> > more than 15000-20000 SSTables depends on node)
>>>> >> > 2) in DC_B last 24h shows up to 50% decrease, which give nice
>>>> >> > prognostics.
>>>> >> > Now I have less then 1000 SSTables
>>>> >> >
>>>> >> > What did you measure during system optimizations? Or do you have
>>>> an idea
>>>> >> > what more should I check?
>>>> >> > 1) I look at CPU Idle (one node is 50% idle, rest 70% idle)
>>>> >> > 2) Disk queue -> mostly is it near zero: avg 0.09. Sometimes there
>>>> are
>>>> >> > spikes
>>>> >> > 3) system RAM usage is almost full
>>>> >> > 4) In Total Bytes Compacted most most lines are below 3MB/s. For
>>>> total
>>>> >> > DC_A
>>>> >> > it is less than 10MB/s, in DC_B it looks much better (avg is like
>>>> >> > 17MB/s)
>>>> >> >
>>>> >> > something else?
>>>> >> >
>>>> >> >
>>>> >> >
>>>> >> > On Wed, Feb 18, 2015 at 1:32 PM, Roni Balthazar
>>>> >> > <ronibaltha...@gmail.com>
>>>> >> > wrote:
>>>> >> >>
>>>> >> >> Hi,
>>>> >> >>
>>>> >> >> You can check if the number of SSTables is decreasing. Look for
>>>> the
>>>> >> >> "SSTable count" information of your tables using "nodetool
>>>> cfstats".
>>>> >> >> The compaction history can be viewed using "nodetool
>>>> >> >> compactionhistory".
>>>> >> >>
>>>> >> >> About the timeouts, check this out:
>>>> >> >>
>>>> >> >>
>>>> http://www.datastax.com/dev/blog/how-cassandra-deals-with-replica-failure
>>>> >> >> Also try to run "nodetool tpstats" to see the threads statistics.
>>>> It
>>>> >> >> can lead you to know if you are having performance problems. If
>>>> you
>>>> >> >> are having too many pending tasks or dropped messages, maybe will
>>>> you
>>>> >> >> need to tune your system (eg: driver's timeout, concurrent reads
>>>> and
>>>> >> >> so on)
>>>> >> >>
>>>> >> >> Regards,
>>>> >> >>
>>>> >> >> Roni Balthazar
>>>> >> >>
>>>> >> >> On Wed, Feb 18, 2015 at 9:51 AM, Ja Sam <ptrstp...@gmail.com>
>>>> wrote:
>>>> >> >> > Hi,
>>>> >> >> > Thanks for your "tip" it looks that something changed - I still
>>>> don't
>>>> >> >> > know
>>>> >> >> > if it is ok.
>>>> >> >> >
>>>> >> >> > My nodes started to do more compaction, but it looks that some
>>>> >> >> > compactions
>>>> >> >> > are really slow.
>>>> >> >> > In IO we have idle, CPU is quite ok (30%-40%). We set
>>>> >> >> > compactionthrouput
>>>> >> >> > to
>>>> >> >> > 999, but I do not see difference.
>>>> >> >> >
>>>> >> >> > Can we check something more? Or do you have any method to
>>>> monitor
>>>> >> >> > progress
>>>> >> >> > with small files?
>>>> >> >> >
>>>> >> >> > Regards
>>>> >> >> >
>>>> >> >> > On Tue, Feb 17, 2015 at 2:43 PM, Roni Balthazar
>>>> >> >> > <ronibaltha...@gmail.com>
>>>> >> >> > wrote:
>>>> >> >> >>
>>>> >> >> >> HI,
>>>> >> >> >>
>>>> >> >> >> Yes... I had the same issue and setting cold_reads_to_omit to
>>>> 0.0
>>>> >> >> >> was
>>>> >> >> >> the solution...
>>>> >> >> >> The number of SSTables decreased from many thousands to a
>>>> number
>>>> >> >> >> below
>>>> >> >> >> a hundred and the SSTables are now much bigger with several
>>>> >> >> >> gigabytes
>>>> >> >> >> (most of them).
>>>> >> >> >>
>>>> >> >> >> Cheers,
>>>> >> >> >>
>>>> >> >> >> Roni Balthazar
>>>> >> >> >>
>>>> >> >> >>
>>>> >> >> >>
>>>> >> >> >> On Tue, Feb 17, 2015 at 11:32 AM, Ja Sam <ptrstp...@gmail.com>
>>>> >> >> >> wrote:
>>>> >> >> >> > After some diagnostic ( we didn't set yet cold_reads_to_omit
>>>> ).
>>>> >> >> >> > Compaction
>>>> >> >> >> > are running but VERY slow with "idle" IO.
>>>> >> >> >> >
>>>> >> >> >> > We had a lot of "Data files" in Cassandra. In DC_A it is
>>>> about
>>>> >> >> >> > ~120000
>>>> >> >> >> > (only
>>>> >> >> >> > xxx-Data.db) in DC_B has only ~4000.
>>>> >> >> >> >
>>>> >> >> >> > I don't know if this change anything but:
>>>> >> >> >> > 1) in DC_A avg size of Data.db file is ~13 mb. I have few a
>>>> really
>>>> >> >> >> > big
>>>> >> >> >> > ones,
>>>> >> >> >> > but most is really small (almost 10000 files are less then
>>>> 100mb).
>>>> >> >> >> > 2) in DC_B avg size of Data.db is much bigger ~260mb.
>>>> >> >> >> >
>>>> >> >> >> > Do you think that above flag will help us?
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >> >> > On Tue, Feb 17, 2015 at 9:04 AM, Ja Sam <ptrstp...@gmail.com
>>>> >
>>>> >> >> >> > wrote:
>>>> >> >> >> >>
>>>> >> >> >> >> I set setcompactionthroughput 999 permanently and it doesn't
>>>> >> >> >> >> change
>>>> >> >> >> >> anything. IO is still same. CPU is idle.
>>>> >> >> >> >>
>>>> >> >> >> >> On Tue, Feb 17, 2015 at 1:15 AM, Roni Balthazar
>>>> >> >> >> >> <ronibaltha...@gmail.com>
>>>> >> >> >> >> wrote:
>>>> >> >> >> >>>
>>>> >> >> >> >>> Hi,
>>>> >> >> >> >>>
>>>> >> >> >> >>> You can run "nodetool compactionstats" to view statistics
>>>> on
>>>> >> >> >> >>> compactions.
>>>> >> >> >> >>> Setting cold_reads_to_omit to 0.0 can help to reduce the
>>>> number
>>>> >> >> >> >>> of
>>>> >> >> >> >>> SSTables when you use Size-Tiered compaction.
>>>> >> >> >> >>> You can also create a cron job to increase the value of
>>>> >> >> >> >>> setcompactionthroughput during the night or when your IO
>>>> is not
>>>> >> >> >> >>> busy.
>>>> >> >> >> >>>
>>>> >> >> >> >>> From http://wiki.apache.org/cassandra/NodeTool:
>>>> >> >> >> >>> 0 0 * * * root nodetool -h `hostname`
>>>> setcompactionthroughput
>>>> >> >> >> >>> 999
>>>> >> >> >> >>> 0 6 * * * root nodetool -h `hostname`
>>>> setcompactionthroughput 16
>>>> >> >> >> >>>
>>>> >> >> >> >>> Cheers,
>>>> >> >> >> >>>
>>>> >> >> >> >>> Roni Balthazar
>>>> >> >> >> >>>
>>>> >> >> >> >>> On Mon, Feb 16, 2015 at 7:47 PM, Ja Sam <
>>>> ptrstp...@gmail.com>
>>>> >> >> >> >>> wrote:
>>>> >> >> >> >>> > One think I do not understand. In my case compaction is
>>>> >> >> >> >>> > running
>>>> >> >> >> >>> > permanently.
>>>> >> >> >> >>> > Is there a way to check which compaction is pending? The
>>>> only
>>>> >> >> >> >>> > information is
>>>> >> >> >> >>> > about total count.
>>>> >> >> >> >>> >
>>>> >> >> >> >>> >
>>>> >> >> >> >>> > On Monday, February 16, 2015, Ja Sam <
>>>> ptrstp...@gmail.com>
>>>> >> >> >> >>> > wrote:
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >> Of couse I made a mistake. I am using 2.1.2. Anyway
>>>> night
>>>> >> >> >> >>> >> build
>>>> >> >> >> >>> >> is
>>>> >> >> >> >>> >> available from
>>>> >> >> >> >>> >> http://cassci.datastax.com/job/cassandra-2.1/
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >> I read about cold_reads_to_omit It looks promising.
>>>> Should I
>>>> >> >> >> >>> >> set
>>>> >> >> >> >>> >> also
>>>> >> >> >> >>> >> compaction throughput?
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >> p.s. I am really sad that I didn't read this before:
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >>
>>>> https://engineering.eventbrite.com/what-version-of-cassandra-should-i-run/
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >> On Monday, February 16, 2015, Carlos Rolo <
>>>> r...@pythian.com>
>>>> >> >> >> >>> >> wrote:
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>> Hi 100% in agreement with Roland,
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>> 2.1.x series is a pain! I would never recommend the
>>>> current
>>>> >> >> >> >>> >>> 2.1.x
>>>> >> >> >> >>> >>> series
>>>> >> >> >> >>> >>> for production.
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>> Clocks is a pain, and check your connectivity! Also
>>>> check
>>>> >> >> >> >>> >>> tpstats
>>>> >> >> >> >>> >>> to
>>>> >> >> >> >>> >>> see
>>>> >> >> >> >>> >>> if your threadpools are being overrun.
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>> Regards,
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>> Carlos Juzarte Rolo
>>>> >> >> >> >>> >>> Cassandra Consultant
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>> Pythian - Love your data
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>> rolo@pythian | Twitter: cjrolo | Linkedin:
>>>> >> >> >> >>> >>> linkedin.com/in/carlosjuzarterolo
>>>> >> >> >> >>> >>> Tel: 1649
>>>> >> >> >> >>> >>> www.pythian.com
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>> On Mon, Feb 16, 2015 at 8:12 PM, Roland Etzenhammer
>>>> >> >> >> >>> >>> <r.etzenham...@t-online.de> wrote:
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>> Hi,
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>> 1) Actual Cassandra 2.1.3, it was upgraded from 2.1.0
>>>> >> >> >> >>> >>>> (suggested
>>>> >> >> >> >>> >>>> by
>>>> >> >> >> >>> >>>> Al
>>>> >> >> >> >>> >>>> Tobey from DataStax)
>>>> >> >> >> >>> >>>> 7) minimal reads (usually none, sometimes few)
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>> those two points keep me repeating an anwser I got.
>>>> First
>>>> >> >> >> >>> >>>> where
>>>> >> >> >> >>> >>>> did
>>>> >> >> >> >>> >>>> you
>>>> >> >> >> >>> >>>> get 2.1.3 from? Maybe I missed it, I will have a
>>>> look. But
>>>> >> >> >> >>> >>>> if
>>>> >> >> >> >>> >>>> it
>>>> >> >> >> >>> >>>> is
>>>> >> >> >> >>> >>>> 2.1.2
>>>> >> >> >> >>> >>>> whis is the latest released version, that version has
>>>> many
>>>> >> >> >> >>> >>>> bugs -
>>>> >> >> >> >>> >>>> most of
>>>> >> >> >> >>> >>>> them I got kicked by while testing 2.1.2. I got many
>>>> >> >> >> >>> >>>> problems
>>>> >> >> >> >>> >>>> with
>>>> >> >> >> >>> >>>> compactions not beeing triggred on column families not
>>>> >> >> >> >>> >>>> beeing
>>>> >> >> >> >>> >>>> read,
>>>> >> >> >> >>> >>>> compactions and repairs not beeing completed.  See
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> https://www.mail-archive.com/search?l=user@cassandra.apache.org&q=subject:%22Re%3A+Compaction+failing+to+trigger%22&o=newest&f=1
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>>
>>>> https://www.mail-archive.com/user%40cassandra.apache.org/msg40768.html
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>> Apart from that, how are those both datacenters
>>>> connected?
>>>> >> >> >> >>> >>>> Maybe
>>>> >> >> >> >>> >>>> there
>>>> >> >> >> >>> >>>> is a bottleneck.
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>> Also do you have ntp up and running on all nodes to
>>>> keep
>>>> >> >> >> >>> >>>> all
>>>> >> >> >> >>> >>>> clocks
>>>> >> >> >> >>> >>>> in
>>>> >> >> >> >>> >>>> thight sync?
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>> Note: I'm no expert (yet) - just sharing my 2 cents.
>>>> >> >> >> >>> >>>>
>>>> >> >> >> >>> >>>> Cheers,
>>>> >> >> >> >>> >>>> Roland
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>> --
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >>>
>>>> >> >> >> >>> >
>>>> >> >> >> >>
>>>> >> >> >> >>
>>>> >> >> >> >
>>>> >> >> >
>>>> >> >> >
>>>> >> >
>>>> >> >
>>>> >
>>>> >
>>>>
>>>
>>>
>>
>


-- 
http://twitter.com/tjake

Reply via email to