Ended up switching the biggest offending column families back to size tiered compaction and pending compactions across the cluster dropped to 0 very quickly.
On Sep 19, 2012, at 10:55 PM, "Michael Kjellman" <mkjell...@barracuda.com> wrote: > After changing my ss_table_size as recommended my pending compactions across > the cluster have leveled off at 34808 but it isn't progressing after 24 hours > at that level. > > As I've already changed the most offending column families I think the only > option I have left is to remove the .json files from all of the column > families and do another rolling restart... > > Developing... Thanks for the help so far > > On Sep 19, 2012, at 10:35 PM, "Віталій Тимчишин" > <tiv...@gmail.com<mailto:tiv...@gmail.com>> wrote: > > I did see problems with schema agreement on 1.1.4, but they did go away after > rolling restart (BTW: it would be still good to check describe schema for > unreachable). Same rolling restart helped to force compactions after moving > to Leveled compaction. If your compactions still don't go, you can try > removing *.json files from the data directory of the stopped node to force > moving all SSTables to level0. > > Best regards, Vitalii Tymchyshyn > > 2012/9/19 Michael Kjellman > <mkjell...@barracuda.com<mailto:mkjell...@barracuda.com>> > Potentially the pending compactions are a symptom and not the root > cause/problem. > > When updating a 3rd column family with a larger sstable_size_in_mb it > looks like the schema may not be in a good state > > [default@xxxx] UPDATE COLUMN FAMILY screenshots WITH > compaction_strategy=LeveledCompactionStrategy AND > compaction_strategy_options={sstable_size_in_mb: 200}; > 290cf619-57b0-3ad1-9ae3-e313290de9c9 > Waiting for schema agreement... > Warning: unreachable nodes 10.8.30.102The schema has not settled in 10 > seconds; further migrations are ill-advised until it does. > Versions are UNREACHABLE:[10.8.30.102], > 290cf619-57b0-3ad1-9ae3-e313290de9c9:[10.8.30.15, 10.8.30.14, 10.8.30.13, > 10.8.30.103, 10.8.30.104, 10.8.30.105, 10.8.30.106], > f1de54f5-8830-31a6-9cdd-aaa6220cccd1:[10.8.30.101] > > > However, tpstats looks good. And the schema changes eventually do get > applied on *all* the nodes (even the ones that seem to have different > schema versions). There are no communications issues between the nodes and > they are all in the same rack > > root@xxxx:~# nodetool tpstats > Pool Name Active Pending Completed Blocked > All time blocked > ReadStage 0 0 1254592 0 > 0 > RequestResponseStage 0 0 9480827 0 > 0 > MutationStage 0 0 8662263 0 > 0 > ReadRepairStage 0 0 339158 0 > 0 > ReplicateOnWriteStage 0 0 0 0 > 0 > GossipStage 0 0 1469197 0 > 0 > AntiEntropyStage 0 0 0 0 > 0 > MigrationStage 0 0 1808 0 > 0 > MemtablePostFlusher 0 0 248 0 > 0 > StreamStage 0 0 0 0 > 0 > FlushWriter 0 0 248 0 > 4 > MiscStage 0 0 0 0 > 0 > commitlog_archiver 0 0 0 0 > 0 > InternalResponseStage 0 0 5286 0 > 0 > HintedHandoff 0 0 21 0 > 0 > > Message type Dropped > RANGE_SLICE 0 > READ_REPAIR 0 > BINARY 0 > READ 0 > MUTATION 0 > REQUEST_RESPONSE 0 > > So I'm guessing maybe the different schema versions may be potentially > stopping compactions? Will compactions still happen if there are different > versions of the schema? > > > > > > On 9/18/12 11:38 AM, "Michael Kjellman" > <mkjell...@barracuda.com<mailto:mkjell...@barracuda.com>> wrote: > >> Thanks, I just modified the schema on the worse offending column family >> (as determined by the .json) from 10MB to 200MB. >> >> Should I kick off a compaction on this cf now/repair?/scrub? >> >> Thanks >> >> -michael >> >> From: Віталій Тимчишин >> <tiv...@gmail.com<mailto:tiv...@gmail.com><mailto:tiv...@gmail.com<mailto:tiv...@gmail.com>>> >> Reply-To: >> "user@cassandra.apache.org<mailto:user@cassandra.apache.org><mailto:user@cassandra.apache.org<mailto:user@cassandra.apache.org>>" >> <user@cassandra.apache.org<mailto:user@cassandra.apache.org><mailto:user@cassandra.apache.org<mailto:user@cassandra.apache.org>>> >> To: >> "user@cassandra.apache.org<mailto:user@cassandra.apache.org><mailto:user@cassandra.apache.org<mailto:user@cassandra.apache.org>>" >> <user@cassandra.apache.org<mailto:user@cassandra.apache.org><mailto:user@cassandra.apache.org<mailto:user@cassandra.apache.org>>> >> Subject: Re: persistent compaction issue (1.1.4 and 1.1.5) >> >> I've started to use LeveledCompaction some time ago and from my >> experience this indicates some SST on lower levels than they should be. >> The compaction is going, moving them up level by level, but total count >> does not change as new data goes in. >> The numbers are pretty high as for me. Such numbers mean a lot of files >> (over 100K in single directory) and a lot of thinking for compaction >> executor to decide what to compact next. I can see numbers like 5K-10K >> and still thing this is high number. If I were you, I'd increase >> sstable_size_in_mb 10-20 times it is now. >> >> 2012/9/17 Michael Kjellman >> <mkjell...@barracuda.com<mailto:mkjell...@barracuda.com><mailto:mkjell...@barracuda.com<mailto:mkjell...@barracuda.com>>> >> Hi All, >> >> I have an issue where each one of my nodes (currently all running at >> 1.1.5) is reporting around 30,000 pending compactions. I understand that >> a pending compaction doesn't necessarily mean it is a scheduled task >> however I'm confused why this behavior is occurring. It is the same on >> all nodes, occasionally goes down 5k pending compaction tasks, and then >> returns to 25,000-35,000 compaction tasks pending. >> >> I have tried a repair operation/scrub operation on two of the nodes and >> while compactions initially happen the number of pending compactions does >> not decrease. >> >> Any ideas? Thanks for your time. >> >> Best, >> michael >> >> >> 'Like' us on Facebook for exclusive content and other resources on all >> Barracuda Networks solutions. >> >> Visit http://barracudanetworks.com/facebook >> >> >> >> >> >> >> >> -- >> Best regards, >> Vitalii Tymchyshyn >> >> 'Like' us on Facebook for exclusive content and other resources on all >> Barracuda Networks solutions. >> >> Visit http://barracudanetworks.com/facebook > > > 'Like' us on Facebook for exclusive content and other resources on all > Barracuda Networks solutions. > > Visit http://barracudanetworks.com/facebook > > > > > > > > -- > Best regards, > Vitalii Tymchyshyn > > 'Like' us on Facebook for exclusive content and other resources on all > Barracuda Networks solutions. > Visit http://barracudanetworks.com/facebook > > 'Like' us on Facebook for exclusive content and other resources on all Barracuda Networks solutions. Visit http://barracudanetworks.com/facebook