Correct. On Fri, Oct 21, 2011 at 6:47 AM, Jeremiah Jordan <jeremiah.jor...@morningstar.com> wrote: > I could be totally wrong here, but If you are doing a QUORUM read and there > is a bad value encountered from the QUORUM won't a repair happen? I thought > read_repair_chance 0 just means it won't query extra nodes to check for bad > values. > > -Jeremiah > > On Oct 17, 2011, at 4:22 PM, Jeremy Hanna wrote: > >> Even after disabling hinted handoff and setting read_repair_chance to 0 on >> all our column families, we were still experiencing massive writes. >> Apparently the read_repair_chance is completely ignored at any CL higher >> than CL.ONE. So we were doing CL.QUORUM on reads and writes and seeing >> massive writes still. It was because of the background read repairs being >> done. We did extensive logging and checking and that's all it could be as >> no mutations were coming in via thrift to those column families. >> >> In any case, just wanted to give some follow-up here as it's been an >> inexplicable rock in our backpack and hopefully clears up where that setting >> is actually used. I'll update the storage configuration wiki to include >> that caveat as well. >> >> On Sep 10, 2011, at 5:14 PM, Jeremy Hanna wrote: >> >>> Thanks for the insights. I may first try disabling hinted handoff for one >>> run of our data pipeline and see if it exhibits the same behavior. Will >>> post back if I see anything enlightening there. >>> >>> On Sep 10, 2011, at 5:04 PM, Chris Goffinet wrote: >>> >>>> You could tail the commit log with `strings` to see what keys are being >>>> inserted. >>>> >>>> On Sat, Sep 10, 2011 at 2:24 PM, Jonathan Ellis <jbel...@gmail.com> wrote: >>>> Two possibilities: >>>> >>>> 1) Hinted handoff (this will show up in the logs on the sending >>>> machine, on the receiving one it will just look like any other write) >>>> >>>> 2) You have something doing writes that you're not aware of, I guess >>>> you could track that down using wireshark to see where the write >>>> messages are coming from >>>> >>>> On Sat, Sep 10, 2011 at 3:56 PM, Jeremy Hanna >>>> <jeremy.hanna1...@gmail.com> wrote: >>>>> Oh and we're running 0.8.4 and the RF is 3. >>>>> >>>>> On Sep 10, 2011, at 3:49 PM, Jeremy Hanna wrote: >>>>> >>>>>> In addition, the mutation stage and the read stage are backed up like: >>>>>> >>>>>> Pool Name Active Pending Blocked >>>>>> ReadStage 32 773 0 >>>>>> RequestResponseStage 0 0 0 >>>>>> ReadRepairStage 0 0 0 >>>>>> MutationStage 158 525918 0 >>>>>> ReplicateOnWriteStage 0 0 0 >>>>>> GossipStage 0 0 0 >>>>>> AntiEntropyStage 0 0 0 >>>>>> MigrationStage 0 0 0 >>>>>> StreamStage 0 0 0 >>>>>> MemtablePostFlusher 1 5 0 >>>>>> FILEUTILS-DELETE-POOL 0 0 0 >>>>>> FlushWriter 2 5 0 >>>>>> MiscStage 0 0 0 >>>>>> FlushSorter 0 0 0 >>>>>> InternalResponseStage 0 0 0 >>>>>> HintedHandoff 0 0 0 >>>>>> CompactionManager n/a 29 >>>>>> MessagingService n/a 0,34 >>>>>> >>>>>> On Sep 10, 2011, at 3:38 PM, Jeremy Hanna wrote: >>>>>> >>>>>>> We are experiencing massive writes to column families when only doing >>>>>>> reads from Cassandra. A set of 5 hadoop jobs are reading from >>>>>>> Cassandra and then writing out to hdfs. That is the only thing >>>>>>> operating on the cluster. We are reading at CL.QUORUM with hadoop and >>>>>>> have written with CL.QUORUM. Read repair chance is set to 0.0 on all >>>>>>> column families. However, in the logs, I'm seeing flush after flush of >>>>>>> memtables and compactions taking place. Is there something else that >>>>>>> would be writing based on the above description? >>>>>>> >>>>>>> Jeremy >>>>>> >>>>> >>>>> >>>> >>>> >>>> >>>> -- >>>> Jonathan Ellis >>>> Project Chair, Apache Cassandra >>>> co-founder of DataStax, the source for professional Cassandra support >>>> http://www.datastax.com >>>> >>> >> > >
-- Jonathan Ellis Project Chair, Apache Cassandra co-founder of DataStax, the source for professional Cassandra support http://www.datastax.com