Can I infer from this that if I have 3 replicas, then running repair without -pr won 1 node will repair the other 2 replicas as well.
-Raj On Sat, Apr 14, 2012 at 2:54 AM, Zhu Han <han...@nutstore.net> wrote: > > On Sat, Apr 14, 2012 at 1:57 PM, Igor <i...@4friends.od.ua> wrote: > >> Hi! >> >> What is the difference between 'repair' and '-pr repair'? Simple repair >> touch all token ranges (for all nodes) and -pr touch only range for which >> given node responsible? >> >> > -pr only touches the primary range of the node. If you executes -pr > against all nodes in replica groups, then all ranges are repaired. > >> >> >> On 04/12/2012 05:59 PM, Sylvain Lebresne wrote: >> >>> On Thu, Apr 12, 2012 at 4:06 PM, Frank Ng<buzzt...@gmail.com> wrote: >>> >>>> I also noticed that if I use the -pr option, the repair process went >>>> down >>>> from 30 hours to 9 hours. Is the -pr option safe to use if I want to >>>> run >>>> repair processes in parallel on nodes that are not replication peers? >>>> >>> There is pretty much two use case for repair: >>> 1) to rebuild a node: if say a node has lost some data due to a hard >>> drive corruption or the like and you want to to rebuild what's missing >>> 2) the periodic repairs to avoid problem with deleted data coming back >>> from the dead (basically: >>> http://wiki.apache.org/**cassandra/Operations#** >>> Frequency_of_nodetool_repair<http://wiki.apache.org/cassandra/Operations#Frequency_of_nodetool_repair> >>> ) >>> >>> In case 1) you want to run 'nodetool repair' (without -pr) against the >>> node to rebuild. >>> In case 2) (which I suspect is the case your talking now), you *want* >>> to use 'nodetool repair -pr' on *every* node of the cluster. I.e. >>> that's the most efficient way to do it. The only reason not to use -pr >>> in this case would be that it's not available because you're using an >>> old version of Cassandra. And yes, it's is safe to run with -pr in >>> parallel on nodes that are not replication peers. >>> >>> -- >>> Sylvain >>> >>> >>> thanks >>>> >>>> >>>> On Thu, Apr 12, 2012 at 12:06 AM, Frank Ng<berryt...@gmail.com> wrote: >>>> >>>>> Thank you for confirming that the per node data size is most likely >>>>> causing the long repair process. I have tried a repair on smaller >>>>> column >>>>> families and it was significantly faster. >>>>> >>>>> On Wed, Apr 11, 2012 at 9:55 PM, aaron morton<aa...@thelastpickle.com* >>>>> *> >>>>> wrote: >>>>> >>>>>> If you have 1TB of data it will take a long time to repair. Every bit >>>>>> of >>>>>> data has to be read and a hash generated. This is one of the reasons >>>>>> we >>>>>> often suggest that around 300 to 400Gb per node is a good load in the >>>>>> general case. >>>>>> >>>>>> Look at nodetool compactionstats .Is there a validation compaction >>>>>> running ? If so it is still building the merkle hash tree. >>>>>> >>>>>> Look at nodetool netstats . Is it streaming data ? If so all hash >>>>>> trees >>>>>> have been calculated. >>>>>> >>>>>> Cheers >>>>>> >>>>>> >>>>>> ----------------- >>>>>> Aaron Morton >>>>>> Freelance Developer >>>>>> @aaronmorton >>>>>> http://www.thelastpickle.com >>>>>> >>>>>> On 12/04/2012, at 2:16 AM, Frank Ng wrote: >>>>>> >>>>>> Can you expand further on your issue? Were you using Random >>>>>> Patitioner? >>>>>> >>>>>> thanks >>>>>> >>>>>> On Tue, Apr 10, 2012 at 5:35 PM, David Leimbach<leim...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> I had this happen when I had really poorly generated tokens for the >>>>>>> ring. Cassandra seems to accept numbers that are too big. You get >>>>>>> hot >>>>>>> spots when you think you should be balanced and repair never ends (I >>>>>>> think >>>>>>> there is a 48 hour timeout). >>>>>>> >>>>>>> >>>>>>> On Tuesday, April 10, 2012, Frank Ng wrote: >>>>>>> >>>>>>>> I am not using tier-sized compaction. >>>>>>>> >>>>>>>> >>>>>>>> On Tue, Apr 10, 2012 at 12:56 PM, Jonathan Rhone<rh...@tinyco.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Data size, number of nodes, RF? >>>>>>>>> >>>>>>>>> Are you using size-tiered compaction on any of the column families >>>>>>>>> that hold a lot of your data? >>>>>>>>> >>>>>>>>> Do your cassandra logs say you are streaming a lot of ranges? >>>>>>>>> zgrep -E "(Performing streaming repair|out of sync)" >>>>>>>>> >>>>>>>>> >>>>>>>>> On Tue, Apr 10, 2012 at 9:45 AM, Igor<i...@4friends.od.ua> wrote: >>>>>>>>> >>>>>>>>>> On 04/10/2012 07:16 PM, Frank Ng wrote: >>>>>>>>>> >>>>>>>>>> Short answer - yes. >>>>>>>>>> But you are asking wrong question. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> I think both processes are taking a while. When it starts up, >>>>>>>>>> netstats and compactionstats show nothing. Anyone out there >>>>>>>>>> successfully >>>>>>>>>> using ext3 and their repair processes are faster than this? >>>>>>>>>> >>>>>>>>>> On Tue, Apr 10, 2012 at 10:42 AM, Igor<i...@4friends.od.ua> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> Hi >>>>>>>>>>> >>>>>>>>>>> You can check with nodetool which part of repair process is >>>>>>>>>>> slow - >>>>>>>>>>> network streams or verify compactions. use nodetool netstats or >>>>>>>>>>> compactionstats. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On 04/10/2012 05:16 PM, Frank Ng wrote: >>>>>>>>>>> >>>>>>>>>>>> Hello, >>>>>>>>>>>> >>>>>>>>>>>> I am on Cassandra 1.0.7. My repair processes are taking over 30 >>>>>>>>>>>> hours to complete. Is it normal for the repair process to take >>>>>>>>>>>> this long? >>>>>>>>>>>> I wonder if it's because I am using the ext3 file system. >>>>>>>>>>>> >>>>>>>>>>>> thanks >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Jonathan Rhone >>>>>>>>> Software Engineer >>>>>>>>> >>>>>>>>> TinyCo >>>>>>>>> 800 Market St., Fl 6 >>>>>>>>> San Francisco, CA 94102 >>>>>>>>> www.tinyco.com >>>>>>>>> >>>>>>>>> >>>>>> >> >