Repairs generate new files that then need to be compacted. Maybe that's where the temporary extra volume comes from? Le 21 avr. 2012 20:43, "Igor" <i...@4friends.od.ua> a écrit :
> Hi > > I can't understand the repair behavior in my case. I have 12 nodes ring > (all 1.0.7): > > 10.254.237.2 LA ADS-LA-1 Up Normal 50.92 GB > 0.00% 0 > 10.254.238.2 TX TX-24-RACK Up Normal 33.29 GB > 0.00% 1 > 10.254.236.2 VA ADS-VA-1 Up Normal 50.07 GB > 0.00% 2 > 10.254.93.2 IL R1 Up Normal 49.29 GB > 0.00% 3 > 10.253.4.2 AZ R1 Up Normal 37.83 GB > 0.00% 5 > 10.254.180.2 GB GB-1 Up Normal 42.86 GB > 50.00% 850705917302346158658436518579**42052863 > 10.254.191.2 LA ADS-LA-1 Up Normal 47.64 GB > 0.00% 850705917302346158658436518579**42052864 > 10.254.221.2 TX TX-24-RACK Up Normal 43.42 GB > 0.00% 850705917302346158658436518579**42052865 > 10.254.217.2 VA ADS-VA-1 Up Normal 38.44 GB > 0.00% 850705917302346158658436518579**42052866 > 10.254.94.2 IL R1 Up Normal 49.31 GB > 0.00% 850705917302346158658436518579**42052867 > 10.253.5.2 AZ R1 Up Normal 49.01 GB > 0.00% 850705917302346158658436518579**42052869 > 10.254.179.2 GB GB-1 Up Normal 27.08 GB > 50.00% 170141183460469231731687303715**884105727 > > I have single keyspace 'meter' and two column families (one 'ids' is > small, and second is bigger). The strange thing happened today when I try > to run > "nodetool -h 10.254.180.2 -pr meter ids" > two times one after another. First repair finished successfully > > INFO 16:33:02,492 [repair #db582370-8bba-11e1-0000-**5b777f708bff] ids > is fully synced > INFO 16:33:02,526 [repair #db582370-8bba-11e1-0000-**5b777f708bff] > session completed successfully > > after moving near 50G of data, and I started second session one hour later: > > INFO 17:44:37,842 [repair #aa415d00-8bd9-11e1-0000-**5b777f708bff] new > session: will sync localhost/1 > 0.254.180.2, /10.254.221.2, /10.254.191.2, /10.254.217.2, /10.253.5.2, / > 10.254.94.2 on range (5,8507 > 059173023461586584365185794205**2863] for meter.[ids] > > What is strange - when streams for the second repair starts they have the > same or even bigger total volume, and I expected that second run will move > less data (or even no data at all). > > Is it OK? Or should I fix something? > > Thanks! > >