but after repair all nodes should be in sync regardless of whether new files were compacted or not.
Do you suggest major compaction after repair? I'd like to avoid it.

On 04/22/2012 11:52 AM, Philippe wrote:

Repairs generate new files that then need to be compacted.
Maybe that's where the temporary extra volume comes from?

Le 21 avr. 2012 20:43, "Igor" <i...@4friends.od.ua <mailto:i...@4friends.od.ua>> a écrit :

    Hi

    I can't understand the repair behavior in my case. I have 12 nodes
    ring (all 1.0.7):

10.254.237.2 LA ADS-LA-1 Up Normal 50.92 GB 0.00% 0 10.254.238.2 TX TX-24-RACK Up Normal 33.29 GB 0.00% 1 10.254.236.2 VA ADS-VA-1 Up Normal 50.07 GB 0.00% 2 10.254.93.2 IL R1 Up Normal 49.29 GB 0.00% 3 10.253.4.2 AZ R1 Up Normal 37.83 GB 0.00% 5 10.254.180.2 GB GB-1 Up Normal 42.86 GB 50.00% 85070591730234615865843651857942052863 10.254.191.2 LA ADS-LA-1 Up Normal 47.64 GB 0.00% 85070591730234615865843651857942052864 10.254.221.2 TX TX-24-RACK Up Normal 43.42 GB 0.00% 85070591730234615865843651857942052865 10.254.217.2 VA ADS-VA-1 Up Normal 38.44 GB 0.00% 85070591730234615865843651857942052866 10.254.94.2 IL R1 Up Normal 49.31 GB 0.00% 85070591730234615865843651857942052867 10.253.5.2 AZ R1 Up Normal 49.01 GB 0.00% 85070591730234615865843651857942052869 10.254.179.2 GB GB-1 Up Normal 27.08 GB 50.00% 170141183460469231731687303715884105727

    I have single keyspace 'meter' and two column families (one 'ids'
    is small, and second is bigger). The strange thing happened today
    when I try to run
    "nodetool -h 10.254.180.2 -pr meter ids"
    two times one after another. First repair finished successfully

     INFO 16:33:02,492 [repair #db582370-8bba-11e1-0000-5b777f708bff]
    ids is fully synced
     INFO 16:33:02,526 [repair #db582370-8bba-11e1-0000-5b777f708bff]
    session completed successfully

    after moving near 50G of data, and I started second session one
    hour later:

    INFO 17:44:37,842 [repair #aa415d00-8bd9-11e1-0000-5b777f708bff]
    new session: will sync localhost/1
    0.254.180.2, /10.254.221.2 <http://10.254.221.2>, /10.254.191.2
    <http://10.254.191.2>, /10.254.217.2 <http://10.254.217.2>,
    /10.253.5.2 <http://10.253.5.2>, /10.254.94.2 <http://10.254.94.2>
    on range (5,8507
    0591730234615865843651857942052863] for meter.[ids]

    What is strange - when streams for the second repair starts they
    have the same or even bigger total volume, and I expected that
    second run will move less data (or even no data at all).

    Is it OK? Or should I fix something?

    Thanks!


Reply via email to