Hi experts,

I have a 6 node cluster spread across 2 DCs.

    DC          Rack        Status State   Load            Owns    Token

 113427455640312814857969558651062452225
    DC1         RAC13       Up     Normal  95.98 GB        33.33%  0
    DC2         RAC5        Up     Normal  50.79 GB        0.00%   1
    DC1         RAC18       Up     Normal  50.83 GB        33.33%
 56713727820156407428984779325531226112
    DC2         RAC7        Up     Normal  50.74 GB        0.00%
56713727820156407428984779325531226113
    DC1         RAC19       Up     Normal  61.72 GB        33.33%
 113427455640312814857969558651062452224
    DC2         RAC9        Up     Normal  50.83 GB        0.00%
113427455640312814857969558651062452225

They are all replicas of each other. All reads and writes are done at
LOCAL_QUORUM. We are on Cassandra 0.8.4. I see that our weekend nodetool
repair runs for more than 12 hours. Especially on the first one which has
96 GB data. Is this usual? We are using 500 GB SAS drives with ext4 file
system. This gets worse every week. This week its even worse, the nodetool
repair has been running for the last 15 hours just on the first node and
when I run nodetool compactionstats I constantly see this -

pending tasks: 3

and nothing else. Looks like its just stuck. There's nothing substantial in
the logs as well. I also dont understand if all these nodes are replicas of
each other why is that the first node has almost double the data. Any help
will be really appreciated.

Thanks
-Raj

Reply via email to