Thanks for the reply Aaron. By compaction being on, do you mean if run
nodetool compact, then the answer is no. I haven't set any explicit
compaction_thresholds which means it should be using the default, min 4 and
max 32. Having said that to solve the problem, I just did a full cluster
restart and ran nodetool repair again. The entire cluster of 6 nodes was
repaired in 10 hours. I am also contemplating since all the 6 nodes are
replicas of each other, do I even need to run repair on all the nodes.
Wouldn't running it on the first node suffice since it will repair all the
ranges its responsible for(which is everything). So unless I upgrade to
1.0.+, where I can use the -pr option is it advisable to just run repair on
the first node.

-Raj

On Tue, May 22, 2012 at 5:05 AM, aaron morton <aa...@thelastpickle.com>wrote:

> I also dont understand if all these nodes are replicas of each other why
> is that the first node has almost double the data.
>
> Have you performed any token moves ? Old data is not deleted unless you
> run nodetool cleanup.
> Another possibility is things like a lot of hints. Admittedly it would
> have to be a *lot* of hints.
> The third is that compaction has fallen behind.
>
> This week its even worse, the nodetool repair has been running for the
> last 15 hours just on the first node and when I run nodetool
> compactionstats I constantly see this -
>
> pending tasks: 3
>
> First check the logs for errors.
>
> Repair will first calculate the differences, you can see this as a
> validation compaction in nodetool compactionstats.
> Then it will stream the data, you can watch that with nodetool netstats.
>
> Try to work out which part is taking the most time.   15 hours for 50Gb
> sounds like a long time (btw do you have compaction on ?)
>
> Cheers
>
> -----------------
> Aaron Morton
> Freelance Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 20/05/2012, at 3:14 AM, Raj N wrote:
>
> Hi experts,
>
> I have a 6 node cluster spread across 2 DCs.
>
>     DC          Rack        Status State   Load            Owns    Token
>
>  113427455640312814857969558651062452225
>     DC1         RAC13       Up     Normal  95.98 GB        33.33%  0
>     DC2         RAC5        Up     Normal  50.79 GB        0.00%   1
>     DC1         RAC18       Up     Normal  50.83 GB        33.33%
>  56713727820156407428984779325531226112
>     DC2         RAC7        Up     Normal  50.74 GB        0.00%
> 56713727820156407428984779325531226113
>     DC1         RAC19       Up     Normal  61.72 GB        33.33%
>  113427455640312814857969558651062452224
>     DC2         RAC9        Up     Normal  50.83 GB        0.00%
> 113427455640312814857969558651062452225
>
> They are all replicas of each other. All reads and writes are done at
> LOCAL_QUORUM. We are on Cassandra 0.8.4. I see that our weekend nodetool
> repair runs for more than 12 hours. Especially on the first one which has
> 96 GB data. Is this usual? We are using 500 GB SAS drives with ext4 file
> system. This gets worse every week. This week its even worse, the nodetool
> repair has been running for the last 15 hours just on the first node and
> when I run nodetool compactionstats I constantly see this -
>
> pending tasks: 3
>
> and nothing else. Looks like its just stuck. There's nothing substantial
> in the logs as well. I also dont understand if all these nodes are replicas
> of each other why is that the first node has almost double the data. Any
> help will be really appreciated.
>
> Thanks
> -Raj
>
>
>

Reply via email to