On Tue, Oct 29, 2013 at 4:09 PM, Piavlo <lolitus...@gmail.com> wrote:

> I can't understand how statistically it's possible that with random
> partitioner some nodes have x2 more keys than others with 50-100 millions
> of keys per node.
>

1) you have very high row variance and really really bad luck with the
distribution of your super-huge rows
OR
2) you have repaired some nodes and not others, and have dramatically
over-repaired the repaired ones, possibly due to f/e TTL + repair issues in
1.0.x

The simple way to see if #2 is the case is to run a major compaction on a
good and a bad node. They should end up the same size.

There are operational implications to doing a major compaction, however, so
you might want to investigate the use of checksstablegarbage [1].

=Rob
 [1] https://github.com/cloudian/support-tools

Reply via email to