No, we're not using columns with TTL, and I performed a major compaction before the repair, so there shouldn't be vast amounts of tombstones moving around.
And the increase happened during the repair, the nodes gained ~20-30GB each. /Henrik On Thu, Nov 8, 2012 at 12:40 PM, horschi <hors...@gmail.com> wrote: > Hi, > > is it possible that your repair is overrepairing due to any of the issues > discussed here: > http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/repair-compaction-and-tombstone-rows-td7583481.html? > > > I've seen repair increasing the load on my cluster, but what you are > describing sounds like a lot to me. > > Does this increase happen due to repair entirely? Or was the load maybe > increasing gradually over the week and you just checked for the first time? > > cheers, > Christian > > > > On Thu, Nov 8, 2012 at 11:55 AM, Henrik Schröder <skro...@gmail.com>wrote: > >> Hi, >> >> We recently ran a major compaction across our cluster, which reduced the >> storage used by about 50%. This is fine, since we do a lot of updates to >> existing data, so that's the expected result. >> >> The day after, we ran a full repair -pr across the cluster, and when that >> finished, each storage node was at about the same size as before the major >> compaction. Why does that happen? What gets transferred to other nodes, and >> why does it suddenly take up a lot of space again? >> >> We haven't run repair -pr regularly, so is this just something that >> happens on the first weekly run, and can we expect a different result next >> week? Or does repair always cause the data to grow on each node? To me it >> just doesn't seem proportional? >> >> >> /Henrik >> > >