> I recently ran into a problem during a repair operation where my nodes
> completely ran out of space and my whole cluster was... well, clusterfucked.
>
> I want to make sure how to prevent this problem in the future.

Depending on which version you're on, you may be seeing this:

   https://issues.apache.org/jira/browse/CASSANDRA-1674

But regardless, disk space variations is a fact of life with
cassandra. Off the top of my head I'm not ready to say what the
expectations are with respect to repair under all circumstances.
Anyone?

> Should I make sure that at all times every node is under 50% of its disk
> space? Are there any normal day-to-day operations that would cause the any
> one node to double in size that I should be aware of? If on or more nodes to
> surpass the 50% mark, what should I plan to do?

Major compactions can potentially double the amount of disk, if you
have a single large column family that contributes almost all disk
space. For such clusters, regular background compaction can indeed
cause a doubling when the compaction happened to be a major one (i.e.,
happened to include all sstables).

-- 
/ Peter Schuller

Reply via email to