We occasionally (twice now on a 40 node cluster over the last 6-8 months) see this. My best guess is that Cassandra can fail to mark an SSTable for cleanup somehow. Forced GC's or reboots don't clear them out. We disable thrift and gossip; drain; snapshot; shutdown; clear data/Keyspace/Table/*.db and restore (hard-linking back into place to avoid data transfer) from the just created snapshot; restart.
On Mar 28, 2013, at 10:12 AM, Ben Chobot <be...@instructure.com> wrote: > Some of my cassandra nodes in my 1.1.5 cluster show a large discrepancy > between what cassandra says the SSTables should sum up to, and what df and du > claim exist. During repairs, this is almost always pretty bad, but > post-repair compactions tend to bring those numbers to within a few percent > of each other... usually. Sometimes they remain much further apart after > compactions have finished - for instance, I'm looking at one node now that > claims to have 205GB of SSTables, but actually has 450GB of files living in > that CF's data directory. No pending compactions, and the most recent > compaction for this CF finished just a few hours ago. > > nodetool cleanup has no effect. > > What could be causing these extra bytes, and how to get them to go away? I'm > ok with a few extra GB of unexplained data, but an extra 245GB (more than all > the data this node is supposed to have!) is a little extreme.