2010/5/19 Maxim Kramarenko <maxi...@trackstudio.com>: > Hi! > > We have mail archive application, so we have a lot of data (30TB on multiple > nodes) and should delete data after a few months of storing. > > Questions are: > > 1) Compaction require extra space to process. What happend if node have no > extra space for compaction ? Will it crash, or just stop compaction process > ?
Stop compaction. > 2) Is it possible to limit max SSTable file size ? I am worry about > following situation: we have 1TB disk, 600 GB of data in single file and > should delete 50 GB of outdated data. This can lead to another 550 GB data > file generation, which cannot fit on the disk. Not currently. You have to manage this operationally. > 3) If we have 30 TB in data and replicas, how much disk space required to > handle this, including adding new data, deleting old, compaction, etc ? All of that depends on the cardinality of those operations (how much is deleted, etc.). You're going to have to benchmark it. > 4) What occurs, if we run decommission, but target node have not enough disk > space ? I don't know. Please let me know when you find out. :) -ryan