I am having an issue on multiple machines where it's simply filling up the disk space during what I can only assume is a compaction. For example, the average node cluster-wide is around 900GB according to DSE OpsCenter--however, after coming in after the three day weekend, I noticed that there were 13 hosts that were down in the cluster.
When I investigated, I see several huge sstables that appeared to be in the middle of compaction when the host ran out of usable disk space: 81G auditing-audit-jb-17623-Data.db > 189G auditing-audit-jb-19863-Data.db > 182G auditing-audit-jb-25298-Data.db > 196G auditing-audit-jb-30791-Data.db > 13G auditing-audit-jb-31003-Data.db > 12G auditing-audit-jb-31341-Data.db > 12G auditing-audit-jb-31678-Data.db > 12G auditing-audit-jb-32019-Data.db > 766M auditing-audit-jb-32039-Data.db > 791M auditing-audit-jb-32060-Data.db > 199M auditing-audit-jb-32065-Data.db > 52M auditing-audit-jb-32066-Data.db > 175G auditing-audit-jb-8179-Data.db > > *643G auditing-audit-tmp-jb-31207-Data.db32G > auditing-audit-tmp-jb-32030-Data.db* > >From what I can tell, it is reaching the point where it is trying to compact the ~190G ones into a combined bigger one, and is failing because of the lack of disk space. How do I work around this? Would adjusting the maximum sstables before a compaction is performed help this situation? I am currently using the default values provided by SizeTieredCompactionStrategy in C* 2.0.6. Or is there a better option for a continuous-write operation (with TTLs for dropping off old data)? Thank you for your help! Andrew