Hello Yatong "If I restart the node or using 'cleanup', it will resume to normal." --> what does df -hl shows for /data6 when you restart or cleanup the node ?
By the way, a single SSTable of 3.6Tb is kind of huge. Do you perform manual repair frequently ? On Sun, May 4, 2014 at 1:51 AM, Yatong Zhang <bluefl...@gmail.com> wrote: > My Cassandra cluster has plenty of free space, for now only about 30% of > space are used > > > On Sun, May 4, 2014 at 6:36 AM, Yatong Zhang <bluefl...@gmail.com> wrote: > >> Hi there, >> >> It was strange that the 'xxx-tmp-xxx.db' file kept increasing until >> Cassandra throw exceptions with 'No space left on device'. I am using CQL 3 >> to create a table to store data about 200K ~ 500K per record. I have 6 >> harddisks per node and cassandra was configured with 6 data >> directories(ext4 file systems, Centos 6.5): >> >> data_file_directories: >>> - /data1/cass >>> - /data2/cass >>> - /data3/cass >>> - /data4/cass >>> - /data5/cass >>> - /data6/cass >>> >> >> And every directory is on a standalone disk. But I just found when the >> error occurred: >> >> [root@node5 images]# ll -hl >>> total 3.6T >>> drwxr-xr-x 4 root root 4.0K Jan 20 09:44 snapshots >>> -rw-r--r-- 1 root root 456M Apr 30 13:42 >>> mydb-images-tmp-jb-91068-CompressionInfo.db >>> -rw-r--r-- 1 root root 3.5T Apr 30 13:42 mydb-images-tmp-jb-91068-Data.db >>> -rw-r--r-- 1 root root 0 Apr 30 13:42 >>> mydb-images-tmp-jb-91068-Filter.db >>> -rw-r--r-- 1 root root 2.0G Apr 30 13:42 >>> mydb-images-tmp-jb-91068-Index.db >>> >> >> [root@node5 images]# df -hl >> Filesystem Size Used Avail Use% Mounted on >> /dev/sda1 49G 7.5G 39G 17% / >> tmpfs 7.8G 0 7.8G 0% /dev/shm >> /dev/sda3 3.6T 1.3T 2.1T 38% /data1 >> /dev/sdb1 3.6T 1.4T 2.1T 39% /data2 >> /dev/sdc1 3.6T 466G 3.0T 14% /data3 >> /dev/sdd1 3.6T 1.3T 2.2T 38% /data4 >> /dev/sde1 3.6T 1.3T 2.2T 38% /data5 >> /dev/sdf1 3.6T 3.6T 0 100% /data6 >> >> *mydb-images-tmp-jb-91068-Data.db *almost occupied all the disk space >> (4T harddisk with 3.6T actual usable size) and the error looks like: >> >> INFO [FlushWriter:4174] 2014-05-04 05:15:15,744 Memtable.java (line 403) >>> Completed flushing >>> /data3/cass/system/compactions_in_progress/system-compactions_in_progress-jb-16942-Data.db >>> (42 bytes) for commitlog position ReplayPosition(segmentId=1398900356204, >>> position=25024609) >>> INFO [CompactionExecutor:3689] 2014-05-04 05:15:15,745 >>> CompactionTask.java (line 115) Compacting >>> [SSTableReader(path='/data3/cass/system/compactions_in_progress/system-compactions_in_progress-jb-16940-Data.db'), >>> SSTableReader(path='/data3/cass/system/compactions_in_progress/system-compactions_in_progress-jb-16942-Data.db'), >>> SSTableReader(path='/data3/cass/system/compactions_in_progress/system-compactions_in_progress-jb-16941-Data.db'), >>> SSTableReader(path='/data3/cass/system/compactions_in_progress/system-compactions_in_progress-jb-16939-Data.db')] >>> ERROR [CompactionExecutor:1245] 2014-05-04 05:15:15,745 >>> CassandraDaemon.java (line 198) Exception in thread >>> Thread[CompactionExecutor:1245,1,main] >>> FSWriteError in >>> /data2/cass/mydb/images/mydb-images-tmp-jb-92181-Filter.db >>> at >>> org.apache.cassandra.io.sstable.SSTableWriter$IndexWriter.close(SSTableWriter.java:475) >>> at >>> org.apache.cassandra.io.util.FileUtils.closeQuietly(FileUtils.java:212) >>> at >>> org.apache.cassandra.io.sstable.SSTableWriter.abort(SSTableWriter.java:301) >>> at >>> org.apache.cassandra.db.compaction.CompactionTask.runWith(CompactionTask.java:209) >>> at >>> org.apache.cassandra.io.util.DiskAwareRunnable.runMayThrow(DiskAwareRunnable.java:48) >>> at >>> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) >>> at >>> org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:60) >>> at >>> org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:59) >>> at >>> org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:197) >>> at >>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:744) >>> Caused by: java.io.IOException: No space left on device >>> at java.io.FileOutputStream.write(Native Method) >>> at java.io.FileOutputStream.write(FileOutputStream.java:295) >>> at java.io.DataOutputStream.writeInt(DataOutputStream.java:197) >>> at >>> org.apache.cassandra.utils.BloomFilterSerializer.serialize(BloomFilterSerializer.java:34) >>> at >>> org.apache.cassandra.utils.Murmur3BloomFilter$Murmur3BloomFilterSerializer.serialize(Murmur3BloomFilter.java:44) >>> at >>> org.apache.cassandra.utils.FilterFactory.serialize(FilterFactory.java:41) >>> at >>> org.apache.cassandra.io.sstable.SSTableWriter$IndexWriter.close(SSTableWriter.java:468) >>> ... 13 more >>> ERROR [CompactionExecutor:1245] 2014-05-04 05:15:15,800 >>> StorageService.java (line 367) Stopping gossiper >>> WARN [CompactionExecutor:1245] 2014-05-04 05:15:15,800 >>> StorageService.java (line 281) Stopping gossip by operator request >>> INFO [CompactionExecutor:1245] 2014-05-04 05:15:15,800 Gossiper.java >>> (line 1271) Announcing shutdown >>> >> >> >> I have changed my table to "LeveledCompactionStrategy" to reduce the disk >> size needed when compaction, with: >> >> ALTER TABLE images WITH compaction = { 'class' : >>> 'LeveledCompactionStrategy', 'sstable_size_in_mb' : '192' }; >>> >> >> But the problem still exists: the file keep increasing, and after about 2 >> or 3 days cassandra will fail due to 'No space left on device' error. If I >> restart the node or using 'cleanup', it will resume to normal. >> >> I don't know is it because my configuration or it's just a bug, so would >> any one please help to solve this issue? >> >> Thanks >> > >