I actually wrote my own compactor that deals with this problem. Anne
From: cem [mailto:cayiro...@gmail.com] Sent: Thursday, July 25, 2013 11:59 AM To: user@cassandra.apache.org Subject: Re: maximum storage per node You will suffer from long compactions if you are planning to get rid of from old records by TTL. Best Regards, Cem. On Thu, Jul 25, 2013 at 5:51 PM, Kanwar Sangha <kan...@mavenir.com<mailto:kan...@mavenir.com>> wrote: Issues with large data nodes would be - * Nodetool repair will be impossible to run * Your read i/o will suffer since you will almost always go to disk (each read will take 3 IOPS worst case) * Boot-straping the node in case of failures will take days/weeks From: Pruner, Anne (Anne) [mailto:pru...@avaya.com<mailto:pru...@avaya.com>] Sent: 25 July 2013 10:45 To: user@cassandra.apache.org<mailto:user@cassandra.apache.org> Subject: RE: maximum storage per node We're storing fairly large files (about 1MB apiece) for a few months and then deleting the oldest to get more space to add new ones. We have large requirements (maybe up to 100 TB), so having a 1TB limit would be unworkable. What is the reason for the limit? Does something fail after that? If there are hardware issues, what's recommended? BTW, we're using Cassandra 1.2 Anne From: cem [mailto:cayiro...@gmail.com] Sent: Thursday, July 25, 2013 11:41 AM To: user@cassandra.apache.org<mailto:user@cassandra.apache.org> Subject: Re: maximum storage per node Between 500GB - 1TB is recommended. But it depends also your hardware, traffic characteristics and requirements. Can you give some details on that? Best Regards, Cem On Thu, Jul 25, 2013 at 5:35 PM, Pruner, Anne (Anne) <pru...@avaya.com<mailto:pru...@avaya.com>> wrote: Does anyone have opinions on the maximum amount of data reasonable to store on one Cassandra node? If there are limitations, what are the reasons for it? Thanks, Anne