I think it's case by case, depended on chasing read performance or write performance, or both. Ours are used for application, read request 10 times larger than writing, application wants read performance and doesn't care writing, we use 4 SSD each 380 GB for each node (total 1.5T a node), read latency is 0.4 ms/op. Of course, if cassandra used for reporting, that is different.
Thanks, James On Sat, Mar 10, 2018 at 7:55 AM, Rahul Singh <rahul.xavier.si...@gmail.com> wrote: > My 1.5T bound is for high throughput for read and write with hundreds of > nodes — specifically with needs for quick bootstrap / repairs when adding / > replacing nodes. > > Lower the density the faster it is to add nodes. > > -- > Rahul Singh > rahul.si...@anant.us > > Anant Corporation > > On Mar 9, 2018, 11:30 AM -0500, Jon Haddad <j...@jonhaddad.com>, wrote: > > I agree with Jeff - I usually advise teams to cap their density around > 3TB, especially with TWCS. Read heavy workloads tend to use smaller > datasets and ring size ends up being a function of performance tuning. > > Since 2.2 bootstrap can now be resumed, which helps quite a bit with the > streaming problem, see CASSANDRA-8838. > > Jon > > > On Mar 9, 2018, at 7:39 AM, Jeff Jirsa <jji...@gmail.com> wrote: > > 1.5 TB sounds very very conservative - 3-4T is where I set the limit at > past jobs. Have heard of people doing twice that (6-8T). > > -- > Jeff Jirsa > > > On Mar 8, 2018, at 11:09 PM, Niclas Hedhman <nic...@apache.org> wrote: > > I am curious about the side comment; "Depending on your usecase you may not > want to have a data density over 1.5 TB per node." > > Why is that? I am planning much bigger than that, and now you give me > pause... > > > Cheers > Niclas > > On Wed, Mar 7, 2018 at 6:59 PM, Rahul Singh <rahul.xavier.si...@gmail.com> > wrote: > >> Are you putting both the commitlogs and the Sstables on the adds? >> Consider moving your snapshots often if that’s also taking up space. Maybe >> able to save some space before you add drives. >> >> You should be able to add these new drives and mount them without an >> issue. Try to avoid different number of data dirs across nodes. It makes >> automation of operational processes a little harder. >> >> As an aside, Depending on your usecase you may not want to have a data >> density over 1.5 TB per node. >> >> -- >> Rahul Singh >> rahul.si...@anant.us >> >> Anant Corporation >> >> On Mar 7, 2018, 1:26 AM -0500, Eunsu Kim <eunsu.bil...@gmail.com>, wrote: >> >> Hello, >> >> I use 5 nodes to create a cluster of Cassandra. (SSD 1TB) >> >> I'm trying to mount an additional disk(SSD 1TB) on each node because each >> disk usage growth rate is higher than I expected. Then I will add the the >> directory to data_file_directories in cassanra.yaml >> >> Can I get advice from who have experienced this situation? >> If we go through the above steps one by one, will we be able to complete >> the upgrade without losing data? >> The replication strategy is SimpleStrategy, RF 2. >> >> Thank you in advance >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org >> For additional commands, e-mail: user-h...@cassandra.apache.org >> >> > > > -- > Niclas Hedhman, Software Developer > http://zest.apache.org - New Energy for Java > > >