My 1.5T bound is for high throughput for read and write with hundreds of nodes 
— specifically with needs for quick bootstrap / repairs when adding / replacing 
nodes.

Lower the density the faster it is to add nodes.

--
Rahul Singh
rahul.si...@anant.us

Anant Corporation

On Mar 9, 2018, 11:30 AM -0500, Jon Haddad <j...@jonhaddad.com>, wrote:
> I agree with Jeff - I usually advise teams to cap their density around 3TB, 
> especially with TWCS.  Read heavy workloads tend to use smaller datasets and 
> ring size ends up being a function of performance tuning.
>
> Since 2.2 bootstrap can now be resumed, which helps quite a bit with the 
> streaming problem, see CASSANDRA-8838.
>
> Jon
>
>
> > On Mar 9, 2018, at 7:39 AM, Jeff Jirsa <jji...@gmail.com> wrote:
> >
> > 1.5 TB sounds very very conservative - 3-4T is where I set the limit at 
> > past jobs. Have heard of people doing twice that (6-8T).
> >
> > --
> > Jeff Jirsa
> >
> >
> > On Mar 8, 2018, at 11:09 PM, Niclas Hedhman <nic...@apache.org> wrote:
> >
> > > I am curious about the side comment; "Depending on your usecase you may 
> > > not
> > > want to have a data density over 1.5 TB per node."
> > >
> > > Why is that? I am planning much bigger than that, and now you give me
> > > pause...
> > >
> > >
> > > Cheers
> > > Niclas
> > >
> > > > On Wed, Mar 7, 2018 at 6:59 PM, Rahul Singh 
> > > > <rahul.xavier.si...@gmail.com> wrote:
> > > > > Are you putting both the commitlogs and the Sstables on the adds? 
> > > > > Consider moving your snapshots often if that’s also taking up space. 
> > > > > Maybe able to save some space before you add drives.
> > > > >
> > > > > You should be able to add these new drives and mount them without an 
> > > > > issue. Try to avoid different number of data dirs across nodes. It 
> > > > > makes automation of operational processes a little harder.
> > > > >
> > > > > As an aside, Depending on your usecase you may not want to have a 
> > > > > data density over 1.5 TB per node.
> > > > >
> > > > > --
> > > > > Rahul Singh
> > > > > rahul.si...@anant.us
> > > > >
> > > > > Anant Corporation
> > > > >
> > > > > On Mar 7, 2018, 1:26 AM -0500, Eunsu Kim <eunsu.bil...@gmail.com>, 
> > > > > wrote:
> > > > > > Hello,
> > > > > >
> > > > > > I use 5 nodes to create a cluster of Cassandra. (SSD 1TB)
> > > > > >
> > > > > > I'm trying to mount an additional disk(SSD 1TB) on each node 
> > > > > > because each disk usage growth rate is higher than I expected. Then 
> > > > > > I will add the the directory to data_file_directories in 
> > > > > > cassanra.yaml
> > > > > >
> > > > > > Can I get advice from who have experienced this situation?
> > > > > > If we go through the above steps one by one, will we be able to 
> > > > > > complete the upgrade without losing data?
> > > > > > The replication strategy is SimpleStrategy, RF 2.
> > > > > >
> > > > > > Thank you in advance
> > > > > > ---------------------------------------------------------------------
> > > > > > To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
> > > > > > For additional commands, e-mail: user-h...@cassandra.apache.org
> > > > > >
> > >
> > >
> > >
> > > --
> > > Niclas Hedhman, Software Developer
> > > http://zest.apache.org - New Energy for Java
>

Reply via email to