For Slender Cassandra (an 18 node cluster for illustrating best practices) to 
communicate between regions I seem to have three choices for Amazon Web 
Services:
1. VPN
2. VPN Peering
3. Over the world wide web

What is the correct choice?

This article on best practices for running Cassandra on AWS just came out by 
the way: 
https://aws.amazon.com/blogs/big-data/best-practices-for-running-apache-cassandra-on-amazon-ec2/
I'm not sure all of the information in it is correct.  Is it?

Kenneth Brotman

-----Original Message-----
From: Kenneth Brotman [mailto:kenbrot...@yahoo.com] 
Sent: Wednesday, January 31, 2018 4:20 PM
To: 'user@cassandra.apache.org'
Subject: RE: Slender Cassandra Cluster Project

Thank you Yuri and Michael for the suggestion.  Yes, a Terraform version makes 
sense.  Will do.

Kenneth Brotman

-----Original Message-----
From: Yuri Subach [mailto:ysub...@gmail.com]
Sent: Wednesday, January 31, 2018 7:20 AM
To: user@cassandra.apache.org
Subject: Re: Slender Cassandra Cluster Project

Yes, I'd prefer Terraform too.

On 2018-01-31 06:32:21, Michael Mior <mm...@uwaterloo.ca> wrote:
> While whatever format this comes out in would be helpful, you might 
> want to consider Terraform. 1Password recently published a blog post 
> on their experience with Terraform vs. CloudFormation.
> 
> https://blog.agilebits.com/2018/01/25/terraforming-1password/
> 
> --
> Michael Mior
> mm...@apache.org
> 
> 2018-01-31 2:34 GMT-05:00 Kenneth Brotman <kenbrot...@yahoo.com.invalid>:
> 
> > Hi Yuri,
> >
> > If possible I will do everything with AWS Cloudformation.  I'm 
> > working on it now.  Nothing published yet.
> >
> > Kenneth Brotman
> >
> > -----Original Message-----
> > From: Yuri Subach [mailto:ysub...@gmail.com]
> > Sent: Tuesday, January 30, 2018 7:02 PM
> > To: user@cassandra.apache.org
> > Subject: RE: Slender Cassandra Cluster Project
> >
> > Hi Kenneth,
> >
> > I like this project idea!
> >
> > A couple of questions:
> > - What tools are you going to use for AWS cluster setup?
> > - Do you have anything published already (github)?
> >
> > On 2018-01-22 22:42:11, Kenneth Brotman 
> > <kenbrot...@yahoo.com.INVALID>
> > wrote:
> > > Thanks Anthony!  I’ve made a note to include that information in 
> > > the
> > documentation. You’re right.  It won’t work as intended unless that 
> > is configured properly.
> > >
> > >
> > >
> > > I’m also favoring a couple other guidelines for Slender Cassandra:
> > >
> > > 1.       SSD’s only, no spinning disks
> > >
> > > 2.       At least two cores per node
> > >
> > >
> > >
> > > For AWS, I’m favoring the c3.large on Linux.  It’s available in 
> > > these
> > regions: US-East, US-West and US-West2.  The specifications are listed as:
> > >
> > > ·         Two (2) vCPU’s
> > >
> > > ·         3.7 Gib Memory
> > >
> > > ·         Two (2) 16 GB SSD’s
> > >
> > > ·         Moderate I/O
> > >
> > >
> > >
> > > It’s going to be hard to beat the inexpensive cost of operating a
> > Slender Cluster on demand in the cloud – and it fits a lot of the 
> > use cases
> > well:
> > >
> > >
> > >
> > > ·         For under a $100 a month, in current pricing for EC2
> > instances, you can operate an eighteen (18) node Slender Cluster for 
> > five
> > (5) hours a day, ten (10) days a month.  That’s fine for 
> > demonstrations, teaching or experiments that last half a day or less.
> > >
> > > ·         For under $20, you can have that Slender Cluster up all day
> > long, up to ten (10) hours, for whatever demonstrations or 
> > experiments you want it for.
> > >
> > >
> > >
> > > As always, feedback is encouraged.
> > >
> > >
> > >
> > > Thanks,
> > >
> > >
> > >
> > > Kenneth Brotman
> > >
> > >
> > >
> > > From: Anthony Grasso [mailto:anthony.gra...@gmail.com]
> > > Sent: Sunday, January 21, 2018 3:57 PM
> > > To: user
> > > Subject: Re: Slender Cassandra Cluster Project
> > >
> > >
> > >
> > > Hi Kenneth,
> > >
> > >
> > >
> > > Fantastic idea!
> > >
> > >
> > >
> > > One thing that came to mind from my reading of the proposed setup 
> > > was
> > rack awareness of each node. Given that the proposed setup contains 
> > three DCs, I assume that each node will be made rack aware? If not, 
> > consider defining three racks for each DC and placing two nodes in 
> > each rack. This will ensure that all the nodes in a single rack 
> > contain at most one replica of the data.
> > >
> > >
> > >
> > > Regards,
> > >
> > > Anthony
> > >
> > >
> > >
> > > On 17 January 2018 at 11:24, Kenneth Brotman
> > <kenbrot...@yahoo.com.invalid> wrote:
> > >
> > > Sure.  That takes the project from awesome to 10X awesome.  I 
> > > absolutely
> > would be willing to do that.  Thanks Kurt!
> > >
> > >
> > >
> > > Regarding your comment on the keyspaces, I agree.  There should be 
> > > a few
> > simple examples one way or the other that can be duplicated and 
> > observed, and then an example to duplicate and play with that has a 
> > nice real world mix, with some keyspaces that replicate over only a 
> > subset of DC’s and some that replicate to all DC’s.
> > >
> > >
> > >
> > > Kenneth Brotman
> > >
> > >
> > >
> > > From: kurt greaves [mailto:k...@instaclustr.com]
> > > Sent: Tuesday, January 16, 2018 1:31 PM
> > > To: User
> > > Subject: Re: Slender Cassandra Cluster Project
> > >
> > >
> > >
> > > Sounds like a great idea. Probably would be valuable to add to the
> > official docs as an example set up if you're willing.
> > >
> > >
> > >
> > > Only thing I'd add is that you should have keyspaces that 
> > > replicate over
> > only a subset of DC's, plus one/some replicated to all DC's
> > >
> > >
> > >
> > > On 17 Jan. 2018 03:26, "Kenneth Brotman" 
> > > <kenbrot...@yahoo.com.invalid>
> > wrote:
> > >
> > > I’ve begun working on a reference project intended to provide 
> > > guidance
> > on configuring and operating a modest Cassandra cluster of about 18 
> > nodes suitable for the economic study, demonstration, 
> > experimentation and testing of a Cassandra cluster.
> > >
> > >
> > >
> > > The slender cluster would be designed to be as inexpensive as 
> > > possible
> > while still using real world hardware in order to lower the cost to 
> > those with limited initial resources. Sorry no Raspberry Pi’s for this 
> > project.
> > >
> > >
> > >
> > > There would be an on-premises version and a cloud version.  
> > > Guidance
> > would be provided on configuring the cluster, on demonstrating key 
> > Cassandra behaviors, on files sizes, capacity to use with the 
> > Slender Cassandra Cluster, and so on.
> > >
> > >
> > >
> > > Why about eighteen nodes? I tried to figure out what the minimum 
> > > number
> > of nodes needed for Cassandra to be Cassandra is?  Here were my
> > considerations:
> > >
> > >
> > >
> > > •             A user wouldn’t run Cassandra in just one data center; so
> > at least two datacenters.
> > >
> > > •             A user probably would want a third data center available
> > for analytics.
> > >
> > > •             There needs to be enough nodes for enough parallelism to
> > observe Cassandra’s distributed nature.
> > >
> > > •             The cluster should have enough nodes that one gets a sense
> > of the need for cluster wide management tools to do things like 
> > repairs, snapshots and cluster monitoring.
> > >
> > > •             The cluster should be able to demonstrate a RF=3 with
> > local quorum.  If replicated in all three data centers, one write 
> > would impact half the 18 nodes, 3 datacenters X 3 nodes per data 
> > center = 9 nodes of 18 nodes  If replicated in two of the data 
> > centers, one write would still impact one third of the 18 nodes, 2 
> > DC’s X 3 nodes per DC = 6 of the
> > 18 nodes.
> > >
> > >
> > >
> > > So eighteen seems like the minimum number of nodes needed.  That’s 
> > > six
> > nodes in each of three data centers.
> > >
> > >
> > >
> > > Before I get too carried away with this project, I’m looking for 
> > > some
> > feedback on whether this project would indeed be helpful to others? 
> > Also, should the project be changed in any way?
> > >
> > >
> > >
> > > It’s always a pleasure to connect with the Cassandra users’ community.
> > Thanks for all the hard work, the expertise, the civil dialog.
> > >
> > >
> > >
> > > Kenneth Brotman
> > >
> > >
> > >
> >
> > --------------------------------------------------------------------
> > - To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
> > For additional commands, e-mail: user-h...@cassandra.apache.org
> >
> >
> > --------------------------------------------------------------------
> > - To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
> > For additional commands, e-mail: user-h...@cassandra.apache.org
> >
> >

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org

Reply via email to