Sorry - should have been clear I was speaking in terms of route optimizing,
not bandwidth. No idea as to the implementation (probably instance
specific) and I doubt it actually doubles bandwidth.

Specifically: having an ENI dedicated to API traffic did smooth out some
recent load tests we did for a client. It could be that overall throughput
increases where more a function of cleaner traffic segmentation/smoother
routing. We werent being terribly scientific - was more an artifact of
testing network segmentation.

I'm just going to say that "using an ENI will make things better" (since
traffic segmentation is always good practice anyway :)  YMMV.



On Thu, Jun 19, 2014 at 3:39 PM, Russell Bradberry <rbradbe...@gmail.com>
wrote:

> does an elastic network interface really use a different physical network
> interface? or is it just to give the ability for multiple ip addresses?
>
>
>
> On June 19, 2014 at 3:56:34 PM, Nate McCall (n...@thelastpickle.com)
> wrote:
>
> If someone really wanted to try this it, I recommend adding an Elastic
> Network Interface or two for gossip and client/API traffic. This lets EBS
> and management traffic have the pre-configured network.
>
>
> On Thu, Jun 19, 2014 at 6:54 AM, Benedict Elliott Smith <
> belliottsm...@datastax.com> wrote:
>
>> I would say this is worth benchmarking before jumping to conclusions. The
>> network being a bottleneck (or latency causing) for EBS is, to my
>> knowledge, supposition, and instances can be started with direct
>> connections to EBS if this is a concern. The blog post below shows that
>> even without SSDs the EBS-optimised provisioned-IOPS instances show pretty
>> consistent latency numbers, although those latencies are higher than you
>> would typically expect from locally attached storage.
>>
>>
>> http://blog.parse.com/2012/09/17/parse-databases-upgraded-to-amazon-provisioned-iops/
>>
>> Note, I'm not endorsing the use of EBS. Cassandra is designed to scale up
>> with number of nodes, not with depth of nodes (as Ben mentions, saturating
>> a single node's data capacity is pretty easy these days. CPUs rapidly
>> become the bottleneck as you try to go deep). However the argument that EBS
>> cannot provide consistent performance seems overly pessimistic, and should
>> probably be empirically determined for your use case.
>>
>>
>> On Thu, Jun 19, 2014 at 9:50 AM, Alain RODRIGUEZ <arodr...@gmail.com>
>> wrote:
>>
>>> Ok, looks fair enough.
>>>
>>> Thanks guys. I would be great to be able to add disks when amount of
>>> data raises and add nodes when throughput increases... :)
>>>
>>>
>>> 2014-06-19 5:27 GMT+02:00 Ben Bromhead <b...@instaclustr.com>:
>>>
>>>
>>>> http://www.datastax.com/documentation/cassandra/1.2/cassandra/architecture/architecturePlanningEC2_c.html
>>>>
>>>> From the link:
>>>>
>>>> EBS volumes are not recommended for Cassandra data volumes for the
>>>> following reasons:
>>>>
>>>> • EBS volumes contend directly for network throughput with standard
>>>> packets. This means that EBS throughput is likely to fail if you saturate a
>>>> network link.
>>>> • EBS volumes have unreliable performance. I/O performance can be
>>>> exceptionally slow, causing the system to back load reads and writes until
>>>> the entire cluster becomes unresponsive.
>>>> • Adding capacity by increasing the number of EBS volumes per host does
>>>> not scale. You can easily surpass the ability of the system to keep
>>>> effective buffer caches and concurrently serve requests for all of the data
>>>> it is responsible for managing.
>>>>
>>>> Still applies, especially the network contention and latency issues.
>>>>
>>>>   Ben Bromhead
>>>>  Instaclustr | www.instaclustr.com | @instaclustr
>>>> <http://twitter.com/instaclustr> | +61 415 936 359
>>>>
>>>>  On 18 Jun 2014, at 7:18 pm, Daniel Chia <danc...@coursera.org> wrote:
>>>>
>>>>  While they guarantee IOPS, they don't really make any guarantees
>>>> about latency. Since EBS goes over the network, there's so many things in
>>>> the path of getting at your data, I would be concerned with random latency
>>>> spikes, unless proven otherwise.
>>>>
>>>> Thanks,
>>>> Daniel
>>>>
>>>>
>>>> On Wed, Jun 18, 2014 at 1:58 AM, Alain RODRIGUEZ <arodr...@gmail.com>
>>>> wrote:
>>>>
>>>>> In this document it is said :
>>>>>
>>>>>
>>>>>    - Provisioned IOPS (SSD) - Volumes of this type are ideal for the
>>>>>    most demanding I/O intensive, transactional workloads and large 
>>>>> relational
>>>>>    or NoSQL databases. This volume type provides the most consistent
>>>>>    performance and allows you to provision the exact level of performance 
>>>>> you
>>>>>    need with the most predictable and consistent performance. With this 
>>>>> type
>>>>>    of volume you provision exactly what you need, and pay for what you
>>>>>    provision. Once again, you can achieve up to 48,000 IOPS by connecting
>>>>>    multiple volumes together using RAID.
>>>>>
>>>>>
>>>>>
>>>>> 2014-06-18 10:57 GMT+02:00 Alain RODRIGUEZ <arodr...@gmail.com>:
>>>>>
>>>>>  Hi,
>>>>>>
>>>>>> I just saw this :
>>>>>> http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/
>>>>>>
>>>>>> Since the problem with EBS was the network, there is no chance that
>>>>>> this hardware architecture might be useful alongside Cassandra, right ?
>>>>>>
>>>>>> Alain
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>
>
>
> --
> -----------------
> Nate McCall
> Austin, TX
> @zznate
>
> Co-Founder & Sr. Technical Consultant
> Apache Cassandra Consulting
> http://www.thelastpickle.com
>
>


-- 
-----------------
Nate McCall
Austin, TX
@zznate

Co-Founder & Sr. Technical Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

Reply via email to