Of course, I definitely hope to see it merged into 5.0.x as soon as possible

Jordan West <jw...@apache.org> 于2025年2月13日周四 10:48写道:

> Regarding the buffer size, it is configurable. My personal take is that
> we’ve tested this on a variety of hardware (from laptops to large instance
> sizes) already, as well as a few different disk configs (it’s also been run
> internally, in test, at a few places) and that it has been reviewed by four
> committers and another contributor. Always love to see more numbers. if
> folks want to take it for a spin on Alibaba cloud, azure, etc and determine
> the best buffer size that’s awesome. We could document which is suggested
> for the community. I don’t think it’s necessary to block on that however.
>
> Also I am of course +1 to including this in 5.0.
>
> Jordan
>
> On Wed, Feb 12, 2025 at 19:50 guo Maxwell <cclive1...@gmail.com> wrote:
>
>> What I understand is that there will be some differences in block storage
>> among various cloud platforms. More intuitively, the default read-ahead
>> size will be the same. For example, AWS ebs seems to be 256K, and Alibaba
>> Cloud seems to be 512K(If I remember correctly).
>>
>> Just like 19488, give the test method, see who can assist in the test ,
>> and provide the results.
>>
>> Jon Haddad <j...@rustyrazorblade.com> 于2025年2月13日周四 08:30写道:
>>
>>> Can you elaborate why?  This would be several hundred hours of work and
>>> would cost me thousands of $$ to perform.
>>>
>>> Filesystems and block devices are well understood.  Could you give me an
>>> example of what you think might be different here?  This is already one of
>>> the most well tested and documented performance patches ever contributed to
>>> the project.
>>>
>>> On Wed, Feb 12, 2025 at 4:26 PM guo Maxwell <cclive1...@gmail.com>
>>> wrote:
>>>
>>>>  I think it should be tested on most cloud platforms(at least
>>>> aws、azure、gcp) before merged into 5.0 . Just like  CASSANDRA-19488.
>>>>
>>>> Paulo Motta <pa...@apache.org>于2025年2月13日 周四上午6:10写道:
>>>>
>>>>> I'm looking forward to these improvements, compaction needs tlc. :-)
>>>>> A couple of questions:
>>>>>
>>>>> Has this been tested only on EBS, or also EC2/bare-metal/Azure/etc? My
>>>>> only concern is if this is an optimization for EBS that can be a
>>>>> deoptimization for other environments.
>>>>>
>>>>> Are there reproducible scripts that anyone can run to verify the
>>>>> improvements in their own environments ? This could help alleviate any
>>>>> concerns and gain confidence to introduce a perf. improvement in a
>>>>> patch release.
>>>>>
>>>>> I have not read the ticket in detail, so apologies if this was already
>>>>> discussed there or elsewhere.
>>>>>
>>>>> On Wed, Feb 12, 2025 at 3:01 PM Jon Haddad <j...@rustyrazorblade.com>
>>>>> wrote:
>>>>> >
>>>>> > Hey folks,
>>>>> >
>>>>> > Over the last 9 months Jordan and I have worked on CASSANDRA-15452
>>>>> [1].  The TL;DR is that we're internalizing a read ahead buffer to allow 
>>>>> us
>>>>> to do fewer requests to disk during compaction and range reads.  This
>>>>> results in far fewer system calls (roughly 16x reduction) and on systems
>>>>> with higher read latency, a significant improvement in compaction
>>>>> throughput.  We've tested several different EBS configurations and found 
>>>>> it
>>>>> delivers up to a 10x improvement when read ahead is optimized to minimize
>>>>> read latency.  I worked with AWS and the EBS team directly on this and the
>>>>> Best Practices for C* on EBS [2] I wrote for them.  I've performance 
>>>>> tested
>>>>> this patch extensively with hundreds of billions of operations across
>>>>> several clusters and thousands of compactions.  It has less of an impact 
>>>>> on
>>>>> local NVMe, since the p99 latency is already 10-30x less than what you see
>>>>> on EBS (100micros vs 1-3ms), and you can do hundreds of thousands of IOPS
>>>>> vs a max of 16K.
>>>>> >
>>>>> > Related to this, Branimir wrote CASSANDRA-20092 [3], which
>>>>> significantly improves compaction by avoiding reading the partition index.
>>>>> CASSANDRA-20092 has been merged to trunk already [4].
>>>>> >
>>>>> > I think we should merge both of these patches into 5.0, as the perf
>>>>> improvement should allow teams to increase density of EBS backed C*
>>>>> clusters by 2-5x, driving cost way down.  There's a lot of teams running 
>>>>> C*
>>>>> on EBS now.  I'm currently working with one that's bottlenecked on maxed
>>>>> out EBS GP3 storage.  I propose we merge both, because without
>>>>> CASSANDRA-20092, we won't get the performance improvements in
>>>>> CASSANDRA-15452 with BTI, only BIG format.  I've tested BTI in other
>>>>> situations and found it to be far more performant than BIG.
>>>>> >
>>>>> > If we were looking at a small win, I wouldn't care much, but since
>>>>> these patches, combined with UCS, allows more teams to run C* on EBS at >
>>>>> 10TB / node, I think it's worth doing now.
>>>>> >
>>>>> > Thanks in advance,
>>>>> > Jon
>>>>> >
>>>>> > [1] https://issues.apache.org/jira/browse/CASSANDRA-15452
>>>>> > [2]
>>>>> https://aws.amazon.com/blogs/database/best-practices-for-running-apache-cassandra-with-amazon-ebs/
>>>>> > [3] https://issues.apache.org/jira/browse/CASSANDRA-20092
>>>>> > [4]
>>>>> https://github.com/apache/cassandra/commit/3078aea1cfc70092a185bab8ac5dc8a35627330f
>>>>> >
>>>>>
>>>>

Reply via email to