Having correlated graphs of CPU and block device usage would be helpful.

To my cynical eye this looks like a clear regression in CPU usage, which was 
always bottlenecking pure-SSD OSDs, and now got worse.
The gains are from doing less IO on IO-saturated HDDs.

Regression of 70% in 16-32K random writes is the most troubling, that's 
coincidentaly the average IO size for a DB2, and the biggest bottleneck to its 
performance I've seen (other databases will be similiar).
It's great 

Btw readahead is not dependant on filesystem (it's a mechanism in the IO 
scheduler), so it should be present even on a block device, I think? 

Jan
 
 
> On 22 Apr 2016, at 17:35, Mark Nelson <mnel...@redhat.com> wrote:
> 
> Hi Guys,
> 
> Now that folks are starting to dig into bluestore with the Jewel release, I 
> wanted to share some of our on-going performance test data. These are from 
> 10.1.0, so almost, but not quite, Jewel.  Generally bluestore is looking very 
> good on HDDs, but there are a couple of strange things to watch out for, 
> especially with NVMe devices.  Mainly:
> 
> 1) in HDD+NVMe configurations performance increases dramatically when 
> replacing the stock CentOS7 kernel with Kernel 4.5.1.
> 
> 2) In NVMe only configurations performance is often lower at middle-sized 
> IOs.  Kernel 4.5.1 doesn't really help here.  In fact it seems to amplify 
> both the cases where bluestore is faster and where it is slower.
> 
> 3) Medium sized sequential reads are where bluestore consistently tends to be 
> slower than filestore.  It's not clear yet if this is simply due to Bluestore 
> not doing read ahead at the OSD (ie being entirely dependent on client read 
> ahead) or something else as well.
> 
> I wanted to post this so other folks have some ideas of what to look for as 
> they do their own bluestore testing.  This data is shown as percentage 
> differences vs filestore, but I can also release the raw throughput values if 
> people are interested in those as well.
> 
> https://drive.google.com/file/d/0B2gTBZrkrnpZOTVQNkV0M2tIWkk/view?usp=sharing
> 
> Thanks!
> Mark
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to