I performed this kernel upgrade (to 3.19.30) over the weekend on my
cluster, and my before / after benchmarks were very close to each other,
about 500MB/s each.

On Tue, Oct 6, 2015 at 3:15 PM, Nick Fisk <n...@fisk.me.uk> wrote:

> I'm wondering if you are hitting the "bug" with the readahead changes?
>
> I know the changes to limit readahead to 2MB was introduced in 3.15, but I
> don't know if it was back ported into 3.13 or not. I have a feeling this
> may
> also limit maximum request size to 2MB as well.
>
> If you look in iostat do you see different request sizes between the two
> kernels?
>
> There is a 4.2 kernel with the readahead change reverted, it might be worth
> testing it.
>
>
> http://gitbuilder.ceph.com/kernel-deb-precise-x86_64-basic/ref/ra-bring-back
> /
>
>
>
> > -----Original Message-----
> > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
> > MailingLists - EWS
> > Sent: 06 October 2015 18:12
> > To: ceph-users@lists.ceph.com
> > Subject: Re: [ceph-users] Poor Read Performance with Ubuntu 14.04 LTS
> > 3.19.0-30 Kernel
> >
> > > Hi,
> > >
> > > Very interesting!  Did you upgrade the kernel on both the OSDs and
> > > clients
> > or
> > > just some of them?  I remember there were some kernel performance
> > > regressions a little while back.  You might try running perf during
> > > your
> > tests
> > > and look for differences.  Also, iperf might be worth trying to see if
> > it's a
> > > network regression.
> > >
> > > I also have a script that compares output from sysctl which might be
> > > worth trying to see if any defaults changes.
> > >
> > > https://github.com/ceph/cbt/blob/master/tools/compare_sysctl.py
> > >
> > > basically just save systctl -a with both kernels and pass them as
> > arguments to
> > > the python script.
> > >
> > > Mark
> >
> > Mark,
> >
> > The testing was done with 3.19 on the client with 3.13 on the OSD nodes
> > using "rados bench -p bench 50 seq" with an initial "rados bench -p bench
> 50
> > write --no-cleanup". We suspected the network as well and tested with
> iperf
> > as one of our first steps and saw expected speeds (9.9Gb/s as we are
> using
> > bonded X540-T2 interfaces) on both kernels. As an added data point, we
> > have no problem with write performance to the same pool with the same
> > kernel configuration (~1GB/s). We also checked the values of
> > read_ahead_kb of the block devices but both were shown to be the default
> > of 128 (we have since changed these to 4096 in our configuration, but the
> > results were seen with the default of 128).
> >
> > We are in the process of rebuilding the entire cluster to use 3.13 and a
> > completely fresh installation of Ceph to make sure nothing else is at
> play
> > here.
> >
> > We did check a few things in iostat and collectl, but we didn't see any
> read IO
> > against the OSDs, so I am leaning towards something further up the stack.
> >
> > Just a little more background on the cluster configuration:
> >
> > Specific pool created just for benchmarking, using 512 pgs and pgps and 2
> > replicas. Using 3 OSD nodes (also handling MON duties) with 8 SATA 7.2K
> > RPM OSDs and 2 NVMe journals (4 OSD to 1 Journal ratio). 1 x Hex core
> CPUs
> > with 32GB of RAM per OSD node.
> >
> > Tom
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to