Anthony, 

I doubt the manufacturer reported 315MB/s for 4K block size. Most likely 
they've used 1M or 4M as the block size to achieve the 300MB/s+ speeds 

Andrei 
----- Original Message -----

> From: "Alexandre DERUMIER" <aderum...@odiso.com>
> To: "Anthony Levesque" <aleves...@gtcomm.net>
> Cc: "ceph-users" <ceph-users@lists.ceph.com>
> Sent: Saturday, 25 April, 2015 5:32:30 PM
> Subject: Re: [ceph-users] Possible improvements for a slow write
> speed (excluding independent SSD journals)

> I'm able to reach around 20000-25000iops with 4k block with s3500
> (with o_dsync) (so yes, around 80-100MB/S).

> I'l bench new s3610 soon to compare.

> ----- Mail original -----
> De: "Anthony Levesque" <aleves...@gtcomm.net>
> À: "Christian Balzer" <ch...@gol.com>
> Cc: "ceph-users" <ceph-users@lists.ceph.com>
> Envoyé: Vendredi 24 Avril 2015 22:00:44
> Objet: Re: [ceph-users] Possible improvements for a slow write speed
> (excluding independent SSD journals)

> Hi Christian,

> We tested some DC S3500 300GB using dd if=randfile of=/dev/sda bs=4k
> count=100000 oflag=direct,dsync

> we got 96 MB/s which is far from the 315 MB/s from the website.

> Can I ask you or anyone on the mailing list how you are testing the
> write speed for journals?

> Thanks
> ---
> Anthony Lévesque
> GloboTech Communications
> Phone: 1-514-907-0050 x 208
> Toll Free: 1-(888)-GTCOMM1 x 208
> Phone Urgency: 1-(514) 907-0047
> 1-(866)-500-1555
> Fax: 1-(514)-907-0750
> aleves...@gtcomm.net
> http://www.gtcomm.net

> On Apr 23, 2015, at 9:05 PM, Christian Balzer < ch...@gol.com >
> wrote:

> Hello,

> On Thu, 23 Apr 2015 18:40:38 -0400 Anthony Levesque wrote:

> BQ_BEGIN
> To update you on the current test in our lab:

> 1.We tested the Samsung OSD in Recovery mode and the speed was able
> to
> maxout 2x 10GbE port(transferring data at 2200+ MB/s during
> recovery).
> So for normal write operation without O_DSYNC writes Samsung drives
> seem
> ok.

> 2.We then tested a couple of different model of SSD we had in stock
> with
> the following command:

> dd if=randfile of=/dev/sda bs=4k count=100000 oflag=direct,dsync

> This was from a blog written by Sebastien Han and I think should be
> able
> to show how the drives would perform in O_DSYNC writes. For people
> interested in some result of what we tested here they are:

> Intel DC S3500 120GB = 114 MB/s
> Samsung Pro 128GB = 2.4 MB/s
> WD Black 1TB (HDD) = 409 KB/s
> Intel 330 120GB = 105 MB/s
> Intel 520 120GB = 9.4 MB/s
> Intel 335 80GB = 9.4 MB/s
> Samsung EVO 1TB = 2.5 MB/s
> Intel 320 120GB = 78 MB/s
> OCZ Revo Drive 240GB = 60.8 MB/s
> 4x Samsung EVO 1TB LSI RAID0 HW + BBU = 28.4 MB/s

> No real surprises here, but a nice summary nonetheless.

> You _really_ want to avoid consumer SSDs for journals and have a good
> idea
> on how much data you'll write per day and how long you expect your
> SSDs to
> last (the TBW/$ ratio).

> BQ_BEGIN
> Please let us know if the command we ran was not optimal to test
> O_DSYNC
> writes

> We order larger drive from Intel DC series to see if we could get
> more
> than 200 MB/s per SSD. We will keep you posted on tests if that
> interested you guys. We dint test multiple parallel test yet (to
> simulate multiple journal on one SSD).

> BQ_END
> You can totally trust the numbers on Intel's site:
> http://ark.intel.com/products/family/83425/Data-Center-SSDs

> The S3500s are by far the slowest and have the lowest endurance.
> Again, depending on your expected write level the S3610 or S3700
> models
> are going to be a better fit regarding price/performance.
> Especially when you consider that loosing a journal SSD will result
> in
> several dead OSDs.

> BQ_BEGIN
> 3.We remove the Journal from all Samsung OSD and put 2x Intel 330
> 120GB
> on all 6 Node to test. The overall speed we were getting from the
> rados
> bench went from 1000 MB/s(approx.) to 450 MB/s which might only be
> because the intel cannot do too much in term of journaling (was
> tested
> at around 100 MB/s). It will be interesting to test with bigger Intel
> DC S3500 drives(and more journals) per node to see if I can back up
> to
> 1000MB/s or even surpass it.

> We also wanted to test if the CPU could be a huge bottle neck so we
> swap
> the Dual E5-2620v2 from node #6 and replace them with Dual
> E5-2609v2(Which are much smaller in core and speed) and the 450 MB/s
> we
> got from he rados bench went even lower to 180 MB/s.

> BQ_END
> You really don't have to swap CPUs around, monitor things with atop
> or
> other tools to see where your bottlenecks are.

> BQ_BEGIN
> So Im wondering if the 1000MB/s we got when the Journal was shared on
> the OSD SSD was not limited by the CPUs (even though the samsung are
> not
> good for journals on the long run) and not just by the fact Samsung
> SSD
> are bad in O_DSYNC writes(or maybe both). It is probable that 16 SSD
> OSD per node in a full SSD cluster is too much and the major
> bottleneck
> will be from the CPU.

> BQ_END
> That's what I kept saying. ^.^

> BQ_BEGIN
> 4.Im wondering if we find good SSD for the journal and keep the
> samsung
> for normal writes and read(We can saturate 20GbE easy with read
> benchmark. We will test 40GbE soon) if the cluster will keep healthy
> since Samsung seem to get burnt from O_DSYNC writes.

> BQ_END
> They will get burned, as in have their cells worn out by any and all
> writes.

> BQ_BEGIN
> 5.In term of HBA controller, did you guys have made any test for a
> full
> SSD cluster or even just for SSD Journal.

> BQ_END
> If you have separate journals and OSDs, it often makes good sense to
> have
> them on separate controllers as well.
> It all depends on density of your setup and capabilities of the
> controllers.
> LSI HBAs in IT mode are a known and working entity.

> Christian
> --
> Christian Balzer Network/Systems Engineer
> ch...@gol.com Global OnLine Japan/Fusion Communications
> http://www.gol.com/

> BQ_END

> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to