I dunno, I think I just go into my Lotus and mull this over ;) (I wish)

This is a storage for a KVM, and we have quite a few boxes.  While right
now none are suffering from IO load, I am seeing slowdown personally and
know that sooner or later others will notice as well.

I think what I will do is remove the SSD from the cluster, and put journals
on it.

On Tue, Nov 24, 2015 at 11:42 AM, Nick Fisk <n...@fisk.me.uk> wrote:

> Separate would be best, but as with many things in life we are not all
> driving around in sports cars!!
>
>
>
> Moving the journals to the SSD’s that are also OSD’s themselves will be
> fine. SSD’s tend to be more bandwidth limited than IOPs and the reverse is
> true for Disks, so you will get maybe 2x improvement for the disk pool and
> you probably won’t even notice the impact on the SSD pool.
>
>
>
> Can I just ask what your workload will be? There maybe other things that
> can be done.
>
>
>
> *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf
> Of *Marek Dohojda
> *Sent:* 24 November 2015 18:32
> *To:* Alan Johnson <al...@supermicro.com>
> *Cc:* ceph-users@lists.ceph.com; Nick Fisk <n...@fisk.me.uk>
>
> *Subject:* Re: [ceph-users] Performance question
>
>
>
> Thank you! I will do that.  Would you suggest getting another SSD drive or
> move the journal to the SSD OSD?
>
>
>
> (Sorry for a stupid question, if that is such).
>
>
>
> On Tue, Nov 24, 2015 at 11:25 AM, Alan Johnson <al...@supermicro.com>
> wrote:
>
> Or separate the journals as this will bring the workload down on the
> spinners to 3Xrather than 6X
>
>
>
> *From:* Marek Dohojda [mailto:mdoho...@altitudedigital.com]
> *Sent:* Tuesday, November 24, 2015 1:24 PM
> *To:* Nick Fisk
> *Cc:* Alan Johnson; ceph-users@lists.ceph.com
>
>
> *Subject:* Re: [ceph-users] Performance question
>
>
>
> Crad I think you are 100% correct:
>
>
>
> rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz
> await r_await w_await  svctm  %util
>
>
>
>  0.00   369.00   33.00 1405.00   132.00 135656.00   188.86     5.61
>  4.02   21.94    3.60   0.70 100.00
>
>
>
> I was kinda wondering that this maybe the case, which is why I was
> wondering if I should be doing too much in terms of troubleshooting.
>
>
>
> So basically what you are saying I need to wait for new version?
>
>
>
>
>
> Thank you very much everybody!
>
>
>
>
>
> On Tue, Nov 24, 2015 at 9:35 AM, Nick Fisk <n...@fisk.me.uk> wrote:
>
> You haven’t stated what size replication you are running. Keep in mind
> that with a replication factor of 3, you will be writing 6x the amount of
> data down to disks than what the benchmark says (3x replication x2 for
> data+journal write).
>
>
>
> You might actually be near the hardware maximums. What does iostat looks
> like whilst you are running rados bench, are the disks getting maxed out?
>
>
>
> *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf
> Of *Marek Dohojda
> *Sent:* 24 November 2015 16:27
> *To:* Alan Johnson <al...@supermicro.com>
>
>
> *Cc:* ceph-users@lists.ceph.com
> *Subject:* Re: [ceph-users] Performance question
>
>
>
> 7 total servers, 20 GIG pipe between servers, both reads and writes.  The
> network itself has plenty of pipe left, it is averaging 40Mbits/s
>
>
>
> Rados Bench SAS 30 writes
>
>  Total time run:         30.591927
>
> Total writes made:      386
>
> Write size:             4194304
>
> Bandwidth (MB/sec):     50.471
>
>
>
> Stddev Bandwidth:       48.1052
>
> Max bandwidth (MB/sec): 160
>
> Min bandwidth (MB/sec): 0
>
> Average Latency:        1.25908
>
> Stddev Latency:         2.62018
>
> Max latency:            21.2809
>
> Min latency:            0.029227
>
>
>
> Rados Bench SSD writes
>
>  Total time run:         20.425192
>
> Total writes made:      1405
>
> Write size:             4194304
>
> Bandwidth (MB/sec):     275.150
>
>
>
> Stddev Bandwidth:       122.565
>
> Max bandwidth (MB/sec): 576
>
> Min bandwidth (MB/sec): 0
>
> Average Latency:        0.231803
>
> Stddev Latency:         0.190978
>
> Max latency:            0.981022
>
> Min latency:            0.0265421
>
>
>
>
>
> As you can see SSD is better but not as much as I would expect SSD to be.
>
>
>
>
>
>
>
> On Tue, Nov 24, 2015 at 9:10 AM, Alan Johnson <al...@supermicro.com>
> wrote:
>
> Hard to know without more config details such as no of servers, network  –
> GigE or !0 GigE, also not sure how you are measuring, (reads or writes) you
> could try RADOS bench as a baseline, I would expect more performance with 7
> X 10K spinners journaled to SSDs. The fact that SSDs did not perform much
> better may mean to a bottleneck elsewhere – network perhaps?
>
> *From:* Marek Dohojda [mailto:mdoho...@altitudedigital.com]
> *Sent:* Tuesday, November 24, 2015 10:37 AM
> *To:* Alan Johnson
> *Cc:* Haomai Wang; ceph-users@lists.ceph.com
>
>
> *Subject:* Re: [ceph-users] Performance question
>
>
>
> Yeah they are, that is one thing I was planning on changing, What I am
> really interested at the moment, is vague expected performance.  I mean is
> 100MB around normal, very low, or "could be better"?
>
>
>
> On Tue, Nov 24, 2015 at 8:02 AM, Alan Johnson <al...@supermicro.com>
> wrote:
>
> Are the journals on the same device – it might be better to use the SSDs
> for journaling since you are not getting better performance with SSDs?
>
>
>
> *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf
> Of *Marek Dohojda
> *Sent:* Monday, November 23, 2015 10:24 PM
> *To:* Haomai Wang
> *Cc:* ceph-users@lists.ceph.com
> *Subject:* Re: [ceph-users] Performance question
>
>
>
>  Sorry I should have specified SAS is the 100 MB :) , but to be honest SSD
> isn't much faster.
>
>
>
> On Mon, Nov 23, 2015 at 7:38 PM, Haomai Wang <haomaiw...@gmail.com> wrote:
>
> On Tue, Nov 24, 2015 at 10:35 AM, Marek Dohojda
> <mdoho...@altitudedigital.com> wrote:
> > No SSD and SAS are in two separate pools.
> >
> > On Mon, Nov 23, 2015 at 7:30 PM, Haomai Wang <haomaiw...@gmail.com>
> wrote:
> >>
> >> On Tue, Nov 24, 2015 at 10:23 AM, Marek Dohojda
> >> <mdoho...@altitudedigital.com> wrote:
> >> > I have a Hammer Ceph cluster on 7 nodes with total 14 OSDs.  7 of
> which
> >> > are
> >> > SSD and 7 of which are SAS 10K drives.  I get typically about 100MB IO
> >> > rates
> >> > on this cluster.
>
> So which pool you get with 100 MB?
>
>
> >>
> >> You mixed up sas and ssd in one pool?
> >>
> >> >
> >> > I have a simple question.  Is 100MB within my configuration what I
> >> > should
> >> > expect, or should it be higher? I am not sure if I should be looking
> for
> >> > issues, or just accept what I have.
> >> >
> >> > _______________________________________________
> >> > ceph-users mailing list
> >> > ceph-users@lists.ceph.com
>
> >> >
> http://xo4t.mj.am/link/xo4t/rslwlms/1/BMAuqvTZa9PuDgefDPxnDw/aHR0cDovL3hvNHQubWouYW0vbGluay94bzR0L3JzeGppdDEvMS9ObEVxaHVhMnJPSHhtWGRpT0NMX3dBL2FIUjBjRG92TDJ4cGMzUnpMbU5sY0dndVkyOXRMMnhwYzNScGJtWnZMbU5uYVM5alpYQm9MWFZ6WlhKekxXTmxjR2d1WTI5dA
> >> >
> >>
> >>
> >>
> >> --
> >> Best Regards,
> >>
> >> Wheat
> >
> >
>
> --
> Best Regards,
>
> Wheat
>
>
>
>
>
>
>
>
> [image: Image removed by sender.]
>
>
>
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to