> -----Original Message-----
> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of 
> William Josefsson
> Sent: 20 October 2016 10:25
> To: Nick Fisk <n...@fisk.me.uk>
> Cc: ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] RBD with SSD journals and SAS OSDs
> 
> On Mon, Oct 17, 2016 at 6:16 PM, Nick Fisk <n...@fisk.me.uk> wrote:
> > Did you also set /check the c-states, this can have a large impact as well?
> 
> Hi Nick. I did try intel_idle.max_cstate=0, and I've got quite a significant 
> improvement as attached below. Thanks for this
advice!
> This is still with DIRECT=1, SYNC=1, BS=4k, RW=WRITE.

Excellent, glad it worked for you. It surprising what the power saving features 
can do to bursty performance.

> 
> I wanted also to ask you about Numa. Some argue it should be disabled for 
> high performance. My hosts are Dual Socket, 2x2630v4
> 2.2Ghz. Do you have any suggestions around whether enable or disable numa and 
> what would be the Impact? Thx will

I don't have much experience around this area other than that I know that numa 
can impact performance, hence the reason all my
recent OSD nodes have been single socket. I took the easy option :-)

There are two things you need to be aware of I think

1. Storage and network controllers could be connected via different sockets 
causing data to be dragged over the interconnect bus.
There isn't much you can do about this, apart from carefully placement of pci-e 
cards, but 1 socket will always suffer.

2. OSD processes flipping between sockets. I think this has been discussed here 
in the past. I believe some gains could be achieved
by pinning the OSD process to certain cores, but I'm afraid your best bet would 
be to search the archives as I can't really offer
much advice.


> 
> 
> 
> simple-write-62: (groupid=14, jobs=62): err= 0: pid=2133: Thu Oct 20
> 12:47:58 2016
>   write: io=1213.8MB, bw=41421KB/s, iops=10355, runt= 30006msec
>     clat (msec): min=2, max=81, avg= 5.99, stdev= 2.72
>      lat (msec): min=2, max=81, avg= 5.99, stdev= 2.72
>     clat percentiles (usec):
>      |  1.00th=[ 2864],  5.00th=[ 3184], 10.00th=[ 3376], 20.00th=[ 3696],
>      | 30.00th=[ 3984], 40.00th=[ 4576], 50.00th=[ 6048], 60.00th=[ 6688],
>      | 70.00th=[ 7264], 80.00th=[ 7712], 90.00th=[ 8640], 95.00th=[ 9920],
>      | 99.00th=[12480], 99.50th=[13248], 99.90th=[38656], 99.95th=[41728],
>      | 99.99th=[81408]
>     bw (KB  /s): min=  343, max= 1051, per=1.62%, avg=669.64, stdev=160.04
>     lat (msec) : 4=30.01%, 10=65.31%, 20=4.53%, 50=0.12%, 100=0.02%
>   cpu          : usr=0.04%, sys=0.54%, ctx=636287, majf=0, minf=1905
>   IO depths    : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
>      submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, 
> >=64=0.0%
>      complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, 
> >=64=0.0%
>      issued    : total=r=0/w=310721/d=0, short=r=0/w=0/d=0
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to