> On Jan 13, 2017, at 1:34 PM, Wido den Hollander <w...@42on.com> wrote:
> 
>> 
>> Op 13 januari 2017 om 18:50 schreef Mohammed Naser <mna...@vexxhost.com>:
>> 
>> 
>> 
>>> On Jan 13, 2017, at 12:41 PM, Wido den Hollander <w...@42on.com> wrote:
>>> 
>>> 
>>>> Op 13 januari 2017 om 18:39 schreef Mohammed Naser <mna...@vexxhost.com>:
>>>> 
>>>> 
>>>> 
>>>>> On Jan 13, 2017, at 12:37 PM, Wido den Hollander <w...@42on.com> wrote:
>>>>> 
>>>>> 
>>>>>> Op 13 januari 2017 om 18:18 schreef Mohammed Naser <mna...@vexxhost.com>:
>>>>>> 
>>>>>> 
>>>>>> Hi everyone,
>>>>>> 
>>>>>> We have a deployment with 90 OSDs at the moment which is all SSD that’s 
>>>>>> not hitting quite the performance that it should be in my opinion, a 
>>>>>> `rados bench` run gives something along these numbers:
>>>>>> 
>>>>>> Maintaining 16 concurrent writes of 4194304 bytes to objects of size 
>>>>>> 4194304 for up to 10 seconds or 0 objects
>>>>>> Object prefix: benchmark_data_bench.vexxhost._30340
>>>>>> sec Cur ops   started  finished  avg MB/s  cur MB/s last lat(s)  avg 
>>>>>> lat(s)
>>>>>>  0       0         0         0         0         0           -           >>>>>> 0
>>>>>>  1      16       158       142   568.513       568   0.0965336   
>>>>>> 0.0939971
>>>>>>  2      16       287       271   542.191       516   0.0291494    
>>>>>> 0.107503
>>>>>>  3      16       375       359    478.75       352   0.0892724    
>>>>>> 0.118463
>>>>>>  4      16       477       461   461.042       408   0.0243493    
>>>>>> 0.126649
>>>>>>  5      16       540       524   419.216       252    0.239123    
>>>>>> 0.132195
>>>>>>  6      16       644       628    418.67       416    0.347606    
>>>>>> 0.146832
>>>>>>  7      16       734       718   410.281       360   0.0534447    
>>>>>> 0.147413
>>>>>>  8      16       811       795   397.487       308   0.0311927     
>>>>>> 0.15004
>>>>>>  9      16       879       863   383.537       272   0.0894534    
>>>>>> 0.158513
>>>>>> 10      16       980       964   385.578       404   0.0969865    
>>>>>> 0.162121
>>>>>> 11       3       981       978   355.613        56    0.798949    
>>>>>> 0.171779
>>>>>> Total time run:         11.063482
>>>>>> Total writes made:      981
>>>>>> Write size:             4194304
>>>>>> Object size:            4194304
>>>>>> Bandwidth (MB/sec):     354.68
>>>>>> Stddev Bandwidth:       137.608
>>>>>> Max bandwidth (MB/sec): 568
>>>>>> Min bandwidth (MB/sec): 56
>>>>>> Average IOPS:           88
>>>>>> Stddev IOPS:            34
>>>>>> Max IOPS:               142
>>>>>> Min IOPS:               14
>>>>>> Average Latency(s):     0.175273
>>>>>> Stddev Latency(s):      0.294736
>>>>>> Max latency(s):         1.97781
>>>>>> Min latency(s):         0.0205769
>>>>>> Cleaning up (deleting benchmark objects)
>>>>>> Clean up completed and total clean up time :3.895293
>>>>>> 
>>>>>> We’ve verified the network by running `iperf` across both replication 
>>>>>> and public networks and it resulted in 9.8Gb/s (10G links for both).  
>>>>>> The machine that’s running the benchmark doesn’t even saturate it’s 
>>>>>> port.  The SSDs are S3520 960GB drives which we’ve benchmarked and they 
>>>>>> can handle the load using fio/etc.  At this point, not really sure where 
>>>>>> to look next.. anyone running all SSD clusters that might be able to 
>>>>>> share their experience?
>>>>> 
>>>>> I suggest that you search a bit on the ceph-users list since this topic 
>>>>> has been discussed multiple times in the past and even recently.
>>>>> 
>>>>> Ceph isn't your average storage system and you have to keep that in mind. 
>>>>> Nothing is free in this world. Ceph provides excellent consistency and 
>>>>> distribution of data, but that also means that you have things like 
>>>>> network and CPU latency.
>>>>> 
>>>>> However, I suggest you look up a few threads on this list which have 
>>>>> valuable tips.
>>>>> 
>>>>> Wido
>>>> 
>>>> Thanks for the reply, I’ve actually done quite a lot of research and went 
>>>> through many of the previous posts. While I agree a 100% with your 
>>>> statement, I’ve found that other people with similar setups have been able 
>>>> to reach numbers that I cannot, which leads me to believe that there is 
>>>> actually an issue in here.  They have been able to max out at 1200 MB/s 
>>>> which is the maximum of their benchmarking host.  We’d like to reach that 
>>>> and I think that given the specifications of the cluster, it can do it 
>>>> with no problems.
>>> 
>>> A few tips:
>>> 
>>> - Disable all logging in Ceph (debug_osd, debug_ms, debug_auth, etc, etc)
>> 
>> All logging is configured to default settings, should those be turned down?
> 
> Yes, disable all logging improves performance.

I’ll look into disabling it.

> 
>> 
>>> - Disable power saving on the CPUs
>> 
>> All disabled as well, everything running on `performance` mode.
>> 
>>> 
>>> Can you also share how the 90 OSDs are distributed in the cluster and what 
>>> CPUs you have?
>> 
>> There are 45 machines with 2 OSDs each.   The servers they’re located on on 
>> average have 24 core ~3GHz Intel CPUs.  Both OSDs are pinned to two cores on 
>> the system.
>> 
> 
> So 45 machines in total with 2 OSDs/SSDs each.
> 
> What is the network? 10GbE? What is the latency for a 8k packet? (ping -s 
> 8192)

It is a 10GbE network, the latency is on average 0.217ms

> 
> Also try running rados bench with more threads, 16 isn't that much. Try 
> running with 128 or so from multiple clients.

With 128 threads, I’m able to get an average of 900.  Every drive seems to 
average out to ~20MBps on that peak.  Looking at running it multiple times 
seems to introduce very odd issues with extra data… is multiple rados bench 
runs not supported?

> 
> Wido
> 
>>> 
>>> Wido
>>> 
>>>> 
>>>>>> 
>>>>>> Thanks,
>>>>>> Mohammed
>>>>>> _______________________________________________
>>>>>> ceph-users mailing list
>>>>>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
>>>>>> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to