Sure ..thanks for your willingness to help Identical servers
Hardware DELL R620, 6 cores, 64GB RAM, 2 x 10 GB ports, Enterprise HDD 600GB( Seagate ST600MM0006), Enterprise grade SSD 340GB (Toshiba PX05SMB040Y) All tests done with the following command rados bench -p rbd 50 write --no-cleanup && rados bench -p rbd 50 seq ceph osd pool ls detail "pool_name": "rbd", "flags": 1, "flags_names": "hashpspool", "type": 1, "size": 2, "min_size": 1, "crush_rule": 1, "object_hash": 2, "pg_num": 64, "pg_placement_num": 64, "crash_replay_interval": 0, "last_change": "354", "last_force_op_resend": "0", "last_force_op_resend_preluminous": "0", "auid": 0, "snap_mode": "selfmanaged", "snap_seq": 0, "snap_epoch": 0, "pool_snaps": [], "removed_snaps": "[]", "quota_max_bytes": 0, "quota_max_objects": 0, "tiers": [], "tier_of": -1, "read_tier": -1, "write_tier": -1, "cache_mode": "none", "target_max_bytes": 0, "target_max_objects": 0, "cache_target_dirty_ratio_micro": 400000, "cache_target_dirty_high_ratio_micro": 600000, "cache_target_full_ratio_micro": 800000, "cache_min_flush_age": 0, "cache_min_evict_age": 0, "erasure_code_profile": "", "hit_set_params": { "type": "none" }, "hit_set_period": 0, "hit_set_count": 0, "use_gmt_hitset": true, "min_read_recency_for_promote": 0, "min_write_recency_for_promote": 0, "hit_set_grade_decay_rate": 0, "hit_set_search_last_n": 0, "grade_table": [], "stripe_width": 0, "expected_num_objects": 0, "fast_read": false, "options": {}, "application_metadata": {} } ceph osd crush rule dump [ { "rule_id": 0, "rule_name": "replicated_rule", "ruleset": 0, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -1, "item_name": "default" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] }, { "rule_id": 1, "rule_name": "rbd", "ruleset": 1, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -9, "item_name": "sas" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] } ] 2 servers, 2 OSD ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -9 4.00000 root sas -10 1.00000 host osd01-sas 2 hdd 1.00000 osd.2 up 0 1.00000 -11 1.00000 host osd02-sas 3 hdd 1.00000 osd.3 up 0 1.00000 -12 1.00000 host osd03-sas 5 hdd 1.00000 osd.5 up 1.00000 1.00000 -19 1.00000 host osd04-sas 6 hdd 1.00000 osd.6 up 1.00000 1.00000 2018-04-19 09:19:01.266010 min lat: 0.0412473 max lat: 1.03227 avg lat: 0.331163 sec Cur ops started finished avg MB/s cur MB/s last lat(s) avg lat(s) 40 16 1941 1925 192.478 192 0.315461 0.331163 41 16 1984 1968 191.978 172 0.262268 0.331529 42 16 2032 2016 191.978 192 0.326608 0.332061 43 16 2081 2065 192.071 196 0.345757 0.332389 44 16 2123 2107 191.524 168 0.307759 0.332745 45 16 2166 2150 191.09 172 0.318577 0.333613 46 16 2214 2198 191.109 192 0.329559 0.333703 47 16 2257 2241 190.702 172 0.423664 0.33427 48 16 2305 2289 190.729 192 0.357342 0.334386 49 16 2348 2332 190.346 172 0.30218 0.334735 50 16 2396 2380 190.379 192 0.318226 0.334981 Total time run: 50.281886 Total writes made: 2397 Write size: 4194304 Object size: 4194304 Bandwidth (MB/sec): 190.685 Stddev Bandwidth: 24.5781 Max bandwidth (MB/sec): 340 Min bandwidth (MB/sec): 164 Average IOPS: 47 Stddev IOPS: 6 Max IOPS: 85 Min IOPS: 41 Average Latency(s): 0.335515 Stddev Latency(s): 0.0867836 Max latency(s): 1.03227 Min latency(s): 0.0412473 2018-04-19 09:19:52.340092 min lat: 0.0209445 max lat: 14.9208 avg lat: 1.31352 sec Cur ops started finished avg MB/s cur MB/s last lat(s) avg lat(s) 40 16 296 280 27.9973 0 - 1.31352 41 16 296 280 27.3144 0 - 1.31352 42 16 296 280 26.664 0 - 1.31352 43 16 323 307 28.5553 9 0.0429661 2.20267 44 16 323 307 27.9063 0 - 2.20267 45 16 363 347 30.8414 80 0.0922424 2.05975 46 16 370 354 30.7795 28 0.0302223 2.02055 47 16 370 354 30.1246 0 - 2.02055 48 16 386 370 30.8303 32 2.72624 2.06407 49 16 386 370 30.2011 0 - 2.06407 50 16 400 384 30.7169 28 2.10543 2.07055 51 16 401 385 30.1931 4 2.53183 2.07175 52 16 401 385 29.6124 0 - 2.07175 53 16 401 385 29.0537 0 - 2.07175 54 16 401 385 28.5157 0 - 2.07175 55 16 401 385 27.9972 0 - 2.07175 56 16 401 385 27.4972 0 - 2.07175 Total time run: 56.042520 Total reads made: 401 Read size: 4194304 Object size: 4194304 Bandwidth (MB/sec): 28.6211 Average IOPS: 7 Stddev IOPS: 11 Max IOPS: 47 Min IOPS: 0 Average Latency(s): 2.23525 Max latency(s): 29.5553 Min latency(s): 0.0209445 4 servers, 4 osds ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -9 4.00000 root sas -10 1.00000 host osd01-sas 2 hdd 1.00000 osd.2 up 1.00000 1.00000 -11 1.00000 host osd02-sas 3 hdd 1.00000 osd.3 up 1.00000 1.00000 -12 1.00000 host osd03-sas 5 hdd 1.00000 osd.5 up 1.00000 1.00000 -19 1.00000 host osd04-sas 6 hdd 1.00000 osd.6 up 1.00000 1.00000 2018-04-19 09:35:43.558843 min lat: 0.0141657 max lat: 11.3013 avg lat: 1.25618 sec Cur ops started finished avg MB/s cur MB/s last lat(s) avg lat(s) 40 16 482 466 46.5956 0 - 1.25618 41 16 488 472 46.0444 12 0.0175485 1.25181 42 16 488 472 44.9481 0 - 1.25181 43 16 488 472 43.9028 0 - 1.25181 44 16 562 546 49.6316 98.6667 0.0150341 1.26385 45 16 569 553 49.1508 28 0.0151556 1.25516 46 16 569 553 48.0823 0 - 1.25516 47 16 569 553 47.0593 0 - 1.25516 48 16 569 553 46.0789 0 - 1.25516 49 16 569 553 45.1386 0 - 1.25516 50 16 569 553 44.2358 0 - 1.25516 51 16 569 553 43.3684 0 - 1.25516 Total time run: 51.724920 Total writes made: 570 Write size: 4194304 Object size: 4194304 Bandwidth (MB/sec): 44.0793 Stddev Bandwidth: 55.3843 Max bandwidth (MB/sec): 232 Min bandwidth (MB/sec): 0 Average IOPS: 11 Stddev IOPS: 13 Max IOPS: 58 Min IOPS: 0 Average Latency(s): 1.45175 Stddev Latency(s): 2.9411 Max latency(s): 11.3013 Min latency(s): 0.0141657 2018-04-19 09:36:35.633624 min lat: 0.00804825 max lat: 10.2583 avg lat: 1.03388 sec Cur ops started finished avg MB/s cur MB/s last lat(s) avg lat(s) 40 16 479 463 46.2955 0 - 1.03388 41 16 540 524 51.1169 24.4 0.00913275 1.23193 42 16 540 524 49.8999 0 - 1.23193 43 16 541 525 48.8324 2 2.31401 1.23399 44 16 541 525 47.7226 0 - 1.23399 45 16 541 525 46.6621 0 - 1.23399 46 16 541 525 45.6477 0 - 1.23399 47 16 541 525 44.6765 0 - 1.23399 48 16 541 525 43.7458 0 - 1.23399 49 16 541 525 42.853 0 - 1.23399 50 16 541 525 41.996 0 - 1.23399 51 16 541 525 41.1725 0 - 1.23399 Total time run: 51.530655 Total reads made: 542 Read size: 4194304 Object size: 4194304 Bandwidth (MB/sec): 42.072 Average IOPS: 10 Stddev IOPS: 15 Max IOPS: 62 Min IOPS: 0 Average Latency(s): 1.5204 Max latency(s): 11.4841 Min latency(s): 0.00627081 Many thanks Steven On Thu, 19 Apr 2018 at 08:42, Hans van den Bogert <hansbog...@gmail.com> wrote: > Hi Steven, > > There is only one bench. Could you show multiple benches of the different > scenarios you discussed? Also provide hardware details. > > Hans > > On Apr 19, 2018 13:11, "Steven Vacaroaia" <ste...@gmail.com> wrote: > > Hi, > > Any idea why 2 servers with one OSD each will provide better performance > than 3 ? > > Servers are identical > Performance is impacted irrespective if I used SSD for WAL/DB or not > Basically, I am getting lots of cur MB/s zero > > Network is separate 10 GB for public and private > I tested it with iperf and I am getting 9.3 Gbs > > I have tried replication by 2 and 3 with same results ( much better for 2 > servers than 3 ) > > reinstalled CEPH multiple times > ceph.conf very simple - no major customization ( see below) > I am out of ideas - any hint will be TRULY appreciated > > Steven > > > > auth_cluster_required = cephx > auth_service_required = cephx > auth_client_required = cephx > > > public_network = 10.10.30.0/24 > cluster_network = 192.168.0.0/24 > > > osd_pool_default_size = 2 > osd_pool_default_min_size = 1 # Allow writing 1 copy in a degraded state > osd_crush_chooseleaf_type = 1 > > > [mon] > mon_allow_pool_delete = true > mon_osd_min_down_reporters = 1 > > [osd] > osd_mkfs_type = xfs > osd_mount_options_xfs = > "rw,noatime,nodiratime,attr2,logbufs=8,logbsize=256k,largeio,inode64,swalloc,allocsize=4M" > osd_mkfs_options_xfs = "-f -i size=2048" > bluestore_block_db_size = 32212254720 > bluestore_block_wal_size = 1073741824 > > rados bench -p rbd 120 write --no-cleanup && rados bench -p rbd 120 seq > hints = 1 > Maintaining 16 concurrent writes of 4194304 bytes to objects of size > 4194304 for up to 120 seconds or 0 objects > Object prefix: benchmark_data_osd01_383626 > sec Cur ops started finished avg MB/s cur MB/s last lat(s) avg > lat(s) > 0 0 0 0 0 0 - > 0 > 1 16 57 41 163.991 164 0.197929 > 0.065543 > 2 16 57 41 81.992 0 - > 0.065543 > 3 16 67 51 67.9936 20 0.0164632 > 0.249939 > 4 16 67 51 50.9951 0 - > 0.249939 > 5 16 71 55 43.9958 8 0.0171439 > 0.319973 > 6 16 181 165 109.989 440 0.0159057 > 0.563746 > 7 16 182 166 94.8476 4 0.221421 > 0.561684 > 8 16 182 166 82.9917 0 - > 0.561684 > 9 16 240 224 99.5458 116 0.0232989 > 0.638292 > 10 16 264 248 99.1901 96 0.0222669 > 0.583336 > 11 16 264 248 90.1729 0 - > 0.583336 > 12 16 285 269 89.6579 42 0.0165706 > 0.600606 > 13 16 285 269 82.7611 0 - > 0.600606 > 14 16 310 294 83.9918 50 0.0254241 > 0.756351 > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com