HI  
1.
2 sw  iscsi gateways(deploy on osd/monitor ) using lrbd to create,the iscsi 
target is LIO 
configuration:
{
  "auth": [
    {
      "target": "iqn.2016-07.org.linux-iscsi.iscsi.x86:testvol", 
      "authentication": "none"
    }
  ], 
  "targets": [
    {
      "target": "iqn.2016-07.org.linux-iscsi.iscsi.x86:testvol", 
      "hosts": [
        {
          "host": "node2", 
          "portal": "east"
        }, 
        {
          "host": "node3", 
          "portal": "west"
        }
      ]
    }
  ], 
  "portals": [
    {
      "name": "east", 
      "addresses": [
        "10.0.52.92"
      ]
    }, 
    {
      "name": "west", 
      "addresses": [
        "10.0.52.93"
      ]
    }
  ], 
  "pools": [
    {
      "pool": "rbd", 
      "gateways": [
        {
          "target": "iqn.2016-07.org.linux-iscsi.iscsi.x86:testvol", 
          "tpg": [
            {
              "image": "testvol"
            }
          ]
        }
      ]
    }
  ]
}

2 the ceph cluster itself’s performance is ok. i create a rbd on one of ceph 
node. fio results is nice: 4K randwrite IOPS=3013 bw=100MB/s.
so i think the ceph cluster have no bottleneck.

3  Intel S3510 SSD 480G enterprise not consumer

new test :clone a VM in wmware can reach 100MB/s. but fio and dd test in vm 
still poor.


> 在 2016年7月1日,下午4:18,Christian Balzer <ch...@gol.com> 写道:
> 
> 
> Hello,
> 
> On Fri, 1 Jul 2016 13:04:45 +0800 mq wrote:
> 
>> Hi list
>> I have tested suse enterprise storage3 using 2 iscsi  gateway attached
>> to  vmware. The performance is bad.  
> 
> First off, it's somewhat funny that you're testing the repackaged SUSE
> Ceph, but asking for help here (with Ceph being owned by Red Hat).
> 
> Aside from that, you're not telling us what these 2 iSCSI gateways are
> (SW, HW specs/configuration).
> 
> Having iSCSI on top of Ceph is by the very nature of things going to be
> slower than native Ceph.
> 
> Use "rbd bench" or a VM client with RBD to get a base number of what your
> Ceph cluster is capable of, this will help identifying where the slowdown
> is.
> 
>> I have turn off  VAAI following the
>> (https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665
>>  
>> <https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665>)
>> <https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665
>>  
>> <https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665>)>.
>> My cluster 3 ceph nodes :2*E5-2620 64G , mem 2*1Gbps (3*10K SAS, 1*480G
>> SSD) per node, SSD as journal 1 vmware node  2*E5-2620 64G , mem 2*1Gbps 
> 
> That's a slow (latency wise) network, but not your problem.
> What SSD model? 
> A 480GB size suggests a consumer model and that would explain a lot.
> 
> Check you storage nodes with atop during the fio runs and see if you can
> spot a bottleneck.
> 
> Christian
> 
>> # ceph -s
>>    cluster 0199f68d-a745-4da3-9670-15f2981e7a15
>>     health HEALTH_OK
>>     monmap e1: 3 mons at
>> {node1=192.168.50.91:6789/0,node2=192.168.50.92:6789/0,node3=192.168.50.93:6789/0}
>> election epoch 22, quorum 0,1,2 node1,node2,node3 osdmap e200: 9 osds: 9
>> up, 9 in flags sortbitwise
>>      pgmap v1162: 448 pgs, 1 pools, 14337 MB data, 4935 objects
>>            18339 MB used, 5005 GB / 5023 GB avail
>>                 448 active+clean
>>  client io 87438 kB/s wr, 0 op/s rd, 213 op/s wr
>> 
>> sudo ceph osd tree
>> ID WEIGHT  TYPE NAME      UP/DOWN REWEIGHT PRIMARY-AFFINITY
>> -1 4.90581 root default
>> -2 1.63527     host node1
>> 0 0.54509         osd.0       up  1.00000          1.00000
>> 1 0.54509         osd.1       up  1.00000          1.00000
>> 2 0.54509         osd.2       up  1.00000          1.00000
>> -3 1.63527     host node2
>> 3 0.54509         osd.3       up  1.00000          1.00000
>> 4 0.54509         osd.4       up  1.00000          1.00000
>> 5 0.54509         osd.5       up  1.00000          1.00000
>> -4 1.63527     host node3
>> 6 0.54509         osd.6       up  1.00000          1.00000
>> 7 0.54509         osd.7       up  1.00000          1.00000
>> 8 0.54509         osd.8       up  1.00000          1.00000
>> 
>> 
>> 
>> An linux vm in vmmare, running fio.  4k randwrite result just 64 IOPS
>> lantency is high,dd test just 11MB/s. 
>> fio -ioengine=libaio -bs=4k -direct=1 -thread -rw=randwrite -size=100G
>> -filename=/dev/sdb  -name="EBS 4KB randwrite test" -iodepth=32
>> -runtime=60 EBS 4KB randwrite test: (g=0): rw=randwrite,
>> bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=32 fio-2.0.13 Starting 1
>> thread Jobs: 1 (f=1): [w] [100.0% done] [0K/131K/0K /s] [0 /32 /0  iops]
>> [eta 00m:00s] EBS 4KB randwrite test: (groupid=0, jobs=1): err= 0:
>> pid=6766: Wed Jun 29 21:28:06 2016 write: io=15696KB, bw=264627 B/s,
>> iops=64 , runt= 60737msec slat (usec): min=10 , max=213 , avg=35.54,
>> stdev=16.41 clat (msec): min=1 , max=31368 , avg=495.01, stdev=1862.52
>>     lat (msec): min=2 , max=31368 , avg=495.04, stdev=1862.52
>>    clat percentiles (msec):
>>     |  1.00th=[    7],  5.00th=[    8], 10.00th=[    8],
>> 20.00th=[    9], | 30.00th=[    9], 40.00th=[   10], 50.00th=[  198],
>> 60.00th=[  204], | 70.00th=[  208], 80.00th=[  217], 90.00th=[  799],
>> 95.00th=[ 1795], | 99.00th=[ 7177], 99.50th=[12649], 99.90th=[16712],
>> 99.95th=[16712], | 99.99th=[16712]
>>    bw (KB/s)  : min=   36, max=11960, per=100.00%, avg=264.77,
>> stdev=1110.81 lat (msec) : 2=0.03%, 4=0.23%, 10=40.93%, 20=0.48%,
>> 50=0.03% lat (msec) : 100=0.08%, 250=39.55%, 500=5.63%, 750=2.91%,
>> 1000=1.35% lat (msec) : 2000=4.03%, >=2000=4.77%
>>  cpu          : usr=0.02%, sys=0.22%, ctx=2973, majf=0,
>> minf=18446744073709538907 IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%,
>> 16=0.4%, 32=99.2%, >=64=0.0% submit    : 0=0.0%, 4=100.0%, 8=0.0%,
>> 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% complete  : 0=0.0%, 4=100.0%,
>> 8=0.0%, 16=0.0%, 32=0.1%, 64=0.0%, >=64=0.0% issued    :
>> total=r=0/w=3924/d=0, short=r=0/w=0/d=0 
>> Run status group 0 (all jobs):
>>  WRITE: io=15696KB, aggrb=258KB/s, minb=258KB/s, maxb=258KB/s,
>> mint=60737msec, maxt=60737msec 
>> Disk stats (read/write):
>>  sdb: ios=83/3921, merge=0/0, ticks=60/1903085, in_queue=1931694,
>> util=100.00%
>> 
>> anyone can give me some suggestion to improve the performance ?
>> 
>> Regards
>> 
>> MQ
>> 
>> 
> 
> 
> -- 
> Christian Balzer        Network/Systems Engineer                
> ch...@gol.com <mailto:ch...@gol.com>          Global OnLine Japan/Rakuten 
> Communications
> http://www.gol.com/ <http://www.gol.com/>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to