Hi, 
same environment, after a test script, the io latency(get from sudo ceph 
--admin-daemon /run/ceph/guests/ceph-client.*.asok per dump) 
increase from about 4ms to 7.3ms

qemu version: debian 2.1.2
kernel:3.10.45-openstack-amd64
system: debian 7.8
ceph: 0.94.5
VM CPU number: 4  (cpu MHz : 2599.998)
VM memory size: 16GB
9 OSD storage servers, with 4 SSD OSD on each, total 36 OSDs.

Test scripts in VM:
# cat reproduce.sh
#!/bin/bash

times=20
for((i=1;i<=$times;i++))
do
    tmpdate=`date "+%F-%T"`
    echo "=======================$tmpdate($i/$times)======================="
    tmp=$((i%2))
    if [[ $tmp -eq 0 ]];then
        echo "############### fio /root/vdb.cfg ###############"
        fio /root/vdb.cfg
    else
        echo "############### fio /root/vdc.cfg ###############"
        fio /root/vdc.cfg
    fi
done


tmpdate=`date "+%F-%T"`
echo "############### [$tmpdate] fio /root/vde.cfg ###############"
fio /root/vde.cfg


# cat vdb.cfg 
[global]
rw=randwrite
direct=1
numjobs=64
ioengine=sync
bsrange=4k-4k
runtime=180
group_reporting

[disk01]
filename=/dev/vdb


# cat vdc.cfg 
[global]
rw=randwrite
direct=1
numjobs=64
ioengine=sync
bsrange=4k-4k
runtime=180
group_reporting

[disk01]
filename=/dev/vdc

# cat vdd.cfg 
[global]
rw=randwrite
direct=1
numjobs=64
ioengine=sync
bsrange=4k-4k
runtime=180
group_reporting

[disk01]
filename=/dev/vdd

# cat vde.cfg 
[global]
rw=randwrite
direct=1
numjobs=64
ioengine=sync
bsrange=4k-4k
runtime=180
group_reporting

[disk01]
filename=/dev/vde               

After run the scripts reproduce.sh, the disks in the VM's IOPS cutdown from 12k 
to 5k, the latency increase from 4ms to 7.3ms.

run steps:
1. create a VM
2. create four volumes and attatch to the VM
3. sh reproduce.sh
4. in the runtime of  reproduce.sh, run "fio vdd.cfg" or "fio vde.cfg" to 
checkt the performance
        
After reproduce.sh finished, performance down.


Anyone has the same problem or has some ideas about this?

Thanks!
--------------
hzwulibin
2015-11-02
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to