Hi,

I seem to have a bad edge effect in my setup, don't know if it's a RBD
problem or a Xen problem.

So, I have one Ceph cluster, in which I setup 2 different storage
pools : one on SSD and one on SAS. With appropriate CRUSH rules, those
pools are complety separated, only MON are commons.

Then, on a Xen host A, I run "VMSSD" and "VMSAS". If I launch a big
reballance on the "SSD pool", then the "VMSSD" *and* "VMSAS" will slow
down (a lot of IOWait). But if I move the "VMSAS" on a different Xen
host (B), then "VMSSD" will still be slow, but the "VMSAS" will be fast
again.

The first thing I checked is the network of the Xen host A, but I didn't
find any problem.

So, is there a queue, shared by all RBD kernel clients running on a same
host ? Or something which can explain this edge effect ?


Olivier

PS : one precision, I have about 60 RBD mapped on the "Xen host A",
don't know if it can be the key of the problem.

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to