Hi 

I've just had a warning ( from ceph -s) that one of the osds is near full. 
Having investigated the warning, i've located that osd.6 is 86% full. The data 
distribution is nowhere near to being equal on my osds as you can see from the 
df command output below: 

/dev/sdj1 2.8T 2.4T 413G 86% /var/lib/ceph/osd/ceph-6 
/dev/sdb1 2.8T 2.1T 625G 78% /var/lib/ceph/osd/ceph-0 
/dev/sdc1 2.8T 2.0T 824G 71% /var/lib/ceph/osd/ceph-1 
/dev/sdd1 2.8T 1.5T 1.3T 55% /var/lib/ceph/osd/ceph-2 
/dev/sde1 2.8T 1.7T 1.1T 63% /var/lib/ceph/osd/ceph-3 
/dev/sdh1 2.8T 1.7T 1.1T 62% /var/lib/ceph/osd/ceph-4 
/dev/sdf1 2.8T 1.9T 932G 67% /var/lib/ceph/osd/ceph-8 
/dev/sdi1 2.8T 1.9T 880G 69% /var/lib/ceph/osd/ceph-5 
/dev/sdg1 2.8T 2.0T 798G 72% /var/lib/ceph/osd/ceph-7 

I seem to have the spread of over 30% disk utilisation between the osds, 
despite all my osds having the identical weight (ceph osd tree output): 


-2 24.56999 host arh-ibstorage1-ib 
1 2.73000 osd.1 up 1.00000 1.00000 
3 2.73000 osd.3 up 1.00000 1.00000 
5 2.73000 osd.5 up 1.00000 1.00000 
6 2.73000 osd.6 up 1.00000 1.00000 
7 2.73000 osd.7 up 1.00000 1.00000 
8 2.73000 osd.8 up 1.00000 1.00000 
4 2.73000 osd.4 up 1.00000 1.00000 
0 2.73000 osd.0 up 1.00000 1.00000 
2 2.73000 osd.2 up 1.00000 1.00000 


What would be the best way to correct the issue without having significant 
impact on the cluster IO? 

Many thanks 

Andrei 
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to