Hello, I added 1 disk to the cluster and after rebalancing, it shows 1 PG is in remapped state. How can I correct it ?
(I had to restart some osds during the rebalancing as there were some slow requests) $ ceph pg dump | grep remapped dumped all 3.4 981 0 0 0 0 2655009792 1535 1535 active+clean+remapped 2017-12-15 22:07:21.663964 2824'785115 2824:2297888 [0,2,NONE,4,12,10,5,1] 0 [0,2,1,4,12,10,5,1] 0 2288'767367 2017-12-14 11:00:15.576741 417'518549 2017-12-08 03:56:14.006982 That PG belongs to an erasure pool with k=5, m =3 profile, failure domain is host. =========== $ ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 16.94565 root default -3 2.73788 host ceph-a1 0 ssd 1.86469 osd.0 up 1.00000 1.00000 14 ssd 0.87320 osd.14 up 1.00000 1.00000 -5 2.73788 host ceph-a2 1 ssd 1.86469 osd.1 up 1.00000 1.00000 15 ssd 0.87320 osd.15 up 1.00000 1.00000 -7 1.86469 host ceph-a3 2 ssd 1.86469 osd.2 up 1.00000 1.00000 -9 1.74640 host ceph-a4 3 ssd 0.87320 osd.3 up 1.00000 1.00000 4 ssd 0.87320 osd.4 up 1.00000 1.00000 -11 1.74640 host ceph-a5 5 ssd 0.87320 osd.5 up 1.00000 1.00000 6 ssd 0.87320 osd.6 up 1.00000 1.00000 -13 1.74640 host ceph-a6 7 ssd 0.87320 osd.7 up 1.00000 1.00000 8 ssd 0.87320 osd.8 up 1.00000 1.00000 -15 1.74640 host ceph-a7 9 ssd 0.87320 osd.9 up 1.00000 1.00000 10 ssd 0.87320 osd.10 up 1.00000 1.00000 -17 2.61960 host ceph-a8 11 ssd 0.87320 osd.11 up 1.00000 1.00000 12 ssd 0.87320 osd.12 up 1.00000 1.00000 13 ssd 0.87320 osd.13 up 1.00000 1.00000 Karun
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
