Hi Cary, No, I didnt try to repair it. I am comparatively new in ceph. Is it okay to try to repair it ? Or should I take any precautions while doing it ?
Karun Josy On Sat, Dec 16, 2017 at 2:08 PM, Cary <dynamic.c...@gmail.com> wrote: > Karun, > > Did you attempt a "ceph pg repair <pgid>"? Replace <pgid> with the pg > ID that needs repaired, 3.4. > > Cary > -D123 > > On Sat, Dec 16, 2017 at 8:24 AM, Karun Josy <karunjo...@gmail.com> wrote: > > Hello, > > > > I added 1 disk to the cluster and after rebalancing, it shows 1 PG is in > > remapped state. How can I correct it ? > > > > (I had to restart some osds during the rebalancing as there were some > slow > > requests) > > > > $ ceph pg dump | grep remapped > > dumped all > > 3.4 981 0 0 0 0 2655009792 > > 1535 1535 active+clean+remapped 2017-12-15 22:07:21.663964 > 2824'785115 > > 2824:2297888 [0,2,NONE,4,12,10,5,1] 0 [0,2,1,4,12,10,5,1] > > 0 2288'767367 2017-12-14 11:00:15.576741 417'518549 2017-12-08 > > 03:56:14.006982 > > > > That PG belongs to an erasure pool with k=5, m =3 profile, failure > domain is > > host. > > > > =========== > > > > $ ceph osd tree > > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > > -1 16.94565 root default > > -3 2.73788 host ceph-a1 > > 0 ssd 1.86469 osd.0 up 1.00000 1.00000 > > 14 ssd 0.87320 osd.14 up 1.00000 1.00000 > > -5 2.73788 host ceph-a2 > > 1 ssd 1.86469 osd.1 up 1.00000 1.00000 > > 15 ssd 0.87320 osd.15 up 1.00000 1.00000 > > -7 1.86469 host ceph-a3 > > 2 ssd 1.86469 osd.2 up 1.00000 1.00000 > > -9 1.74640 host ceph-a4 > > 3 ssd 0.87320 osd.3 up 1.00000 1.00000 > > 4 ssd 0.87320 osd.4 up 1.00000 1.00000 > > -11 1.74640 host ceph-a5 > > 5 ssd 0.87320 osd.5 up 1.00000 1.00000 > > 6 ssd 0.87320 osd.6 up 1.00000 1.00000 > > -13 1.74640 host ceph-a6 > > 7 ssd 0.87320 osd.7 up 1.00000 1.00000 > > 8 ssd 0.87320 osd.8 up 1.00000 1.00000 > > -15 1.74640 host ceph-a7 > > 9 ssd 0.87320 osd.9 up 1.00000 1.00000 > > 10 ssd 0.87320 osd.10 up 1.00000 1.00000 > > -17 2.61960 host ceph-a8 > > 11 ssd 0.87320 osd.11 up 1.00000 1.00000 > > 12 ssd 0.87320 osd.12 up 1.00000 1.00000 > > 13 ssd 0.87320 osd.13 up 1.00000 1.00000 > > > > > > > > Karun > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com