Thank you Mykola. The issue is that I/we strongly suggested to add OSD for many times, but we are not the decision maker. For now, I just want to mount the ceph drive again, even in read only mode, so that they can read the data. Any idea on how to achieve this?
Thanks, X On Tue, Feb 2, 2016 at 9:57 AM, Mykola Dvornik <mykola.dvor...@gmail.com> wrote: > I would strongly(!) suggest you to add few more OSDs to cluster before > things get worse / corrupted. > > -Mykola > > > On Tue, Feb 2, 2016 at 6:45 PM, Zhao Xu <xuzh....@gmail.com> wrote: > > Hi All, > Recently our ceph storage is running at low performance. Today, we can > not write to the folder. We tried to unmount the ceph storage then to > re-mount it, however, we can not even mount it now: > > # mount -v -t ceph igc-head,is1,i1,i2,i3:6789:/ /mnt/igcfs/ -o > name=admin,secretfile=/etc/admin.secret > parsing options: rw,name=admin,secretfile=/etc/admin.secret > mount error 5 = Input/output error > > Previously there are some nearly full osd, so we did the "ceph osd > reweight-by-utilization" to rebalance the usage. The ceph health is not > ideal but it should still alive. Please help me to mount the disk again. > > [root@igc-head ~]# ceph -s > cluster debdcfe9-20d3-404b-921c-2210534454e1 > health HEALTH_WARN > 39 pgs degraded > 39 pgs stuck degraded > 3 pgs stuck inactive > 332 pgs stuck unclean > 39 pgs stuck undersized > 39 pgs undersized > 48 requests are blocked > 32 sec > recovery 129755/8053623 objects degraded (1.611%) > recovery 965837/8053623 objects misplaced (11.993%) > mds0: Behind on trimming (455/30) > clock skew detected on mon.i1, mon.i2, mon.i3 > monmap e1: 5 mons at {i1= > 10.1.10.11:6789/0,i2=10.1.10.12:6789/0,i3=10.1.10.13:6789/0,igc-head=10.1.10.1:6789/0,is1=10.1.10.100:6789/0 > } > election epoch 1314, quorum 0,1,2,3,4 igc-head,i1,i2,i3,is1 > mdsmap e1602: 1/1/1 up {0=igc-head=up:active} > osdmap e8007: 17 osds: 17 up, 17 in; 298 remapped pgs > pgmap v5726326: 1088 pgs, 3 pools, 7442 GB data, 2621 kobjects > 22228 GB used, 18652 GB / 40881 GB avail > 129755/8053623 objects degraded (1.611%) > 965837/8053623 objects misplaced (11.993%) > 755 active+clean > 293 active+remapped > 31 active+undersized+degraded > 5 active+undersized+degraded+remapped > 3 undersized+degraded+peered > 1 active+clean+scrubbing > > [root@igc-head ~]# ceph osd tree > ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY > -1 39.86992 root default > -2 18.14995 host is1 > 0 3.62999 osd.0 up 1.00000 1.00000 > 1 3.62999 osd.1 up 1.00000 1.00000 > 2 3.62999 osd.2 up 1.00000 1.00000 > 3 3.62999 osd.3 up 1.00000 1.00000 > 4 3.62999 osd.4 up 1.00000 1.00000 > -3 7.23999 host i1 > 5 1.81000 osd.5 up 0.44101 1.00000 > 6 1.81000 osd.6 up 0.40675 1.00000 > 7 1.81000 osd.7 up 0.60754 1.00000 > 8 1.81000 osd.8 up 0.50868 1.00000 > -4 7.23999 host i2 > 9 1.81000 osd.9 up 0.54956 1.00000 > 10 1.81000 osd.10 up 0.44815 1.00000 > 11 1.81000 osd.11 up 0.53262 1.00000 > 12 1.81000 osd.12 up 0.47197 1.00000 > -5 7.23999 host i3 > 13 1.81000 osd.13 up 0.55557 1.00000 > 14 1.81000 osd.14 up 0.65874 1.00000 > 15 1.81000 osd.15 up 0.49663 1.00000 > 16 1.81000 osd.16 up 0.50136 1.00000 > > > Thanks, > X > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com