Hi X Have you tried to inspect the mds for problematic sessions still connected from those clients?
To check which sessions are still connected to the mds, do (in ceph 9.2.0, the command might be different or even do not exist in other older versions) ceph daemon mds.<name> session ls Cheers G. ________________________________ From: ceph-users [ceph-users-boun...@lists.ceph.com] on behalf of Zhao Xu [xuzh....@gmail.com] Sent: 03 February 2016 08:10 To: Mykola Dvornik Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] Urgent help needed for ceph storage "mount error 5 = Input/output error" I am not lucky on the ceph-fuse [root@igc-head ~]# ceph-fuse -d -m igc-head,is1,i1,i2,i3:6789 /mnt/igcfs/ 2016-02-03 04:55:08.756420 7fe3f7437780 0 ceph version 0.94.5 (9764da52395923e0b32908d83a9f7304401fee43), process ceph-fuse, pid 5822 ceph-fuse[5822]: starting ceph client 2016-02-03 04:55:08.794920 7fe3f7437780 -1 init, newargv = 0x3292e20 newargc=11 ceph-fuse[5822]: ceph mount failed with (110) Connection timed out 2016-02-03 05:00:08.829498 7fe3e77fe700 0 client.274164 ms_handle_reset on 10.1.10.1:6800/2641<http://10.1.10.1:6800/2641> [root@igc-head ~]# ceph-fuse -d -m igc-head:6789 /mnt/igcfs/ 2016-02-03 05:00:47.029698 7f1ec270a780 0 ceph version 0.94.5 (9764da52395923e0b32908d83a9f7304401fee43), process ceph-fuse, pid 5939 ceph-fuse[5939]: starting ceph client 2016-02-03 05:00:47.067364 7f1ec270a780 -1 init, newargv = 0x2dd9330 newargc=11 ceph-fuse[5939]: ceph mount failed with (110) Connection timed out 2016-02-03 05:05:47.100815 7f1eb67fc700 0 client.274125 ms_handle_reset on 10.1.10.1:6800/2641<http://10.1.10.1:6800/2641> Any log file I should provide here? Thanks, X On Tue, Feb 2, 2016 at 11:47 AM, Mykola Dvornik <mykola.dvor...@gmail.com<mailto:mykola.dvor...@gmail.com>> wrote: Try to mount with ceph-fuse. It worked for me when I've faced the same sort of issues you are now dealing with. -Mykola On Tue, Feb 2, 2016 at 8:42 PM, Zhao Xu <xuzh....@gmail.com<mailto:xuzh....@gmail.com>> wrote: Thank you Mykola. The issue is that I/we strongly suggested to add OSD for many times, but we are not the decision maker. For now, I just want to mount the ceph drive again, even in read only mode, so that they can read the data. Any idea on how to achieve this? Thanks, X On Tue, Feb 2, 2016 at 9:57 AM, Mykola Dvornik <mykola.dvor...@gmail.com<mailto:mykola.dvor...@gmail.com>> wrote: I would strongly(!) suggest you to add few more OSDs to cluster before things get worse / corrupted. -Mykola On Tue, Feb 2, 2016 at 6:45 PM, Zhao Xu <xuzh....@gmail.com<mailto:xuzh....@gmail.com>> wrote: Hi All, Recently our ceph storage is running at low performance. Today, we can not write to the folder. We tried to unmount the ceph storage then to re-mount it, however, we can not even mount it now: # mount -v -t ceph igc-head,is1,i1,i2,i3:6789:/ /mnt/igcfs/ -o name=admin,secretfile=/etc/admin.secret parsing options: rw,name=admin,secretfile=/etc/admin.secret mount error 5 = Input/output error Previously there are some nearly full osd, so we did the "ceph osd reweight-by-utilization" to rebalance the usage. The ceph health is not ideal but it should still alive. Please help me to mount the disk again. [root@igc-head ~]# ceph -s cluster debdcfe9-20d3-404b-921c-2210534454e1 health HEALTH_WARN 39 pgs degraded 39 pgs stuck degraded 3 pgs stuck inactive 332 pgs stuck unclean 39 pgs stuck undersized 39 pgs undersized 48 requests are blocked > 32 sec recovery 129755/8053623 objects degraded (1.611%) recovery 965837/8053623 objects misplaced (11.993%) mds0: Behind on trimming (455/30) clock skew detected on mon.i1, mon.i2, mon.i3 monmap e1: 5 mons at {i1=10.1.10.11:6789/0,i2=10.1.10.12:6789/0,i3=10.1.10.13:6789/0,igc-head=10.1.10.1:6789/0,is1=10.1.10.100:6789/0<http://10.1.10.11:6789/0,i2=10.1.10.12:6789/0,i3=10.1.10.13:6789/0,igc-head=10.1.10.1:6789/0,is1=10.1.10.100:6789/0>} election epoch 1314, quorum 0,1,2,3,4 igc-head,i1,i2,i3,is1 mdsmap e1602: 1/1/1 up {0=igc-head=up:active} osdmap e8007: 17 osds: 17 up, 17 in; 298 remapped pgs pgmap v5726326: 1088 pgs, 3 pools, 7442 GB data, 2621 kobjects 22228 GB used, 18652 GB / 40881 GB avail 129755/8053623 objects degraded (1.611%) 965837/8053623 objects misplaced (11.993%) 755 active+clean 293 active+remapped 31 active+undersized+degraded 5 active+undersized+degraded+remapped 3 undersized+degraded+peered 1 active+clean+scrubbing [root@igc-head ~]# ceph osd tree ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY -1 39.86992 root default -2 18.14995 host is1 0 3.62999 osd.0 up 1.00000 1.00000 1 3.62999 osd.1 up 1.00000 1.00000 2 3.62999 osd.2 up 1.00000 1.00000 3 3.62999 osd.3 up 1.00000 1.00000 4 3.62999 osd.4 up 1.00000 1.00000 -3 7.23999 host i1 5 1.81000 osd.5 up 0.44101 1.00000 6 1.81000 osd.6 up 0.40675 1.00000 7 1.81000 osd.7 up 0.60754 1.00000 8 1.81000 osd.8 up 0.50868 1.00000 -4 7.23999 host i2 9 1.81000 osd.9 up 0.54956 1.00000 10 1.81000 osd.10 up 0.44815 1.00000 11 1.81000 osd.11 up 0.53262 1.00000 12 1.81000 osd.12 up 0.47197 1.00000 -5 7.23999 host i3 13 1.81000 osd.13 up 0.55557 1.00000 14 1.81000 osd.14 up 0.65874 1.00000 15 1.81000 osd.15 up 0.49663 1.00000 16 1.81000 osd.16 up 0.50136 1.00000 Thanks, X
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com