Here is the updated direct copy/paste dump eph@ceph-node1:~$ ceph osd dump epoch 25 fsid 17bea68b-1634-4cd1-8b2a-00a60ef4761d created 2015-02-08 16:59:07.050875 modified 2015-02-09 22:35:33.191218 flags pool 0 'data' replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 128 pgp_num 64 last_change 24 flags hashpspool crash_replay_interval 45 stripe_width 0 pool 1 'metadata' replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool stripe_width 0 pool 2 'rbd' replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool stripe_width 0 max_osd 6 osd.0 up in weight 1 up_from 4 up_thru 17 down_at 0 last_clean_interval [0,0) 172.31.0.84:6800/11739 172.31.0.84:6801/11739 172.31.0.84:6802/11739 172.31.0.84:6803/11739 exists,up 765f5066-d13e-4a9e-a446-8630ee06e596 osd.1 up in weight 1 up_from 7 up_thru 0 down_at 0 last_clean_interval [0,0) 172.31.0.84:6805/12279 172.31.0.84:6806/12279 172.31.0.84:6807/12279 172.31.0.84:6808/12279 exists,up e1d073e5-9397-4b63-8b7c-a4064e430f7a osd.2 up in weight 1 up_from 10 up_thru 0 down_at 0 last_clean_interval [0,0) 172.31.3.57:6800/5517 172.31.3.57:6801/5517 172.31.3.57:6802/5517 172.31.3.57:6803/5517 exists,up 5af5deed-7a6d-4251-aa3c-819393901d1f osd.3 up in weight 1 up_from 13 up_thru 0 down_at 0 last_clean_interval [0,0) 172.31.3.57:6805/6043 172.31.3.57:6806/6043 172.31.3.57:6807/6043 172.31.3.57:6808/6043 exists,up 958f37ab-b434-40bd-87ab-3acbd3118f92 osd.4 up in weight 1 up_from 16 up_thru 0 down_at 0 last_clean_interval [0,0) 172.31.3.56:6800/5106 172.31.3.56:6801/5106 172.31.3.56:6802/5106 172.31.3.56:6803/5106 exists,up ce5c0b86-96be-408a-8022-6397c78032be osd.5 up in weight 1 up_from 22 up_thru 0 down_at 0 last_clean_interval [0,0) 172.31.3.56:6805/7019 172.31.3.56:6806/7019 172.31.3.56:6807/7019 172.31.3.56:6808/7019 exists,up da67b604-b32a-44a0-9920-df0774ad2ef3
> On Feb 10, 2015, at 12:55 PM, B L <super.itera...@gmail.com> wrote: > > >> On Feb 10, 2015, at 12:37 PM, B L <super.itera...@gmail.com >> <mailto:super.itera...@gmail.com>> wrote: >> >> Hi Vickie, >> >> Thanks for your reply! >> >> You can find the dump in this link: >> >> https://gist.github.com/anonymous/706d4a1ec81c93fd1eca >> <https://gist.github.com/anonymous/706d4a1ec81c93fd1eca> >> >> Thanks! >> B. >> >> >>> On Feb 10, 2015, at 12:23 PM, Vickie ch <mika.leaf...@gmail.com >>> <mailto:mika.leaf...@gmail.com>> wrote: >>> >>> Hi Beanos: >>> Would you post the reult of "$ceph osd dump"? >>> >>> Best wishes, >>> Vickie >>> >>> 2015-02-10 16:36 GMT+08:00 B L <super.itera...@gmail.com >>> <mailto:super.itera...@gmail.com>>: >>> Having problem with my fresh non-healthy cluster, my cluster status summary >>> shows this: >>> >>> ceph@ceph-node1:~$ ceph -s >>> >>> cluster 17bea68b-1634-4cd1-8b2a-00a60ef4761d >>> health HEALTH_WARN 256 pgs incomplete; 256 pgs stuck inactive; 256 pgs >>> stuck unclean; pool data pg_num 128 > pgp_num 64 >>> monmap e1: 1 mons at {ceph-node1=172.31.0.84:6789/0 >>> <http://172.31.0.84:6789/0>}, election epoch 2, quorum 0 ceph-node1 >>> osdmap e25: 6 osds: 6 up, 6 in >>> pgmap v82: 256 pgs, 3 pools, 0 bytes data, 0 objects >>> 198 MB used, 18167 MB / 18365 MB avail >>> 192 incomplete >>> 64 creating+incomplete >>> >>> >>> Where shall I start troubleshooting this? >>> >>> P.S. I’m new to CEPH. >>> >>> Thanks! >>> Beanos >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com> >>> >>> >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com