*# ceph health detail* HEALTH_OK *# ceph osd stat* 48 osds: 48 up, 48 in
*# ceph pg stat* 3200 pgs: 3200 active+clean; 5336 MB data, 79455 MB used, 53572 GB / 53650 GB avail *German* 2017-09-13 13:24 GMT-03:00 dE <de.tec...@gmail.com>: > On 09/13/2017 09:08 PM, German Anders wrote: > > Hi cephers, > > I'm having an issue with a newly created cluster 12.2.0 ( > 32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc). Basically when I > reboot one of the nodes, and when it come back, it come outside of the root > type on the tree: > > root@cpm01:~# ceph osd tree > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > -15 12.00000 *root default* > * 36 nvme 1.00000 osd.36 up 1.00000 1.00000* > * 37 nvme 1.00000 osd.37 up 1.00000 1.00000* > * 38 nvme 1.00000 osd.38 up 1.00000 1.00000* > * 39 nvme 1.00000 osd.39 up 1.00000 1.00000* > * 40 nvme 1.00000 osd.40 up 1.00000 1.00000* > * 41 nvme 1.00000 osd.41 up 1.00000 1.00000* > * 42 nvme 1.00000 osd.42 up 1.00000 1.00000* > * 43 nvme 1.00000 osd.43 up 1.00000 1.00000* > * 44 nvme 1.00000 osd.44 up 1.00000 1.00000* > * 45 nvme 1.00000 osd.45 up 1.00000 1.00000* > * 46 nvme 1.00000 osd.46 up 1.00000 1.00000* > * 47 nvme 1.00000 osd.47 up 1.00000 1.00000* > -7 36.00000 *root root* > -5 24.00000 rack rack1 > -1 12.00000 node cpn01 > 0 1.00000 osd.0 up 1.00000 1.00000 > 1 1.00000 osd.1 up 1.00000 1.00000 > 2 1.00000 osd.2 up 1.00000 1.00000 > 3 1.00000 osd.3 up 1.00000 1.00000 > 4 1.00000 osd.4 up 1.00000 1.00000 > 5 1.00000 osd.5 up 1.00000 1.00000 > 6 1.00000 osd.6 up 1.00000 1.00000 > 7 1.00000 osd.7 up 1.00000 1.00000 > 8 1.00000 osd.8 up 1.00000 1.00000 > 9 1.00000 osd.9 up 1.00000 1.00000 > 10 1.00000 osd.10 up 1.00000 1.00000 > 11 1.00000 osd.11 up 1.00000 1.00000 > -3 12.00000 node cpn03 > 24 1.00000 osd.24 up 1.00000 1.00000 > 25 1.00000 osd.25 up 1.00000 1.00000 > 26 1.00000 osd.26 up 1.00000 1.00000 > 27 1.00000 osd.27 up 1.00000 1.00000 > 28 1.00000 osd.28 up 1.00000 1.00000 > 29 1.00000 osd.29 up 1.00000 1.00000 > 30 1.00000 osd.30 up 1.00000 1.00000 > 31 1.00000 osd.31 up 1.00000 1.00000 > 32 1.00000 osd.32 up 1.00000 1.00000 > 33 1.00000 osd.33 up 1.00000 1.00000 > 34 1.00000 osd.34 up 1.00000 1.00000 > 35 1.00000 osd.35 up 1.00000 1.00000 > -6 12.00000 rack rack2 > -2 12.00000 node cpn02 > 12 1.00000 osd.12 up 1.00000 1.00000 > 13 1.00000 osd.13 up 1.00000 1.00000 > 14 1.00000 osd.14 up 1.00000 1.00000 > 15 1.00000 osd.15 up 1.00000 1.00000 > 16 1.00000 osd.16 up 1.00000 1.00000 > 17 1.00000 osd.17 up 1.00000 1.00000 > 18 1.00000 osd.18 up 1.00000 1.00000 > 19 1.00000 osd.19 up 1.00000 1.00000 > 20 1.00000 osd.20 up 1.00000 1.00000 > 21 1.00000 osd.21 up 1.00000 1.00000 > 22 1.00000 osd.22 up 1.00000 1.00000 > 23 1.00000 osd.23 up 1.00000 1.00000 > * -4 0 node cpn04* > > Any ideas of why this happen? and how can I fix it? It supposed to be > inside rack2 > > Thanks in advance, > > Best, > > *German* > > > _______________________________________________ > ceph-users mailing > listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > Can we see the output of ceph health detail. Maybe they're under the > process of recovery. > > Also post the output of ceph osd stat so we can see what nodes are up/in > etc... and ceph pg stat to see the status of various PGs (a pointer to the > recovery process). >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com