On 09/13/2017 09:08 PM, German Anders wrote:
Hi cephers,

I'm having an issue with a newly created cluster 12.2.0 (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc). Basically when I reboot one of the nodes, and when it come back, it come outside of the root type on the tree:

root@cpm01:~# ceph osd tree
ID  CLASS WEIGHT   TYPE NAME    STATUS REWEIGHT PRI-AFF
-15       12.00000 *root default*
* 36  nvme  1.00000     osd.36           up  1.00000 1.00000*
* 37  nvme  1.00000     osd.37           up  1.00000 1.00000*
* 38  nvme  1.00000     osd.38           up  1.00000 1.00000*
* 39  nvme  1.00000     osd.39           up  1.00000 1.00000*
* 40  nvme  1.00000     osd.40           up  1.00000 1.00000*
* 41  nvme  1.00000     osd.41           up  1.00000 1.00000*
* 42  nvme  1.00000     osd.42           up  1.00000 1.00000*
* 43  nvme  1.00000     osd.43           up  1.00000 1.00000*
* 44  nvme  1.00000     osd.44           up  1.00000 1.00000*
* 45  nvme  1.00000     osd.45           up  1.00000 1.00000*
* 46  nvme  1.00000     osd.46           up  1.00000 1.00000*
* 47  nvme  1.00000     osd.47           up  1.00000 1.00000*
 -7       36.00000 *root root*
 -5       24.00000     rack rack1
 -1       12.00000         node cpn01
  0        1.00000 osd.0      up  1.00000 1.00000
  1        1.00000 osd.1      up  1.00000 1.00000
  2        1.00000 osd.2      up  1.00000 1.00000
  3        1.00000 osd.3      up  1.00000 1.00000
  4        1.00000 osd.4      up  1.00000 1.00000
  5        1.00000 osd.5      up  1.00000 1.00000
  6        1.00000 osd.6      up  1.00000 1.00000
  7        1.00000 osd.7      up  1.00000 1.00000
  8        1.00000 osd.8      up  1.00000 1.00000
  9        1.00000 osd.9      up  1.00000 1.00000
 10        1.00000 osd.10     up  1.00000 1.00000
 11        1.00000 osd.11     up  1.00000 1.00000
 -3       12.00000         node cpn03
 24        1.00000 osd.24     up  1.00000 1.00000
 25        1.00000 osd.25     up  1.00000 1.00000
 26        1.00000 osd.26     up  1.00000 1.00000
 27        1.00000 osd.27     up  1.00000 1.00000
 28        1.00000 osd.28     up  1.00000 1.00000
 29        1.00000 osd.29     up  1.00000 1.00000
 30        1.00000 osd.30     up  1.00000 1.00000
 31        1.00000 osd.31     up  1.00000 1.00000
 32        1.00000 osd.32     up  1.00000 1.00000
 33        1.00000 osd.33     up  1.00000 1.00000
 34        1.00000 osd.34     up  1.00000 1.00000
 35        1.00000 osd.35     up  1.00000 1.00000
 -6       12.00000     rack rack2
 -2       12.00000         node cpn02
 12        1.00000 osd.12     up  1.00000 1.00000
 13        1.00000 osd.13     up  1.00000 1.00000
 14        1.00000 osd.14     up  1.00000 1.00000
 15        1.00000 osd.15     up  1.00000 1.00000
 16        1.00000 osd.16     up  1.00000 1.00000
 17        1.00000 osd.17     up  1.00000 1.00000
 18        1.00000 osd.18     up  1.00000 1.00000
 19        1.00000 osd.19     up  1.00000 1.00000
 20        1.00000 osd.20     up  1.00000 1.00000
 21        1.00000 osd.21     up  1.00000 1.00000
 22        1.00000 osd.22     up  1.00000 1.00000
 23        1.00000 osd.23     up  1.00000 1.00000
* -4              0         node cpn04*

Any ideas of why this happen? and how can I fix it? It supposed to be inside rack2

Thanks in advance,

Best,

**

*German*


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Can we see the output of ceph health detail. Maybe they're under the process of recovery.

Also post the output of ceph osd stat so we can see what nodes are up/in etc... and ceph pg stat to see the status of various PGs (a pointer to the recovery process).

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to