On 09/13/2017 09:08 PM, German Anders wrote:
Hi cephers,
I'm having an issue with a newly created cluster 12.2.0
(32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc). Basically
when I reboot one of the nodes, and when it come back, it come outside
of the root type on the tree:
root@cpm01:~# ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-15 12.00000 *root default*
* 36 nvme 1.00000 osd.36 up 1.00000 1.00000*
* 37 nvme 1.00000 osd.37 up 1.00000 1.00000*
* 38 nvme 1.00000 osd.38 up 1.00000 1.00000*
* 39 nvme 1.00000 osd.39 up 1.00000 1.00000*
* 40 nvme 1.00000 osd.40 up 1.00000 1.00000*
* 41 nvme 1.00000 osd.41 up 1.00000 1.00000*
* 42 nvme 1.00000 osd.42 up 1.00000 1.00000*
* 43 nvme 1.00000 osd.43 up 1.00000 1.00000*
* 44 nvme 1.00000 osd.44 up 1.00000 1.00000*
* 45 nvme 1.00000 osd.45 up 1.00000 1.00000*
* 46 nvme 1.00000 osd.46 up 1.00000 1.00000*
* 47 nvme 1.00000 osd.47 up 1.00000 1.00000*
-7 36.00000 *root root*
-5 24.00000 rack rack1
-1 12.00000 node cpn01
0 1.00000 osd.0 up 1.00000 1.00000
1 1.00000 osd.1 up 1.00000 1.00000
2 1.00000 osd.2 up 1.00000 1.00000
3 1.00000 osd.3 up 1.00000 1.00000
4 1.00000 osd.4 up 1.00000 1.00000
5 1.00000 osd.5 up 1.00000 1.00000
6 1.00000 osd.6 up 1.00000 1.00000
7 1.00000 osd.7 up 1.00000 1.00000
8 1.00000 osd.8 up 1.00000 1.00000
9 1.00000 osd.9 up 1.00000 1.00000
10 1.00000 osd.10 up 1.00000 1.00000
11 1.00000 osd.11 up 1.00000 1.00000
-3 12.00000 node cpn03
24 1.00000 osd.24 up 1.00000 1.00000
25 1.00000 osd.25 up 1.00000 1.00000
26 1.00000 osd.26 up 1.00000 1.00000
27 1.00000 osd.27 up 1.00000 1.00000
28 1.00000 osd.28 up 1.00000 1.00000
29 1.00000 osd.29 up 1.00000 1.00000
30 1.00000 osd.30 up 1.00000 1.00000
31 1.00000 osd.31 up 1.00000 1.00000
32 1.00000 osd.32 up 1.00000 1.00000
33 1.00000 osd.33 up 1.00000 1.00000
34 1.00000 osd.34 up 1.00000 1.00000
35 1.00000 osd.35 up 1.00000 1.00000
-6 12.00000 rack rack2
-2 12.00000 node cpn02
12 1.00000 osd.12 up 1.00000 1.00000
13 1.00000 osd.13 up 1.00000 1.00000
14 1.00000 osd.14 up 1.00000 1.00000
15 1.00000 osd.15 up 1.00000 1.00000
16 1.00000 osd.16 up 1.00000 1.00000
17 1.00000 osd.17 up 1.00000 1.00000
18 1.00000 osd.18 up 1.00000 1.00000
19 1.00000 osd.19 up 1.00000 1.00000
20 1.00000 osd.20 up 1.00000 1.00000
21 1.00000 osd.21 up 1.00000 1.00000
22 1.00000 osd.22 up 1.00000 1.00000
23 1.00000 osd.23 up 1.00000 1.00000
* -4 0 node cpn04*
Any ideas of why this happen? and how can I fix it? It supposed to be
inside rack2
Thanks in advance,
Best,
**
*German*
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Can we see the output of ceph health detail. Maybe they're under the
process of recovery.
Also post the output of ceph osd stat so we can see what nodes are up/in
etc... and ceph pg stat to see the status of various PGs (a pointer to
the recovery process).
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com