*# ceph health detail*
HEALTH_OK

*# ceph osd stat*
48 osds: 48 up, 48 in

*# ceph pg stat*
3200 pgs: 3200 active+clean; 5336 MB data, 79455 MB used, 53572 GB / 53650
GB avail


*German*

2017-09-13 13:24 GMT-03:00 dE <de.tec...@gmail.com>:

> On 09/13/2017 09:08 PM, German Anders wrote:
>
> Hi cephers,
>
> I'm having an issue with a newly created cluster 12.2.0 (
> 32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc). Basically when I
> reboot one of the nodes, and when it come back, it come outside of the root
> type on the tree:
>
> root@cpm01:~# ceph osd tree
> ID  CLASS WEIGHT   TYPE NAME          STATUS REWEIGHT PRI-AFF
> -15       12.00000 *root default*
> * 36  nvme  1.00000     osd.36             up  1.00000 1.00000*
> * 37  nvme  1.00000     osd.37             up  1.00000 1.00000*
> * 38  nvme  1.00000     osd.38             up  1.00000 1.00000*
> * 39  nvme  1.00000     osd.39             up  1.00000 1.00000*
> * 40  nvme  1.00000     osd.40             up  1.00000 1.00000*
> * 41  nvme  1.00000     osd.41             up  1.00000 1.00000*
> * 42  nvme  1.00000     osd.42             up  1.00000 1.00000*
> * 43  nvme  1.00000     osd.43             up  1.00000 1.00000*
> * 44  nvme  1.00000     osd.44             up  1.00000 1.00000*
> * 45  nvme  1.00000     osd.45             up  1.00000 1.00000*
> * 46  nvme  1.00000     osd.46             up  1.00000 1.00000*
> * 47  nvme  1.00000     osd.47             up  1.00000 1.00000*
>  -7       36.00000 *root root*
>  -5       24.00000     rack rack1
>  -1       12.00000         node cpn01
>   0        1.00000             osd.0      up  1.00000 1.00000
>   1        1.00000             osd.1      up  1.00000 1.00000
>   2        1.00000             osd.2      up  1.00000 1.00000
>   3        1.00000             osd.3      up  1.00000 1.00000
>   4        1.00000             osd.4      up  1.00000 1.00000
>   5        1.00000             osd.5      up  1.00000 1.00000
>   6        1.00000             osd.6      up  1.00000 1.00000
>   7        1.00000             osd.7      up  1.00000 1.00000
>   8        1.00000             osd.8      up  1.00000 1.00000
>   9        1.00000             osd.9      up  1.00000 1.00000
>  10        1.00000             osd.10     up  1.00000 1.00000
>  11        1.00000             osd.11     up  1.00000 1.00000
>  -3       12.00000         node cpn03
>  24        1.00000             osd.24     up  1.00000 1.00000
>  25        1.00000             osd.25     up  1.00000 1.00000
>  26        1.00000             osd.26     up  1.00000 1.00000
>  27        1.00000             osd.27     up  1.00000 1.00000
>  28        1.00000             osd.28     up  1.00000 1.00000
>  29        1.00000             osd.29     up  1.00000 1.00000
>  30        1.00000             osd.30     up  1.00000 1.00000
>  31        1.00000             osd.31     up  1.00000 1.00000
>  32        1.00000             osd.32     up  1.00000 1.00000
>  33        1.00000             osd.33     up  1.00000 1.00000
>  34        1.00000             osd.34     up  1.00000 1.00000
>  35        1.00000             osd.35     up  1.00000 1.00000
>  -6       12.00000     rack rack2
>  -2       12.00000         node cpn02
>  12        1.00000             osd.12     up  1.00000 1.00000
>  13        1.00000             osd.13     up  1.00000 1.00000
>  14        1.00000             osd.14     up  1.00000 1.00000
>  15        1.00000             osd.15     up  1.00000 1.00000
>  16        1.00000             osd.16     up  1.00000 1.00000
>  17        1.00000             osd.17     up  1.00000 1.00000
>  18        1.00000             osd.18     up  1.00000 1.00000
>  19        1.00000             osd.19     up  1.00000 1.00000
>  20        1.00000             osd.20     up  1.00000 1.00000
>  21        1.00000             osd.21     up  1.00000 1.00000
>  22        1.00000             osd.22     up  1.00000 1.00000
>  23        1.00000             osd.23     up  1.00000 1.00000
> * -4              0         node cpn04*
>
> Any ideas of why this happen? and how can I fix it? It supposed to be
> inside rack2
>
> Thanks in advance,
>
> Best,
>
> *German*
>
>
> _______________________________________________
> ceph-users mailing 
> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
> Can we see the output of ceph health detail. Maybe they're under the
> process of recovery.
>
> Also post the output of ceph osd stat so we can see what nodes are up/in
> etc... and ceph pg stat to see the status of various PGs (a pointer to the
> recovery process).
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to