Hi to All,

  I've the following warning message (WARN) in my cluster:

ceph@ceph-node04:~$ sudo ceph status
   cluster 50ae3778-dfe3-4492-9628-54a8918ede92
    health HEALTH_WARN too few pgs per osd (3 < min 20)
monmap e1: 1 mons at {ceph-node01=10.1.1.151:6789/0}, election epoch 2, quorum 0 ceph-node01
    osdmap e259: 55 osds: 55 up, 55 in
     pgmap v703: 192 pgs, 3 pools, 0 bytes data, 0 objects
           2579 MB used, 7328 GB / 7331 GB avail
                192 active+clean

ceph@ceph-node04:~$


It's a new cluster setup, the OSD tree is the following:

ceph@ceph-node04:~$ sudo ceph osd tree
# id    weight    type name    up/down    reweight
-1    7.27    root default
-2    1.15        host ceph-node01
12    0.06999            osd.12    up    1
13    0.06999            osd.13    up    1
14    0.06999            osd.14    up    1
15    0.06999            osd.15    up    1
16    0.06999            osd.16    up    1
17    0.06999            osd.17    up    1
18    0.06999            osd.18    up    1
19    0.06999            osd.19    up    1
20    0.06999            osd.20    up    1
21    0.45            osd.21    up    1
22    0.06999            osd.22    up    1
-3    1.53        host ceph-node02
23    0.06999            osd.23    up    1
24    0.06999            osd.24    up    1
25    0.06999            osd.25    up    1
26    0.06999            osd.26    up    1
27    0.06999            osd.27    up    1
28    0.06999            osd.28    up    1
29    0.06999            osd.29    up    1
30    0.06999            osd.30    up    1
31    0.06999            osd.31    up    1
32    0.45            osd.32    up    1
33    0.45            osd.33    up    1
-4    1.53        host ceph-node03
34    0.06999            osd.34    up    1
35    0.06999            osd.35    up    1
36    0.06999            osd.36    up    1
37    0.06999            osd.37    up    1
38    0.06999            osd.38    up    1
39    0.06999            osd.39    up    1
40    0.06999            osd.40    up    1
41    0.06999            osd.41    up    1
42    0.06999            osd.42    up    1
43    0.45            osd.43    up    1
44    0.45            osd.44    up    1
-5    1.53        host ceph-node04
0    0.06999            osd.0    up    1
1    0.06999            osd.1    up    1
2    0.06999            osd.2    up    1
3    0.06999            osd.3    up    1
4    0.06999            osd.4    up    1
5    0.06999            osd.5    up    1
6    0.06999            osd.6    up    1
7    0.06999            osd.7    up    1
8    0.06999            osd.8    up    1
9    0.45            osd.9    up    1
10    0.45            osd.10    up    1
-6    1.53        host ceph-node05
11    0.06999            osd.11    up    1
45    0.06999            osd.45    up    1
46    0.06999            osd.46    up    1
47    0.06999            osd.47    up    1
48    0.06999            osd.48    up    1
49    0.06999            osd.49    up    1
50    0.06999            osd.50    up    1
51    0.06999            osd.51    up    1
52    0.06999            osd.52    up    1
53    0.45            osd.53    up    1
54    0.45            osd.54    up    1

ceph@ceph-node04:~$


Someone could give me a hand to resolved this situation.

German Anders








_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to