Hi,

Le 18/04/2014 13:14, Ирек Фасихов a écrit :
> Show command please: ceph osd tree.

Sure :

root@node1:~# ceph osd tree
# id    weight    type name    up/down    reweight
-1    3    root default
-2    3        host node1
0    1            osd.0    up    1   
1    1            osd.1    up    1   
2    1            osd.2    up    1   

>
> 2014-04-18 14:51 GMT+04:00 Cedric Lemarchand <ced...@yipikai.org
> <mailto:ced...@yipikai.org>>:
>
>     Hi,
>
>     I am facing a strange behaviour where a pool is stucked, I have no
>     idea how this pool appear in the cluster in the way I have not
>     played with pool creation, *yet*.
>
>     ##### root@node1:~# ceph -s
>         cluster 1b147882-722c-43d8-8dfb-38b78d9fbec3
>          health HEALTH_WARN 333 pgs degraded; 333 pgs stuck unclean;
>     pool .rgw.buckets has too few pgs
>          monmap e1: 1 mons at {node1=127.0.0.1:6789/0
>     <http://127.0.0.1:6789/0>}, election epoch 1, quorum 0 node1
>          osdmap e154: 3 osds: 3 up, 3 in
>           pgmap v16812: 3855 pgs, 14 pools, 41193 MB data, 24792 objects
>                 57236 MB used, 644 GB / 738 GB avail
>                     3522 active+clean
>                      333 active+degraded
>
>     ##### root@node1:/etc/ceph# ceph osd dump
>     epoch 154
>     fsid 1b147882-722c-43d8-8dfb-38b78d9fbec3
>     created 2014-04-16 20:46:46.516403
>     modified 2014-04-18 12:14:29.052231
>     flags
>
>     pool 0 'data' rep size 1 min_size 1 crush_ruleset 0 object_hash
>     rjenkins pg_num 64 pgp_num 64 last_change 1 owner 0
>     crash_replay_interval 45
>     pool 1 'metadata' rep size 1 min_size 1 crush_ruleset 1
>     object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 owner 0
>     pool 2 'rbd' rep size 1 min_size 1 crush_ruleset 2 object_hash
>     rjenkins pg_num 64 pgp_num 64 last_change 1 owner 0
>     pool 3 '.rgw.root' rep size 1 min_size 1 crush_ruleset 0
>     object_hash rjenkins pg_num 333 pgp_num 333 last_change 16 owner 0
>     pool 4 '.rgw.control' rep size 1 min_size 1 crush_ruleset 0
>     object_hash rjenkins pg_num 333 pgp_num 333 last_change 18 owner 0
>     pool 5 '.rgw' rep size 1 min_size 1 crush_ruleset 0 object_hash
>     rjenkins pg_num 333 pgp_num 333 last_change 20 owner 0
>     pool 6 '.rgw.gc' rep size 1 min_size 1 crush_ruleset 0 object_hash
>     rjenkins pg_num 333 pgp_num 333 last_change 21 owner 0
>     pool 7 '.users.uid' rep size 1 min_size 1 crush_ruleset 0
>     object_hash rjenkins pg_num 333 pgp_num 333 last_change 22 owner 0
>     pool 8 '.users' rep size 1 min_size 1 crush_ruleset 0 object_hash
>     rjenkins pg_num 333 pgp_num 333 last_change 26 owner 0
>     pool 9 '.users.swift' rep size 1 min_size 1 crush_ruleset 0
>     object_hash rjenkins pg_num 333 pgp_num 333 last_change 28 owner 0
>     pool 10 '.users.email' rep size 1 min_size 1 crush_ruleset 0
>     object_hash rjenkins pg_num 333 pgp_num 333 last_change 56 owner 0
>     pool 11 '.rgw.buckets.index' rep size 1 min_size 1 crush_ruleset 0
>     object_hash rjenkins pg_num 333 pgp_num 333 last_change 58 owner
>     18446744073709551615
>     pool 12 '.rgw.buckets' rep size 1 min_size 1 crush_ruleset 0
>     object_hash rjenkins pg_num 333 pgp_num 333 last_change 60 owner
>     18446744073709551615
>     pool 13 '' rep size 2 min_size 1 crush_ruleset 0 object_hash
>     rjenkins pg_num 333 pgp_num 333 last_change 146 owner
>     18446744073709551615
>
>     max_osd 5
>     osd.0 up   in  weight 1 up_from 151 up_thru 151 down_at 148
>     last_clean_interval [144,147) 192.168.1.18:6800/26681
>     <http://192.168.1.18:6800/26681> 192.168.1.18:6801/26681
>     <http://192.168.1.18:6801/26681> 192.168.1.18:6802/26681
>     <http://192.168.1.18:6802/26681> 192.168.1.18:6803/26681
>     <http://192.168.1.18:6803/26681> exists,up
>     f6f63e8a-42af-4dda-b523-ffb835165420
>     osd.1 up   in  weight 1 up_from 149 up_thru 149 down_at 148
>     last_clean_interval [139,147) 192.168.1.18:6805/26685
>     <http://192.168.1.18:6805/26685> 192.168.1.18:6806/26685
>     <http://192.168.1.18:6806/26685> 192.168.1.18:6807/26685
>     <http://192.168.1.18:6807/26685> 192.168.1.18:6808/26685
>     <http://192.168.1.18:6808/26685> exists,up
>     fa4689ac-e0ca-4ec3-ab2a-6afa57cc7498
>     osd.2 up   in  weight 1 up_from 153 up_thru 153 down_at 148
>     last_clean_interval [141,147) 192.168.1.18:6810/26691
>     <http://192.168.1.18:6810/26691> 192.168.1.18:6811/26691
>     <http://192.168.1.18:6811/26691> 192.168.1.18:6812/26691
>     <http://192.168.1.18:6812/26691> 192.168.1.18:6813/26691
>     <http://192.168.1.18:6813/26691> exists,up
>     6b2f7e3f-619c-4922-bdf9-bb0f2eee7413
>
>     ##### root@node1:/etc/ceph# ceph pg dump_stuck unclean |sort
>     13.0    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:28.438523    0'0    154:13    [0]    [0]   
>     0'0    2014-04-18 11:12:05.322855    0'0    2014-04-18 11:12:05.322855
>     13.100    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:26.110633    0'0    154:13    [0]    [0]   
>     0'0    2014-04-18 11:12:06.318159    0'0    2014-04-18 11:12:06.318159
>     13.10    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:37.081087    0'0    154:12    [2]    [2]   
>     0'0    2014-04-18 11:12:05.642317    0'0    2014-04-18 11:12:05.642317
>     13.1    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:20.874829    0'0    154:13    [1]    [1]   
>     0'0    2014-04-18 11:12:05.580874    0'0    2014-04-18 11:12:05.580874
>     13.101    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:16.723100    0'0    154:14    [1]    [1]   
>     0'0    2014-04-18 11:12:06.540975    0'0    2014-04-18 11:12:06.540975
>     13.102    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:35.795491    0'0    154:12    [2]    [2]   
>     0'0    2014-04-18 11:12:06.543846    0'0    2014-04-18 11:12:06.543846
>     13.103    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:35.809492    0'0    154:12    [2]    [2]   
>     0'0    2014-04-18 11:12:06.561542    0'0    2014-04-18 11:12:06.561542
>     13.104    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:35.817750    0'0    154:12    [2]    [2]   
>     0'0    2014-04-18 11:12:06.569706    0'0    2014-04-18 11:12:06.569706
>     13.105    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:35.840668    0'0    154:12    [2]    [2]   
>     0'0    2014-04-18 11:12:06.602826    0'0    2014-04-18 11:12:06.602826
>
>     [...]
>
>     13.f7    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:16.990648    0'0    154:14    [1]    [1]   
>     0'0    2014-04-18 11:12:06.483859    0'0    2014-04-18 11:12:06.483859
>     13.f8    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:35.947686    0'0    154:12    [2]    [2]   
>     0'0    2014-04-18 11:12:06.481459    0'0    2014-04-18 11:12:06.481459
>     13.f9    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:35.961392    0'0    154:12    [2]    [2]   
>     0'0    2014-04-18 11:12:06.505039    0'0    2014-04-18 11:12:06.505039
>     13.fa    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:17.062254    0'0    154:14    [1]    [1]   
>     0'0    2014-04-18 11:12:06.493605    0'0    2014-04-18 11:12:06.493605
>     13.fb    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:17.058748    0'0    154:14    [1]    [1]   
>     0'0    2014-04-18 11:12:06.526013    0'0    2014-04-18 11:12:06.526013
>     13.fc    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:26.277414    0'0    154:13    [0]    [0]   
>     0'0    2014-04-18 11:12:06.243714    0'0    2014-04-18 11:12:06.243714
>     13.fd    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:26.312618    0'0    154:13    [0]    [0]   
>     0'0    2014-04-18 11:12:06.263824    0'0    2014-04-18 11:12:06.263824
>     13.fe    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:35.977273    0'0    154:12    [2]    [2]   
>     0'0    2014-04-18 11:12:06.511879    0'0    2014-04-18 11:12:06.511879
>     13.ff    0    0    0    0    0    0    0    active+degraded   
>     2014-04-18 12:14:26.262810    0'0    154:13    [0]    [0]   
>     0'0    2014-04-18 11:12:06.289603    0'0    2014-04-18 11:12:06.289603
>     pg_stat    objects    mip    degr    unf    bytes    log   
>     disklog    state    state_stamp    v    reported    up   
>     acting    last_scrub    scrub_stamp    last_deep_scrub   
>     deep_scrub_stamp
>
>
>     ##### root@node1:~# rados df
>     pool name       category                 KB      objects      
>     clones     degraded      unfound           rd        rd
>     KB           wr        wr KB
>                     -                          0           
>     0            0            0           0            0           
>     0            0            0
>     .rgw            -                          1           
>     5            0            0           0           31          
>     23           17            6
>     .rgw.buckets    -                   42182267       
>     24733            0            0           0         4485       
>     17420       163372     50559394
>     .rgw.buckets.index -                          0           
>     3            0            0           0        47113      
>     105894        44735            0
>     .rgw.control    -                          0           
>     8            0            0           0            0           
>     0            0            0
>     .rgw.gc         -                          0          
>     32            0            0           0         7114        
>     7704         8524            0
>     .rgw.root       -                          1           
>     3            0            0           0           16          
>     10            3            3
>     .users          -                          1           
>     2            0            0           0            0           
>     0            2            2
>     .users.email    -                          1           
>     1            0            0           0            0           
>     0            1            1
>     .users.swift    -                          1           
>     2            0            0           0            5           
>     3            2            2
>     .users.uid      -                          1           
>     3            0            0           0           52          
>     46           16            6
>     data            -                          0           
>     0            0            0           0            0           
>     0            0            0
>     metadata        -                          0           
>     0            0            0           0            0           
>     0            0            0
>     rbd             -                          0           
>     0            0            0           0            0           
>     0            0            0
>       total used        58610648        24792
>       total avail      676160692
>       total space      774092940
>
>
>     The pool seams empty, so I have tried to removed it but the
>     command complain about the empty name. The last modification that
>     have been done was changing the "osd pool default size" in
>     ceph.conf from 1 to 2 and restart the whole cluster (mon + osd),
>     AFAICR the cluster was healtly before doing that.
>
>     This is a small bed test so every thing can be trashed, but I am
>     still a bit curious of what happens and how it could be fixed ?
>
>     Cheers
>
>     -- 
>     Cédric
>
>
>     _______________________________________________
>     ceph-users mailing list
>     ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
>
> -- 
> С уважением, Фасихов Ирек Нургаязович
> Моб.: +79229045757

-- 
Cédric

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to