Thanks your reply, Haomai. What I don't understand is that, why the stuck 
unclean pgs keep the same numbers after 12 hours. It's the common behavior or 
not?


Wei Cao (Buddy)

-----Original Message-----
From: Haomai Wang [mailto:haomaiw...@gmail.com] 
Sent: Wednesday, April 30, 2014 11:36 AM
To: Cao, Buddy
Cc: ceph-users@lists.ceph.com
Subject: Re: [ceph-users] mkcephfs questions

The result of "ceph -s" should tell you the reason. There only exists
21 OSD up but we need 24 OSDs

On Wed, Apr 30, 2014 at 11:21 AM, Cao, Buddy <buddy....@intel.com> wrote:
> Hi,
>
>
>
> I setup ceph cluster thru mkcephfs command, after I enter “ceph –s”, 
> it always returns 4950 stuck unclean pgs. I tried the same “ceph -s” 
> after 12 hrs,  there still returns the same unclean pgs number, nothing 
> changed.
> Does mkcephfs always has the problem or I did something wrong? I 
> attached the result of “ceph -s”, “ceph osd tree” and ceph.conf I 
> have, please kindly help.
>
>
>
>
>
> [root@ceph]# ceph -s
>
>     cluster 99fd4ff8-0fb8-47b9-8179-fefbba1c2503
>
>      health HEALTH_WARN 4950 pgs degraded; 4950 pgs stuck unclean; 
> recovery
> 21/42 objects degraded (50.000%); 3/24 in osds are down; clock skew 
> detected on mon.1, mon.2
>
>      monmap e1: 3 mons at
> {0=192.168.0.2:6789/0,1=192.168.0.3:6789/0,2=192.168.0.4:6789/0}, 
> election epoch 6, quorum 0,1,2 0,1,2
>
>      mdsmap e4: 1/1/1 up {0=0=up:active}
>
>      osdmap e6019: 24 osds: 21 up, 24 in
>
>       pgmap v16445: 4950 pgs, 6 pools, 9470 bytes data, 21 objects
>
>             4900 MB used, 93118 MB / 98019 MB avail
>
>             21/42 objects degraded (50.000%)
>
>                 4950 active+degraded
>
>
>
> [root@ceph]# ceph osd tree //part of returns
>
> # id    weight  type name       up/down reweight
>
> -36     25      root vsm
>
> -31     3.2             storage_group ssd
>
> -16     3                       zone zone_a_ssd
>
> -1      1                               host vsm2_ssd_zone_a
>
> 2       1                                       osd.2   up      1
>
> -6      1                               host vsm3_ssd_zone_a
>
> 10      1                                       osd.10  up      1
>
> -11     1                               host vsm4_ssd_zone_a
>
> 18      1                                       osd.18  up      1
>
> -21     0.09999                 zone zone_c_ssd
>
> -26     0.09999                 zone zone_b_ssd
>
> -33     3.2             storage_group sata
>
> -18     3                       zone zone_a_sata
>
> -3      1                               host vsm2_sata_zone_a
>
> 1       1                                       osd.1   up      1
>
> -8      1                               host vsm3_sata_zone_a
>
> 9       1                                       osd.9   up      1
>
> -13     1                               host vsm4_sata_zone_a
>
> 17      1                                       osd.17  up      1
>
> -23     0.09999                 zone zone_c_sata
>
> -28     0.09999                 zone zone_b_sata
>
>
>
>
>
> Wei Cao (Buddy)
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>



--
Best Regards,

Wheat
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to