I found "osd journal size = 0" in your ceph.conf?
Do you really run mkcephfs with this? I think it will be fail.

On Wed, Apr 30, 2014 at 2:42 PM, Cao, Buddy <buddy....@intel.com> wrote:
> Here you go... I did not see any stuck clean related log...
>
>
>
> Wei Cao (Buddy)
>
> -----Original Message-----
> From: Haomai Wang [mailto:haomaiw...@gmail.com]
> Sent: Wednesday, April 30, 2014 2:12 PM
> To: Cao, Buddy
> Cc: ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] mkcephfs questions
>
> Hmm, it should be another problem plays. Maybe more logs could explain it.
>
> ceph.log
> ceph-mon.log
>
> On Wed, Apr 30, 2014 at 12:06 PM, Cao, Buddy <buddy....@intel.com> wrote:
>> Thanks your reply, Haomai. What I don't understand is that, why the stuck 
>> unclean pgs keep the same numbers after 12 hours. It's the common behavior 
>> or not?
>>
>>
>> Wei Cao (Buddy)
>>
>> -----Original Message-----
>> From: Haomai Wang [mailto:haomaiw...@gmail.com]
>> Sent: Wednesday, April 30, 2014 11:36 AM
>> To: Cao, Buddy
>> Cc: ceph-users@lists.ceph.com
>> Subject: Re: [ceph-users] mkcephfs questions
>>
>> The result of "ceph -s" should tell you the reason. There only exists
>> 21 OSD up but we need 24 OSDs
>>
>> On Wed, Apr 30, 2014 at 11:21 AM, Cao, Buddy <buddy....@intel.com> wrote:
>>> Hi,
>>>
>>>
>>>
>>> I setup ceph cluster thru mkcephfs command, after I enter “ceph –s”,
>>> it always returns 4950 stuck unclean pgs. I tried the same “ceph -s”
>>> after 12 hrs,  there still returns the same unclean pgs number, nothing 
>>> changed.
>>> Does mkcephfs always has the problem or I did something wrong? I
>>> attached the result of “ceph -s”, “ceph osd tree” and ceph.conf I
>>> have, please kindly help.
>>>
>>>
>>>
>>>
>>>
>>> [root@ceph]# ceph -s
>>>
>>>     cluster 99fd4ff8-0fb8-47b9-8179-fefbba1c2503
>>>
>>>      health HEALTH_WARN 4950 pgs degraded; 4950 pgs stuck unclean;
>>> recovery
>>> 21/42 objects degraded (50.000%); 3/24 in osds are down; clock skew
>>> detected on mon.1, mon.2
>>>
>>>      monmap e1: 3 mons at
>>> {0=192.168.0.2:6789/0,1=192.168.0.3:6789/0,2=192.168.0.4:6789/0},
>>> election epoch 6, quorum 0,1,2 0,1,2
>>>
>>>      mdsmap e4: 1/1/1 up {0=0=up:active}
>>>
>>>      osdmap e6019: 24 osds: 21 up, 24 in
>>>
>>>       pgmap v16445: 4950 pgs, 6 pools, 9470 bytes data, 21 objects
>>>
>>>             4900 MB used, 93118 MB / 98019 MB avail
>>>
>>>             21/42 objects degraded (50.000%)
>>>
>>>                 4950 active+degraded
>>>
>>>
>>>
>>> [root@ceph]# ceph osd tree //part of returns
>>>
>>> # id    weight  type name       up/down reweight
>>>
>>> -36     25      root vsm
>>>
>>> -31     3.2             storage_group ssd
>>>
>>> -16     3                       zone zone_a_ssd
>>>
>>> -1      1                               host vsm2_ssd_zone_a
>>>
>>> 2       1                                       osd.2   up      1
>>>
>>> -6      1                               host vsm3_ssd_zone_a
>>>
>>> 10      1                                       osd.10  up      1
>>>
>>> -11     1                               host vsm4_ssd_zone_a
>>>
>>> 18      1                                       osd.18  up      1
>>>
>>> -21     0.09999                 zone zone_c_ssd
>>>
>>> -26     0.09999                 zone zone_b_ssd
>>>
>>> -33     3.2             storage_group sata
>>>
>>> -18     3                       zone zone_a_sata
>>>
>>> -3      1                               host vsm2_sata_zone_a
>>>
>>> 1       1                                       osd.1   up      1
>>>
>>> -8      1                               host vsm3_sata_zone_a
>>>
>>> 9       1                                       osd.9   up      1
>>>
>>> -13     1                               host vsm4_sata_zone_a
>>>
>>> 17      1                                       osd.17  up      1
>>>
>>> -23     0.09999                 zone zone_c_sata
>>>
>>> -28     0.09999                 zone zone_b_sata
>>>
>>>
>>>
>>>
>>>
>>> Wei Cao (Buddy)
>>>
>>>
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>
>>
>>
>> --
>> Best Regards,
>>
>> Wheat
>
>
>
> --
> Best Regards,
>
> Wheat



-- 
Best Regards,

Wheat
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to