Please paste 'ceph osd tree'.

Robert LeBlanc

Sent from a mobile device please excuse any typos.
On Oct 28, 2015 6:54 PM, "Wah Peng" <wah_p...@yahoo.com.sg> wrote:

> Hello,
>
> Just did it, but still no good health. can you help? thanks.
>
> ceph@ceph:~/my-cluster$ ceph osd stat
>      osdmap e24: 3 osds: 3 up, 3 in
>
> ceph@ceph:~/my-cluster$ ceph health
> HEALTH_WARN 89 pgs degraded; 67 pgs incomplete; 67 pgs stuck inactive; 192
> pgs stuck unclean
>
>
> On 2015/10/29 星期四 8:38, Lindsay Mathieson wrote:
>
>>
>> On 29 October 2015 at 10:29, Wah Peng <wah_p...@yahoo.com.sg
>> <mailto:wah_p...@yahoo.com.sg>> wrote:
>>
>>     $ ceph osd stat
>>           osdmap e18: 2 osds: 2 up, 2 in
>>
>>     this is what it shows.
>>     does it mean I need to add up to 3 osds? I just use  the default
>> setup.
>>
>>
>> If you went with the defaults then your pool size will be 3, meaning it
>> needs 3 copies of the data (replica 3) to be valid - as you only have
>> two nodes/osd's that can never happen :)
>>
>> Your options are:
>> - Add another node and osd.
>> or
>> - reduce the size to 2.(ceph osd set <poolname> size 2)
>>
>>
>>
>> --
>> Lindsay
>>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to