You need to change the CRUSH map to select osd instead of host.

Robert LeBlanc

Sent from a mobile device please excuse any typos.
On Oct 28, 2015 7:00 PM, "Wah Peng" <wah_p...@yahoo.com.sg> wrote:

> $ ceph osd tree
> # id    weight  type name       up/down reweight
> -1      0.24    root default
> -2      0.24            host ceph2
> 0       0.07999                 osd.0   up      1
> 1       0.07999                 osd.1   up      1
> 2       0.07999                 osd.2   up      1
>
>
> On 2015/10/29 星期四 8:55, Robert LeBlanc wrote:
>
>> Please paste 'ceph osd tree'.
>>
>> Robert LeBlanc
>>
>> Sent from a mobile device please excuse any typos.
>>
>> On Oct 28, 2015 6:54 PM, "Wah Peng" <wah_p...@yahoo.com.sg
>> <mailto:wah_p...@yahoo.com.sg>> wrote:
>>
>>     Hello,
>>
>>     Just did it, but still no good health. can you help? thanks.
>>
>>     ceph@ceph:~/my-cluster$ ceph osd stat
>>           osdmap e24: 3 osds: 3 up, 3 in
>>
>>     ceph@ceph:~/my-cluster$ ceph health
>>     HEALTH_WARN 89 pgs degraded; 67 pgs incomplete; 67 pgs stuck
>>     inactive; 192 pgs stuck unclean
>>
>>
>>     On 2015/10/29 星期四 8:38, Lindsay Mathieson wrote:
>>
>>
>>         On 29 October 2015 at 10:29, Wah Peng <wah_p...@yahoo.com.sg
>>         <mailto:wah_p...@yahoo.com.sg>
>>         <mailto:wah_p...@yahoo.com.sg <mailto:wah_p...@yahoo.com.sg>>>
>>         wrote:
>>
>>              $ ceph osd stat
>>                    osdmap e18: 2 osds: 2 up, 2 in
>>
>>              this is what it shows.
>>              does it mean I need to add up to 3 osds? I just use  the
>>         default setup.
>>
>>
>>         If you went with the defaults then your pool size will be 3,
>>         meaning it
>>         needs 3 copies of the data (replica 3) to be valid - as you only
>>         have
>>         two nodes/osd's that can never happen :)
>>
>>         Your options are:
>>         - Add another node and osd.
>>         or
>>         - reduce the size to 2.(ceph osd set <poolname> size 2)
>>
>>
>>
>>         --
>>         Lindsay
>>
>>     _______________________________________________
>>     ceph-users mailing list
>>     ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>>     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to