A Google search should have lead you the rest of the way. Follow this [1] and in the rule section on step choose leaf change host to osd. You won't need to change the configuration this way, it is saved in the CRUSH map.
[1] http://docs.ceph.com/docs/master/rados/operations/crush-map/#editing-a-crush-map Robert LeBlanc Sent from a mobile device please excuse any typos. On Oct 28, 2015 7:46 PM, "Wah Peng" <wah_p...@yahoo.com.sg> wrote: > Is there a ceph sub-command existing instead of changing the config file? > :) > > > On 2015/10/29 星期四 9:24, Li, Chengyuan wrote: > >> Try " osd crush chooseleaf type = 0" in /etc/ceph/<clustername>.conf >> >> >> Regards, >> CY. >> >> -----Original Message----- >> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of >> Wah Peng >> Sent: 2015年10月29日 9:14 >> To: Robert LeBlanc >> Cc: Lindsay Mathieson; Gurjar, Unmesh; ceph-users@lists.ceph.com >> Subject: Re: [ceph-users] creating+incomplete issues >> >> wow this sounds hard to me. can you show the details? >> thanks a lot. >> >> >> On 2015/10/29 星期四 9:01, Robert LeBlanc wrote: >> >>> You need to change the CRUSH map to select osd instead of host. >>> >>> Robert LeBlanc >>> >>> Sent from a mobile device please excuse any typos. >>> >>> On Oct 28, 2015 7:00 PM, "Wah Peng" <wah_p...@yahoo.com.sg >>> <mailto:wah_p...@yahoo.com.sg>> wrote: >>> >>> $ ceph osd tree >>> # id weight type name up/down reweight >>> -1 0.24 root default >>> -2 0.24 host ceph2 >>> 0 0.07999 osd.0 up 1 >>> 1 0.07999 osd.1 up 1 >>> 2 0.07999 osd.2 up 1 >>> >>> >>> On 2015/10/29 星期四 8:55, Robert LeBlanc wrote: >>> >>> Please paste 'ceph osd tree'. >>> >>> Robert LeBlanc >>> >>> Sent from a mobile device please excuse any typos. >>> >>> On Oct 28, 2015 6:54 PM, "Wah Peng" <wah_p...@yahoo.com.sg >>> <mailto:wah_p...@yahoo.com.sg> >>> <mailto:wah_p...@yahoo.com.sg <mailto:wah_p...@yahoo.com.sg>>> >>> wrote: >>> >>> Hello, >>> >>> Just did it, but still no good health. can you help? >>> thanks. >>> >>> ceph@ceph:~/my-cluster$ ceph osd stat >>> osdmap e24: 3 osds: 3 up, 3 in >>> >>> ceph@ceph:~/my-cluster$ ceph health >>> HEALTH_WARN 89 pgs degraded; 67 pgs incomplete; 67 pgs >>> stuck >>> inactive; 192 pgs stuck unclean >>> >>> >>> On 2015/10/29 星期四 8:38, Lindsay Mathieson wrote: >>> >>> >>> On 29 October 2015 at 10:29, Wah Peng >>> <wah_p...@yahoo.com.sg <mailto:wah_p...@yahoo.com.sg> >>> <mailto:wah_p...@yahoo.com.sg >>> <mailto:wah_p...@yahoo.com.sg>> >>> <mailto:wah_p...@yahoo.com.sg >>> <mailto:wah_p...@yahoo.com.sg> <mailto:wah_p...@yahoo.com.sg >>> <mailto:wah_p...@yahoo.com.sg>>>> >>> wrote: >>> >>> $ ceph osd stat >>> osdmap e18: 2 osds: 2 up, 2 in >>> >>> this is what it shows. >>> does it mean I need to add up to 3 osds? I just >>> use the >>> default setup. >>> >>> >>> If you went with the defaults then your pool size will >>> be 3, >>> meaning it >>> needs 3 copies of the data (replica 3) to be valid - as >>> you only >>> have >>> two nodes/osd's that can never happen :) >>> >>> Your options are: >>> - Add another node and osd. >>> or >>> - reduce the size to 2.(ceph osd set <poolname> size >>> 2) >>> >>> >>> >>> -- >>> Lindsay >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> >>> <mailto:ceph-users@lists.ceph.com >>> <mailto:ceph-users@lists.ceph.com>> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com