Hi Oliver,
can you post here on the mailing list the steps taken?
>From the IRC logs you said "if I use "choose .... osd", it works -- but
"chooseleaf ... host" doesn't work"
So, to have data balanced between 2 rooms, is the rule "step chooseleaf
firstn 0 type room" correct?

Thanks

-- 
Marco


2013/3/8 Olivier Bonvalet <ceph.l...@daevel.fr>

>
> >
> > Thanks for your answer. So I made some tests on a dedicated spool, I was
> > able to move data from «platter» to «SSD» very well, it's great.
> >
> > But I can't obtain that per "network" neither per "host" setup :
> > with 2 hosts, each one with 2 OSD, and with a pool with use only 1
> > replica (so, 2 copies), I tried this rule :
> >
> >         rule rbdperhost {
> >               ruleset 5
> >               type replicated
> >               min_size 1
> >               max_size 10
> >               step take default
> >               step chooseleaf firstn 0 type host
> >               step emit
> >         }
> >
> >
> > As a result I obtain some PG which stuck in «active+remapped» state.
> > When querying one of this PG, I see that CRUSH find only one OSD up for
> > this one and can't find an other OSD to set replica.
> >
> > If I well understand, in this case the "chooseleaf firstn 0 type host"
> > say to Ceph to choose 2 differents hosts, then in each of them choose
> > one OSD. So with 2 hosts, it should works, no ?
> >
> > Thanks,
> > Olivier B.
> >
> >
>
> So, as said on IRC, it's solved. My rules were not working, and after
> use of «tunables» it's ok.
>
> I love that feature of changing data spreading on live !
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to