Yes, I am trying it over luminous.

Well the bug has been going for 8 month and it hasn't been merged yet.
Idk if that is whats preventing me to make it work. Tomorrow I will try
to prove it again.


El 19/10/2017 a las 23:00, David Turner escribió:
> Running a cluster on various versions of Hammer and Jewel I haven't
> had any problems.  I haven't upgraded to Luminous quite yet, but I'd
> be surprised if there is that severe of a regression especially since
> they did so many improvements to Erasure Coding.
>
> On Thu, Oct 19, 2017 at 4:59 PM Jorge Pinilla López <jorp...@unizar.es
> <mailto:jorp...@unizar.es>> wrote:
>
>     Well I was trying it some days ago and it didn't work for me.
>
>     maybe because of this:
>
>     http://tracker.ceph.com/issues/18749
>
>     https://github.com/ceph/ceph/pull/17619
>
>     I don't know if now it's actually working
>
>
>     El 19/10/2017 a las 22:55, David Turner escribió:
>>     In a 3 node cluster with EC k=2 m=1, you can turn off one of the
>>     nodes and the cluster will still operate normally.  If you lose a
>>     disk during this state or another server goes offline, then you
>>     lose access to your data.  But assuming that you bring up the
>>     third node and let it finish backfilling/recovering before
>>     restarting any other nodes, then you're fine.
>>
>>     On Thu, Oct 19, 2017 at 4:49 PM Jorge Pinilla López
>>     <jorp...@unizar.es <mailto:jorp...@unizar.es>> wrote:
>>
>>         Imagine we have a 3 OSDs cluster and I make an erasure pool
>>         with k=2 m=1.
>>
>>         If I have an OSD fail, we can rebuild the data but (I think)
>>         the hole cluster won't be able to perform IOS.
>>
>>         Wouldn't be possible to make the cluster work in a degraded
>>         mode?
>>         I think it would be a good idea to make the cluster work on
>>         degraded mode and promise to re balance/re build whenever a
>>         third OSD comes alive.
>>         On reads, it could serve the data using the live data chunks
>>         and rebuilding (if necessary) the missing ones(using cpu to
>>         calculate the data before serving// with 0 RTA) or trying to
>>         rebuild the missing parts so it actually has the 2 data
>>         chunks on the 2 live OSDs (with some RTA and space usage) or
>>         even doing both things at the same time (with high network
>>         and cpu and storage cost).
>>         On writes, it could write the 2 data parts into the live OSDs
>>         and whenever the third OSD comes up, the cluster could re
>>         balance rebuilding the parity chunk and re positioning the
>>         parts so all OSDs have the same amount of data/work.
>>
>>         would this be possible?
>>
>>         
>> ------------------------------------------------------------------------
>>         *Jorge Pinilla López*
>>         jorp...@unizar.es <mailto:jorp...@unizar.es>
>>         Estudiante de ingenieria informática
>>         Becario del area de sistemas (SICUZ)
>>         Universidad de Zaragoza
>>         PGP-KeyID: A34331932EBC715A
>>         
>> <http://pgp.rediris.es:11371/pks/lookup?op=get&search=0xA34331932EBC715A>
>>         
>> ------------------------------------------------------------------------
>>         _______________________________________________
>>         ceph-users mailing list
>>         ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>>         http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>     -- 
>     ------------------------------------------------------------------------
>     *Jorge Pinilla López*
>     jorp...@unizar.es <mailto:jorp...@unizar.es>
>     Estudiante de ingenieria informática
>     Becario del area de sistemas (SICUZ)
>     Universidad de Zaragoza
>     PGP-KeyID: A34331932EBC715A
>     <http://pgp.rediris.es:11371/pks/lookup?op=get&search=0xA34331932EBC715A>
>     ------------------------------------------------------------------------
>

-- 
------------------------------------------------------------------------
*Jorge Pinilla López*
jorp...@unizar.es
Estudiante de ingenieria informática
Becario del area de sistemas (SICUZ)
Universidad de Zaragoza
PGP-KeyID: A34331932EBC715A
<http://pgp.rediris.es:11371/pks/lookup?op=get&search=0xA34331932EBC715A>
------------------------------------------------------------------------
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to