Unless your min_size is set to 3, then you are not hitting the bug in the
tracker you linked.  Most likely you are running with a min_size of 2 which
means that bug is not relevant to your cluster.  Upload this if you
wouldn't mind.  `ceph osd pool get {pool_name} all`

On Thu, Oct 19, 2017 at 5:03 PM Jorge Pinilla López <jorp...@unizar.es>
wrote:

> Yes, I am trying it over luminous.
>
> Well the bug has been going for 8 month and it hasn't been merged yet. Idk
> if that is whats preventing me to make it work. Tomorrow I will try to
> prove it again.
>
> El 19/10/2017 a las 23:00, David Turner escribió:
>
> Running a cluster on various versions of Hammer and Jewel I haven't had
> any problems.  I haven't upgraded to Luminous quite yet, but I'd be
> surprised if there is that severe of a regression especially since they did
> so many improvements to Erasure Coding.
>
> On Thu, Oct 19, 2017 at 4:59 PM Jorge Pinilla López <jorp...@unizar.es>
> wrote:
>
>> Well I was trying it some days ago and it didn't work for me.
>>
>> maybe because of this:
>>
>> http://tracker.ceph.com/issues/18749
>>
>> https://github.com/ceph/ceph/pull/17619
>>
>> I don't know if now it's actually working
>>
>> El 19/10/2017 a las 22:55, David Turner escribió:
>>
>> In a 3 node cluster with EC k=2 m=1, you can turn off one of the nodes
>> and the cluster will still operate normally.  If you lose a disk during
>> this state or another server goes offline, then you lose access to your
>> data.  But assuming that you bring up the third node and let it finish
>> backfilling/recovering before restarting any other nodes, then you're fine.
>>
>> On Thu, Oct 19, 2017 at 4:49 PM Jorge Pinilla López <jorp...@unizar.es>
>> wrote:
>>
>>> Imagine we have a 3 OSDs cluster and I make an erasure pool with k=2 m=1.
>>>
>>> If I have an OSD fail, we can rebuild the data but (I think) the hole
>>> cluster won't be able to perform IOS.
>>>
>>> Wouldn't be possible to make the cluster work in a degraded mode?
>>> I think it would be a good idea to make the cluster work on degraded
>>> mode and promise to re balance/re build whenever a third OSD comes alive.
>>> On reads, it could serve the data using the live data chunks and
>>> rebuilding (if necessary) the missing ones(using cpu to calculate the data
>>> before serving// with 0 RTA) or trying to rebuild the missing parts so it
>>> actually has the 2 data chunks on the 2 live OSDs (with some RTA and space
>>> usage) or even doing both things at the same time (with high network and
>>> cpu and storage cost).
>>> On writes, it could write the 2 data parts into the live OSDs and
>>> whenever the third OSD comes up, the cluster could re balance rebuilding
>>> the parity chunk and re positioning the parts so all OSDs have the same
>>> amount of data/work.
>>>
>>> would this be possible?
>>>
>>> ------------------------------
>>> *Jorge Pinilla López*
>>> jorp...@unizar.es
>>> Estudiante de ingenieria informática
>>> Becario del area de sistemas (SICUZ)
>>> Universidad de Zaragoza
>>> PGP-KeyID: A34331932EBC715A
>>> <http://pgp.rediris.es:11371/pks/lookup?op=get&search=0xA34331932EBC715A>
>>> ------------------------------
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>
>> --
>> ------------------------------
>> *Jorge Pinilla López*
>> jorp...@unizar.es
>> Estudiante de ingenieria informática
>> Becario del area de sistemas (SICUZ)
>> Universidad de Zaragoza
>> PGP-KeyID: A34331932EBC715A
>> <http://pgp.rediris.es:11371/pks/lookup?op=get&search=0xA34331932EBC715A>
>> ------------------------------
>>
>
> --
> ------------------------------
> *Jorge Pinilla López*
> jorp...@unizar.es
> Estudiante de ingenieria informática
> Becario del area de sistemas (SICUZ)
> Universidad de Zaragoza
> PGP-KeyID: A34331932EBC715A
> <http://pgp.rediris.es:11371/pks/lookup?op=get&search=0xA34331932EBC715A>
> ------------------------------
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to