Running a cluster on various versions of Hammer and Jewel I haven't had any
problems.  I haven't upgraded to Luminous quite yet, but I'd be surprised
if there is that severe of a regression especially since they did so many
improvements to Erasure Coding.

On Thu, Oct 19, 2017 at 4:59 PM Jorge Pinilla López <jorp...@unizar.es>
wrote:

> Well I was trying it some days ago and it didn't work for me.
>
> maybe because of this:
>
> http://tracker.ceph.com/issues/18749
>
> https://github.com/ceph/ceph/pull/17619
>
> I don't know if now it's actually working
>
> El 19/10/2017 a las 22:55, David Turner escribió:
>
> In a 3 node cluster with EC k=2 m=1, you can turn off one of the nodes and
> the cluster will still operate normally.  If you lose a disk during this
> state or another server goes offline, then you lose access to your data.
> But assuming that you bring up the third node and let it finish
> backfilling/recovering before restarting any other nodes, then you're fine.
>
> On Thu, Oct 19, 2017 at 4:49 PM Jorge Pinilla López <jorp...@unizar.es>
> wrote:
>
>> Imagine we have a 3 OSDs cluster and I make an erasure pool with k=2 m=1.
>>
>> If I have an OSD fail, we can rebuild the data but (I think) the hole
>> cluster won't be able to perform IOS.
>>
>> Wouldn't be possible to make the cluster work in a degraded mode?
>> I think it would be a good idea to make the cluster work on degraded mode
>> and promise to re balance/re build whenever a third OSD comes alive.
>> On reads, it could serve the data using the live data chunks and
>> rebuilding (if necessary) the missing ones(using cpu to calculate the data
>> before serving// with 0 RTA) or trying to rebuild the missing parts so it
>> actually has the 2 data chunks on the 2 live OSDs (with some RTA and space
>> usage) or even doing both things at the same time (with high network and
>> cpu and storage cost).
>> On writes, it could write the 2 data parts into the live OSDs and
>> whenever the third OSD comes up, the cluster could re balance rebuilding
>> the parity chunk and re positioning the parts so all OSDs have the same
>> amount of data/work.
>>
>> would this be possible?
>>
>> ------------------------------
>> *Jorge Pinilla López*
>> jorp...@unizar.es
>> Estudiante de ingenieria informática
>> Becario del area de sistemas (SICUZ)
>> Universidad de Zaragoza
>> PGP-KeyID: A34331932EBC715A
>> <http://pgp.rediris.es:11371/pks/lookup?op=get&search=0xA34331932EBC715A>
>> ------------------------------
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
> --
> ------------------------------
> *Jorge Pinilla López*
> jorp...@unizar.es
> Estudiante de ingenieria informática
> Becario del area de sistemas (SICUZ)
> Universidad de Zaragoza
> PGP-KeyID: A34331932EBC715A
> <http://pgp.rediris.es:11371/pks/lookup?op=get&search=0xA34331932EBC715A>
> ------------------------------
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to