You talked about "using default settings wherever possible"... Well, Ceph's
default settings everywhere they exist, is to not allow you to write while
you don't have at least 1 more copy that you can lose without data loss.
If your bosses require you to be able to lose 2 servers and still serve
customers, then tell them that Ceph requires you to have 3 parity copies of
the data.

Why do you want to change your one and only copy of the data while you
already have a degraded system?  And not just a degraded system, but a
system where 2/5's of your servers are down... That sounds awful, terrible,
and just plain bad.

To directly answer your question about min_size, min_size does not affect
where data is placed.  It only affects when a PG claims to not have enough
copies online to be able to receive read or write requests.

On Tue, May 8, 2018 at 7:47 AM Janne Johansson <icepic...@gmail.com> wrote:

> 2018-05-08 1:46 GMT+02:00 Maciej Puzio <mkp37...@gmail.com>:
>
>> Paul, many thanks for your reply.
>> Thinking about it, I can't decide if I'd prefer to operate the storage
>> server without redundancy, or have it automatically force a downtime,
>> subjecting me to a rage of my users and my boss.
>> But I think that the typical expectation is that system serves the
>> data while it is able to do so.
>
>
> If you want to prevent angry bosses, you would have made 10 OSD hosts
> or some other large number so that ceph cloud place PGs over more places
> so that 2 lost hosts would not impact so much, but also so it can recover
> into
> each PG into one of the 10 ( minus two broken minus the three that already
> hold data you want to spread out) other OSDs and get back into full service
> even with two lost hosts.
>
> It's fun to test assumptions and "how low can I go", but if you REALLY
> wanted
> a cluster with resilience to planned and unplanned maintenance,
> you would have redundancy, just like that Raid6 disk box would
> presumably have a fair amount of hot and perhaps cold spares nearby to kick
> in if lots of disks started go missing.
>
> --
> May the most significant bit of your life be positive.
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to