To be pedantic … backfill usually means copying data in toto, so like normal 
write replication it necessarily has to traverse the WAN.

Recovery of just a lost shard/replica in theory with the LRC plugin, but as 
noted that doesn’t seem like a good choice.  With the default EC plugin, there 
*may* be some read locality preference but it’s not something I would bank on.

Stretch clusters are great when you need zero RPO when you really need a single 
cluster and can manage client endpoint use accordingly.  But with tradeoffs, in 
many cases two clusters with async replication can be a better solution, 
depends on needs and what you’re solving for.

> On May 7, 2025, at 5:06 AM, Janne Johansson <icepic...@gmail.com> wrote:
> 
> Den ons 7 maj 2025 kl 10:59 skrev Torkil Svensgaard <tor...@drcmr.dk>:
>> We are looking at a cluster split between two DCs with the DCs as
>> failure domains.
>> 
>> Am I right in assuming that any recovery or backfill taking place should
>> largely happen inside each DC and not between them? Or can no such
>> assumptions be made?
>> Pools would be EC 4+8, if that matters.
> 
> Unless I am mistaken, the first/primary of each PG is the one "doing"
> the backfills, so if the primaries are evenly distributed between the
> sites, the source of all backfills would be in the remote DC in 50% of
> the cases.
> I do not think the backfills are going to calculate how it can use
> only "local" pieces to rebuild a missing/degraded PG piece without
> going over the DC-DC link even if it is theoretically possible.
> 
> -- 
> May the most significant bit of your life be positive.

It’s good to be 8-bit-clean, if you aren’t , then Kermit can compensate.

> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to