Hi there !
Being suspicious about replication between two clusters,
I've done a radosgw-admin data sync init on the secondary zone.
Since then, after a lot of activity, I'm stuck with recovering
shards, nothing moves. Incremental sync still work.
Wondering if I had a bad state also on the primary, I also did a data sync init
on
the primary...
And now, it's also stuck with recovering shard !
In sync error list, I can find some "failed to sync bucket
instance: (125) Operation canceled" errors
I also tried to rewrite some buckets chown in those errors, but
nothing changes. Strange, in those errors, the objects names are
not real objects, example : "name": "replic_cfn_rec/cfb0047:aefd400
3-1866-4b16-b1b3-2f308075cd1c.20298566.4:11[0]"
I wonder what is this ending ":10[0]".
I also tried to remove stale instances, but nothing.
I've still not retry a data sync init on secondary, perhaps I should, but the
generated activity is impactfull.
Can we reduce that resync activity priority ?
ah, My primary cluster is on Reef 18.2.4, the secondary still on 18.2.2 (needs
OS upgrade, Ubuntu 18.04).
--
Gilles
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io