Casey, This was the result of trying 'data sync init':
root@c2-rgw1:~# radosgw-admin data sync init ERROR: source zone not specified root@c2-rgw1:~# radosgw-admin data sync init --source-zone=<master zone uuid> WARNING: cannot find source zone id for name=<master zone uuid> ERROR: sync.init_sync_status() returned ret=-2 root@c2-rgw1:~# radosgw-admin data sync init --source-zone=c1-zone ERROR: sync.init() returned ret=-5 2019-03-06 10:14:59.815735 7fecb214fe40 0 data sync: ERROR: failed to fetch datalog info root@c2-rgw1:~# Do you have any further advice or info? Thanks again, Trey On Wed, Mar 6, 2019 at 11:47 AM Casey Bodley <cbod...@redhat.com> wrote: > Hi Trey, > > I think it's more likely that these stale metadata entries are from > deleted buckets, rather than accidental bucket reshards. When a bucket > is deleted in a multisite configuration, we don't delete its bucket > instance because other zones may still need to sync the object deletes - > and they can't make progress on sync if the bucket metadata disappears. > These leftover bucket instances look the same to the 'reshard > stale-instances' commands, but I'd be cautious about using that to > remove them in multisite, as it may cause more sync errors and > potentially leak storage if they still contain objects. > > Regarding 'datalog trim', that alone isn't safe because it could trim > entries that hadn't been applied on other zones yet, causing them to > miss some updates. What you can do is run 'data sync init' on each zone, > and restart gateways. This will restart with a data full sync (which > will scan all buckets for changes), and skip past any datalog entries > from before the full sync. I was concerned that the bug in error > handling (ie "ERROR: init sync on...") would also affect full sync, but > that doesn't appear to be the case - so I do think that's worth trying. > > On 3/5/19 6:24 PM, Trey Palmer wrote: > > Casey, > > > > Thanks very much for the reply! > > > > We definitely have lots of errors on sync-disabled buckets and the > > workaround for that is obvious (most of them are empty anyway). > > > > Our second form of error is stale buckets. We had dynamic resharding > > enabled but have now disabled it (having discovered it was on by > > default, and not supported in multisite). > > > > We removed several hundred stale buckets via 'radosgw-admin sharding > > stale-instances rm', but they are still giving us sync errors. > > > > I have found that these buckets do have entries in 'radosgw-admin > > datalog list', and my guess is this could be fixed by doing a > > 'radosgw-admin datalog trim' for each entry on the master zone. > > > > Does that sound right? :-) > > > > Thanks again for the detailed explanation, > > > > Trey Palmer > > > > On Tue, Mar 5, 2019 at 5:55 PM Casey Bodley <cbod...@redhat.com > > <mailto:cbod...@redhat.com>> wrote: > > > > Hi Christian, > > > > I think you've correctly intuited that the issues are related to > > the use > > of 'bucket sync disable'. There was a bug fix for that feature in > > http://tracker.ceph.com/issues/26895, and I recently found that a > > block > > of code was missing from its luminous backport. That missing code is > > what handled those "ERROR: init sync on <bucket instance> failed, > > retcode=-2" errors. > > > > I included a fix for that in a later backport > > (https://github.com/ceph/ceph/pull/26549), which I'm still working > to > > get through qa. I'm afraid I can't really recommend a workaround > > for the > > issue in the meantime. > > > > Looking forward though, we do plan to support something like s3's > > cross > > region replication so you can enable replication on a specific bucket > > without having to enable it globally. > > > > Casey > > > > > > On 3/5/19 2:32 PM, Christian Rice wrote: > > > > > > Much appreciated. We’ll continue to poke around and certainly will > > > disable the dynamic resharding. > > > > > > We started with 12.2.8 in production. We definitely did not > > have it > > > enabled in ceph.conf > > > > > > *From: *Matthew H <matthew.he...@hotmail.com > > <mailto:matthew.he...@hotmail.com>> > > > *Date: *Tuesday, March 5, 2019 at 11:22 AM > > > *To: *Christian Rice <cr...@pandora.com > > <mailto:cr...@pandora.com>>, ceph-users > > > <ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>> > > > *Cc: *Trey Palmer <nerdmagic...@gmail.com > > <mailto:nerdmagic...@gmail.com>> > > > *Subject: *Re: radosgw sync falling behind regularly > > > > > > Hi Christian, > > > > > > To be on the safe side and future proof yourself will want to go > > ahead > > > and set the following in your ceph.conf file, and then issue a > > restart > > > to your RGW instances. > > > > > > rgw_dynamic_resharding = false > > > > > > There are a number of issues with dynamic resharding, multisite rgw > > > problems being just one of them. However I thought it was disabled > > > automatically when multisite rgw is used (but I will have to double > > > check the code on that). What version of Ceph did you initially > > > install the cluster with? Prior to v12.2.2 this feature was > > enabled by > > > default for all rgw use cases. > > > > > > Thanks, > > > > > > > > > ------------------------------------------------------------------------ > > > > > > *From:*Christian Rice <cr...@pandora.com <mailto:cr...@pandora.com > >> > > > *Sent:* Tuesday, March 5, 2019 2:07 PM > > > *To:* Matthew H; ceph-users > > > *Subject:* Re: radosgw sync falling behind regularly > > > > > > Matthew, first of all, let me say we very much appreciate your > help! > > > > > > So I don’t think we turned dynamic resharding on, nor did we > > manually > > > reshard buckets. Seems like it defaults to on for luminous but the > > > mimic docs say it’s not supported in multisite. So do we need to > > > disable it manually via tell and ceph.conf? > > > > > > Also, after running the command you suggested, all the stale > > instances > > > are gone…these from my examples were in output: > > > > > > "bucket_instance": > > > > > > "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.303", > > > > > > "bucket_instance": > > > > > > "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299", > > > > > > "bucket_instance": > > > > > > "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.301", > > > > > > Though we still get lots of log messages like so in rgw: > > > > > > 2019-03-05 11:01:09.526120 7f64120ae700 0 ERROR: failed to get > > bucket > > > instance info for > > > > > > > .bucket.meta.sysad_task:sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299 > > > > > > 2019-03-05 11:01:09.528664 7f63e5016700 1 civetweb: > > 0x55976f1c2000: > > > 172.17.136.17 - - [05/Mar/2019:10:54:06 -0800] "GET > > > > > > > /admin/metadata/bucket.instance/sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299?key=sysad_task%2Fsysad-task%3A1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299&rgwx-zonegroup=de6af748-1a2f-44a1-9d44-30799cf1313e > > > > > HTTP/1.1" 404 0 - - > > > > > > 2019-03-05 11:01:09.529648 7f64130b0700 0 meta sync: ERROR: can't > > > remove key: > > > > > > > bucket.instance:sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299 > > > > > ret=-2 > > > > > > 2019-03-05 11:01:09.530324 7f64138b1700 0 ERROR: failed to get > > bucket > > > instance info for > > > > > > > .bucket.meta.sysad_task:sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299 > > > > > > 2019-03-05 11:01:09.530345 7f6405094700 0 data sync: ERROR: > > failed to > > > retrieve bucket info for > > > > > > bucket=sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299 > > > > > > 2019-03-05 11:01:09.531774 7f6405094700 0 data sync: WARNING: > > > skipping data log entry for missing bucket > > > > sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299 > > > > > > 2019-03-05 11:01:09.571680 7f6405094700 0 data sync: ERROR: > > init sync > > > on > > > > > sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302 > > > failed, retcode=-2 > > > > > > 2019-03-05 11:01:09.573179 7f6405094700 0 data sync: WARNING: > > > skipping data log entry for missing bucket > > > > sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302 > > > > > > 2019-03-05 11:01:13.504308 7f63f903e700 1 civetweb: > > 0x55976f0f2000: > > > 10.105.18.20 - - [05/Mar/2019:11:00:57 -0800] "GET > > > > > > > /admin/metadata/bucket.instance/sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299?key=sysad_task%2Fsysad-task%3A1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299&rgwx-zonegroup=de6af748-1a2f-44a1-9d44-30799cf1313e > > > > > HTTP/1.1" 404 0 - - > > > > > > *From: *Matthew H <matthew.he...@hotmail.com > > <mailto:matthew.he...@hotmail.com>> > > > *Date: *Tuesday, March 5, 2019 at 10:03 AM > > > *To: *Christian Rice <cr...@pandora.com > > <mailto:cr...@pandora.com>>, ceph-users > > > <ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>> > > > *Subject: *Re: radosgw sync falling behind regularly > > > > > > Hi Christian, > > > > > > You have stale bucket instances that need to be clean up, which is > > > what 'radosgw-admin reshard stale-instances list' is showing > > you. Have > > > you or were you manually resharding your buckets? The errors you > > are > > > seeing in the logs are related to these stale instances being kept > > > around. > > > > > > In v12.2.11 this command along with 'radosgw-admin reshard > > > stale-instance rm' was introduced [1]. > > > > > > Hopefully this helps. > > > > > > [1] > > > > > > https://ceph.com/releases/v12-2-11-luminous-released/ > > > > > < > https://urldefense.proofpoint.com/v2/url?u=https-3A__ceph.com_releases_v12-2D2-2D11-2Dluminous-2Dreleased_&d=DwMF-g&c=gFTBenQ7Vj71sUi1A4CkFnmPzqwDo07QsHw-JRepxyw&r=NE1NbWtVhgG-K7YvLdoLZigfFx8zGPwOGk6HWpYK04I&m=vdtYIn6lEKaWD9wW297aHjQLpmQdHZrOVpOhmCBqkqo&s=nGCpS4p5jnaSpPUFlziSi3Y3pFijhVDy6e3867jA9BE&e= > > > > > > > > /"There have been fixes to RGW dynamic and manual resharding, > > which no > > > longer > > > leaves behind stale bucket instances to be removed manually. For > > > finding and > > > cleaning up older instances from a reshard a radosgw-admin > > command reshard > > > stale-instances list and reshard stale-instances rm should do the > > > necessary > > > cleanup."/ > > > > > > > > > ------------------------------------------------------------------------ > > > > > > *From:*Christian Rice <cr...@pandora.com <mailto:cr...@pandora.com > >> > > > *Sent:* Tuesday, March 5, 2019 11:34 AM > > > *To:* Matthew H; ceph-users > > > *Subject:* Re: radosgw sync falling behind regularly > > > > > > The output of “radosgw-admin reshard stale-instances list” shows > > 242 > > > entries, which might embed too much proprietary info for me to > > list, > > > but here’s a tiny sample: > > > > > > > > > "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.303", > > > > > > > > > "sysad_task/sysad_task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.281", > > > > > > > > > "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299", > > > > > > > > > "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.301", > > > > > > Some of appear repeatedly in the radosgw error logs like so: > > > > > > 2019-03-05 08:13:08.929206 7f6405094700 0 data sync: ERROR: > > init sync > > > on > > > > > sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302 > > > failed, retcode=-2 > > > > > > 2019-03-05 08:13:08.930581 7f6405094700 0 data sync: WARNING: > > > skipping data log entry for missing bucket > > > > sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302 > > > > > > 2019-03-05 08:13:08.972053 7f6405094700 0 data sync: ERROR: > > init sync > > > on > > > > > sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299 > > > failed, retcode=-2 > > > > > > 2019-03-05 08:13:08.973442 7f6405094700 0 data sync: WARNING: > > > skipping data log entry for missing bucket > > > > sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299 > > > > > > 2019-03-05 08:13:19.528295 7f6406897700 0 data sync: ERROR: > > init sync > > > on > > > > > sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299 > > > failed, retcode=-2 > > > > > > Notably, “Sync is disabled for bucket sysad-task.” We use “bucket > > > sync disable” A LOT. It’s the only way we’ve been able to use > > > multisite and a single namespace and not replicate things that are > > > unneeded to every zone. Perhaps there’s a bug in the > > implementation > > > that’s tripping us up now, with the new sync multisite sync code > > from > > > 12.2.9 onward? > > > > > > What might we do with stale bucket instances, then? > > > > > > Of note, our master zone endpoint, which was timing out health > > checks > > > most of the day after the upgrade (was running but overworked by > > > cluster confusion, so we couldn’t create new buckets or do user > > ops), > > > has returned to availability late last night. There’s a lot of > > data > > > to look at, but in my estimation, due to lack of user complaints > > (or > > > their unawareness of specific issues), it seems the zones are > > > nominally available, even with all the errors and warnings being > > > logged. We’ve tested simple zone replication by creating a few > > files > > > in one zone and seeing them show up elsewhere… > > > > > > here’s “period get” output: > > > > > > sv5-ceph-rgw1 > > > > > > { > > > > > > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53", > > > > > > "epoch": 16, > > > > > > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb", > > > > > > "sync_status": [], > > > > > > "period_map": { > > > > > > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53", > > > > > > "zonegroups": [ > > > > > > { > > > > > > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e", > > > > > > "name": "us", > > > > > > "api_name": "us", > > > > > > "is_master": "true", > > > > > > "endpoints": [ > > > > > > "http://sv5-ceph-rgw1.savagebeast.com:8080" > > > > > > ], > > > > > > "hostnames": [], > > > > > > "hostnames_s3website": [], > > > > > > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "zones": [ > > > > > > { > > > > > > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839", > > > > > > "name": "dc11-prod", > > > > > > "endpoints": [ > > > > > > "http://dc11-ceph-rgw1:8080" > > > > > > ], > > > > > > "log_meta": "false", > > > > > > "log_data": "true", > > > > > > "bucket_index_max_shards": 0, > > > > > > "read_only": "false", > > > > > > "tier_type": "", > > > > > > "sync_from_all": "true", > > > > > > "sync_from": [] > > > > > > }, > > > > > > { > > > > > > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "name": "sv5-corp", > > > > > > "endpoints": [ > > > > > > "http://sv5-ceph-rgw1.savagebeast.com:8080" > > > > > > ], > > > > > > "log_meta": "false", > > > > > > "log_data": "true", > > > > > > "bucket_index_max_shards": 0, > > > > > > "read_only": "false", > > > > > > "tier_type": "", > > > > > > "sync_from_all": "true", > > > > > > "sync_from": [] > > > > > > }, > > > > > > { > > > > > > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8", > > > > > > "name": "sv3-prod", > > > > > > "endpoints": [ > > > > > > "http://sv3-ceph-rgw1:8080" > > > > > > ], > > > > > > "log_meta": "false", > > > > > > "log_data": "true", > > > > > > "bucket_index_max_shards": 0, > > > > > > "read_only": "false", > > > > > > "tier_type": "", > > > > > > "sync_from_all": "true", > > > > > > "sync_from": [] > > > > > > } > > > > > > ], > > > > > > "placement_targets": [ > > > > > > { > > > > > > "name": "default-placement", > > > > > > "tags": [] > > > > > > } > > > > > > ], > > > > > > "default_placement": "default-placement", > > > > > > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd" > > > > > > } > > > > > > ], > > > > > > "short_zone_ids": [ > > > > > > { > > > > > > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839", > > > > > > "val": 1720993486 > > > > > > }, > > > > > > { > > > > > > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "val": 2301637458 > > > > > > }, > > > > > > { > > > > > > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8", > > > > > > "val": 1449486239 > > > > > > } > > > > > > ] > > > > > > }, > > > > > > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e", > > > > > > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "period_config": { > > > > > > "bucket_quota": { > > > > > > "enabled": false, > > > > > > "check_on_raw": false, > > > > > > "max_size": -1, > > > > > > "max_size_kb": 0, > > > > > > "max_objects": -1 > > > > > > }, > > > > > > "user_quota": { > > > > > > "enabled": false, > > > > > > "check_on_raw": false, > > > > > > "max_size": -1, > > > > > > "max_size_kb": 0, > > > > > > "max_objects": -1 > > > > > > } > > > > > > }, > > > > > > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd", > > > > > > "realm_name": "savagebucket", > > > > > > "realm_epoch": 2 > > > > > > } > > > > > > sv3-ceph-rgw1 > > > > > > { > > > > > > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53", > > > > > > "epoch": 16, > > > > > > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb", > > > > > > "sync_status": [], > > > > > > "period_map": { > > > > > > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53", > > > > > > "zonegroups": [ > > > > > > { > > > > > > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e", > > > > > > "name": "us", > > > > > > "api_name": "us", > > > > > > "is_master": "true", > > > > > > "endpoints": [ > > > > > > "http://sv5-ceph-rgw1.savagebeast.com:8080" > > > > > > ], > > > > > > "hostnames": [], > > > > > > "hostnames_s3website": [], > > > > > > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "zones": [ > > > > > > { > > > > > > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839", > > > > > > "name": "dc11-prod", > > > > > > "endpoints": [ > > > > > > "http://dc11-ceph-rgw1:8080" > > > > > > ], > > > > > > "log_meta": "false", > > > > > > "log_data": "true", > > > > > > "bucket_index_max_shards": 0, > > > > > > "read_only": "false", > > > > > > "tier_type": "", > > > > > > "sync_from_all": "true", > > > > > > "sync_from": [] > > > > > > }, > > > > > > { > > > > > > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "name": "sv5-corp", > > > > > > "endpoints": [ > > > > > > "http://sv5-ceph-rgw1.savagebeast.com:8080" > > > > > > ], > > > > > > "log_meta": "false", > > > > > > "log_data": "true", > > > > > > "bucket_index_max_shards": 0, > > > > > > "read_only": "false", > > > > > > "tier_type": "", > > > > > > "sync_from_all": "true", > > > > > > "sync_from": [] > > > > > > }, > > > > > > { > > > > > > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8", > > > > > > "name": "sv3-prod", > > > > > > "endpoints": [ > > > > > > "http://sv3-ceph-rgw1:8080" > > > > > > ], > > > > > > "log_meta": "false", > > > > > > "log_data": "true", > > > > > > "bucket_index_max_shards": 0, > > > > > > "read_only": "false", > > > > > > "tier_type": "", > > > > > > "sync_from_all": "true", > > > > > > "sync_from": [] > > > > > > } > > > > > > ], > > > > > > "placement_targets": [ > > > > > > { > > > > > > "name": "default-placement", > > > > > > "tags": [] > > > > > > } > > > > > > ], > > > > > > "default_placement": "default-placement", > > > > > > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd" > > > > > > } > > > > > > ], > > > > > > "short_zone_ids": [ > > > > > > { > > > > > > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839", > > > > > > "val": 1720993486 > > > > > > }, > > > > > > { > > > > > > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "val": 2301637458 > > > > > > }, > > > > > > { > > > > > > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8", > > > > > > "val": 1449486239 > > > > > > } > > > > > > ] > > > > > > }, > > > > > > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e", > > > > > > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "period_config": { > > > > > > "bucket_quota": { > > > > > > "enabled": false, > > > > > > "check_on_raw": false, > > > > > > "max_size": -1, > > > > > > "max_size_kb": 0, > > > > > > "max_objects": -1 > > > > > > }, > > > > > > "user_quota": { > > > > > > "enabled": false, > > > > > > "check_on_raw": false, > > > > > > "max_size": -1, > > > > > > "max_size_kb": 0, > > > > > > "max_objects": -1 > > > > > > } > > > > > > }, > > > > > > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd", > > > > > > "realm_name": "savagebucket", > > > > > > "realm_epoch": 2 > > > > > > } > > > > > > dc11-ceph-rgw1 > > > > > > { > > > > > > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53", > > > > > > "epoch": 16, > > > > > > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb", > > > > > > "sync_status": [], > > > > > > "period_map": { > > > > > > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53", > > > > > > "zonegroups": [ > > > > > > { > > > > > > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e", > > > > > > "name": "us", > > > > > > "api_name": "us", > > > > > > "is_master": "true", > > > > > > "endpoints": [ > > > > > > "http://sv5-ceph-rgw1.savagebeast.com:8080" > > > > > > ], > > > > > > "hostnames": [], > > > > > > "hostnames_s3website": [], > > > > > > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "zones": [ > > > > > > { > > > > > > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839", > > > > > > "name": "dc11-prod", > > > > > > "endpoints": [ > > > > > > "http://dc11-ceph-rgw1:8080" > > > > > > ], > > > > > > "log_meta": "false", > > > > > > "log_data": "true", > > > > > > "bucket_index_max_shards": 0, > > > > > > "read_only": "false", > > > > > > "tier_type": "", > > > > > > "sync_from_all": "true", > > > > > > "sync_from": [] > > > > > > }, > > > > > > { > > > > > > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "name": "sv5-corp", > > > > > > "endpoints": [ > > > > > > "http://sv5-ceph-rgw1.savagebeast.com:8080" > > > > > > ], > > > > > > "log_meta": "false", > > > > > > "log_data": "true", > > > > > > "bucket_index_max_shards": 0, > > > > > > "read_only": "false", > > > > > > "tier_type": "", > > > > > > "sync_from_all": "true", > > > > > > "sync_from": [] > > > > > > }, > > > > > > { > > > > > > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8", > > > > > > "name": "sv3-prod", > > > > > > "endpoints": [ > > > > > > "http://sv3-ceph-rgw1:8080" > > > > > > ], > > > > > > "log_meta": "false", > > > > > > "log_data": "true", > > > > > > "bucket_index_max_shards": 0, > > > > > > "read_only": "false", > > > > > > "tier_type": "", > > > > > > "sync_from_all": "true", > > > > > > "sync_from": [] > > > > > > } > > > > > > ], > > > > > > "placement_targets": [ > > > > > > { > > > > > > "name": "default-placement", > > > > > > "tags": [] > > > > > > } > > > > > > ], > > > > > > "default_placement": "default-placement", > > > > > > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd" > > > > > > } > > > > > > ], > > > > > > "short_zone_ids": [ > > > > > > { > > > > > > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839", > > > > > > "val": 1720993486 > > > > > > }, > > > > > > { > > > > > > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "val": 2301637458 > > > > > > }, > > > > > > { > > > > > > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8", > > > > > > "val": 1449486239 > > > > > > } > > > > > > ] > > > > > > }, > > > > > > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e", > > > > > > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04", > > > > > > "period_config": { > > > > > > "bucket_quota": { > > > > > > "enabled": false, > > > > > > "check_on_raw": false, > > > > > > "max_size": -1, > > > > > > "max_size_kb": 0, > > > > > > "max_objects": -1 > > > > > > }, > > > > > > "user_quota": { > > > > > > "enabled": false, > > > > > > "check_on_raw": false, > > > > > > "max_size": -1, > > > > > > "max_size_kb": 0, > > > > > > "max_objects": -1 > > > > > > } > > > > > > }, > > > > > > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd", > > > > > > "realm_name": "savagebucket", > > > > > > "realm_epoch": 2 > > > > > > } > > > > > > *From: *Matthew H <matthew.he...@hotmail.com > > <mailto:matthew.he...@hotmail.com>> > > > *Date: *Tuesday, March 5, 2019 at 4:31 AM > > > *To: *Christian Rice <cr...@pandora.com > > <mailto:cr...@pandora.com>>, ceph-users > > > <ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>> > > > *Subject: *Re: radosgw sync falling behind regularly > > > > > > Hi Christian, > > > > > > You haven't resharded any of your buckets have you? You can run > > the > > > command below in v12.2.11 to list stale bucket instances. > > > > > > radosgw-admin reshard stale-instances list > > > > > > Can you also send the output from the following command on each > rgw? > > > > > > radosgw-admin period get > > > > > > > > > _______________________________________________ > > > ceph-users mailing list > > > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com