14.2.1 Thanks, I'll try that. On Thu, Jul 25, 2019 at 2:54 PM Casey Bodley <cbod...@redhat.com> wrote:
> What ceph version is this cluster running? Luminous or later should not > be writing any new meta.log entries when it detects a single-zone > configuration. > > I'd recommend editing your zonegroup configuration (via 'radosgw-admin > zonegroup get' and 'put') to set both log_meta and log_data to false, > then commit the change with 'radosgw-admin period update --commit'. > > You can then delete any meta.log.* and data_log.* objects from your log > pool using the rados tool. > > On 7/25/19 2:30 PM, Brett Chancellor wrote: > > Casey, > > These clusters were setup with the intention of one day doing multi > > site replication. That has never happened. The cluster has a single > > realm, which contains a single zonegroup, and that zonegroup contains > > a single zone. > > > > -Brett > > > > On Thu, Jul 25, 2019 at 2:16 PM Casey Bodley <cbod...@redhat.com > > <mailto:cbod...@redhat.com>> wrote: > > > > Hi Brett, > > > > These meta.log objects store the replication logs for metadata > > sync in > > multisite. Log entries are trimmed automatically once all other zones > > have processed them. Can you verify that all zones in the multisite > > configuration are reachable and syncing? Does 'radosgw-admin sync > > status' on any zone show that it's stuck behind on metadata sync? > > That > > would prevent these logs from being trimmed and result in these large > > omap warnings. > > > > On 7/25/19 1:59 PM, Brett Chancellor wrote: > > > I'm having an issue similar to > > > > > > http://lists.ceph.com/pipermail/ceph-users-ceph.com/2019-March/033611.html > . > > > > > I don't see where any solution was proposed. > > > > > > $ ceph health detail > > > HEALTH_WARN 1 large omap objects > > > LARGE_OMAP_OBJECTS 1 large omap objects > > > 1 large objects found in pool 'us-prd-1.rgw.log' > > > Search the cluster log for 'Large omap object found' for > > more details. > > > > > > $ grep "Large omap object" /var/log/ceph/ceph.log > > > 2019-07-25 14:58:21.758321 osd.3 (osd.3) 15 : cluster [WRN] > > Large omap > > > object found. Object: > > > 51:61eb35fe:::meta.log.e557cf47-46df-4b45-988e-9a94c5004a2e.19:head > > > Key count: 3382154 Size (bytes): 611384043 > > > > > > $ rados -p us-prd-1.rgw.log listomapkeys > > > meta.log.e557cf47-46df-4b45-988e-9a94c5004a2e.19 |wc -l > > > 3382154 > > > > > > $ rados -p us-prd-1.rgw.log listomapvals > > > meta.log.e557cf47-46df-4b45-988e-9a94c5004a2e.19 > > > This returns entries from almost every bucket, across multiple > > > tenants. Several of the entries are from buckets that no longer > > exist > > > on the system. > > > > > > $ ceph df |egrep 'OBJECTS|.rgw.log' > > > POOL ID STORED OBJECTS USED %USED MAX > > > AVAIL > > > us-prd-1.rgw.log 51 758 MiB 228 758 MiB > > > 0 102 TiB > > > > > > Thanks, > > > > > > -Brett > > > > > > _______________________________________________ > > > ceph-users mailing list > > > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com