I was able to answer my own question. For future interested parties, I
initiated a deep scrub on the placement group, which cleared the error.

On Tue, Jul 30, 2019 at 1:48 PM Brett Chancellor <[email protected]>
wrote:

> I was able to remove the meta objects, but the cluster is still in WARN
> state
> HEALTH_WARN 1 large omap objects
> LARGE_OMAP_OBJECTS 1 large omap objects
>     1 large objects found in pool 'us-prd-1.rgw.log'
>     Search the cluster log for 'Large omap object found' for more details.
>
> How do I go about clearing it out? I don't see any other references to
> large omap in any of the logs.  I've tried restarted the mgr's, the
> monitors, and even the osd that reported the issue.
>
> -Brett
>
> On Thu, Jul 25, 2019 at 2:55 PM Brett Chancellor <
> [email protected]> wrote:
>
>> 14.2.1
>> Thanks, I'll try that.
>>
>> On Thu, Jul 25, 2019 at 2:54 PM Casey Bodley <[email protected]> wrote:
>>
>>> What ceph version is this cluster running? Luminous or later should not
>>> be writing any new meta.log entries when it detects a single-zone
>>> configuration.
>>>
>>> I'd recommend editing your zonegroup configuration (via 'radosgw-admin
>>> zonegroup get' and 'put') to set both log_meta and log_data to false,
>>> then commit the change with 'radosgw-admin period update --commit'.
>>>
>>> You can then delete any meta.log.* and data_log.* objects from your log
>>> pool using the rados tool.
>>>
>>> On 7/25/19 2:30 PM, Brett Chancellor wrote:
>>> > Casey,
>>> >   These clusters were setup with the intention of one day doing multi
>>> > site replication. That has never happened. The cluster has a single
>>> > realm, which contains a single zonegroup, and that zonegroup contains
>>> > a single zone.
>>> >
>>> > -Brett
>>> >
>>> > On Thu, Jul 25, 2019 at 2:16 PM Casey Bodley <[email protected]
>>> > <mailto:[email protected]>> wrote:
>>> >
>>> >     Hi Brett,
>>> >
>>> >     These meta.log objects store the replication logs for metadata
>>> >     sync in
>>> >     multisite. Log entries are trimmed automatically once all other
>>> zones
>>> >     have processed them. Can you verify that all zones in the multisite
>>> >     configuration are reachable and syncing? Does 'radosgw-admin sync
>>> >     status' on any zone show that it's stuck behind on metadata sync?
>>> >     That
>>> >     would prevent these logs from being trimmed and result in these
>>> large
>>> >     omap warnings.
>>> >
>>> >     On 7/25/19 1:59 PM, Brett Chancellor wrote:
>>> >     > I'm having an issue similar to
>>> >     >
>>> >
>>> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2019-March/033611.html
>>>  .
>>> >
>>> >     > I don't see where any solution was proposed.
>>> >     >
>>> >     > $ ceph health detail
>>> >     > HEALTH_WARN 1 large omap objects
>>> >     > LARGE_OMAP_OBJECTS 1 large omap objects
>>> >     >     1 large objects found in pool 'us-prd-1.rgw.log'
>>> >     >     Search the cluster log for 'Large omap object found' for
>>> >     more details.
>>> >     >
>>> >     > $ grep "Large omap object" /var/log/ceph/ceph.log
>>> >     > 2019-07-25 14:58:21.758321 osd.3 (osd.3) 15 : cluster [WRN]
>>> >     Large omap
>>> >     > object found. Object:
>>> >     >
>>> 51:61eb35fe:::meta.log.e557cf47-46df-4b45-988e-9a94c5004a2e.19:head
>>> >     > Key count: 3382154 Size (bytes): 611384043
>>> >     >
>>> >     > $ rados -p us-prd-1.rgw.log listomapkeys
>>> >     > meta.log.e557cf47-46df-4b45-988e-9a94c5004a2e.19 |wc -l
>>> >     > 3382154
>>> >     >
>>> >     > $ rados -p us-prd-1.rgw.log listomapvals
>>> >     > meta.log.e557cf47-46df-4b45-988e-9a94c5004a2e.19
>>> >     > This returns entries from almost every bucket, across multiple
>>> >     > tenants. Several of the entries are from buckets that no longer
>>> >     exist
>>> >     > on the system.
>>> >     >
>>> >     > $ ceph df |egrep 'OBJECTS|.rgw.log'
>>> >     >     POOL        ID      STORED      OBJECTS     USED    %USED MAX
>>> >     > AVAIL
>>> >     >     us-prd-1.rgw.log                 51     758 MiB 228   758 MiB
>>> >     >       0       102 TiB
>>> >     >
>>> >     > Thanks,
>>> >     >
>>> >     > -Brett
>>> >     >
>>> >     > _______________________________________________
>>> >     > ceph-users mailing list
>>> >     > [email protected] <mailto:[email protected]>
>>> >     > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> >     _______________________________________________
>>> >     ceph-users mailing list
>>> >     [email protected] <mailto:[email protected]>
>>> >     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> >
>>>
>>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to