Hi Danish, 

The "unable to find head object data pool for..." could be an incorrect warning 
since it pops out for 'most of the objects'. [1] 

Regarding the only object named 'cursor.png' that fails to sync, one thing you 
could try (since you can't delete it with an s3 client) is to rewrite it with 
an s3 client (copy) and then retry the delete. 
If it fails with an s3 client, you could try with 'radosgw-admin object put' 
and/or 'radosgw-admin object rm'. 

It that still fails, then here's what you can do to (at least deal with the 
bucket synchronization issue) to remove this object from the index: 

1/ Set bucket_name, index_pool_name, and bucket_id (jq command is required) 

$ bucket_name="bucket-test" 
$ index_pool_name=".rgw.sme.index" 
$ bucket_id=$(radosgw-admin bucket stats --bucket=${bucket_name} | jq -r .id) 

2/ Retrieve all index shards along with their omap keys 

$ mkdir "$bucket_id" 
$ for i in $(rados -p $index_pool_name ls | grep "$bucket_id"); do echo $i ; 
rados -p $index_pool_name listomapkeys $i > "${bucket_id}/${i}" ; done 

3/ Identify in which shard the 'cursor.png' object listed (be sure to identify 
the right object. You may have several WP using the same image...) 

$ grep 'cursor.png' ${bucket_id}/.dir.${bucket_id}* | sed -e 
"s/^${bucket_id}\///g" > remove_from_index.txt 

4/ Make sure the content of remove_from_index.txt file only has one line 
corresponding to the object you want to remove from the index : 

$ cat remove_from_index.txt 
.dir.0f448533-3c6c-4cb8-bde9-c9763ac17751.738183.1.6:48/wp-content/plugins/plugins/yellow-pencil-visual-theme-customizer/images/cursor.png
 

5/ Remove the object from the index shard 

while IFS=':' read -r object key ; do echo "Removing Key ${key}" ; rados -p 
${index_pool_name} rmomapkey "${object}" "${key}" ; done < 
remove_from_index.txt 

Restart both RGWs and check the sync state again. 

Next, you might want to check for inconsistencies between the index and the 
actual data. You could use the rgw-orphan-list script for this [2]. And of 
course, upgrade your cluster. 

Hope this helps, 

Regards, 
Frédéric. 

[1] [ https://bugzilla.redhat.com/show_bug.cgi?id=2126787 | 
https://bugzilla.redhat.com/show_bug.cgi?id=2126787 ] 
[2] [ 
https://www.ibm.com/docs/en/storage-ceph/8.0?topic=management-finding-orphan-leaky-objects
 | 
https://www.ibm.com/docs/en/storage-ceph/8.0?topic=management-finding-orphan-leaky-objects
 ] 

----- Le 26 Mar 25, à 6:12, Danish Khan <danish52....@gmail.com> a écrit : 



Dear Frédéric, 

Unfortunately, I am still using Octopus version and these commands are showing 
unrecognized. 

Versioning is also not enabled on the bucket. 
I tried running : 
radosgw-admin bucket check --bucket=<bucket> --fix 

which run for few minutes giving lot of output, which contained below lines for 
most of the objects: 
WARNING: unable to find head object data pool for 
"<bucket>:wp-content/uploads/sites/74/2025/03/mutation-house-no-629.pdf", not 
updating version pool/epoch 

Is this issue fixable in octopus or should I plan to upgrade ceph cluster till 
Quincy version? 

Regards, 
Danish 


On Wed, Mar 26, 2025 at 2:41 AM Frédéric Nass < [ 
mailto:frederic.n...@univ-lorraine.fr | frederic.n...@univ-lorraine.fr ] > 
wrote: 

BQ_BEGIN

Hi Danish, 

Can you specify the version of Ceph used and whether versioning is enabled on 
this bucket? 

There are 2 ways to clean up orphan entries in a bucket index that I'm aware of 
: 

- One (the preferable way) is to rely on radosgw-admin command to check and 
hopefully fix the issue, cleaning up the index from orphan entries or even 
rebuilding the index entirely if necessary. 

There's been new radosgw-admin commands coded recently [1] to cleanup leftover 
OLH index entries and unlinked instance objects within versioned buckets. 

If this bucket is versioned, I would advise you try and run the new check / fix 
commands mentioned in this [2] release note : 
radosgw-admin bucket check unlinked [--fix] 

radosgw-admin bucket check olh [--fix] 
- Another one (as a second chance) is to act at the rados layer, identifying in 
which shard the orphan index entry is listed (listomapkeys) and remove it from 
the specified shard (rmomapkey). I could elaborate on that later if needed. 

Regards, 
Frédéric. 

[1] [ https://tracker.ceph.com/issues/62075 | 
https://tracker.ceph.com/issues/62075 ] 
[2] [ https://ceph.io/en/news/blog/2023/v18-2-1-reef-released/ | 
https://ceph.io/en/news/blog/2023/v18-2-1-reef-released/ ] 




De : Danish Khan < [ mailto:danish52....@gmail.com | danish52....@gmail.com ] > 
Envoyé : mardi 25 mars 2025 17:16 
À : Frédéric Nass 
Cc: ceph-users 
Objet : Re: [ceph-users] ceph-ansible LARGE OMAP in RGW pool 

Hi Frédéric, 

Thank you for replying. 

I followed the steps mentioned in [ https://tracker.ceph.com/issues/62845 | 
https://tracker.ceph.com/issues/62845 ] and was able to trim all the errors. 

Everything seemed to be working fine until the same error appeared again. 

I am still assuming the main culprit of this issue is one missing object and 
all the errors are showing this object only. 

I am able to list this object using s3cmd tool but I am unable to perform any 
action on this object, I am even unable to delete it, overwrite it or get it. 

I tried stopping all RGWs one by one and even tried after stopping all the RGWS 
but recovery is still not getting completed. 

And the LARGE OMAP is now only increasing. 

Is there a way I can delete it from index or from ceph end directly from pool 
so that it don't try to recover it? 

Regards, 
Danish 



On Tue, Mar 25, 2025 at 11:29 AM Frédéric Nass < [ 
mailto:frederic.n...@univ-lorraine.fr | frederic.n...@univ-lorraine.fr ] > 
wrote: 

BQ_BEGIN

Hi Danish, 

While reviewing the backports for upcoming v18.2.5, I came across this [1]. 
Could be your issue. 

Can you try the suggested workaround (--marker=9) and report back? 

Regards, 
Frédéric. 

[1] [ https://tracker.ceph.com/issues/62845 | 
https://tracker.ceph.com/issues/62845 ] 


De : Danish Khan < [ mailto:danish52....@gmail.com | danish52....@gmail.com ] > 
Envoyé : vendredi 14 mars 2025 23:11 
À : Frédéric Nass 
Cc: ceph-users 
Objet : Re: [ceph-users] ceph-ansible LARGE OMAP in RGW pool 

Dear Frédéric, 

1/ Identify the shards with the most sync errors log entries: 

I have identified the shard which is causing the issue is shard 31, but almost 
all the error shows only one object of a bucket. And the object exists in the 
master zone. but I'm not sure why the replication site is unable to sync it. 

2/ For each shard, list every sync error log entry along with their ids: 

radosgw-admin sync error list --shard-id=X 

The output of this command shows same shard and same objects mostly (shard 31 
and object 
/plugins/plugins/yellow-pencil-visual-theme-customizer/images/cursor.png) 

3/ Remove them **except the last one** with: 

radosgw-admin sync error trim --shard-id=X --marker=1_1682101321.201434_8669.1 
Trimming did remove a few entries from the error log. But still there are many 
error logs for the same object which I am unable to trim. 

Now the trim command is executing successfully but not doing anything. 

I am still getting error about the object which is not syncing in radosgw log: 

2025-03-15T03:05:48.060+0530 7fee2affd700 0 
RGW-SYNC:data:sync:shard[80]:entry[mbackup:70134e66-872072ee2d32.2205852207.1:48]:bucket_sync_sources[target=:[]):source_bucket=:[]):source_zone=872072ee2d32]:bucket[mbackup:70134e66-872072ee2d32.2205852207.1:48<-mod-backup:70134e66-872072ee2d32.2205852207.1:48]:full_sync[mod-backup:70134e66-872072ee2d32.2205852207.1:48]:entry[wp-content/plugins/plugins/yellow-pencil-visual-theme-customizer/images/cursor.png]:
 ERROR: failed to sync object: 
mbackup:70134e66-872072ee2d32.2205852207.1:48/wp-content/plugins/plugins/yellow-pencil-visual-theme-customizer/images/cursor.png
 

I am getting this error from appox two months, And if I remember correctly, we 
are getting LARGE OMAP warning from then only. 

I will try to delete this object from the Master zone on Monday and will see if 
this fixes the issue. 

Do you have any other suggestions on this, which I should consider? 

Regards, 
Danish 






On Thu, Mar 13, 2025 at 6:07 PM Frédéric Nass < [ 
mailto:frederic.n...@univ-lorraine.fr | frederic.n...@univ-lorraine.fr ] > 
wrote: 

BQ_BEGIN
Hi Danish, 

Can you access this KB article [1]? A free developer account should allow you 
to. 

It pretty much describes what you're facing and suggests to trim the sync error 
log of recovering shards. Actually, every log entry **except the last one**. 

1/ Identify the shards with the most sync errors log entries: 

radosgw-admin sync error list --max-entries=1000000 | grep shard_id | sort -n | 
uniq -c | sort -h 

2/ For each shard, list every sync error log entry along with their ids: 

radosgw-admin sync error list --shard-id=X 

3/ Remove them **except the last one** with: 

radosgw-admin sync error trim --shard-id=X --marker=1_1682101321.201434_8669.1 

the --marker above being the log entry id. 

Are the replication threads running on the same RGWs that S3 clients are using? 

If so, using dedicated RGWs for the sync job might help you avoid 
non-recovering shards in the future, as described in Matthew's post [2] 

Regards, 
Frédéric. 

[1] [ https://access.redhat.com/solutions/7023912 | 
https://access.redhat.com/solutions/7023912 ] 
[2] [ https://www.spinics.net/lists/ceph-users/msg83988.html | 
https://www.spinics.net/lists/ceph-users/msg83988.html ] 

----- Le 12 Mar 25, à 11:15, Danish Khan [ mailto:danish52....@gmail.com | 
danish52....@gmail.com ] a écrit : 

> Dear All, 
> 
> My ceph cluster is giving Large OMAP warning from approx 2-3 Months. I 
> tried a few things like : 
> *Deep scrub of PGs* 
> *Compact OSDs* 
> *Trim log* 
> But these didn't work out. 
> 
> I guess the main issue is that 4 shards in replication site are always 
> recovering from 2-3 months. 
> 
> Any suggestions are highly appreciated. 
> 
> Sync status: 
> root@drhost1:~# radosgw-admin sync status 
> realm e259e0a92 (object-storage) 
> zonegroup 7a8606d2 (staas) 
> zone c8022ad1 (repstaas) 
> metadata sync syncing 
> full sync: 0/64 shards 
> incremental sync: 64/64 shards 
> metadata is caught up with master 
> data sync source: 2072ee2d32 (masterstaas) 
> syncing 
> full sync: 0/128 shards 
> incremental sync: 128/128 shards 
> data is behind on 3 shards 
> behind shards: [7,90,100] 
> oldest incremental change not applied: 
> 2025-03-12T13:14:10.268469+0530 [7] 
> 4 shards are recovering 
> recovering shards: [31,41,55,80] 
> 
> 
> Master site: 
> 1. *root@master1:~# for obj in $(rados ls -p masterstaas.rgw.log); do echo 
> "$(rados listomapkeys -p masterstaas.rgw.log $obj | wc -l) $obj";done | 
> sort -nr | head -10* 
> 1225387 data_log.91 
> 1225065 data_log.86 
> 1224662 data_log.87 
> 1224448 data_log.92 
> 1224018 data_log.89 
> 1222156 data_log.93 
> 1201489 data_log.83 
> 1174125 data_log.90 
> 363498 data_log.84 
> 258709 data_log.6 
> 
> 
> 2. *root@master1:~# for obj in data_log.91 data_log.86 data_log.87 
> data_log.92 data_log.89 data_log.93 data_log.83 data_log.90; do rados stat 
> -p masterstaas.rgw.log $obj; done* 
> masterstaas.rgw.log/data_log.91 mtime 2025-02-24T15:09:25.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.86 mtime 2025-02-24T15:01:25.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.87 mtime 2025-02-24T15:02:25.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.92 mtime 2025-02-24T15:11:01.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.89 mtime 2025-02-24T14:54:55.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.93 mtime 2025-02-24T14:53:25.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.83 mtime 2025-02-24T14:16:21.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.90 mtime 2025-02-24T15:05:25.000000+0530, size 
> 0 
> 
> *3. ceph cluster log :* 
> 2025-02-22T04:18:27.324886+0530 osd.173 (osd.173) 19 : cluster [WRN] Large 
> omap object found. Object: 124:b2ddf551:::data_log.93:head PG: 124.8aafbb4d 
> (124.d) Key count: 1218170 Size (bytes): 297085860 
> 2025-02-22T04:18:28.735886+0530 osd.65 (osd.65) 308 : cluster [WRN] Large 
> omap object found. Object: 124:f2081d70:::data_log.92:head PG: 124.eb8104f 
> (124.f) Key count: 1220420 Size (bytes): 295240028 
> 2025-02-22T04:18:30.668884+0530 mon.master1 (mon.0) 7974038 : cluster [WRN] 
> Health check update: 3 large omap objects (LARGE_OMAP_OBJECTS) 
> 2025-02-22T04:18:31.127585+0530 osd.18 (osd.18) 224 : cluster [WRN] Large 
> omap object found. Object: 124:d1061236:::data_log.86:head PG: 124.6c48608b 
> (124.b) Key count: 1221047 Size (bytes): 295398274 
> 2025-02-22T04:18:33.189561+0530 osd.37 (osd.37) 32665 : cluster [WRN] Large 
> omap object found. Object: 124:9a2e04b7:::data_log.87:head PG: 124.ed207459 
> (124.19) Key count: 1220584 Size (bytes): 295290366 
> 2025-02-22T04:18:35.007117+0530 osd.77 (osd.77) 135 : cluster [WRN] Large 
> omap object found. Object: 124:6b9e929a:::data_log.89:head PG: 124.594979d6 
> (124.16) Key count: 1219913 Size (bytes): 295127816 
> 2025-02-22T04:18:36.189141+0530 mon.master1 (mon.0) 7974039 : cluster [WRN] 
> Health check update: 5 large omap objects (LARGE_OMAP_OBJECTS) 
> 2025-02-22T04:18:36.340247+0530 osd.112 (osd.112) 259 : cluster [WRN] Large 
> omap object found. Object: 124:0958bece:::data_log.83:head PG: 124.737d1a90 
> (124.10) Key count: 1200406 Size (bytes): 290280292 
> 2025-02-22T04:18:38.523766+0530 osd.73 (osd.73) 1064 : cluster [WRN] Large 
> omap object found. Object: 124:fddd971f:::data_log.91:head PG: 124.f8e9bbbf 
> (124.3f) Key count: 1221183 Size (bytes): 295425320 
> 2025-02-22T04:18:42.619926+0530 osd.92 (osd.92) 285 : cluster [WRN] Large 
> omap object found. Object: 124:7dc404fa:::data_log.90:head PG: 124.5f2023be 
> (124.3e) Key count: 1169895 Size (bytes): 283025576 
> 2025-02-22T04:18:44.242655+0530 mon.master1 (mon.0) 7974043 : cluster [WRN] 
> Health check update: 8 large omap objects (LARGE_OMAP_OBJECTS) 
> 
> Replica site: 
> 1. *for obj in $(rados ls -p repstaas.rgw.log); do echo "$(rados 
> listomapkeys -p repstaas.rgw.log $obj | wc -l) $obj";done | sort -nr | head 
> -10* 
> 
> 432850 data_log.91 
> 432384 data_log.87 
> 432323 data_log.93 
> 431783 data_log.86 
> 431510 data_log.92 
> 427959 data_log.89 
> 414522 data_log.90 
> 407571 data_log.83 
> 151015 data_log.84 
> 109790 data_log.4 
> 
> 
> 2. *ceph cluster log:* 
> grep -ir "Large omap object found" /var/log/ceph/ 
> /var/log/ceph/ceph-mon.drhost1.log:2025-03-12T14:49:59.997+0530 
> 7fc4ad544700 0 log_channel(cluster) log [WRN] : Search the cluster log 
> for 'Large omap object found' for more details. 
> /var/log/ceph/ceph.log:2025-03-12T14:49:02.078108+0530 osd.10 (osd.10) 21 : 
> cluster [WRN] Large omap object found. Object: 
> 6:b2ddf551:::data_log.93:head PG: 6.8aafbb4d (6.d) Key count: 432323 Size 
> (bytes): 105505884 
> /var/log/ceph/ceph.log:2025-03-12T14:49:02.389288+0530 osd.48 (osd.48) 37 : 
> cluster [WRN] Large omap object found. Object: 
> 6:d1061236:::data_log.86:head PG: 6.6c48608b (6.b) Key count: 431782 Size 
> (bytes): 104564674 
> /var/log/ceph/ceph.log:2025-03-12T14:49:07.166954+0530 osd.24 (osd.24) 24 : 
> cluster [WRN] Large omap object found. Object: 
> 6:0958bece:::data_log.83:head PG: 6.737d1a90 (6.10) Key count: 407571 Size 
> (bytes): 98635522 
> /var/log/ceph/ceph.log:2025-03-12T14:49:09.100110+0530 osd.63 (osd.63) 5 : 
> cluster [WRN] Large omap object found. Object: 
> 6:9a2e04b7:::data_log.87:head PG: 6.ed207459 (6.19) Key count: 432384 Size 
> (bytes): 104712350 
> /var/log/ceph/ceph.log:2025-03-12T14:49:08.703760+0530 osd.59 (osd.59) 11 : 
> cluster [WRN] Large omap object found. Object: 
> 6:6b9e929a:::data_log.89:head PG: 6.594979d6 (6.16) Key count: 427959 Size 
> (bytes): 103773777 
> /var/log/ceph/ceph.log:2025-03-12T14:49:11.126132+0530 osd.40 (osd.40) 24 : 
> cluster [WRN] Large omap object found. Object: 
> 6:f2081d70:::data_log.92:head PG: 6.eb8104f (6.f) Key count: 431508 Size 
> (bytes): 104520406 
> /var/log/ceph/ceph.log:2025-03-12T14:49:13.799473+0530 osd.43 (osd.43) 61 : 
> cluster [WRN] Large omap object found. Object: 
> 6:fddd971f:::data_log.91:head PG: 6.f8e9bbbf (6.1f) Key count: 432850 Size 
> (bytes): 104418869 
> /var/log/ceph/ceph.log:2025-03-12T14:49:14.398480+0530 osd.3 (osd.3) 55 : 
> cluster [WRN] Large omap object found. Object: 
> 6:7dc404fa:::data_log.90:head PG: 6.5f2023be (6.1e) Key count: 414521 Size 
> (bytes): 100396561 
> /var/log/ceph/ceph.log:2025-03-12T14:50:00.000484+0530 mon.drhost1 (mon.0) 
> 207423 : cluster [WRN] Search the cluster log for 'Large omap object 
> found' for more details. 
> 
> Regards, 
> Danish 
> _______________________________________________ 
> ceph-users mailing list -- [ mailto:ceph-users@ceph.io | ceph-users@ceph.io ] 
> To unsubscribe send an email to [ mailto:ceph-users-le...@ceph.io | 
> ceph-users-le...@ceph.io ] 




BQ_END


De : Danish Khan < [ mailto:danish52....@gmail.com | danish52....@gmail.com ] > 
Envoyé : mardi 25 mars 2025 17:16 
À : Frédéric Nass 
Cc: ceph-users 
Objet : Re: [ceph-users] ceph-ansible LARGE OMAP in RGW pool 


Hi Frédéric, 

Thank you for replying. 

I followed the steps mentioned in [ https://tracker.ceph.com/issues/62845 | 
https://tracker.ceph.com/issues/62845 ] and was able to trim all the errors. 

Everything seemed to be working fine until the same error appeared again. 

I am still assuming the main culprit of this issue is one missing object and 
all the errors are showing this object only. 

I am able to list this object using s3cmd tool but I am unable to perform any 
action on this object, I am even unable to delete it, overwrite it or get it. 

I tried stopping all RGWs one by one and even tried after stopping all the RGWS 
but recovery is still not getting completed. 

And the LARGE OMAP is now only increasing. 

Is there a way I can delete it from index or from ceph end directly from pool 
so that it don't try to recover it? 

Regards, 
Danish 



On Tue, Mar 25, 2025 at 11:29 AM Frédéric Nass < [ 
mailto:frederic.n...@univ-lorraine.fr | frederic.n...@univ-lorraine.fr ] > 
wrote: 

BQ_BEGIN

Hi Danish, 

While reviewing the backports for upcoming v18.2.5, I came across this [1]. 
Could be your issue. 

Can you try the suggested workaround (--marker=9) and report back? 

Regards, 
Frédéric. 

[1] [ https://tracker.ceph.com/issues/62845 | 
https://tracker.ceph.com/issues/62845 ] 


De : Danish Khan < [ mailto:danish52....@gmail.com | danish52....@gmail.com ] > 
Envoyé : vendredi 14 mars 2025 23:11 
À : Frédéric Nass 
Cc: ceph-users 
Objet : Re: [ceph-users] ceph-ansible LARGE OMAP in RGW pool 


Dear Frédéric, 

1/ Identify the shards with the most sync errors log entries: 

I have identified the shard which is causing the issue is shard 31, but almost 
all the error shows only one object of a bucket. And the object exists in the 
master zone. but I'm not sure why the replication site is unable to sync it. 

2/ For each shard, list every sync error log entry along with their ids: 

radosgw-admin sync error list --shard-id=X 

The output of this command shows same shard and same objects mostly (shard 31 
and object 
/plugins/plugins/yellow-pencil-visual-theme-customizer/images/cursor.png) 

3/ Remove them **except the last one** with: 

radosgw-admin sync error trim --shard-id=X --marker=1_1682101321.201434_8669.1 
Trimming did remove a few entries from the error log. But still there are many 
error logs for the same object which I am unable to trim. 

Now the trim command is executing successfully but not doing anything. 

I am still getting error about the object which is not syncing in radosgw log: 

2025-03-15T03:05:48.060+0530 7fee2affd700 0 
RGW-SYNC:data:sync:shard[80]:entry[mbackup:70134e66-872072ee2d32.2205852207.1:48]:bucket_sync_sources[target=:[]):source_bucket=:[]):source_zone=872072ee2d32]:bucket[mbackup:70134e66-872072ee2d32.2205852207.1:48<-mod-backup:70134e66-872072ee2d32.2205852207.1:48]:full_sync[mod-backup:70134e66-872072ee2d32.2205852207.1:48]:entry[wp-content/plugins/plugins/yellow-pencil-visual-theme-customizer/images/cursor.png]:
 ERROR: failed to sync object: 
mbackup:70134e66-872072ee2d32.2205852207.1:48/wp-content/plugins/plugins/yellow-pencil-visual-theme-customizer/images/cursor.png
 

I am getting this error from appox two months, And if I remember correctly, we 
are getting LARGE OMAP warning from then only. 

I will try to delete this object from the Master zone on Monday and will see if 
this fixes the issue. 

Do you have any other suggestions on this, which I should consider? 

Regards, 
Danish 






On Thu, Mar 13, 2025 at 6:07 PM Frédéric Nass < [ 
mailto:frederic.n...@univ-lorraine.fr | frederic.n...@univ-lorraine.fr ] > 
wrote: 

BQ_BEGIN
Hi Danish, 

Can you access this KB article [1]? A free developer account should allow you 
to. 

It pretty much describes what you're facing and suggests to trim the sync error 
log of recovering shards. Actually, every log entry **except the last one**. 

1/ Identify the shards with the most sync errors log entries: 

radosgw-admin sync error list --max-entries=1000000 | grep shard_id | sort -n | 
uniq -c | sort -h 

2/ For each shard, list every sync error log entry along with their ids: 

radosgw-admin sync error list --shard-id=X 

3/ Remove them **except the last one** with: 

radosgw-admin sync error trim --shard-id=X --marker=1_1682101321.201434_8669.1 

the --marker above being the log entry id. 

Are the replication threads running on the same RGWs that S3 clients are using? 

If so, using dedicated RGWs for the sync job might help you avoid 
non-recovering shards in the future, as described in Matthew's post [2] 

Regards, 
Frédéric. 

[1] [ https://access.redhat.com/solutions/7023912 | 
https://access.redhat.com/solutions/7023912 ] 
[2] [ https://www.spinics.net/lists/ceph-users/msg83988.html | 
https://www.spinics.net/lists/ceph-users/msg83988.html ] 

----- Le 12 Mar 25, à 11:15, Danish Khan [ mailto:danish52....@gmail.com | 
danish52....@gmail.com ] a écrit : 

> Dear All, 
> 
> My ceph cluster is giving Large OMAP warning from approx 2-3 Months. I 
> tried a few things like : 
> *Deep scrub of PGs* 
> *Compact OSDs* 
> *Trim log* 
> But these didn't work out. 
> 
> I guess the main issue is that 4 shards in replication site are always 
> recovering from 2-3 months. 
> 
> Any suggestions are highly appreciated. 
> 
> Sync status: 
> root@drhost1:~# radosgw-admin sync status 
> realm e259e0a92 (object-storage) 
> zonegroup 7a8606d2 (staas) 
> zone c8022ad1 (repstaas) 
> metadata sync syncing 
> full sync: 0/64 shards 
> incremental sync: 64/64 shards 
> metadata is caught up with master 
> data sync source: 2072ee2d32 (masterstaas) 
> syncing 
> full sync: 0/128 shards 
> incremental sync: 128/128 shards 
> data is behind on 3 shards 
> behind shards: [7,90,100] 
> oldest incremental change not applied: 
> 2025-03-12T13:14:10.268469+0530 [7] 
> 4 shards are recovering 
> recovering shards: [31,41,55,80] 
> 
> 
> Master site: 
> 1. *root@master1:~# for obj in $(rados ls -p masterstaas.rgw.log); do echo 
> "$(rados listomapkeys -p masterstaas.rgw.log $obj | wc -l) $obj";done | 
> sort -nr | head -10* 
> 1225387 data_log.91 
> 1225065 data_log.86 
> 1224662 data_log.87 
> 1224448 data_log.92 
> 1224018 data_log.89 
> 1222156 data_log.93 
> 1201489 data_log.83 
> 1174125 data_log.90 
> 363498 data_log.84 
> 258709 data_log.6 
> 
> 
> 2. *root@master1:~# for obj in data_log.91 data_log.86 data_log.87 
> data_log.92 data_log.89 data_log.93 data_log.83 data_log.90; do rados stat 
> -p masterstaas.rgw.log $obj; done* 
> masterstaas.rgw.log/data_log.91 mtime 2025-02-24T15:09:25.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.86 mtime 2025-02-24T15:01:25.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.87 mtime 2025-02-24T15:02:25.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.92 mtime 2025-02-24T15:11:01.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.89 mtime 2025-02-24T14:54:55.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.93 mtime 2025-02-24T14:53:25.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.83 mtime 2025-02-24T14:16:21.000000+0530, size 
> 0 
> masterstaas.rgw.log/data_log.90 mtime 2025-02-24T15:05:25.000000+0530, size 
> 0 
> 
> *3. ceph cluster log :* 
> 2025-02-22T04:18:27.324886+0530 osd.173 (osd.173) 19 : cluster [WRN] Large 
> omap object found. Object: 124:b2ddf551:::data_log.93:head PG: 124.8aafbb4d 
> (124.d) Key count: 1218170 Size (bytes): 297085860 
> 2025-02-22T04:18:28.735886+0530 osd.65 (osd.65) 308 : cluster [WRN] Large 
> omap object found. Object: 124:f2081d70:::data_log.92:head PG: 124.eb8104f 
> (124.f) Key count: 1220420 Size (bytes): 295240028 
> 2025-02-22T04:18:30.668884+0530 mon.master1 (mon.0) 7974038 : cluster [WRN] 
> Health check update: 3 large omap objects (LARGE_OMAP_OBJECTS) 
> 2025-02-22T04:18:31.127585+0530 osd.18 (osd.18) 224 : cluster [WRN] Large 
> omap object found. Object: 124:d1061236:::data_log.86:head PG: 124.6c48608b 
> (124.b) Key count: 1221047 Size (bytes): 295398274 
> 2025-02-22T04:18:33.189561+0530 osd.37 (osd.37) 32665 : cluster [WRN] Large 
> omap object found. Object: 124:9a2e04b7:::data_log.87:head PG: 124.ed207459 
> (124.19) Key count: 1220584 Size (bytes): 295290366 
> 2025-02-22T04:18:35.007117+0530 osd.77 (osd.77) 135 : cluster [WRN] Large 
> omap object found. Object: 124:6b9e929a:::data_log.89:head PG: 124.594979d6 
> (124.16) Key count: 1219913 Size (bytes): 295127816 
> 2025-02-22T04:18:36.189141+0530 mon.master1 (mon.0) 7974039 : cluster [WRN] 
> Health check update: 5 large omap objects (LARGE_OMAP_OBJECTS) 
> 2025-02-22T04:18:36.340247+0530 osd.112 (osd.112) 259 : cluster [WRN] Large 
> omap object found. Object: 124:0958bece:::data_log.83:head PG: 124.737d1a90 
> (124.10) Key count: 1200406 Size (bytes): 290280292 
> 2025-02-22T04:18:38.523766+0530 osd.73 (osd.73) 1064 : cluster [WRN] Large 
> omap object found. Object: 124:fddd971f:::data_log.91:head PG: 124.f8e9bbbf 
> (124.3f) Key count: 1221183 Size (bytes): 295425320 
> 2025-02-22T04:18:42.619926+0530 osd.92 (osd.92) 285 : cluster [WRN] Large 
> omap object found. Object: 124:7dc404fa:::data_log.90:head PG: 124.5f2023be 
> (124.3e) Key count: 1169895 Size (bytes): 283025576 
> 2025-02-22T04:18:44.242655+0530 mon.master1 (mon.0) 7974043 : cluster [WRN] 
> Health check update: 8 large omap objects (LARGE_OMAP_OBJECTS) 
> 
> Replica site: 
> 1. *for obj in $(rados ls -p repstaas.rgw.log); do echo "$(rados 
> listomapkeys -p repstaas.rgw.log $obj | wc -l) $obj";done | sort -nr | head 
> -10* 
> 
> 432850 data_log.91 
> 432384 data_log.87 
> 432323 data_log.93 
> 431783 data_log.86 
> 431510 data_log.92 
> 427959 data_log.89 
> 414522 data_log.90 
> 407571 data_log.83 
> 151015 data_log.84 
> 109790 data_log.4 
> 
> 
> 2. *ceph cluster log:* 
> grep -ir "Large omap object found" /var/log/ceph/ 
> /var/log/ceph/ceph-mon.drhost1.log:2025-03-12T14:49:59.997+0530 
> 7fc4ad544700 0 log_channel(cluster) log [WRN] : Search the cluster log 
> for 'Large omap object found' for more details. 
> /var/log/ceph/ceph.log:2025-03-12T14:49:02.078108+0530 osd.10 (osd.10) 21 : 
> cluster [WRN] Large omap object found. Object: 
> 6:b2ddf551:::data_log.93:head PG: 6.8aafbb4d (6.d) Key count: 432323 Size 
> (bytes): 105505884 
> /var/log/ceph/ceph.log:2025-03-12T14:49:02.389288+0530 osd.48 (osd.48) 37 : 
> cluster [WRN] Large omap object found. Object: 
> 6:d1061236:::data_log.86:head PG: 6.6c48608b (6.b) Key count: 431782 Size 
> (bytes): 104564674 
> /var/log/ceph/ceph.log:2025-03-12T14:49:07.166954+0530 osd.24 (osd.24) 24 : 
> cluster [WRN] Large omap object found. Object: 
> 6:0958bece:::data_log.83:head PG: 6.737d1a90 (6.10) Key count: 407571 Size 
> (bytes): 98635522 
> /var/log/ceph/ceph.log:2025-03-12T14:49:09.100110+0530 osd.63 (osd.63) 5 : 
> cluster [WRN] Large omap object found. Object: 
> 6:9a2e04b7:::data_log.87:head PG: 6.ed207459 (6.19) Key count: 432384 Size 
> (bytes): 104712350 
> /var/log/ceph/ceph.log:2025-03-12T14:49:08.703760+0530 osd.59 (osd.59) 11 : 
> cluster [WRN] Large omap object found. Object: 
> 6:6b9e929a:::data_log.89:head PG: 6.594979d6 (6.16) Key count: 427959 Size 
> (bytes): 103773777 
> /var/log/ceph/ceph.log:2025-03-12T14:49:11.126132+0530 osd.40 (osd.40) 24 : 
> cluster [WRN] Large omap object found. Object: 
> 6:f2081d70:::data_log.92:head PG: 6.eb8104f (6.f) Key count: 431508 Size 
> (bytes): 104520406 
> /var/log/ceph/ceph.log:2025-03-12T14:49:13.799473+0530 osd.43 (osd.43) 61 : 
> cluster [WRN] Large omap object found. Object: 
> 6:fddd971f:::data_log.91:head PG: 6.f8e9bbbf (6.1f) Key count: 432850 Size 
> (bytes): 104418869 
> /var/log/ceph/ceph.log:2025-03-12T14:49:14.398480+0530 osd.3 (osd.3) 55 : 
> cluster [WRN] Large omap object found. Object: 
> 6:7dc404fa:::data_log.90:head PG: 6.5f2023be (6.1e) Key count: 414521 Size 
> (bytes): 100396561 
> /var/log/ceph/ceph.log:2025-03-12T14:50:00.000484+0530 mon.drhost1 (mon.0) 
> 207423 : cluster [WRN] Search the cluster log for 'Large omap object 
> found' for more details. 
> 
> Regards, 
> Danish 
> _______________________________________________ 
> ceph-users mailing list -- [ mailto:ceph-users@ceph.io | ceph-users@ceph.io ] 
> To unsubscribe send an email to [ mailto:ceph-users-le...@ceph.io | 
> ceph-users-le...@ceph.io ] 

BQ_END


BQ_END


BQ_END


BQ_END


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to