Hi Dev,

I am not sure why you formatted osd.19, may be I missed something. Clearly Ceph is very sensitive to network issues on any of the network used. Your priority should be to ensure that your network config is ok between you Ceph servers. This is an OS configuration issue that you need to troubleshoot with the usual OS tools. The problem may be a change in your network infrastructure (switches config for example), a problem with MTU size if you're using Jumbo frames...

If your cluster was kind of ok before the upgrade, for me there is no reason to reformat OSDs or change anything to the cluster config. You need to spot what the problem cause is, may be something outside Ceph, and fix it before trying to restart the upgrade.

Good luck.

Michel
Sent from my mobile
Le 21 juin 2025 23:50:01 Devender Singh <deven...@netskrt.io> a écrit :
Hello Fred

I formatted the osd.19 but facing similar issue on osd.9. I have pause the upgrade.

Below are the logs, another issue found is my osd’s are not using cluster network…. How to deal with it?

root@pl-host04n:/var/lib/ceph/a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac/osd.9# ceph config get mon public_network
10.104.1.0/24
root@pl-host04n:/var/lib/ceph/a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac/osd.9# ceph config get mon cluster_network
10.104.5.0/24
root@pl-host04n:/var/lib/ceph/a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac/osd.9# ceph osd find 9
{
"osd": 9,
"addrs": {
"addrvec": [
{
"type": "v2",
"addr": "10.104.1.124:6802",
"nonce": 2257868117
},
{
"type": "v1",
"addr": "10.104.1.124:6803",
"nonce": 2257868117
}
]
},
"osd_fsid": "4db6e332-9031-4c81-8de0-00fdd6b860f6",
"host": "pl-host04n.phl.example.com",
"crush_location": {
"host": "pl-host04n",
"root": "default"
}
}
root@pl-host04n:/var/lib/ceph/a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac/osd.9# grep -E '\bosd\.9\b' /var/log/syslog-ceph |tail -20 Jun 21 20:10:00 pl-host04n bash[3965687]: debug 2025-06-21T20:09:59.996+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 20:10:00 pl-host04n bash[3965687]: cluster 2025-06-21T20:10:00.000344+0000 mon.pl-host04n (mon.0) 208269 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 20:20:00 pl-host04n bash[3965687]: debug 2025-06-21T20:19:59.994+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 20:20:00 pl-host04n bash[3965687]: cluster 2025-06-21T20:20:00.000282+0000 mon.pl-host04n (mon.0) 208543 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 20:30:00 pl-host04n bash[3965687]: debug 2025-06-21T20:29:59.993+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 20:30:00 pl-host04n bash[3965687]: cluster 2025-06-21T20:30:00.000292+0000 mon.pl-host04n (mon.0) 208802 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 20:40:00 pl-host04n bash[3965687]: debug 2025-06-21T20:39:59.996+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 20:40:00 pl-host04n bash[3965687]: cluster 2025-06-21T20:40:00.000322+0000 mon.pl-host04n (mon.0) 209104 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 20:50:00 pl-host04n bash[3965687]: debug 2025-06-21T20:49:59.994+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 20:50:00 pl-host04n bash[3965687]: cluster 2025-06-21T20:50:00.000342+0000 mon.pl-host04n (mon.0) 209394 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:00:00 pl-host04n bash[3965687]: debug 2025-06-21T20:59:59.993+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:00:00 pl-host04n bash[3965687]: cluster 2025-06-21T21:00:00.000331+0000 mon.pl-host04n (mon.0) 209681 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:10:00 pl-host04n bash[3965687]: debug 2025-06-21T21:09:59.996+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:10:00 pl-host04n bash[3965687]: cluster 2025-06-21T21:10:00.000308+0000 mon.pl-host04n (mon.0) 209967 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:20:00 pl-host04n bash[3965687]: debug 2025-06-21T21:19:59.994+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:20:00 pl-host04n bash[3965687]: cluster 2025-06-21T21:20:00.000268+0000 mon.pl-host04n (mon.0) 210298 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:30:00 pl-host04n bash[3965687]: debug 2025-06-21T21:29:59.993+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:30:00 pl-host04n bash[3965687]: cluster 2025-06-21T21:30:00.000316+0000 mon.pl-host04n (mon.0) 210560 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:40:00 pl-host04n bash[3965687]: debug 2025-06-21T21:39:59.996+0000 7fd05b6b4640 0 log_channel(cluster) log [WRN] : daemon osd.9 on pl-host04n.phl.example.com is in error state Jun 21 21:40:00 pl-host04n bash[3965687]: cluster 2025-06-21T21:40:00.000231+0000 mon.pl-host04n (mon.0) 210817 : cluster [WRN] daemon osd.9 on pl-host04n.phl.example.com is in error state


root@pl-host04n:/var/lib/ceph/a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac/osd.9# systemctl list-units |grep -i osd.9 ● ceph-a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac@osd.9.service loaded failed failed Ceph osd.9 for a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac ● ceph-osd@9.service loaded failed failed Ceph object storage daemon osd.9 root@pl-host04n:/var/lib/ceph/a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac/osd.9# systemctl status ceph-a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac@osd.9.service × ceph-a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac@osd.9.service - Ceph osd.9 for a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac Loaded: loaded (/etc/systemd/system/ceph-a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac@.service; enabled; vendor preset: enabled) Active: failed (Result: exit-code) since Sat 2025-06-21 18:46:01 UTC; 3h 2min ago Process: 3156903 ExecStart=/bin/bash /var/lib/ceph/a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac/osd.9/unit.run (code=exited, status=1/FAI> Process: 3158431 ExecStopPost=/bin/bash /var/lib/ceph/a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac/osd.9/unit.poststop (code=exited, stat>
Main PID: 3156903 (code=exited, status=1/FAILURE)
CPU: 571ms

Jun 21 18:46:01 pl-host04n.phl.example.com systemd[1]: ceph-a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac@osd.9.service: Scheduled > Jun 21 18:46:01 pl-host04n.phl.example.com systemd[1]: Stopped Ceph osd.9 for a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac. Jun 21 18:46:01 pl-host04n.phl.example.com systemd[1]: ceph-a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac@osd.9.service: Start requ> Jun 21 18:46:01 pl-host04n.phl.example.com systemd[1]: ceph-a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac@osd.9.service: Failed wit> Jun 21 18:46:01 pl-host04n.phl.example.com systemd[1]: Failed to start Ceph osd.9 for a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac. root@pl-host04n:/var/lib/ceph/a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac/osd.9# journalctl -u ceph-a0bd51e8-4dfc-11ee-b5a9-3b06e501a0ac@osd.9.service Jun 17 04:44:20 pl-host04n.phl.example.com bash[2152282]: debug 2025-06-17T04:44:20.588+0000 7fded3075640 -1 osd.9 pg_epoc> Jun 17 04:44:24 pl-host04n.phl.example.com bash[2152282]: debug 2025-06-17T04:44:24.076+0000 7fdede08b640 4 rocksdb: [db/> Jun 17 04:45:38 pl-host04n.phl.example.com bash[2152282]: debug 2025-06-17T04:45:38.812+0000 7fdede08b640 4 rocksdb: [db/> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: debug 2025-06-17T04:45:51.776+0000 7fdedd089640 4 rocksdb: [db/> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: debug 2025-06-17T04:45:51.776+0000 7fdedd089640 4 rocksdb: [db/>
Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: ** DB Stats **
Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Uptime(secs): 88200.5 total, 600.0 interval Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Cumulative writes: 4721K writes, 20M keys, 4721K commit groups, > Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Cumulative WAL: 4721K writes, 1745K syncs, 2.71 writes per sync,> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Interval writes: 21K writes, 229K keys, 21K commit groups, 1.0 w> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Interval WAL: 21K writes, 7176 syncs, 2.98 writes per sync, writ> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: ** Compaction Stats [O-1] ** Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: ----------------------------------------------------------------> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: L0 0/0 0.00 KB 0.0 0.0 0.0 0.0 0> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: L1 7/0 393.07 MB 0.4 4.3 0.4 3.9 > Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Sum 7/0 393.07 MB 0.0 4.3 0.4 3.9 > Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: ** Compaction Stats [O-1] ** Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Wri> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: ----------------------------------------------------------------> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Low 0/0 0.00 KB 0.0 4.3 0.4 3.9 4> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: High 0/0 0.00 KB 0.0 0.0 0.0 0.0 0> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, sp> Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Uptime(secs): 88200.5 total, 4800.1 interval Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Flush(GB): cumulative 0.407, interval 0.000 Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: AddFile(GB): cumulative 0.000, interval 0.000 Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: AddFile(Total Files): cumulative 0, interval 0 Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: AddFile(L0 Files): cumulative 0, interval 0 Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: AddFile(Keys): cumulative 0, interval 0 Jun 17 04:45:51 pl-host04n.phl.example.com bash[2152282]: Cumulative compaction: 4.64 GB write, 0.05 MB/s write, 4.26 GB r>

Regards
Dev


On Jun 20, 2025, at 9:57 PM, Frédéric Nass <frederic.n...@univ-lorraine.fr> wrote:

Hi Dev,

Since MGRs and MONs were already upgraded successfully, you should be safe stopping the upgrade, and restart it at a later time.

But before that, you could investigate why osd.19 is not coming up and why ceph-volume inventory times out. Can you ssh from MGR host to osd.19 'host phl-prod-host04n.example.comis'?

I would look into ceph-osd.19.log and /var/log/messages for any hints on why osd.19 didn't start, start it manually and see if the upgrade resumes.


If the upgrade doesn't resume, I would increase cephadm command timeout to 1800 (default is 900):


$ ceph config set global mgr/cephadm/default_cephadm_command_timeout 1800


(This might need a ceph mgr fail but ceph mgr fail will also interrupt the upgrade, IIRC.)

then run

$ ceph orch device ls --hostname=phl-prod-host04n.example.comis --refresh

and see if the upgrade resumes. It if doesn't, check

$ ceph log last 1000 debug cephadm


and run

$ ceph orch upgrade pause
$ ceph orch upgrade resume

again, see if the upgrade resumes. If it still doesn't, then

$ ceph orch upgrade stop
$ ceph mgr fail
$ ceph orch upgrade start --image quay.io/ceph/ceph:v19.2.2

All those commands should be safe to run in the current state of your cluster.

Regards,
Frédéric.

De : Devender Singh <deven...@netskrt.io>
Envoyé : vendredi 20 juin 2025 23:35
À : Anthony D'Atri
Cc: Michel Jouvin; ceph-users
Objet : [ceph-users] Re: CEPH upgrade from 18.2.7 to 19.2.2 -- Hung from last 24h at 66%



Thanks all, what if I stop upgrade, what worst will happen ?

Regards
Dev

On Jun 20, 2025, at 6:41 AM, Anthony D'Atri <a...@dreamsnake.net> wrote:

Or depending on the release in force when the OSDs were created, perhaps shard RocksDB column families?



https://www.ibm.com/docs/en/storage-ceph/8.0.0?topic=bluestore-resharding-rocksdb-database <https://www.google.com/url?q=https://www.ibm.com/docs/en/storage-ceph/8.0.0?topic%3Dbluestore-resharding-rocksdb-database&source=gmail-imap&ust=1751031728000000&usg=AOvVaw2QwtMhyiksMW179R-6g3Ut&gt;
(Playbook from cephadm-ansible)

On Jun 20, 2025, at 1:57 AM, Michel Jouvin <michel.jou...@ijclab.in2p3.fr> wrote:

Hi Dev,

Not sure to understand why there was these service deployment time-out, the log says that one OSD failed, this may explain that that the upgrade is not progressing anymore. The Bluestore slow ops (a new warning so not necessarily something new) on so many OSD seem to suggest that there something not optimal. As suggested in another thread recently it may be an indication that you need to compact OSD.

I am not sure what you adjusted but as long as the cluster works, I would not have changed parameters and try to fix the mentioned problems.

Good luck,

Michel
Sent from my mobile
Le 20 juin 2025 05:13:38 Devender Singh <deven...@netskrt.io> a écrit :

Here is the status

# ceph orch upgrade status
{
"target_image": "quay.io/ceph/ceph@sha256:8214ebff6133ac27d20659038df6962dbf9d77da21c9438a296b2e2059a56af6&quot;,
  "in_progress": true,
  "which": "Upgrading all daemon types on all hosts",
  "services_complete": [
      "crash",
      "mgr",
      "mon"
  ],
  "progress": "74/113 daemons upgraded",
  "message": "",
  "is_paused": false
}

Regards
Dev

On Jun 19, 2025, at 8:06 PM, Devender Singh <deven...@netskrt.io> wrote:


Hello all

I have a cluster where my cluster is in hung state, Some back fills are there but I reduced it to 1 but still upgrade not progressing…
Please help…

```# ceph health detail
HEALTH_WARN 8 OSD(s) experiencing slow operations in BlueStore; Failed to apply 2 service(s): osd.all-available-devices,osd.iops_optimized; 1 failed cephadm daemon(s); failed to probe daemons or devices; noscrub,nodeep-scrub flag(s) set; Degraded data redundancy: 1600150/39365198 objects degraded (4.065%), 93 pgs degraded, 103 pgs undersized; 127 pgs not deep-scrubbed in time [WRN] BLUESTORE_SLOW_OP_ALERT: 8 OSD(s) experiencing slow operations in BlueStore
osd.5 observed slow operation indications in BlueStore
osd.9 observed slow operation indications in BlueStore
osd.18 observed slow operation indications in BlueStore
osd.36 observed slow operation indications in BlueStore
osd.59 observed slow operation indications in BlueStore
osd.66 observed slow operation indications in BlueStore
osd.106 observed slow operation indications in BlueStore
osd.110 observed slow operation indications in BlueStore
[WRN] CEPHADM_APPLY_SPEC_FAIL: Failed to apply 2 service(s): osd.all-available-devices,osd.iops_optimized osd.all-available-devices: Command timed out on host cephadm deploy (osd daemon) (default 900 second timeout) osd.iops_optimized: Command timed out on host cephadm deploy (osd daemon) (default 900 second timeout)
[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s)
daemon osd.19 on phl-prod-host04n.example.comis in error state
[WRN] CEPHADM_REFRESH_FAILED: failed to probe daemons or devices
Command "cephadm ceph-volume -- inventory" timed out on host phl-prod-converged03n.phl.netskrt.org (default 900 second timeout)
[WRN] OSDMAP_FLAGS: noscrub,nodeep-scrub flag(s) set
[WRN] PG_DEGRADED: Degraded data redundancy: 1600150/39365198 objects degraded (4.065%), 93 pgs degraded, 103 pgs undersized pg 25.0 is stuck undersized for 21h, current state active+undersized+degraded+remapped+backfill_wait, last acting [90,77,NONE,38,3] pg 25.1 is stuck undersized for 23h, current state active+undersized+degraded+remapped+backfill_wait, last acting [131,1,66,NONE,72] pg 25.c is stuck undersized for 23h, current state active+undersized+degraded+remapped+backfill_wait, last acting [1,135,110,28,NONE] pg 25.e is active+undersized+degraded+remapped+backfill_wait, acting [18,20,108,101,NONE] pg 25.14 is stuck undersized for 19h, current state active+undersized+degraded+remapped+backfill_wait, last acting [32,13,NONE,65,97] pg 25.17 is stuck undersized for 19h, current state active+undersized+degraded+remapped+backfill_wait, last acting [32,65,110,23,NONE] pg 25.1a is stuck undersized for 23h, current state active+undersized+degraded+remapped+backfill_wait, last acting [97,66,NONE,112,139] pg 25.1c is stuck undersized for 2h, current state active+undersized+degraded+remapped+backfill_wait, last acting [102,136,97,NONE,45] pg 25.1f is stuck undersized for 8h, current state active+undersized+degraded+remapped+backfill_wait, last acting [NONE,66,7,39,109] pg 26.45 is stuck undersized for 23h, current state active+undersized+degraded+remapped+backfill_wait, last acting [32,47] pg 26.48 is stuck undersized for 23h, current state active+undersized+degraded+remapped+backfill_wait, last acting [113,55] pg 26.4b is stuck undersized for 8h, current state active+undersized+degraded+remapped+backfill_wait, last acting [105,7] pg 26.59 is stuck undersized for 23h, current state active+undersized+degraded+remapped+backfill_wait, last acting [35,40] pg 26.6a is stuck undersized for 7h, current state active+undersized+degraded+remapped+backfill_wait, last acting [25,45] pg 26.74 is stuck undersized for 23h, current state active+undersized+degraded+remapped+backfill_wait, last acting [112,131] pg 26.77 is stuck undersized for 2h, current state active+undersized+degraded+remapped+backfill_wait, last acting [17,105] pg 26.9c is stuck undersized for 19h, current state active+undersized+degraded+remapped+backfill_wait, last acting [47,76] pg 26.bf is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [32,26] pg 26.c2 is stuck undersized for 2h, current state active+undersized+degraded+remapped+backfill_wait, last acting [97,135] pg 26.ec is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [26,80] pg 26.f7 is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [49,105] pg 31.12 is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [47,110,131,37,NONE] pg 31.19 is stuck undersized for 2h, current state active+undersized+degraded+remapped+backfill_wait, last acting [NONE,87,3,137,39] pg 31.1b is stuck undersized for 2h, current state active+undersized+degraded+remapped+backfill_wait, last acting [108,NONE,42,102,97] pg 31.1c is stuck undersized for 19h, current state active+undersized+degraded+remapped+backfill_wait, last acting [113,66,NONE,45,72] pg 31.41 is stuck undersized for 2h, current state active+undersized+remapped+backfill_wait, last acting [14,101,NONE,67,18] pg 31.42 is stuck undersized for 7h, current state active+undersized+degraded+remapped+backfill_wait, last acting [37,134,NONE,82,8] pg 31.44 is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [49,NONE,101,3,62] pg 31.46 is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfilling, last acting [95,38,NONE,25,102] pg 31.47 is stuck undersized for 18h, current state active+undersized+degraded+remapped+backfill_wait, last acting [60,NONE,130,72,110] pg 31.4a is stuck undersized for 4h, current state active+undersized+degraded+remapped+backfilling, last acting [66,NONE,135,82,14] pg 31.4c is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [34,NONE,1,82,18] pg 31.4d is stuck undersized for 2h, current state active+undersized+degraded+remapped+backfill_wait, last acting [101,13,65,30,NONE] pg 31.52 is stuck undersized for 4m, current state active+undersized+remapped+backfill_wait, last acting [26,112,66,NONE,135] pg 31.53 is stuck undersized for 17h, current state active+undersized+degraded+remapped+backfill_wait, last acting [NONE,24,11,42,110] pg 31.55 is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [116,NONE,27,4,117] pg 31.57 is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [17,15,110,NONE,1] pg 31.5a is stuck undersized for 19h, current state active+undersized+degraded+remapped+backfill_wait, last acting [13,24,124,26,NONE] pg 31.5b is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [NONE,31,8,27,42] pg 31.5c is stuck undersized for 2h, current state active+undersized+degraded+remapped+backfill_wait, last acting [NONE,60,28,18,119] pg 31.5d is stuck undersized for 17h, current state active+undersized+degraded+remapped+backfill_wait, last acting [124,NONE,85,6,11] pg 31.5e is stuck undersized for 2h, current state active+undersized+degraded+remapped+backfill_wait, last acting [NONE,4,1,23,138] pg 31.60 is stuck undersized for 18h, current state active+undersized+degraded+remapped+backfilling, last acting [37,11,102,NONE,133] pg 31.64 is stuck undersized for 4m, current state active+undersized+remapped+backfill_wait, last acting [26,106,45,34,NONE] pg 31.65 is stuck undersized for 4h, current state active+undersized+degraded+remapped+backfill_wait, last acting [NONE,14,127,62,3] pg 31.66 is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [34,NONE,35,23,59] pg 31.67 is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [67,24,127,8,NONE] pg 31.68 is stuck undersized for 4m, current state active+undersized+degraded+remapped+backfill_wait, last acting [32,41,18,17,NONE] pg 31.70 is stuck undersized for 19h, current state active+undersized+degraded+remapped+backfilling, last acting [106,38,1,NONE,97] pg 31.75 is stuck undersized for 2h, current state active+undersized+degraded+remapped+backfill_wait, last acting [26,31,34,8,NONE] pg 31.7f is stuck undersized for 2h, current state active+undersized+remapped+backfilling, last acting [NONE,101,110,10,45]
[WRN] PG_NOT_DEEP_SCRUBBED: 127 pgs not deep-scrubbed in time
pg 26.e5 not deep-scrubbed since 2025-06-07T05:59:22.612553+0000
pg 26.ca not deep-scrubbed since 2025-06-06T18:21:07.435823+0000
pg 26.bf not deep-scrubbed since 2025-06-03T07:29:02.791095+0000
pg 26.be not deep-scrubbed since 2025-06-04T07:55:54.389714+0000
pg 26.a5 not deep-scrubbed since 2025-06-07T07:28:56.878096+0000
pg 26.85 not deep-scrubbed since 2025-06-03T15:18:59.395929+0000
pg 26.7f not deep-scrubbed since 2025-06-04T21:29:28.412637+0000
pg 26.7e not deep-scrubbed since 2025-06-03T09:14:19.585388+0000
pg 26.7d not deep-scrubbed since 2025-06-03T18:37:30.931020+0000
pg 26.7c not deep-scrubbed since 2025-06-04T13:38:00.061488+0000
pg 26.73 not deep-scrubbed since 2025-06-03T06:20:15.111819+0000
pg 26.6f not deep-scrubbed since 2025-06-03T13:45:24.880397+0000
pg 26.6e not deep-scrubbed since 2025-05-26T23:15:32.099862+0000
pg 26.6d not deep-scrubbed since 2025-06-04T14:04:10.449101+0000
pg 31.62 not deep-scrubbed since 2025-06-03T13:34:49.518456+0000
pg 26.65 not deep-scrubbed since 2025-06-04T07:56:25.353411+0000
pg 31.66 not deep-scrubbed since 2025-06-03T10:32:05.364424+0000
pg 26.62 not deep-scrubbed since 2025-06-04T09:35:58.267976+0000
pg 31.65 not deep-scrubbed since 2025-06-03T16:04:40.003140+0000
pg 31.5b not deep-scrubbed since 2025-06-03T14:18:18.835477+0000
pg 26.5d not deep-scrubbed since 2025-06-04T15:14:30.870252+0000
pg 31.58 not deep-scrubbed since 2025-06-03T03:09:27.568605+0000
pg 26.5c not deep-scrubbed since 2025-06-03T01:57:27.644129+0000
pg 31.5f not deep-scrubbed since 2025-06-03T05:53:20.860393+0000
pg 31.52 not deep-scrubbed since 2025-05-27T00:01:27.040861+0000
pg 31.53 not deep-scrubbed since 2025-05-24T09:37:58.964829+0000
pg 26.55 not deep-scrubbed since 2025-06-04T21:25:34.135356+0000
pg 26.54 not deep-scrubbed since 2025-06-04T06:07:12.978734+0000
pg 31.56 not deep-scrubbed since 2025-06-04T12:58:17.599712+0000
pg 31.57 not deep-scrubbed since 2025-06-03T07:02:16.859990+0000
pg 26.51 not deep-scrubbed since 2025-06-03T05:42:22.435483+0000
pg 26.4f not deep-scrubbed since 2025-06-03T09:10:22.617328+0000
pg 31.4a not deep-scrubbed since 2025-05-28T00:54:55.246532+0000
pg 26.4e not deep-scrubbed since 2025-06-03T11:16:49.278513+0000
pg 31.4b not deep-scrubbed since 2025-06-03T10:24:24.123351+0000
pg 26.4d not deep-scrubbed since 2025-06-04T19:01:44.614410+0000
pg 31.49 not deep-scrubbed since 2025-05-28T04:56:29.368285+0000
pg 31.42 not deep-scrubbed since 2025-05-28T08:38:57.151865+0000
pg 26.41 not deep-scrubbed since 2025-06-03T05:47:35.443867+0000
pg 26.40 not deep-scrubbed since 2025-06-03T05:43:13.283668+0000
pg 25.17 not deep-scrubbed since 2025-06-03T09:26:52.253625+0000
pg 32.2e not deep-scrubbed since 2025-06-05T13:01:06.175389+0000
pg 22.1a not deep-scrubbed since 2025-06-07T01:40:45.063268+0000
pg 31.13 not deep-scrubbed since 2025-06-03T12:21:17.965218+0000
pg 22.1b not deep-scrubbed since 2025-06-04T12:22:44.947751+0000
pg 25.14 not deep-scrubbed since 2025-05-28T06:26:32.552200+0000
pg 26.17 not deep-scrubbed since 2025-06-03T18:37:26.617483+0000
pg 31.12 not deep-scrubbed since 2025-05-28T08:39:23.271194+0000
pg 31.1c not deep-scrubbed since 2025-06-03T08:17:51.230187+0000
pg 25.1f not deep-scrubbed since 2025-05-28T00:19:23.653883+0000
77 more pgs…

Regards
Dev

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to