Hi Kirby,

Not at all. It still sends the ticks and you can see the last lines of the 
ongoing logs below:

Oct 24 19:58:35 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:35.798+0000 7e2cb3476640 20 osd.11 70297 check_full_status cur 
ratio 0.77673, physical ratio 0.77673, new state none 
Oct 24 19:58:36 ank-backup01 ceph-osd[586118]: osd.11 70297 
tick_without_osd_lock
Oct 24 19:58:36 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:36.106+0000 7e2cc86b8640 10 osd.11 70297 tick_without_osd_lock
Oct 24 19:58:36 ank-backup01 ceph-osd[586118]: osd.11 70297 tick
Oct 24 19:58:36 ank-backup01 ceph-osd[586118]: osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:36 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:36.661+0000 7e2cc8eb9640 10 osd.11 70297 tick
Oct 24 19:58:36 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:36.661+0000 7e2cc8eb9640 20 osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:37 ank-backup01 ceph-osd[586118]: osd.11 70297 
tick_without_osd_lock
Oct 24 19:58:37 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:37.094+0000 7e2cc86b8640 10 osd.11 70297 tick_without_osd_lock
Oct 24 19:58:37 ank-backup01 ceph-osd[586118]: osd.11 70297 tick
Oct 24 19:58:37 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:37.681+0000 7e2cc8eb9640 10 osd.11 70297 tick
Oct 24 19:58:37 ank-backup01 ceph-osd[586118]: osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:37 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:37.681+0000 7e2cc8eb9640 20 osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:38 ank-backup01 ceph-osd[586118]: osd.11 70297 
tick_without_osd_lock
Oct 24 19:58:38 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:38.074+0000 7e2cc86b8640 10 osd.11 70297 tick_without_osd_lock
Oct 24 19:58:38 ank-backup01 ceph-osd[586118]: osd.11 70297 tick
Oct 24 19:58:38 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:38.644+0000 7e2cc8eb9640 10 osd.11 70297 tick
Oct 24 19:58:38 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:38.644+0000 7e2cc8eb9640 20 osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:38 ank-backup01 ceph-osd[586118]: osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:39 ank-backup01 ceph-osd[586118]: osd.11 70297 
tick_without_osd_lock
Oct 24 19:58:39 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:39.064+0000 7e2cc86b8640 10 osd.11 70297 tick_without_osd_lock
Oct 24 19:58:39 ank-backup01 ceph-osd[586118]: osd.11 70297 tick
Oct 24 19:58:39 ank-backup01 ceph-osd[586118]: osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:39 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:39.658+0000 7e2cc8eb9640 10 osd.11 70297 tick
Oct 24 19:58:39 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:39.658+0000 7e2cc8eb9640 20 osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:40 ank-backup01 ceph-osd[586118]: osd.11 70297 
tick_without_osd_lock
Oct 24 19:58:40 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:40.071+0000 7e2cc86b8640 10 osd.11 70297 tick_without_osd_lock
Oct 24 19:58:40 ank-backup01 ceph-osd[586118]: osd.11 70297 reports for 0 
queries
Oct 24 19:58:40 ank-backup01 ceph-osd[586118]: osd.11 70297 collect_pg_stats
Oct 24 19:58:40 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:40.415+0000 7e2cc18ba640 20 osd.11 70297 reports for 0 queries
Oct 24 19:58:40 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:40.415+0000 7e2cc18ba640 15 osd.11 70297 collect_pg_stats
Oct 24 19:58:40 ank-backup01 ceph-osd[586118]: osd.11 70297 tick
Oct 24 19:58:40 ank-backup01 ceph-osd[586118]: osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:40 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:40.624+0000 7e2cc8eb9640 10 osd.11 70297 tick
Oct 24 19:58:40 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:40.624+0000 7e2cc8eb9640 20 osd.11 70297 tick 
last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
2025-10-25T18:30:10.329484+0000
Oct 24 19:58:41 ank-backup01 ceph-osd[586118]: osd.11 70297 
tick_without_osd_lock
Oct 24 19:58:41 ank-backup01 
ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
2025-10-24T16:58:41.048+0000 7e2cc86b8640 10 osd.11 70297 tick_without_osd_lock

Selamlar,
Huseyin Cotuk
[email protected]




> On 24 Oct 2025, at 19:57, Kirby Haze <[email protected]> wrote:
> 
> Are these the end of the logs? This actually looks normal and fine but 
> eventually it will crash? or does the systemd unit stay running and they are 
> stuck on these logs?
> 
> On Fri, Oct 24, 2025 at 9:26 AM Huseyin Cotuk <[email protected] 
> <mailto:[email protected]>> wrote:
>> journalctl logs of a failing OSD starts like in the link: 
>> https://pastes.io/osd11-restart-log
>> 
>> Oct 24 18:50:46 ank-backup01 systemd[1]: Started 
>> [email protected] - Ceph osd.11 for 
>> 4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f.
>> Oct 24 18:50:46 ank-backup01 ceph-osd[586118]: set uid:gid to 167:167 
>> (ceph:ceph)
>> Oct 24 18:50:46 ank-backup01 ceph-osd[586118]: ceph version 19.2.3 
>> (c92aebb279828e9c3c1f5d24613efca272649e62) squid (stable), process ceph-osd, 
>> pid 2
>> Oct 24 18:50:46 ank-backup01 ceph-osd[586118]: pidfile_write: ignore empty 
>> --pid-file
>> 
>> I can see that it runs with the Ceph version 19.2.3. 
>> 
>> After the activation process (it takes some time), it does not crash 
>> actually. AFAIR from the logs, it mounts bluefs, starts the rocksdb version 
>> 7.9.2 and loads the tables. It evens compacts the rocksdb. 
>> 
>> Afterwards,  it triggers the boot process as follows.
>> 
>> Oct 24 18:53:41 ank-backup01 ceph-osd[586118]: osd.11 0 journal looks like 
>> hdd
>> Oct 24 18:53:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
>> 2025-10-24T15:53:41.705+0000 7e2ccf90a740  2 osd.11 0 journal looks like hdd
>> Oct 24 18:53:41 ank-backup01 ceph-osd[586118]: osd.11 0 boot
>> Oct 24 18:53:41 ank-backup01 ceph-osd[586118]: osd.11 0 configured 
>> osd_max_object_name[space]_len looks ok
>> Oct 24 18:53:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
>> 2025-10-24T15:53:41.705+0000 7e2ccf90a740  2 osd.11 0 boot
>> Oct 24 18:53:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
>> 2025-10-24T15:53:41.705+0000 7e2ccf90a740 20 osd.11 0 configured 
>> osd_max_object_name[space]_len looks ok
>> 
>> I noted that at this point, it loads the osdmap 70297
>> 
>> Oct 24 18:53:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
>> 2025-10-24T15:53:41.706+0000 7e2ccf90a740 20 osd.11 0 get_map 70297 - 
>> loading and decoding 0x565a07bb0000
>> 
>> There are lots of missing logs:
>> 
>> Oct 24 18:53:48 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
>> 2025-10-24T15:53:48.603+0000 7e2ccf90a740 20 read_log_and_missing 
>> 68276'8386071 (0'0) modify   
>> 23:bf0d9243:::48f964cb-12c5-4689-8aa4-0b8e573f41d5.37815204.2_001%2f184%2f288%2f052.zip:head
>>  by client.43525885.0:1306616953 2025-10-09T11:06:44.324360+0000 0 
>> ObjectCleanRegions clean_offsets: [(48911, 18446744073709502704)], 
>> clean_omap: true, new_object: false
>> 
>> After all these missing logs, it starts and loads the PGs. 
>> 
>> Oct 24 18:54:05 ank-backup01 ceph-osd[586118]: osd.11 70297 load_pgs opened 
>> 66 pgs
>> Oct 24 18:54:05 ank-backup01 ceph-osd[586118]: osd.11 70297 superblock: I am 
>> osd.11
>> Oct 24 18:54:05 ank-backup01 ceph-osd[586118]: osd.11 70297 heartbeat 
>> osd_stat(store_statfs(0x19ff4006000/0x0/0x74702400000, data 
>> 0x4b4a481b898/0x4b4a4e68000, compress 0x0/0x0/0x0, omap 0x1a50, meta 
>> 0xf26958e5b0), peers [] op hist [])
>> Oct 24 18:54:05 ank-backup01 ceph-osd[586118]: osd.11 70297 done with init, 
>> starting boot process
>> Oct 24 18:54:05 ank-backup01 ceph-osd[586118]: osd.11 70297 start_boot
>> 
>> It periodically dumps stats that shows that OSD is running (up) for xxx 
>> secs. 
>> 
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: Uptime(secs): 
>> 600.6 total, 600.0 interval
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: Flush(GB): 
>> cumulative 0.000, interval 0.000
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: AddFile(GB): 
>> cumulative 0.000, interval 0.000
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: AddFile(Total 
>> Files): cumulative 0, interval 0
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: AddFile(L0 Files): 
>> cumulative 0, interval 0
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: AddFile(Keys): 
>> cumulative 0, interval 0
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: Cumulative 
>> compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 
>> 0.0 seconds
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: Interval 
>> compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 
>> 0.0 seconds
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: Stalls(count): 0 
>> level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 
>> level0_numfiles_with_compaction, 0 stop for pending_compaction_b>
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: Block cache 
>> BinnedLRUCache@0x5659dca05350#2 capacity: 1.06 GB usage: 100.56 MB 
>> table_size: 0 occupancy: 18446744073709551615 collections: 2 last_copies: 8>
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: Block cache entry 
>> stats(count,size,portion): DataBlock(24571,97.85 MB,8.99328%) 
>> FilterBlock(14,1.17 MB,0.107375%) IndexBlock(20,1.54 MB,0.141681%) Misc(1,>
>> Oct 24 19:03:41 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
>> 
>> 
>> The problem is, it can not register itself to the mon as UP and running. 
>> There are lots of ticks and tick_without_osd_lock messages in the log file. 
>> 
>> Oct 24 18:57:14 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
>> 2025-10-24T15:57:14.370+0000 7e2cc8eb9640 10 osd.11 70297 tick
>> Oct 24 18:57:14 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
>> 2025-10-24T15:57:14.370+0000 7e2cc8eb9640 20 osd.11 70297 tick 
>> last_purged_snaps_scrub 2025-10-24T08:38:08.454484+0000 next 
>> 2025-10-25T18:30:10.329484+0000
>> Oct 24 18:57:14 ank-backup01 ceph-osd[586118]: osd.11 70297 
>> tick_without_osd_lock
>> Oct 24 18:57:14 ank-backup01 
>> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[586110]: 
>> 2025-10-24T15:57:14.729+0000 7e2cc86b8640 10 osd.11 70297 
>> tick_without_osd_lock
>> Oct 24 18:57:15 ank-backup01 ceph-osd[586118]: osd.11 70297 reports for 0 
>> queries
>> Oct 24 18:57:15 ank-backup01 ceph-osd[586118]: osd.11 70297 collect_pg_stats
>> 
>> It looks like it stays in a loop that send ticks with the older epoch 
>> (70297). 
>> 
>> Thanks for your help
>> 
>> BR,
>> Huseyin Cotuk
>> [email protected] <mailto:[email protected]>
>> 
>> 
>> 
>> 
>>> On 24 Oct 2025, at 18:36, Kirby Haze <[email protected] 
>>> <mailto:[email protected]>> wrote:
>>> 
>>> if you look at the osd’s logs and follow it while trigger a restart through 
>>> systemctl 
>>> 
>>> systemctl list-units | grep osd
>>> 
>>> journalctl -u <osd unit> 
>>> 
>>> When the osd boots up it spits out the version, I would be curious to see 
>>> how it crashes and the stack trace 
>>> 
>>> On Fri, Oct 24, 2025 at 8:31 AM Huseyin Cotuk <[email protected] 
>>> <mailto:[email protected]>> wrote:
>>>> Hi Kirby,
>>>> 
>>>> Is there any way to check whether these OSDs have completed the upgrade 
>>>> successfully or not? 
>>>> 
>>>> Any help will be appreciated
>>>> 
>>>> BR,
>>>> Huseyin
>>>> [email protected] <mailto:[email protected]>
>>>> 
>>>> 
>>>> 
>>>> 
>>>>> On 24 Oct 2025, at 18:14, Kirby Haze <[email protected] 
>>>>> <mailto:[email protected]>> wrote:
>>>>> 
>>>>> Osds usually try to load the last clean map on boot, but if these osds 
>>>>> finished the upgrade it must have hit an assert I think
>>>>> 
>>>>> On Fri, Oct 24, 2025 at 8:06 AM Huseyin Cotuk <[email protected] 
>>>>> <mailto:[email protected]>> wrote:
>>>>>> Hi Can,
>>>>>> 
>>>>>> No, I did not add or recreate these OSDs. I tried cephadm to run failing 
>>>>>> OSDs with the help of following commands without success:
>>>>>> 
>>>>>> cephadm --image quay.io/ceph/ceph:v19 <http://quay.io/ceph/ceph:v19> 
>>>>>> unit --name osd.$i --fsid 4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f restart
>>>>>> 
>>>>>> cephadm --image aade1b12b8e6  run --name osd.$i --fsid 
>>>>>> 4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f 
>>>>>> 
>>>>>> BR,
>>>>>> Huseyin
>>>>>> [email protected] <mailto:[email protected]>
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> > On 24 Oct 2025, at 17:55, Can Özyurt <[email protected] 
>>>>>> > <mailto:[email protected]>> wrote:
>>>>>> > 
>>>>>> > Hi Huseyin,
>>>>>> > 
>>>>>> > Did you add or recreate these OSDs after the upgrade?
>>>>>> > 
>>>>>> > On Fri, 24 Oct 2025 at 16:42, Huseyin Cotuk <[email protected] 
>>>>>> > <mailto:[email protected]> <mailto:[email protected] 
>>>>>> > <mailto:[email protected]>>> wrote:
>>>>>> >> More debug logs from a failing OSD: 
>>>>>> >> 
>>>>>> >> Oct 24 17:39:52 ank-backup01 
>>>>>> >> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[4100566]: 
>>>>>> >> 2025-10-24T14:39:52.271+0000 77f47e35c640 20 osd.11 70297 reports for 
>>>>>> >> 0 queries
>>>>>> >> Oct 24 17:39:52 ank-backup01 
>>>>>> >> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[4100566]: 
>>>>>> >> 2025-10-24T14:39:52.271+0000 77f47e35c640 15 osd.11 70297 
>>>>>> >> collect_pg_stats
>>>>>> >> Oct 24 17:39:52 ank-backup01 ceph-osd[4100574]: osd.11 pg_epoch: 
>>>>>> >> 70297 pg[23.b9s0( v 69083'8392506 (68272'8387884,69083'8392506] 
>>>>>> >> local-lis/les=69076/69077 n=8314181 ec=21807/21807 lis/c=69076/68833 
>>>>>> >> les/c/f=69077/68834/39143 sis=70280) 
>>>>>> >> [11,NONE,NONE,12,31,21,7,NONE,NONE,26,0]p11(0) r=0 lpr=70297 
>>>>>> >> pi=[68833,70280)/5 crt=69083'8392506 lcod 0'0 mlcod 0'0 unknown 
>>>>>> >> mbc={}] PeeringState::prepare_stats_for_publish reporting 
>>>>>> >> purged_snaps []
>>>>>> >> Oct 24 17:39:52 ank-backup01 
>>>>>> >> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[4100566]: 
>>>>>> >> 2025-10-24T14:39:52.271+0000 77f47e35c640 20 osd.11 pg_epoch: 70297 
>>>>>> >> pg[23.b9s0( v 69083'8392506 (68272'8387884,69083'8392506] 
>>>>>> >> local-lis/les=69076/69077 n=8314181 ec=21807/21807 lis/c=69076/68833 
>>>>>> >> les/c/f=69077/68834/39143 sis=70280) 
>>>>>> >> [11,NONE,NONE,12,31,21,7,NONE,NONE,26,0]p11(0) r=0 lpr=70297 
>>>>>> >> pi=[68833,70280)/5 crt=69083'8392506 lcod 0'0 mlcod 0'0 unknown 
>>>>>> >> mbc={}] PeeringState::prepare_stats_for_publish reporting 
>>>>>> >> purged_snaps []
>>>>>> >> Oct 24 17:39:52 ank-backup01 ceph-osd[4100574]: osd.11 pg_epoch: 
>>>>>> >> 70297 pg[23.b9s0( v 69083'8392506 (68272'8387884,69083'8392506] 
>>>>>> >> local-lis/les=69076/69077 n=8314181 ec=21807/21807 lis/c=69076/68833 
>>>>>> >> les/c/f=69077/68834/39143 sis=70280) 
>>>>>> >> [11,NONE,NONE,12,31,21,7,NONE,NONE,26,0]p11(0) r=0 lpr=70297 
>>>>>> >> pi=[68833,70280)/5 crt=69083'8392506 lcod 0'0 mlcod 0'0 unknown 
>>>>>> >> mbc={}] PeeringState::prepare_stats_for_publish publish_stats_to_osd 
>>>>>> >> 70297:52872808
>>>>>> >> Oct 24 17:39:52 ank-backup01 
>>>>>> >> ceph-4e7e7d1c-22db-49c7-9f24-5a75cd3a3b9f-osd-11[4100566]: 
>>>>>> >> 2025-10-24T14:39:52.271+0000 77f47e35c640 15 osd.11 pg_epoch: 70297 
>>>>>> >> pg[23.b9s0( v 69083'8392506 (68272'8387884,69083'8392506] 
>>>>>> >> local-lis/les=69076/69077 n=8314181 ec=21807/21807 lis/c=69076/68833 
>>>>>> >> les/c/f=69077/68834/39143 sis=70280) 
>>>>>> >> [11,NONE,NONE,12,31,21,7,NONE,NONE,26,0]p11(0) r=0 lpr=70297 
>>>>>> >> pi=[68833,70280)/5 crt=69083'8392506 lcod 0'0 mlcod 0'0 unknown 
>>>>>> >> mbc={}] PeeringState::prepare_stats_for_publish publish_stats_to_osd 
>>>>>> >> 70297:52872808
>>>>>> >> Oct 24 17:39:52 ank-backup01 ceph-osd[4100574]: osd.11 pg_epoch: 
>>>>>> >> 70297 pg[23.5fs2( v 69081'8395094 (68276'8390628,69081'8395094] 
>>>>>> >> local-lis/les=69078/69079 n=8316950 ec=21807/21807 lis/c=69078/68833 
>>>>>> >> les/c/f=69079/68834/39143 sis=70280) 
>>>>>> >> [NONE,NONE,11,8,14,16,0,20,17,18,NONE]p11(2) r=2 lpr=70297 
>>>>>> >> pi=[68833,70280)/7 crt=69081'8395094 lcod 0'0 mlcod 0'0 unknown 
>>>>>> >> mbc={}] PeeringState::prepare_stats_for_publish reporting 
>>>>>> >> purged_snaps []
>>>>>> >> 
>>>>>> >> 
>>>>>> >> Selamlar,
>>>>>> >> Huseyin Cotuk
>>>>>> >> [email protected] <mailto:[email protected]> <mailto:[email protected] 
>>>>>> >> <mailto:[email protected]>>
>>>>>> >> 
>>>>>> >> 
>>>>>> >> 
>>>>>> >> 
>>>>>> >> _______________________________________________
>>>>>> >> ceph-users mailing list -- [email protected] 
>>>>>> >> <mailto:[email protected]> <mailto:[email protected] 
>>>>>> >> <mailto:[email protected]>>
>>>>>> >> To unsubscribe send an email to [email protected] 
>>>>>> >> <mailto:[email protected]> <mailto:[email protected] 
>>>>>> >> <mailto:[email protected]>>
>>>>>> 
>>>>>> _______________________________________________
>>>>>> ceph-users mailing list -- [email protected] <mailto:[email protected]>
>>>>>> To unsubscribe send an email to [email protected] 
>>>>>> <mailto:[email protected]>
>> 

_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to