To clarify, there's no "issue" with the code itself. It's just that the code now reveals a potential "issue" with the OSD's underlying device, as Igor explained.
This warning can pop up starting from Quincy v17.2.8 (PR 59468), Reef v18.2.5 (PR #59466) and Squid v19.2.1 (PR #59464). Regards, Frédéric. ----- Le 2 Mai 25, à 12:36, Eugen Block ebl...@nde.ag a écrit : > The link Frederic shared is for 19.2.1, so yes, the new warning > appeared in 19.2.1 as well. > > Zitat von Laimis Juzeliūnas <laimis.juzeliu...@oxylabs.io>: > >> Hi all, >> >> Could this also be an issue with 19.2.2? >> We have seen few of these warnings right after upgrading from >> 19.2.0. A simple OSD restart removed them, but we haven’t seen them >> before. >> There are some users on the Ceph Slack channels discussing this >> observation in 19.2.2 as well. >> >> Best, >> Laimis J. >> >>> On 2 May 2025, at 13:11, Igor Fedotov <igor.fedo...@croit.io> wrote: >>> >>> Hi Everyone, >>> >>> well, indeed this warning has been introduced in 18.2.6. >>> >>> But I wouldn't say that's not an issue. Having it permanently >>> visible (particularly for a specific OSD only) might indicate some >>> issues with this OSD which could negatively impact overall cluster >>> performance. >>> >>> OSD log to be checked for potential clues and more research on the >>> root cause is recommended. >>> >>> And once again - likely that's not a regression in 18.2.6 but >>> rather some additional diagnostics brought by the release which >>> reveals a potential issue. >>> >>> >>> Thanks, >>> >>> Igor >>> >>> On 02.05.2025 11:19, Frédéric Nass wrote: >>>> Hi Michel, >>>> >>>> This is not an issue. It's a new warning that can be adjusted or >>>> muted. Check this thread [1] and this part [2] of the Reef >>>> documentation about this new alert. >>>> Came to Reef with PR #59466 [3]. >>>> >>>> Cheers, >>>> Frédéric. >>>> >>>> [1] >>>> https://www.google.com/url?q=https://www.spinics.net/lists/ceph-users/msg86131.html&source=gmail-imap&ust=1746785596000000&usg=AOvVaw27M4y8QaoDcRiJBkxDVVoK >>>> [2] >>>> https://www.google.com/url?q=https://docs.ceph.com/en/latest/rados/operations/health-checks/%23bluestore-slow-op-alert&source=gmail-imap&ust=1746785596000000&usg=AOvVaw21VoozoT2KT6FESbkkVJ_w >>>> [3] >>>> https://www.google.com/url?q=https://github.com/ceph/ceph/pull/59466&source=gmail-imap&ust=1746785596000000&usg=AOvVaw0nnpOvrWFLB1lAk0Ekms1i >>>> >>>> ----- Le 2 Mai 25, à 9:44, Michel Jouvin >>>> michel.jou...@ijclab.in2p3.fr a écrit : >>>> >>>>> Hi, >>>>> >>>>> Since our upgrade to 18.2.6 2 days ago, our cluster is reporting the >>>>> warning "1 OSD(s) experiencing slow operations in BlueStore": >>>>> >>>>> [root@dig-osd4 bluestore-slow-ops]# ceph health detail >>>>> HEALTH_WARN 1 OSD(s) experiencing slow operations in BlueStore >>>>> [WRN] BLUESTORE_SLOW_OP_ALERT: 1 OSD(s) experiencing slow operations in >>>>> BlueStore >>>>> osd.247 observed slow operation indications in BlueStore >>>>> >>>>> I have never seen this warning before so I've the feeling it is somehow >>>>> related to the upgrade and it doesn't seem related to the regression >>>>> mentioned in another thread (that should result in an OSD crash). >>>>> Googling quickly, I found this reported on 19.2.1 with SSD where in my >>>>> case it is an HDD. I don't know if the workaround mentioned in the issue >>>>> (bdev_xxx_discard=true) also applies to 18.2.6... >>>>> >>>>> Did somebody saw this in 18.2.x? Any recommandation? Our plan was, >>>>> according to best practicies described recently in another thread to >>>>> move from 18.2.2 to 18.2.6 and then from 18.2.6 to 19.2.2... Will 19.2.2 >>>>> clear this issue (at the risk of others as it is probably not >>>>> widely used)? >>>>> >>>>> Best regards, >>>>> >>>>> Michel >>>>> _______________________________________________ >>>>> ceph-users mailing list -- ceph-users@ceph.io >>>>> To unsubscribe send an email to ceph-users-le...@ceph.io >>>> _______________________________________________ >>>> ceph-users mailing list -- ceph-users@ceph.io >>>> To unsubscribe send an email to ceph-users-le...@ceph.io >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@ceph.io >>> To unsubscribe send an email to ceph-users-le...@ceph.io >> >> _______________________________________________ >> ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an email to ceph-users-le...@ceph.io > > > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io