Ok the two issues I see with reef release v18.2.5

- Subnet check seems to be ipv4 only which leads to e.g "public address is
not in 'fd01:1:f00f:443::/64' subnet" warnings on ipv6 only clusters.


   -

   common/pick_address: check if address in subnet all public address (
   pr#57590 <https://github.com/ceph/ceph/pull/57590>, Nitzan Mordechai)
   -

   osd: Report health error if OSD public address is not within subnet (
   pr#55697 <https://github.com/ceph/ceph/pull/55697>, Prashant D)

- cryptsetup version check isn't working at least in the container image of
v18.2.5 (
https://github.com/ceph/ceph/blob/reef/src/ceph-volume/ceph_volume/util/encryption.py)
which leads to encrypted osds not starting due to "'Error while checking
cryptsetup version.\n', '`cryptsetup --version` output:\n', 'cryptsetup
2.7.2 flags: UDEV BLKID KEYRING FIPS KERNEL_CAPI PWQUALITY '"

Happy to help with logs etc.

BR

Stephan



Am Fr., 11. Apr. 2025 um 09:11 Uhr schrieb Stephan Hohn <
step...@gridscale.io>:

> Hi all,
>
> started an update on our staging cluster from v18.2.4 --> v18.2.5
>
> ~# ceph orch upgrade start --image quay.io/ceph/ceph:v18.2.5 Mons and Mgr
> went fine but osds not coming up with v18.2.5 Apr 11 06:59:56
> 0cc47a6df14e podman[263290]: 2025-04-11 06:59:56.697993041 +0000 UTC
> m=+0.057869056 image pull  quay.io/ceph/ceph:v18.2.5
> Apr 11 06:59:56 0cc47a6df14e podman[263290]: 2025-04-11 06:59:56.778833855
> +0000 UTC m=+0.138709869 container init
> 5db97f7e32705cc0e8fee1bc5741dfbd97ffa430b8fb5a1cfe19b768aed78b23 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20, CEPH_GIT_REPO=
> https://github.com/ceph/ceph.git, OSD_FLAVOR=default,
> org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=
> https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/,
> org.opencontainers.image.documentation=https://docs.ceph.com/,
> CEPH_REF=reef, org.label-schema.vendor=CentOS, ceph=True,
> org.label-schema.name=CentOS Stream 9 Base Image,
> io.buildah.version=1.39.3,
> CEPH_SHA1=a5b0e13f9c96f3b45f596a95ad098f51ca0ccce1, FROM_IMAGE=
> quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph
> Release Team <ceph-maintain...@ceph.io>, org.label-schema.license=GPLv2,
> org.label-schema.build-date=20250325)
> Apr 11 06:59:56 0cc47a6df14e podman[263290]: 2025-04-11 06:59:56.790749299
> +0000 UTC m=+0.150625308 container start
> 5db97f7e32705cc0e8fee1bc5741dfbd97ffa430b8fb5a1cfe19b768aed78b23 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20,
> org.label-schema.build-date=20250325, org.opencontainers.image.authors=Ceph
> Release Team <ceph-maintain...@ceph.io>, org.label-schema.license=GPLv2,
> org.label-schema.schema-version=1.0, ceph=True, CEPH_REF=reef,
> CEPH_GIT_REPO=https://github.com/ceph/ceph.git, OSD_FLAVOR=default,
> org.label-schema.name=CentOS Stream 9 Base Image,
> org.opencontainers.image.documentation=https://docs.ceph.com/,
> GANESHA_REPO_BASEURL=
> https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/,
> io.buildah.version=1.39.3,
> CEPH_SHA1=a5b0e13f9c96f3b45f596a95ad098f51ca0ccce1, FROM_IMAGE=
> quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS)
> Apr 11 06:59:56 0cc47a6df14e bash[263290]:
> 5db97f7e32705cc0e8fee1bc5741dfbd97ffa430b8fb5a1cfe19b768aed78b23
> Apr 11 06:59:56 0cc47a6df14e
> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20[263380]:
> 2025-04-11T06:59:56.817+0000 7b72d0abc740 -1
> bluestore(/var/lib/ceph/osd/ceph-20/block) _read_bdev_label failed to open
> /var/lib/ceph/osd/ceph-20/block: (2) No such file or directory
> Apr 11 06:59:56 0cc47a6df14e
> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20[263380]:
> 2025-04-11T06:59:56.817+0000 7b72d0abc740 -1  ** ERROR: unable to open OSD
> superblock on /var/lib/ceph/osd/ceph-20: (2) No such file or directory
> Apr 11 06:59:56 0cc47a6df14e systemd[1]: Started Ceph osd.20 for
> 03977a23-f00f-4bb0-b9a7-de57f40ba853.
> Apr 11 06:59:56 0cc47a6df14e podman[263399]: 2025-04-11 06:59:56.90105365
> +0000 UTC m=+0.076310419 container died
> 5db97f7e32705cc0e8fee1bc5741dfbd97ffa430b8fb5a1cfe19b768aed78b23 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20)
> Apr 11 06:59:56 0cc47a6df14e podman[263399]: 2025-04-11 06:59:56.948423169
> +0000 UTC m=+0.123679914 container remove
> 5db97f7e32705cc0e8fee1bc5741dfbd97ffa430b8fb5a1cfe19b768aed78b23 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20, ceph=True,
> io.buildah.version=1.39.3, org.label-schema.name=CentOS Stream 9 Base
> Image, org.opencontainers.image.authors=Ceph Release Team <
> ceph-maintain...@ceph.io>, CEPH_REF=reef, GANESHA_REPO_BASEURL=
> https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/,
> CEPH_SHA1=a5b0e13f9c96f3b45f596a95ad098f51ca0ccce1, FROM_IMAGE=
> quay.io/centos/centos:stream9, OSD_FLAVOR=default,
> org.label-schema.build-date=20250325, org.label-schema.vendor=CentOS,
> org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=
> https://docs.ceph.com/, org.label-schema.license=GPLv2, CEPH_GIT_REPO=
> https://github.com/ceph/ceph.git)
> Apr 11 06:59:56 0cc47a6df14e systemd[1]:
> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@osd.20.service: Main process
> exited, code=exited, status=1/FAILURE
> Apr 11 06:59:57 0cc47a6df14e podman[263966]:
> Apr 11 06:59:57 0cc47a6df14e podman[263966]: 2025-04-11 06:59:57.495704469
> +0000 UTC m=+0.105177519 container create
> d96a2746c9b6ac37f42e1beaac9f572d22558c16d662dfaff994d1d90c611ad8 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20-deactivate,
> CEPH_GIT_REPO=https://github.com/ceph/ceph.git,
> CEPH_SHA1=a5b0e13f9c96f3b45f596a95ad098f51ca0ccce1, GANESHA_REPO_BASEURL=
> https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/,
> org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9
> Base Image, FROM_IMAGE=quay.io/centos/centos:stream9,
> io.buildah.version=1.39.3, org.opencontainers.image.documentation=
> https://docs.ceph.com/, org.label-schema.schema-version=1.0,
> org.label-schema.build-date=20250325, ceph=True, CEPH_REF=reef,
> org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph
> Release Team <ceph-maintain...@ceph.io>, OSD_FLAVOR=default)
> Apr 11 06:59:57 0cc47a6df14e podman[263966]: 2025-04-11 06:59:57.447692243
> +0000 UTC m=+0.057165313 image pull  quay.io/ceph/ceph:v18.2.5
> Apr 11 06:59:57 0cc47a6df14e podman[263966]: 2025-04-11 06:59:57.532516757
> +0000 UTC m=+0.141989804 container init
> d96a2746c9b6ac37f42e1beaac9f572d22558c16d662dfaff994d1d90c611ad8 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20-deactivate,
> CEPH_REF=reef, org.label-schema.license=GPLv2,
> org.label-schema.schema-version=1.0, CEPH_GIT_REPO=
> https://github.com/ceph/ceph.git, GANESHA_REPO_BASEURL=
> https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/,
> org.label-schema.vendor=CentOS, OSD_FLAVOR=default,
> org.opencontainers.image.authors=Ceph Release Team <
> ceph-maintain...@ceph.io>,
> CEPH_SHA1=a5b0e13f9c96f3b45f596a95ad098f51ca0ccce1, 
> org.label-schema.name=CentOS
> Stream 9 Base Image, org.opencontainers.image.documentation=
> https://docs.ceph.com/, ceph=True, io.buildah.version=1.39.3,
> org.label-schema.build-date=20250325, FROM_IMAGE=
> quay.io/centos/centos:stream9)
> Apr 11 06:59:57 0cc47a6df14e podman[263966]: 2025-04-11 06:59:57.542542891
> +0000 UTC m=+0.152015938 container start
> d96a2746c9b6ac37f42e1beaac9f572d22558c16d662dfaff994d1d90c611ad8 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20-deactivate,
> GANESHA_REPO_BASEURL=
> https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/,
> org.label-schema.build-date=20250325, OSD_FLAVOR=default, CEPH_REF=reef,
> ceph=True, org.label-schema.name=CentOS Stream 9 Base Image,
> org.opencontainers.image.authors=Ceph Release Team <
> ceph-maintain...@ceph.io>,
> CEPH_SHA1=a5b0e13f9c96f3b45f596a95ad098f51ca0ccce1,
> org.label-schema.vendor=CentOS, io.buildah.version=1.39.3,
> org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=
> quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0,
> org.label-schema.license=GPLv2, CEPH_GIT_REPO=
> https://github.com/ceph/ceph.git)
> Apr 11 06:59:57 0cc47a6df14e podman[263966]: 2025-04-11 06:59:57.542633919
> +0000 UTC m=+0.152106981 container attach
> d96a2746c9b6ac37f42e1beaac9f572d22558c16d662dfaff994d1d90c611ad8 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20-deactivate,
> FROM_IMAGE=quay.io/centos/centos:stream9,
> org.label-schema.schema-version=1.0,
> CEPH_SHA1=a5b0e13f9c96f3b45f596a95ad098f51ca0ccce1, OSD_FLAVOR=default,
> org.opencontainers.image.documentation=https://docs.ceph.com/,
> org.label-schema.vendor=CentOS, io.buildah.version=1.39.3, ceph=True,
> org.label-schema.license=GPLv2, org.label-schema.build-date=20250325,
> org.opencontainers.image.authors=Ceph Release Team <
> ceph-maintain...@ceph.io>, GANESHA_REPO_BASEURL=
> https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/,
> CEPH_GIT_REPO=https://github.com/ceph/ceph.git, CEPH_REF=reef,
> org.label-schema.name=CentOS Stream 9 Base Image)
> Apr 11 06:59:58 0cc47a6df14e
> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20-deactivate[264056]:
> Running command: /usr/sbin/cryptsetup remove
> UL0ODv-5ESt-FfPX-aGAK-0lfX-hRF5-14K71x
> Apr 11 06:59:58 0cc47a6df14e bash[263966]: Running command:
> /usr/sbin/cryptsetup remove UL0ODv-5ESt-FfPX-aGAK-0lfX-hRF5-14K71x
> Apr 11 06:59:58 0cc47a6df14e
> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20-deactivate[264056]:
>  stderr: Device UL0ODv-5ESt-FfPX-aGAK-0lfX-hRF5-14K71x is not active.
> Apr 11 06:59:58 0cc47a6df14e bash[263966]:  stderr: Device
> UL0ODv-5ESt-FfPX-aGAK-0lfX-hRF5-14K71x is not active.
> Apr 11 06:59:58 0cc47a6df14e
> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20-deactivate[264056]: -->
> command returned non-zero exit status: 4
> Apr 11 06:59:58 0cc47a6df14e bash[263966]: --> command returned non-zero
> exit status: 4
> Apr 11 06:59:58 0cc47a6df14e podman[263966]: 2025-04-11 06:59:58.401862967
> +0000 UTC m=+1.011336030 container died
> d96a2746c9b6ac37f42e1beaac9f572d22558c16d662dfaff994d1d90c611ad8 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20-deactivate)
> Apr 11 06:59:58 0cc47a6df14e podman[263966]: 2025-04-11 06:59:58.456490831
> +0000 UTC m=+1.065963881 container remove
> d96a2746c9b6ac37f42e1beaac9f572d22558c16d662dfaff994d1d90c611ad8 (image=
> quay.io/ceph/ceph:v18.2.5,
> name=ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd-20-deactivate,
> org.opencontainers.image.authors=Ceph Release Team <
> ceph-maintain...@ceph.io>, CEPH_GIT_REPO=https://github.com/ceph/ceph.git,
> org.label-schema.build-date=20250325, GANESHA_REPO_BASEURL=
> https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/,
> io.buildah.version=1.39.3, org.label-schema.schema-version=1.0, ceph=True,
> org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=
> https://docs.ceph.com/, OSD_FLAVOR=default, CEPH_REF=reef,
> org.label-schema.license=GPLv2,
> CEPH_SHA1=a5b0e13f9c96f3b45f596a95ad098f51ca0ccce1, 
> org.label-schema.name=CentOS
> Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9)
> Apr 11 06:59:58 0cc47a6df14e systemd[1]:
> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@osd.20.service: Failed with
> result 'exit-code'.
> Apr 11 06:59:58 0cc47a6df14e systemd[1]:
> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@osd.20.service: Consumed 2.252s
> CPU time.
>
>
> I was able to bring osds up again by redewploying with the old container
> image v18.2.4
>
> ~# ceph orch daemon redeploy osd.16 --image quay.io/ceph/ceph:v18.2.4
>
> Everything is up again but seeing the this error now. ~# ceph health
> detail HEALTH_ERR noout flag(s) set; 24 osds(s) are not reachable [WRN]
> OSDMAP_FLAGS: noout flag(s) set [ERR] OSD_UNREACHABLE: 24 osds(s) are not
> reachable osd.0's public address is not in 'fd01:1:f00f:443::/64' subnet
> osd.1's public address is not in 'fd01:1:f00f:443::/64' subnet
> ...
>
> ~# ceph config dump | grep netw global advanced public_network
> fd01:1:f00f:443::/64 * mon advanced public_network fd01:1:f00f:443::/64 *
> ~# ss -tulpn |grep osd tcp LISTEN 0 512 [fd01:1:f00f:443::10]:6828 [::]:*
> users:(("ceph-osd",pid=204834,fd=22)) tcp LISTEN 0 512
> [fd01:1:f00f:443::10]:6829 [::]:* users:(("ceph-osd",pid=204834,fd=23))
>
> Looks like some see the same problem with latest squid release v19.2.2 as
> well.
> BR Stephan
>
>
>
>
>
> Am Do., 10. Apr. 2025 um 22:45 Uhr schrieb Yuri Weinstein <
> ywein...@redhat.com>:
>
>> We're happy to announce the 5th point release in the Reef series.
>>
>> We recommend users to update to this release.
>> For detailed release notes with links & changelog please refer to the
>> official blog entry at
>> https://ceph.io/en/news/blog/2025/v18-2-5-reef-released/
>>
>>
>> Notable Changes
>> ---------------
>> * RBD: The ``try-netlink`` mapping option for rbd-nbd has become the
>> default
>>   and is now deprecated. If the NBD netlink interface is not supported by
>> the
>>   kernel, then the mapping is retried using the legacy ioctl interface.
>>
>> * RADOS: A new command, `ceph osd rm-pg-upmap-primary-all`, has been
>> added that allows
>>   users to clear all pg-upmap-primary mappings in the osdmap when desired.
>>
>>   Related trackers:
>>    - https://tracker.ceph.com/issues/67179
>>    - https://tracker.ceph.com/issues/66867
>>
>> Getting Ceph
>> ------------
>> * Git at git://github.com/ceph/ceph.git
>> * Tarball at https://download.ceph.com/tarballs/ceph_18.2.5.orig.tar.gz
>> * Containers at https://quay.io/repository/ceph/ceph
>> * For packages, see https://docs.ceph.com/en/latest/install/get-packages/
>> * Release git sha1: a5b0e13f9c96f3b45f596a95ad098f51ca0ccce1
>> _______________________________________________
>> Ceph-announce mailing list -- ceph-annou...@ceph.io
>> To unsubscribe send an email to ceph-announce-le...@ceph.io
>>
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to