Hi,
on one of my test clusters the upgrade from 19.2.3 to 20.2.0 failed, I
didn't look too deep into that since it's a single node cluster. So I
decided to bootstrap a fresh cluster:
soc9-ceph:~ # cephadm --image quay.io/ceph/ceph:v20.2.0 bootstrap
--mon-ip 192.168.124.186 --single-host-defaults
--allow-mismatched-release --skip-firewalld --allow-overwrite
--skip-monitoring
There's no error in the terminal output, the bootstrap command
finishes successfully (including adding hosts). But when I tried to
add OSDs, I noticed that the host is not present:
soc9-ceph:~ # ceph orch host label add soc9-ceph osd
host soc9-ceph does not exist
-mgr-soc9-ceph-gssfzo[3578085]:
orchestrator._interface.OrchestratorError: check-host failed:
-mgr-soc9-ceph-gssfzo[3578085]: Traceback (most recent call last):
-mgr-soc9-ceph-gssfzo[3578085]: File
"/usr/lib64/python3.6/runpy.py", line 193, in _run_module_as_main
-mgr-soc9-ceph-gssfzo[3578085]: "__main__", mod_spec)
-mgr-soc9-ceph-gssfzo[3578085]: File
"/usr/lib64/python3.6/runpy.py", line 85, in _run_code
-mgr-soc9-ceph-gssfzo[3578085]: exec(code, run_globals)
-mgr-soc9-ceph-gssfzo[3578085]: File
"/var/lib/ceph/c46111cc-c526-11f0-9577-fa163e2ad8c5/cephadm.ed5a13ad26f7f55dd30e9b63855e4e581fd>
-mgr-soc9-ceph-gssfzo[3578085]: File
"/var/lib/ceph/c46111cc-c526-11f0-9577-fa163e2ad8c5/cephadm.ed5a13ad26f7f55dd30e9b63855e4e581fd>
-mgr-soc9-ceph-gssfzo[3578085]: ImportError: cannot import name 'TypedDict'
-mgr-soc9-ceph-gssfzo[3578085]: debug 2025-11-19T09:05:32.478+0000
7f7c97228640 -1 mgr.server reply reply (22) Invalid argument check->
-mgr-soc9-ceph-gssfzo[3578085]: Traceback (most recent call last):
-mgr-soc9-ceph-gssfzo[3578085]: File
"/usr/lib64/python3.6/runpy.py", line 193, in _run_module_as_main
-mgr-soc9-ceph-gssfzo[3578085]: "__main__", mod_spec)
-mgr-soc9-ceph-gssfzo[3578085]: File
"/usr/lib64/python3.6/runpy.py", line 85, in _run_code
-mgr-soc9-ceph-gssfzo[3578085]: exec(code, run_globals)
-mgr-soc9-ceph-gssfzo[3578085]: File
"/var/lib/ceph/c46111cc-c526-11f0-9577-fa163e2ad8c5/cephadm.ed5a13ad26f7f55dd30e9b63855e4e581fd>
-mgr-soc9-ceph-gssfzo[3578085]: File
"/var/lib/ceph/c46111cc-c526-11f0-9577-fa163e2ad8c5/cephadm.ed5a13ad26f7f55dd30e9b63855e4e581fd>
-mgr-soc9-ceph-gssfzo[3578085]: ImportError: cannot import name 'TypedDict'
Maybe this doesn't work because of python3.6 or something, I don't
really have the time to check right now, but the bootstrap itself
should abort and show an error when some of the steps failed. Should I
create a tracker for this?
Thanks,
Eugen
Zitat von Yuri Weinstein <[email protected]>:
We're very happy to announce the first stable release of the Tentacle
series.
We express our gratitude to all members of the Ceph community who
contributed by proposing pull requests, testing this release,
providing feedback, and offering valuable suggestions.
We would like to especially thank some of our community members who helped
us test upgrades for the pre-released version of 20.2.0.
Your feedback and effort are greatly appreciated!
Note from the Ceph Infrastructure Team:
Part of our Standard Operating Procedure for Ceph releases is typically to
upgrade the Ceph cluster in the Ceph lab to the latest version before
announcing the release.
This is a production cluster, which we colloquially call the Long Running
Cluster (LRC), as it has existed and been updated for the past decade or
even longer.
For this release, we have decided to hold off on updating the LRC until we
have migrated some of the labs from Red Hat data centers to IBM.
We plan to update the LRC to Tentacle in the next calendar year. We believe
this decision should not be viewed as a reflection of our confidence in the
project, but rather as an attempt to focus on other priorities before the
migration.
---Highlights---
CephFS
* Directories may now be configured with case-insensitive or normalized
directory entry names.
* Modifying the FS setting variable ``max_mds`` when a cluster is unhealthy
now requires users to pass the confirmation flag
(``--yes-i-really-mean-it``).
* ``EOPNOTSUPP`` (Operation not supported) is now returned by the CephFS
FUSE
client for ``fallocate`` for the default case (i.e. ``mode == 0``).
Crimson
* SeaStore Tech Preview: SeaStore object store is now deployable
alongside Crimson-OSD, mainly for early testing and experimentation.
Community feedback is encouraged to help with future improvements.
Dashboard
* Support has been added for NVMe/TCP gateway groups and multiple
namespaces, multi-cluster management, OAuth 2.0 integration, and enhanced
RGW/SMB features including multi-site automation, tiering, policies,
lifecycles, notifications, and granular replication.
Integrated SMB support
* Ceph clusters now offer an SMB Manager module that works like the existing
NFS subsystem. The new SMB support allows the Ceph cluster to
automatically
create Samba-backed SMB file shares connected to CephFS. The ``smb``
module
can configure both basic Active Directory domain or standalone user
authentication. The Ceph cluster can host one or more virtual SMB clusters
which can be truly clustered using Samba's CTDB technology. The ``smb``
module requires a cephadm-enabled Ceph cluster and deploys container
images
provided by the ``samba-container`` project. The Ceph dashboard can be
used
to configure SMB clusters and shares. A new ``cephfs-proxy`` daemon is
automatically deployed to improve scalability and memory usage when
connecting
Samba to CephFS.
MGR
* Users now have the ability to force-disable always-on modules.
* The ``restful`` and ``zabbix`` modules (deprecated since 2020) have been
officially removed.
RADOS
* FastEC: Long-anticipated performance and space amplification
optimizations are added for erasure-coded pools.
* BlueStore: Improved compression and a new, faster WAL (write-ahead-log).
* Data Availability Score: Users can now track a data availability score
for each pool in their cluster.
* OMAP: All components have been switched to the faster OMAP iteration
interface, which improves RGW bucket listing and scrub operations.
RBD
* New live migration features: RBD images can now be instantly imported
from another Ceph cluster (native format) or from a wide variety of
external sources/formats.
* There is now support for RBD namespace remapping while mirroring between
Ceph clusters.
* Several commands related to group and group snap info were added or
improved, and ``rbd device map`` command now defaults to ``msgr2``.
RGW
* Added support for S3 ``GetObjectAttributes``.
* For compatibility with AWS S3, ``LastModified`` timestamps are now
truncated
to the second. Note that during upgrade, users may observe these
timestamps
moving backwards as a result.
* Bucket resharding now does most of its processing before it starts to
block
write operations. This should significantly reduce the client-visible
impact
of resharding on large buckets.
* The User Account feature introduced in Squid provides first-class support
for
IAM APIs and policy. Our preliminary STS support was based on tenants, and
exposed some IAM APIs to admins only. This tenant-level IAM functionality
is now
deprecated in favor of accounts. While we'll continue to support the
tenant feature
itself for namespace isolation, the following features will be removed no
sooner
than the V release:
- Tenant-level IAM APIs including CreateRole, PutRolePolicy and
PutUserPolicy,
- Use of tenant names instead of accounts in IAM policy documents,
- Interpretation of IAM policy without cross-account policy evaluation,
- S3 API support for cross-tenant names such as
`Bucket='tenant:bucketname'`
- STS Lite and `sts:GetSessionToken`.
We encourage you to read the full release notes at
https://ceph.io/en/news/blog/2025/v20-2-0-tentacle-released/
Getting Ceph
------------
* Git at git://github.com/ceph/ceph.git
* Tarball at https://download.ceph.com/tarballs/ceph_20.2.0.tar.gz
* Containers at https://quay.io/repository/ceph/ceph
* For packages, see https://docs.ceph.com/docs/master/install/get-packages/
* Release git sha1: 69f84cc2651aa259a15bc192ddaabd3baba07489
Did you know? Every Ceph release is built and tested on resources
funded directly by the non-profit Ceph Foundation.
If you would like to support this and our other efforts, please
consider joining now https://ceph.io/en/foundation/.
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]