We're very happy to announce the first stable release of the Tentacle
series.

We express our gratitude to all members of the Ceph community who
contributed by proposing pull requests, testing this release,
providing feedback, and offering valuable suggestions.

We would like to especially thank some of our community members who helped
us test upgrades for the pre-released version of 20.2.0.
Your feedback and effort are greatly appreciated!

Note from the Ceph Infrastructure Team:

Part of our Standard Operating Procedure for Ceph releases is typically to
upgrade the Ceph cluster in the Ceph lab to the latest version before
announcing the release.
This is a production cluster, which we colloquially call the Long Running
Cluster (LRC), as it has existed and been updated for the past decade or
even longer.
For this release, we have decided to hold off on updating the LRC until we
have migrated some of the labs from Red Hat data centers to IBM.
We plan to update the LRC to Tentacle in the next calendar year. We believe
this decision should not be viewed as a reflection of our confidence in the
project, but rather as an attempt to focus on other priorities before the
migration.

---Highlights---

CephFS

* Directories may now be configured with case-insensitive or normalized
  directory entry names.
* Modifying the FS setting variable ``max_mds`` when a cluster is unhealthy
  now requires users to pass the confirmation flag
(``--yes-i-really-mean-it``).
* ``EOPNOTSUPP`` (Operation not supported) is now returned by the CephFS
FUSE
  client for ``fallocate`` for the default case (i.e. ``mode == 0``).

Crimson

* SeaStore Tech Preview: SeaStore object store is now deployable
  alongside Crimson-OSD, mainly for early testing and experimentation.
  Community feedback is encouraged to help with future improvements.

Dashboard

* Support has been added for NVMe/TCP gateway groups and multiple
  namespaces, multi-cluster management, OAuth 2.0 integration, and enhanced
  RGW/SMB features including multi-site automation, tiering, policies,
  lifecycles, notifications, and granular replication.

Integrated SMB support

* Ceph clusters now offer an SMB Manager module that works like the existing
  NFS subsystem. The new SMB support allows the Ceph cluster to
automatically
  create Samba-backed SMB file shares connected to CephFS. The ``smb``
module
  can configure both basic Active Directory domain or standalone user
  authentication. The Ceph cluster can host one or more virtual SMB clusters
  which can be truly clustered using Samba's CTDB technology. The ``smb``
  module requires a cephadm-enabled Ceph cluster and deploys container
images
  provided by the ``samba-container`` project. The Ceph dashboard can be
used
  to configure SMB clusters and shares. A new ``cephfs-proxy`` daemon is
  automatically deployed to improve scalability and memory usage when
connecting
  Samba to CephFS.

MGR

* Users now have the ability to force-disable always-on modules.
* The ``restful`` and ``zabbix`` modules (deprecated since 2020) have been
  officially removed.

RADOS

* FastEC: Long-anticipated performance and space amplification
  optimizations are added for erasure-coded pools.
* BlueStore: Improved compression and a new, faster WAL (write-ahead-log).
* Data Availability Score: Users can now track a data availability score
  for each pool in their cluster.
* OMAP: All components have been switched to the faster OMAP iteration
  interface, which improves RGW bucket listing and scrub operations.

RBD

* New live migration features: RBD images can now be instantly imported
  from another Ceph cluster (native format) or from a wide variety of
  external sources/formats.
* There is now support for RBD namespace remapping while mirroring between
  Ceph clusters.
* Several commands related to group and group snap info were added or
  improved, and ``rbd device map`` command now defaults to ``msgr2``.

RGW

* Added support for S3 ``GetObjectAttributes``.
* For compatibility with AWS S3, ``LastModified`` timestamps are now
truncated
  to the second. Note that during upgrade, users may observe these
timestamps
  moving backwards as a result.
* Bucket resharding now does most of its processing before it starts to
block
  write operations. This should significantly reduce the client-visible
impact
  of resharding on large buckets.
* The User Account feature introduced in Squid provides first-class support
for
  IAM APIs and policy. Our preliminary STS support was based on tenants, and
  exposed some IAM APIs to admins only. This tenant-level IAM functionality
is now
  deprecated in favor of accounts. While we'll continue to support the
tenant feature
  itself for namespace isolation, the following features will be removed no
sooner
  than the V release:
  - Tenant-level IAM APIs including CreateRole, PutRolePolicy and
PutUserPolicy,
  - Use of tenant names instead of accounts in IAM policy documents,
  - Interpretation of IAM policy without cross-account policy evaluation,
  - S3 API support for cross-tenant names such as
`Bucket='tenant:bucketname'`
  - STS Lite and `sts:GetSessionToken`.

We encourage you to read the full release notes at
https://ceph.io/en/news/blog/2025/v20-2-0-tentacle-released/

Getting Ceph
------------
* Git at git://github.com/ceph/ceph.git
* Tarball at https://download.ceph.com/tarballs/ceph_20.2.0.tar.gz
* Containers at https://quay.io/repository/ceph/ceph
* For packages, see https://docs.ceph.com/docs/master/install/get-packages/
* Release git sha1: 69f84cc2651aa259a15bc192ddaabd3baba07489


Did you know? Every Ceph release is built and tested on resources
funded directly by the non-profit Ceph Foundation.
If you would like to support this and our other efforts, please
consider joining now https://ceph.io/en/foundation/.
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to