Questo contenuto non è disponibile nella lingua selezionata.
Chapter 7. Asynchronous errata updates
This section describes the bug fixes, known issues, and enhancements of the z-stream releases.
7.1. Red Hat Ceph Storage 8.0z1 Copia collegamentoCollegamento copiato negli appunti!
Red Hat Ceph Storage release 8.0z1 is now available. The security updates and bug fixes that are included in the update are listed in the RHSA-2024:10956 and RHSA-2024:10957 advisories.
7.2. Red Hat Ceph Storage 8.0z2 Copia collegamentoCollegamento copiato negli appunti!
Red Hat Ceph Storage release 8.0z2 is now available. The bug fixes that are included in the update are listed in the RHBA-2025:2457 and RHBA-2025:2458 advisories.
7.2.1. Enhancements Copia collegamentoCollegamento copiato negli appunti!
7.2.1.1. Ceph File System Copia collegamentoCollegamento copiato negli appunti!
quota.max_bytes is now set in more understandable size values.
Previously, the quota.max_bytes
value was set in bytes, resulting in often very large size values that were hard to set or change.
With this enhancement, the quota.max_bytes values can now be set with human-friendly values, such as K/Ki, M/Mi, G/Gi, or T/Ti. For example, 10GiB
or 100K
.
Directories configured for Ceph File System mirroring can now be listed using the ceph fs
command.
Previously, to list the directories configured for mirroring, users had to go through the peer list.
With this enhancement, users can now list the directories configured for mirroring using the ceph fs snapshot mirror ls FS_NAME
command. This provides a direct list of the directories configured for mirroring.
7.2.1.2. Ceph Object Gateway Copia collegamentoCollegamento copiato negli appunti!
Increased efficiency for ordered bucket listings with namespaced bucket index entries.
Previously, when ignoring namespaced bucket index entries, the code still accessed the ignored entry. As a result, unnecessary latency occurred in the ignored listings.
With this enhancement, the code now avoids accessing ignored entries, improving the efficiency of ordered bucket listings.
Users can be created without a key using the radosgw-admin command-line
Previously, there was no provision to create Ceph Object Gateway (RGW) users without a key using the radosgw-admin command-line. This feature was only available for the adminops.
With this enhancement, the provision to create RGW users without a key is provided on the command-line when --generate-key false
flag is set to radosgw-admin user create
.
7.2.1.3. RADOS Copia collegamentoCollegamento copiato negli appunti!
The admin socket command now works with default parameters.
Previously, listing allocator histogram = admin socket
with bluestore allocator fragmentation histogram did not work for bluefs-db
and bluefs-wal
with default parameters.
With this enhancement,the admin socket
command works with the default parameters.
Inspection of disk allocator state through the admin socket command
This enhancement provides a middle point between the allocator score, which gives a single number, and the allocator dump, which lists all free chunks.
As a result, the fragmentation histogram groups free chunks by size, giving some approximation of the allocator state. This gives a chance to estimate the severity of current fragmentation. The fragmentation histogram works for block/bluefs-db/bluefs-wal
allocators. The extra parameter disk_alloc
influences the calculation of how many free chunks are unaligned to the disk_alloc
boundary. Extra parameter num_buckets determines the size of the histogram, but the granularity remains the same. For example:
bluestore allocator fragmentation histogram block 4096 12
bluestore allocator fragmentation histogram block 4096 12
7.2.2. Known issues Copia collegamentoCollegamento copiato negli appunti!
7.2.2.1. Ceph Object Gateway Copia collegamentoCollegamento copiato negli appunti!
Ceph Object Gateway crashes when performing multipart upload using AWS CLI.
Currently, during a multipart upload using AWS CLI, RGW crashes due to new checksum algorithms and reporting behavior introduced in AWS S3 and AWS SDKs, specifically the new CRC64NVME checksum algorithm.
As a workaround, use an earlier version of the AWS CLI client.
For more information, see RGW services crashing during multipart upload in RGWPutObj_Cksum
function.
radosgw-admin user stats --reset-stats ERROR: value too large for defined data type.
Currently,the default size of RADOS RETURNVEC
(64 bytes) is too small for some environments running safe user reset stats.
As a workaround, increase osd_max_write_op_reply_len
to 128
(the largest value tested). With this change, user reset stats complete successfully.
7.3. Red Hat Ceph Storage 8.0z3 Copia collegamentoCollegamento copiato negli appunti!
Red Hat Ceph Storage release 8.0z3 is now available. The bug fixes that are included in the update are listed in the RHSA-2025:3635 and RHBA-2025:3638 advisories.
7.3.1. Enhancements Copia collegamentoCollegamento copiato negli appunti!
7.3.1.1. Cephadm utility Copia collegamentoCollegamento copiato negli appunti!
Core dumps for crashed services are now available
Previously, core dumps were not being generated or truncated during a crash and the dumps were hard to reproduce. As a result, valuable debugging information was lost for the user.
With this enhancement, the related core dump and cephadm sizes are increased and users can access core dumps for crashed services, as expected.
7.3.1.2. Ceph Object Gateway Copia collegamentoCollegamento copiato negli appunti!
Bucket notifications now send to a multi-node Kafka cluster
Previously, Ceph Object Gateway could only send a message to a single node Kafka cluster.
With this enhancement, bucket notifications can now send to a multi-node Kafka cluster. With multi-node Kafka cluster support, there is now proper utilization of the cluster’s high availability (HA). In cases that a node is in a down
state with other Kafka nodes being up
, messages can now be sent. In addition, since Ceph Object Gateway is now connected to each node, bucket notification failures no longer occur due to topic partitions not being replicated to all of the nodes.
7.4. Red Hat Ceph Storage 8.0z4 Copia collegamentoCollegamento copiato negli appunti!
Red Hat Ceph Storage release 8.0z4 is now available. The bug fixes that are included in the update are listed in the RHBA-2025:8259 and RHBA-2025:8260 advisories.
7.5. Red Hat Ceph Storage 8.0z5 Copia collegamentoCollegamento copiato negli appunti!
Red Hat Ceph Storage release 8.0z5 is now available. The bug fixes that are included in the update are listed in the RHBA-2025:8694 and RHBA-2025:8695 advisories.