Ceph Tentacle (v20)
This document lists known critical bugs affecting Ceph Tentacle (v20) releases.
OSD Crash When Enabling EC Optimizations on CephFS
Severity: High Affected Versions: 20.2.0 Bug Report: https://tracker.ceph.com/issues/71642
Description
OSDs crash when allow_ec_optimizations is enabled on an existing CephFS Erasure Coded (EC) data pool that does not have allow_ec_overwrites explicitly enabled. The crash occurs in ECTransaction::WritePlanObj when accessing a non-existent transaction key.
Recommendation
- Wait for future tentacle releases
- See the full post mortem for details
New or Recreated OSD Missing DB Device with Hybrid Spec
Severity: Medium
Affected Versions: 20.2.0, 20.2.1
Bug Report: https://tracker.ceph.com/issues/72696
Description
When using a hybrid OSD spec (HDDs for data, SSDs/NVMe for DB devices), newly created or recreated OSDs are deployed without a DB device. The ceph-volume lvm batch command issued by the orchestrator omits the --db-device argument.
The root cause is a regression introduced by the fix for tracker #68576: the ceph_device attribute in ceph-volume inventory JSON output was renamed to ceph_device_lvm. This causes the selector to treat all existing RocksDB volumes as unavailable, so it cannot assign DB devices when recreating OSDs.
Note: This bug also affects Ceph Squid (v19). See the Squid known bugs page for details.
Recommendation
- A patch is available in PR #65986
- Follow the bug tracker for fix and backport updates
FastEC Scrub Errors After Recovery
Severity: Medium Affected Versions: 20.2.0 Bug Report: https://tracker.ceph.com/issues/73184
Description
After recovering from the EC optimization crash, clusters may experience excessive scrub errors with messages like candidate size X info size Y mismatch. This is a secondary issue related to the FastEC code path.
Recommendation
- Wait for future tentacle releases
- Apply patches from PR #65872 and PR #65788
- See the full post mortem