Skip to main content

Ceph Tentacle (v20)

This document lists known critical bugs affecting Ceph Tentacle (v20) releases.

OSD Crash When Enabling EC Optimizations on CephFS

Severity: High Affected Versions: 20.2.0 Bug Report: https://tracker.ceph.com/issues/71642

Description

OSDs crash when allow_ec_optimizations is enabled on an existing CephFS Erasure Coded (EC) data pool that does not have allow_ec_overwrites explicitly enabled. The crash occurs in ECTransaction::WritePlanObj when accessing a non-existent transaction key.

Recommendation

New or Recreated OSD Missing DB Device with Hybrid Spec

Severity: Medium Affected Versions: 20.2.0, 20.2.1
Bug Report: https://tracker.ceph.com/issues/72696

Description

When using a hybrid OSD spec (HDDs for data, SSDs/NVMe for DB devices), newly created or recreated OSDs are deployed without a DB device. The ceph-volume lvm batch command issued by the orchestrator omits the --db-device argument.

The root cause is a regression introduced by the fix for tracker #68576: the ceph_device attribute in ceph-volume inventory JSON output was renamed to ceph_device_lvm. This causes the selector to treat all existing RocksDB volumes as unavailable, so it cannot assign DB devices when recreating OSDs.

Note: This bug also affects Ceph Squid (v19). See the Squid known bugs page for details.

Recommendation

  • A patch is available in PR #65986
  • Follow the bug tracker for fix and backport updates

FastEC Scrub Errors After Recovery

Severity: Medium Affected Versions: 20.2.0 Bug Report: https://tracker.ceph.com/issues/73184

Description

After recovering from the EC optimization crash, clusters may experience excessive scrub errors with messages like candidate size X info size Y mismatch. This is a secondary issue related to the FastEC code path.

Recommendation