ceph bug of the year 2020 - CERN
interesting insights into how dependency on external operating system libraries can affect the operation of Ceph.
interesting insights into how dependency on external operating system libraries can affect the operation of Ceph.
ceph tell 'osd.*' injectargs --osd-max-backfills=2 --osd-recovery-max-active=2
Start in small steps, observe the Ceph status, client IOPs and throughput and then continue to increase in small steps.
In the producton with regard to the applications and hardware infrastructure, we recommend setting these settings back to default as soon as possible.
For some time now there are no more official packages for Debian on the ceph.io site. The reason for this is the switch to a C++ version, which was only supported by Debian with Buster. All the more pleasing is the fact that Bernd Zeimetz has been working on the Ceph package for Debian since 28.11.2019 and is currently maintaining it for the current Nautilus versions from 14.2.4-1 for Bullseye and Buster backports.
See changelog of the project:
ceph daemon mon.clyso-mon1 sessions
If you are looking for the IP addresses for the output of ceph features.
With the Nautilus release PG splitting and merging was introduced and the following default values were set:
"osd_pool_default_pg_num": "8"
"osd_pool_default_pgp_num": "0"
Furthermore, the osd_pool_default_pg_num should be set to a value that makes sense for the respective Ceph cluster.
The value 0 of osd_pool_default_pgp_num now indicates that this value is automatically monitored by the Ceph cluster and adjusted according to the following criteria:
Starting in Nautilus, this second step is no longer necessary: as long as pgp_num and pg_num currently match, pgp_num will automatically track any pg_num changes. More importantly, the adjustment of pgp_num to migrate data and (eventually) converge to pg_num is done gradually to limit the data migration load on the system based on the new target_max_misplaced_ratio config option (which defaults to .05, or 5%). That is, by default, Ceph will try to have no more than 5% of the data in a “misplaced” state and queued for migration, limiting the impact on client workloads. ceph.com/rados/new-in-nautilus-pg-merging-and-autotuning/
Before the Nautilus release, the number of PGs had to be adjusted manually for the respective pools. With Nautilus, the Ceph Manager module pg_autoscaler can take over.
Since version Luminous 12.2.x, pg-upmap is available in the ceph mgr balancer module.
When commissioning new Ceph clusters, our standard tests also include measuring the I/O latency for RBD.
We also always measure the performance values for the entire stack. Over the years, we have seen the results of our hard work in improving ceph osd in various tests.
For our tests, we create a temporary work file and read random blocks with non-cached read operations from it.
We are now measuring latencies of 300 to 600 microseconds.
On the following page you will find the agenda of Cephalocon 2019 in Barcelona. In each subitem you will find the slides of the presentations.