Ceph Quincy, a look into the future

Every year, a new Ceph version is released. In 2019, version 14 (Nautilus) was released, in 2020, version 15 (Octopus) and in 2021, version 16 called Pacific. These versions have an end of life date, so make sure you are up-to-date and operate the same version throughout your Ceph clusters. The end of life date […]

About RBD, Rados Block Device

For anybody enjoying Latin pop in the 2000s, RBD was a popular Mexican pop band from Mexico City labeled by EMI Virgin. The group achieved international success from 2004 until their separation in 2009 and sold over 15 million records worldwide, making them one of the best-selling Latin music artists of all time. The group […]

What about our Ceph Fundamentals training?

Currently, we are planning a new Ceph Fundamentals training course for November or December. Because of this, we thought it would be nice to share some more information on the topics of this training course. The 42on Ceph Fundamentals Training is a 2-day online, instructor-led training course. The training is led by one of our […]

Best practices for CephFS

File storage is a data storage format in which data is stored and managed as files within folders. Why would you choose to use file storage? Well, the advantages of file storage include the following: 1. User-friendly interface: A simple file management and sharing system which are easy to understand for human users which makes […]

How to handle large omap objects

Every once in a while a customer will ask me what to do with these messages: health: HEALTH_WARN1 large omap objects First let’s see what this means: Most of the time when seeing the large omap warning is in relation to the RGW workload. The RGW or Rados Gateway provides a S3 and/or SWIFT compatible […]

Protecting your Ceph pools against removal or property changes

One of the dangers of Ceph was that by accident you could remove a multi TerraByte pool and loose all the data. Although the CLI tools asked you for conformation, librados and all it’s bindings did not. Imagine explaining that you just removed a 200TB pool from your storage system due to a typo in […]

PowerDNS backend for a global RADOS Gateway namespace

At my hosting company PCextreme we are building a cloud offering based on Ceph and CloudStack. We call our cloud services Aurora. Our cloud services are composed out of two components: Compute and Objects. For our Aurora Objects service we use the RADOS Gateway from Ceph and we are using the Federated Config to create […]

Calculating RADOS objects for RBD images

Ceph’s RBD (RADOS Block Device) is just a thin wrapper on top of RADOS, the object store of Ceph. It stripes (by default) over 4MB objects in RADOS. It’s very simple to calculate which RADOS object corresponds with which sector on your RBD image/block device. First you have to find out the block device’s object […]

Changing the region of a RGW bucket

As of Ceph version 0.67 (Dumpling) the Ceph Object Gateway aka RADOS Gateway supports regions. This allows you to create a geo-replicated Amazon S3 compatible service. While working on a setup we decided later in the process that we wanted regions, but we already created about 50 buckets with data in them. We didn’t feel […]

Redundant Ceph monitors with Round Robin DNS

One of the unique features of Ceph is that it can be build without any Single Point of Failure. No single machine will take your cluster down when designed properly. Ceph’s monitors play a crucial part in this. To make them redundant you want a odd number of monitors, where 3 is more then sufficient […]