diff --git a/2024-08-10-recovering-ceph-cluster.md b/2024-08-10-recovering-ceph-cluster.md index e5661ca..d4f4f34 100644 --- a/2024-08-10-recovering-ceph-cluster.md +++ b/2024-08-10-recovering-ceph-cluster.md @@ -33,8 +33,7 @@ the old OSDs from the tree and letting replication work its magic. I noticed although the block pool was replicated, we had lost all our RADOS object storage. - -![](./images/ec-2-1.png) +![](/b/images/ec-2-1.png) The erasure-coding profile was `k=2, m=1`. That meant we could only lose 2 OSDs, which had already happened. @@ -76,7 +75,7 @@ pods, I saw (but did not understand) the problem: `osd/ECUtil.h: 34: FAILED ceph_assert(stripe_width % stripe_size == 0)` -![](./images/ec-3-2.png) +![](/b/images/ec-3-2.png) With the _"fixed"_ configuration, what I had actually done is split the object store pool across _5_ OSDs. We had _3_. @@ -534,6 +533,7 @@ I rebuilt the mon data, using the existing RocksDB kv store. This would have worked without the backup, but I was interested to see the `osdmaps` trimmed due to the other 2 removed OSDs. +``` [root@he-prod-k3s-controlplane-ch-a-1 ceph]# ceph-objectstore-tool --type bluestore --data-path /var/lib/ceph/osd/ceph-0/ --op update-mon-db --mon-store-path /tmp/mon-a/data/ osd.0 : 3099 osdmaps trimmed, 635 osdmaps added. ```