How to remove failed OSD & reuse it?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi guys.

I've browsing through the net in a search of a relatively clear "howto" but I failed to find one. It's rather many, sometimes different notes/thoughts on how to deal with such/similar situation. Having a 3-node containerized cluster which lost osd - it crushed, there is nothing wrong with the node, nothing wrong with the disk, but never mind that.
Is there a howto which covers containerized environment?
One example I followed is: https://docs.redhat.com/en/documentation/red_hat_ceph_storage/1.2.3/html/red_hat_ceph_administration_guide/setting_unsetting_overrides
but it is - to me - clear, what to do with "broken" containers.
I'm got to:
-> $ ceph osd tree
ID  CLASS  WEIGHT   TYPE NAME          STATUS  REWEIGHT  PRI-AFF
-1         0.68359  root default
-3               0      host podster1
-7         0.34180      host podster2
 2    hdd  0.04880          osd.2          up   1.00000  1.00000  4    hdd  0.29300          osd.4          up   1.00000  1.00000
-5         0.34180      host podster3
 1    hdd  0.04880          osd.1          up   1.00000  1.00000  5    hdd  0.29300          osd.5          up   1.00000  1.00000

yet:
-> $ ceph orch ps --daemon-type=osd
NAME   HOST                PORTS  STATUS         REFRESHED  AGE MEM USE  MEM LIM  VERSION    IMAGE ID      CONTAINER ID osd.0  podster1.mine.priv         error             7m ago 3w        -    4096M  <unknown>  <unknown> <unknown> osd.1  podster3.mine.priv         running (25h)     7m ago 3w     942M    4096M  19.2.3     aade1b12b8e6  d71051ea79dc osd.2  podster2.mine.priv         running (6d)      7m ago   3w 1192M    4096M  19.2.3     aade1b12b8e6  e8d05142a73a osd.3  podster1.mine.priv         error             7m ago 2w        -    4096M  <unknown>  <unknown> <unknown> osd.4  podster2.mine.priv         running (6d)      7m ago   2w 3293M    4096M  19.2.3     aade1b12b8e6  6116277f69d1 osd.5  podster3.mine.priv         running (25h)     7m ago   2w 2963M    4096M  19.2.3     aade1b12b8e6  d671bf73cc01

what would be next bits needed to complete such removal&reuse/re-create of osd(s)? p.s. This a 'lab' setup so I'm not worried, but it'd be great to complete this process in a healthy manner.
many thanks, L.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux