Project

General

Profile

Actions

Tasks #51341

open

Steps to recover file system(s) after recovering the Ceph monitor store

Added by Ramana Raja almost 3 years ago. Updated over 2 years ago.

Status:
In Progress
Priority:
High
Assignee:
-
Category:
Administration/Usability
Target version:
-
% Done:

0%

Tags:
Reviewed:
Affected Versions:
Component(FS):
Labels (FS):
multifs, multimds
Pull request ID:

Description

In certain rare cases, all the Ceph Monitors might end up with corrupted Monitor stores. The Monitor stores can be recovered from the OSDs using the monmap tool, and Monitors can be brought back online. MDSMaps however are lost. Additional steps are required to bring back the file system(s) and the MDSs. Steps for recovery will differ between single active MDS and multi active MDS file systems, single and multi file systems in a Ceph cluster

The steps identified to bring back a single active MDS file system post recovery of Monitor stores and Monitor is as follows:

- Ensure all MDSs are stopped on cluster
systemctl stop ceph-mds@<mds-id>

- Force create ceph file system using existing file system pools
ceph fs new <fs-name> <cephfs-metadata-poolname> <cephfs-data-poolname> --force

- Reset file system
ceph fs reset <fs-name> --yes-i-really-mean-it

- Restart MDSs
systemctl start ceph-mds@<mds-id>

For multi MDS file system, it may be possible to recover by marking the new file system unjoinable, setting max_mds, and then marking it joinable.


Related issues 1 (0 open1 closed)

Related to CephFS - Feature #51716: Add option in `fs new` command to start rank 0 in failed stateResolvedRamana Raja

Actions
Actions

Also available in: Atom PDF