Bug #45835
mds: OpenFileTable::prefetch_inodes during rejoin can cause out-of-memory
% Done:
0%
Source:
Community (dev)
Tags:
Backport:
octopus,nautilus
Regression:
No
Severity:
2 - major
Reviewed:
Affected Versions:
ceph-qa-suite:
fs
Component(FS):
MDS
Labels (FS):
Pull request ID:
Crash signature:
Description
We just upgraded from mimic v13.2.6 to nautilus v14.2.9 and the single active MDS was going out-of-memory during the rejoin step.
MDS cache size was 4GB, but the OOM occured on 16GB and even 32GB VMs.
`mds_wipe_sessions = true` did not help.
To recover the cluster I did `rados -p cephfs_metadata_pool rm mds0_openfiles.0` then the MDS could activate quickly.
Related issues
History
#1 Updated by Patrick Donnelly 8 months ago
- Subject changed from OpenFileTable::prefetch_inodes during rejoin can cause out-of-memory to mds: OpenFileTable::prefetch_inodes during rejoin can cause out-of-memory
- Status changed from New to Triaged
- Assignee set to Zheng Yan
- Target version set to v16.0.0
- Source set to Community (dev)
- Backport set to octopus,nautilus
#3 Updated by Dan van der Ster 4 months ago
The fix was merged. Something needed to start the backports process?
#4 Updated by Patrick Donnelly 4 months ago
- Status changed from Fix Under Review to Pending Backport
#5 Updated by Nathan Cutler 4 months ago
- Copied to Backport #47608: octopus: mds: OpenFileTable::prefetch_inodes during rejoin can cause out-of-memory added
#6 Updated by Nathan Cutler 4 months ago
- Copied to Backport #47609: nautilus: mds: OpenFileTable::prefetch_inodes during rejoin can cause out-of-memory added
#7 Updated by Nathan Cutler 4 months ago
Dan van der Ster wrote:
The fix was merged. Something needed to start the backports process?
@Dan, the "backporting process" has started, but it might take some time to finish, because the changeset looks pretty challenging to backport.