Project

General

Profile

Actions

Bug #54296

closed

OSDs using too much memory

Added by Ruben Kerkhof about 2 years ago. Updated almost 2 years ago.

Status:
Resolved
Priority:
Normal
Assignee:
-
Category:
-
Target version:
-
% Done:

0%

Source:
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(RADOS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

One of our customers upgraded from Nautilus to Octopus, and now a lot of his OSDs are using way more ram than allowed by the osd_memory_target.
Here's one:
$ sudo ceph daemon osd.75 config get osd_memory_target {
"osd_memory_target": "4294967296"
}
$ sudo ceph daemon osd.75 dump_mempools | jq '.mempool.total' {
"items": 51529102,
"bytes": 4589310658
}
$ ps -o rss -p $(pgrep -f '/usr/bin/ceph-osd -f --cluster ceph --id 75 --setuser ceph --setgroup ceph')
RSS
9134604

Some more details about his cluster:
All nodes are running 15.2.15.
He has hdd nodes in the default crush root, and ssd nodes in a separate crush tree.
The hdd osds are using approximately 4GB ram, only the SSD osds use double. He's running nightly snap trims only on the SSD-backed pools.

Please let me know what additional details I can provide.


Files

mempools.txt (3.21 KB) mempools.txt dump_mempools output Ruben Kerkhof, 02/16/2022 12:19 PM
pg-dump.txt.gz (224 KB) pg-dump.txt.gz Ruben Kerkhof, 02/17/2022 09:46 AM

Related issues 1 (0 open1 closed)

Is duplicate of RADOS - Bug #53729: ceph-osd takes all memory before oom on bootResolvedNitzan Mordechai

Actions
Actions

Also available in: Atom PDF