Project

General

Profile

Actions

Bug #53542

open

Ceph Metadata Pool disk throughput usage increasing

Added by Andras Sali over 2 years ago. Updated over 2 years ago.

Status:
Fix Under Review
Priority:
Normal
Assignee:
Category:
-
Target version:
% Done:

0%

Source:
Tags:
Backport:
pacific
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(FS):
MDS
Labels (FS):
task(medium)
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

Hi All,

We have been observing that if we let our MDS run for some time, the bandwidth usage of the disks in the metadata pool starts increasing significantly (whilst IOPS is about constant), even though the number of clients, the workloads or anything else doesn't change.

However, after restarting the MDS, the issue goes away for some time and the same workloads require 1/10th of the metadata disk bandwidth whilst doing the same IOPS.

We run our CephFS cluster in a cloud environment where the disk throughput / bandwidth capacity is quite expensive to increase and we are hitting bandwidth / throughput limits, even though we still have a lot of IOPS capacity left.

We suspect that somehow the journaling of the MDS becomes more extensive (i.e. larger journal updates for each operation), but we couldn't really pin down which parameter might affect this.

I attach a plot of how the Bytes / Operation (throughput in bytes per sec / IOPS) evolves over time, when we restart the MDS, it drops to around 32kb (even though the min block size for the metadata pool OSDs is 4kb in our settings) and then increases over time to around 300kb.

Any insight into whether this is "normal" behaviour or how to tune this would be really appreciated.


Files

Bytes_per_op.png (10.6 KB) Bytes_per_op.png Bytes per operation Andras Sali, 12/08/2021 04:54 PM

Related issues 3 (2 open1 closed)

Related to CephFS - Bug #40002: mds: not trim log under heavy loadFix Under ReviewXiubo Li

Actions
Related to CephFS - Bug #52280: Mds crash and fails with assert on prepare_new_inodeResolvedXiubo Li

Actions
Related to CephFS - Bug #53623: mds: LogSegment will only save one ESubtreeMap event if the ESubtreeMap event size is large enough.Fix Under ReviewXiubo Li

Actions
Actions

Also available in: Atom PDF