Project

General

Profile

Actions

Bug #45663

open

luminous to nautilus upgrade

Added by none none almost 4 years ago. Updated almost 2 years ago.

Status:
Triaged
Priority:
Normal
Assignee:
-
Category:
-
Target version:
-
% Done:

0%

Source:
Tags:
Backport:
pacific,octopus,nautilus
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(FS):
Labels (FS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

I have been using snapshots on cephfs since luminous, 1xfs and
1xactivemds and used an rsync on it for backup (mounted cephfs on osd node).
Under luminious I did not encounter any problems with this setup. I
think I was even snapshotting user dirs every 7 days having thousands
of snapshots (which I later heard, is not recommend and one should
stick below 400 or so?)

When upgrading to nautilus, this snapshot feature was disabled (that
is default in the upgrade). Did not notice nor expected this. When I
enabled again snapshotting. I had problems with the rsync backup. So I
reverted back to the slower ceph-fuse mount. I also brought down the
snapshots to 36, but I am still stuck with "clients failing to respond
to capability release", "clients failing to respond to cache pressure"
and "MDSs report slow requests"
Which is odd, since my use did not change since luminous.

I know about the issue of running ceph kernel clients on osd nodes.
But this node has enough memory.

[@ ~]# free -m
total used free shared buff/cache available
Mem: 160989 42452 73562 110 44974 117745
Swap: 10239 534 9705

Since the upgrade, I think I also have problems with nfs-ganesha, I don't think this is related to bad connection.

MDS_CLIENT_LATE_RELEASE 1 clients failing to respond to capability release
mdsa(mds.0): Client c01:cephfs.nfs failing to respond to capability release client_id: 4012632
MDS_CLIENT_RECALL 2 clients failing to respond to cache pressure
mdsa(mds.0): Client c04:cephfs.backup failing to respond to cache pressure client_id: 3905222
mdsa(mds.0): Client c01:cephfs.nfs failing to respond to cache pressure client_id: 4012617
MDS_SLOW_REQUEST 1 MDSs report slow requests
mdsa(mds.0): 1 slow requests are blocked > 30 secs

as requested here, opened tracker ticket
https://www.mail-archive.com/ceph-users@ceph.io/msg03592.html

Actions

Also available in: Atom PDF