Actions
Bug #58597
openThe MDS crashes when deleting a specific file
Status:
New
Priority:
Normal
Assignee:
-
Category:
fsck/damage handling
Target version:
-
% Done:
0%
Source:
Community (user)
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(FS):
MDS
Labels (FS):
crash
Pull request ID:
Crash signature (v1):
Crash signature (v2):
Description
-88> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache oldest old_inode is [28051,28051], done.
-87> 2023-01-29T04:23:09.960+0100 7f429ccd8700 20 mds.0.journal EMetaBlob::add_dir_context(0x55b33ecb9a80) reached unambig auth subtree, don't need at [dir 0x100 ~mds0/ [2,head] auth pv=175065187 v=175065185 cv=0/0 dir_auth=0
-86> 2023-01-29T04:23:09.960+0100 7f429ccd8700 20 mds.0.journal EMetaBlob::add_dir_context final:
-85> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache journal_cow_dentry follows head on [dentry #0x100/stray0 [2,head] auth (dversion lock) pv=175065186 v=175065158 ino=0x600 state=1610612736 | inodepin=1 dirty=1 0x55
-84> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache journal_cow_dentry follows 28921 < first on [inode 0x600 [...28922,head] ~mds0/stray0/ auth v175065158 pv175065186 ap=1 f(v38 m2023-01-29T04:14:47.829119+0100) n(v8
-83> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache journal_cow_dentry follows head on [dentry #0x1/docker/nextcloud-13-nzh/db~/mysql/innodb_index_stats.ibd [10000000031,head] auth (dn xlock x=1 by 0x55b33fa88c00) (d
-82> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache journal_cow_dentry follows 28921 < first on [dentry #0x1/docker/nextcloud-13-nzh/db~/mysql/innodb_index_stats.ibd [10000000031,head] auth (dn xlock x=1 by 0x55b33fa
-81> 2023-01-29T04:23:09.960+0100 7f429ccd8700 -1 ./src/mds/Server.cc: In function 'void Server::_unlink_local(MDRequestRef&, CDentry*, CDentry*)' thread 7f429ccd8700 time 2023-01-29T04:23:09.962505+0100
./src/mds/Server.cc: 7806: FAILED ceph_assert(in->first <= straydn->first)
ceph version 16.2.9 (a569859f5e07da0c4c39da81d5fb5675cd95da49) pacific (stable)
1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x124) [0x7f42a21b8fde]
2: /usr/lib/ceph/libceph-common.so.2(+0x251169) [0x7f42a21b9169]
3: (Server::_unlink_local(boost::intrusive_ptr<MDRequestImpl>&, CDentry*, CDentry*)+0x11ff) [0x55b33c58072f]
4: (MDSContext::complete(int)+0x5b) [0x55b33c80786b]
5: (void finish_contexts<std::vector<MDSContext*, std::allocator<MDSContext*> > >(ceph::common::CephContext*, std::vector<MDSContext*, std::allocator<MDSContext*> >&, int)+0x98) [0x55b33c4d7218]
6: (Locker::eval(CInode*, int, bool)+0x3de) [0x55b33c6d14de]
7: (Locker::handle_client_caps(boost::intrusive_ptr<MClientCaps const> const&)+0x21ef) [0x55b33c6dcddf]
8: (Locker::dispatch(boost::intrusive_ptr<Message const> const&)+0x224) [0x55b33c6ded34]
9: (MDSRank::_dispatch(boost::intrusive_ptr<Message const> const&, bool)+0x5c0) [0x55b33c4f63f0]
10: (MDSRankDispatcher::ms_dispatch(boost::intrusive_ptr<Message const> const&)+0x58) [0x55b33c4f69e8]
11: (MDSDaemon::ms_dispatch2(boost::intrusive_ptr<Message> const&)+0x1bf) [0x55b33c4d0b6f]
12: (Messenger::ms_deliver_dispatch(boost::intrusive_ptr<Message> const&)+0x468) [0x7f42a23e8cb8]
13: (DispatchQueue::entry()+0x5ef) [0x7f42a23e63bf]
14: (DispatchQueue::DispatchThread::entry()+0xd) [0x7f42a24a54bd]
15: /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7) [0x7f42a1f13ea7]
16: clone()
This is perfectly reproducible, unfortunately on a Production System.
Actions