Project

General

Profile

Actions

Bug #58597

open

The MDS crashes when deleting a specific file

Added by Tobias Reinhard over 1 year ago. Updated about 1 year ago.

Status:
New
Priority:
Normal
Assignee:
-
Category:
fsck/damage handling
Target version:
-
% Done:

0%

Source:
Community (user)
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(FS):
MDS
Labels (FS):
crash
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

   -88> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache  oldest old_inode is [28051,28051], done.
   -87> 2023-01-29T04:23:09.960+0100 7f429ccd8700 20 mds.0.journal EMetaBlob::add_dir_context(0x55b33ecb9a80) reached unambig auth subtree, don't need  at [dir 0x100 ~mds0/ [2,head] auth pv=175065187 v=175065185 cv=0/0 dir_auth=0
   -86> 2023-01-29T04:23:09.960+0100 7f429ccd8700 20 mds.0.journal EMetaBlob::add_dir_context final:
   -85> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache journal_cow_dentry follows head on [dentry #0x100/stray0 [2,head] auth (dversion lock) pv=175065186 v=175065158 ino=0x600 state=1610612736 | inodepin=1 dirty=1 0x55
   -84> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache journal_cow_dentry follows 28921 < first on [inode 0x600 [...28922,head] ~mds0/stray0/ auth v175065158 pv175065186 ap=1 f(v38 m2023-01-29T04:14:47.829119+0100) n(v8
   -83> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache journal_cow_dentry follows head on [dentry #0x1/docker/nextcloud-13-nzh/db~/mysql/innodb_index_stats.ibd [10000000031,head] auth (dn xlock x=1 by 0x55b33fa88c00) (d
   -82> 2023-01-29T04:23:09.960+0100 7f429ccd8700 10 mds.0.cache journal_cow_dentry follows 28921 < first on [dentry #0x1/docker/nextcloud-13-nzh/db~/mysql/innodb_index_stats.ibd [10000000031,head] auth (dn xlock x=1 by 0x55b33fa
   -81> 2023-01-29T04:23:09.960+0100 7f429ccd8700 -1 ./src/mds/Server.cc: In function 'void Server::_unlink_local(MDRequestRef&, CDentry*, CDentry*)' thread 7f429ccd8700 time 2023-01-29T04:23:09.962505+0100
./src/mds/Server.cc: 7806: FAILED ceph_assert(in->first <= straydn->first)

 ceph version 16.2.9 (a569859f5e07da0c4c39da81d5fb5675cd95da49) pacific (stable)
 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x124) [0x7f42a21b8fde]
 2: /usr/lib/ceph/libceph-common.so.2(+0x251169) [0x7f42a21b9169]
 3: (Server::_unlink_local(boost::intrusive_ptr<MDRequestImpl>&, CDentry*, CDentry*)+0x11ff) [0x55b33c58072f]
 4: (MDSContext::complete(int)+0x5b) [0x55b33c80786b]
 5: (void finish_contexts<std::vector<MDSContext*, std::allocator<MDSContext*> > >(ceph::common::CephContext*, std::vector<MDSContext*, std::allocator<MDSContext*> >&, int)+0x98) [0x55b33c4d7218]
 6: (Locker::eval(CInode*, int, bool)+0x3de) [0x55b33c6d14de]
 7: (Locker::handle_client_caps(boost::intrusive_ptr<MClientCaps const> const&)+0x21ef) [0x55b33c6dcddf]
 8: (Locker::dispatch(boost::intrusive_ptr<Message const> const&)+0x224) [0x55b33c6ded34]
 9: (MDSRank::_dispatch(boost::intrusive_ptr<Message const> const&, bool)+0x5c0) [0x55b33c4f63f0]
 10: (MDSRankDispatcher::ms_dispatch(boost::intrusive_ptr<Message const> const&)+0x58) [0x55b33c4f69e8]
 11: (MDSDaemon::ms_dispatch2(boost::intrusive_ptr<Message> const&)+0x1bf) [0x55b33c4d0b6f]
 12: (Messenger::ms_deliver_dispatch(boost::intrusive_ptr<Message> const&)+0x468) [0x7f42a23e8cb8]
 13: (DispatchQueue::entry()+0x5ef) [0x7f42a23e63bf]
 14: (DispatchQueue::DispatchThread::entry()+0xd) [0x7f42a24a54bd]
 15: /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7) [0x7f42a1f13ea7]
 16: clone()

This is perfectly reproducible, unfortunately on a Production System.

Actions

Also available in: Atom PDF