Project

General

Profile

Actions

Bug #38452

open

mds: assert crash loop while unlinking file

Added by Jérôme Poulin about 5 years ago. Updated 8 months ago.

Status:
Need More Info
Priority:
Urgent
Assignee:
-
Category:
Correctness/Safety
Target version:
-
% Done:

0%

Source:
Community (user)
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(FS):
MDS
Labels (FS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

Here is the stack trace, it was caused by a Postgresql trying to unlink a file in the log archive.

 ceph version 13.2.4 (b10be4d44915a4d78a8e06aa31919e74927b142e) mimic (stable)
 1: (()+0x4058c0) [0x5588549cd8c0]
 2: (()+0x12890) [0x7fdf788d3890]
 3: (gsignal()+0xc7) [0x7fdf779c6e97]
 4: (abort()+0x141) [0x7fdf779c8801]
 5: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x250) [0x7fdf78fbd530]
 6: (()+0x26d5a7) [0x7fdf78fbd5a7]
 7: (Server::_unlink_local(boost::intrusive_ptr<MDRequestImpl>&, CDentry*, CDentry*)+0x15f8) [0x5588547482e8]
 8: (Server::handle_client_unlink(boost::intrusive_ptr<MDRequestImpl>&)+0x961) [0x558854748cd1]
 9: (Server::handle_client_request(MClientRequest*)+0x49b) [0x55885476305b]
 10: (Server::dispatch(Message*)+0x2db) [0x558854766d1b]
 11: (MDSRank::handle_deferrable_message(Message*)+0x434) [0x5588546da1e4]
 12: (MDSRank::_dispatch(Message*, bool)+0x89b) [0x5588546e7a1b]
 13: (MDSRank::retry_dispatch(Message*)+0x12) [0x5588546e8012]
 14: (MDSInternalContextBase::complete(int)+0x67) [0x55885494df87]
 15: (MDSRank::_advance_queues()+0xf1) [0x5588546e69c1]
 16: (MDSRank::ProgressThread::entry()+0x43) [0x5588546e7043]
 17: (()+0x76db) [0x7fdf788c86db]
 18: (clone()+0x3f) [0x7fdf77aa988f]

Attached a debug-ms=10 file, however I can't reproduce as I fixed it by killing the ceph-fuse client on our Postgres server et restarting the MDS.


Files

ceph-mds.log.gz (225 KB) ceph-mds.log.gz Jérôme Poulin, 02/22/2019 05:06 PM
ceph-mds.log.xz (82.8 KB) ceph-mds.log.xz Jérôme Poulin, 03/04/2019 07:16 AM
ceph-mds-active.log.xz (92.7 KB) ceph-mds-active.log.xz Jérôme Poulin, 03/04/2019 02:54 PM
ceph-mds-0x100000367e5.log (51.9 KB) ceph-mds-0x100000367e5.log Jérôme Poulin, 03/05/2019 02:49 PM
mds_crash.tar.gz (125 KB) mds_crash.tar.gz error scatter_lock Felix Stolte, 12/13/2022 09:28 AM

Related issues 7 (2 open5 closed)

Related to CephFS - Bug #42829: tools/cephfs: linkages injected by cephfs-data-scan have first == headResolvedZheng Yan

Actions
Related to CephFS - Bug #58482: mds: catch damage to CDentry's first member before persistingResolvedPatrick Donnelly

Actions
Related to CephFS - Bug #54546: mds: crash due to corrupt inode and omap entryNewPatrick Donnelly

Actions
Has duplicate CephFS - Bug #41147: mds: crash loop - Server.cc 6835: FAILED ceph_assert(in->first <= straydn->first)Duplicate

Actions
Has duplicate CephFS - Bug #54643: crash: void Server::_unlink_local(MDRequestRef&, CDentry*, CDentry*): assert(in->first <= straydn->first)Duplicate

Actions
Has duplicate CephFS - Bug #53179: Crash when unlink in corrupted cephfsDuplicateVenky Shankar

Actions
Has duplicate CephFS - Bug #60669: crash: void Server::_unlink_local(MDRequestRef&, CDentry*, CDentry*): assert(in->first <= straydn->first)New

Actions
Actions

Also available in: Atom PDF