Project

General

Profile

Actions

Bug #11913

closed

Failure in TestClusterFull.test_barrier

Added by John Spray almost 9 years ago. Updated almost 9 years ago.

Status:
Resolved
Priority:
Normal
Assignee:
Category:
Testing
Target version:
-
% Done:

0%

Source:
other
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(FS):
Labels (FS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

http://pulpito.ceph.com/teuthology-2015-06-05_23:04:02-fs-master---basic-multi/922578/

mount_a is getting the new OSD map unexpectedly soon.

It's hitting the barrier before trying to explicitly do any metadata operations afterwards, because it's getting a cap revoke from the MDS when mount_b does its metadata operations.

Which is weird, because mount_b was already the last guy do to any metadata ops before this point, so he should already have had all the needed caps.

Actions #1

Updated by Greg Farnum almost 9 years ago

(Referring to ceph-qa-suite/tasks/cephfs/test_full.py::test_barrier().)

So mount.a is doing open_no_data("alpha") and that looks to me to be the last thing either client does before the 30 second sleep. Then mount.b touches/opens "bravo", and a here revoke against mount.a definitely looks possible to me?

Actions #2

Updated by John Spray almost 9 years ago

  • Status changed from New to Fix Under Review
Actions #3

Updated by Greg Farnum almost 9 years ago

  • Status changed from Fix Under Review to Resolved

commit:bf9a9a2d9ff2be129b303d535899f60ad49f7c23

Actions

Also available in: Atom PDF