Project

General

Profile

Actions

Bug #614

closed

SEGV loop on _open_lock_pg after rmpool

Added by John Leach over 13 years ago. Updated over 13 years ago.

Status:
Resolved
Priority:
Normal
Assignee:
Category:
-
Target version:
-
% Done:

0%

Source:
Tags:
Backport:
Regression:
Severity:
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

discovered my cosd processes at 100%, possibly following some "rados rmpool" commands to delete some pools. Stopped entire cluster and started on one node only and still get 100% cpu loop. osd debug log attached (starting at the start of the session).


Files

osd.0.log.crash.gz (88.5 KB) osd.0.log.crash.gz John Leach, 11/29/2010 02:23 PM
Actions #1

Updated by Colin McCabe over 13 years ago

  • Status changed from New to Resolved
  • Assignee set to Colin McCabe

In OSD::load_pgs, we weren't checking to make sure that the pool existed when going through all the collections.

Fixed by 7ee50addba035a8f34b77fd6eaa0b8eb345a7caa

Actions #2

Updated by Colin McCabe over 13 years ago

Er, by that I mean:

load_pgs shouldn't try to load a PG that is in a nonexistent pool. This could only happen after the number of pools was reduced while still leaving remnants of the old pools and PGs on-disk.

Actions

Also available in: Atom PDF