Project

General

Profile

Actions

Bug #22796

closed

bluestore gets to ENOSPC with small devices

Added by David Turner over 6 years ago. Updated about 5 years ago.

Status:
Resolved
Priority:
Normal
Assignee:
-
Target version:
-
% Done:

0%

Source:
Community (user)
Tags:
bluestore
Backport:
Regression:
No
Severity:
2 - major
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

I have a 3 node cluster with mon, mds, mgr, and osds all running on each. The steps I've recently performed on my cluster have all gone well until all 3 of my Bluestore SSD OSDs started crashing with the titled segfault.

I upgraded to 12.2.2 from 10.2.10.
Migrated my 9 HDD OSDs to bluestore (without flash media for rocksdb or WAL).
Configured my crush rules to specifically use class HDD.
Failed to be able to remove the previously required cache tier on top of an EC cephfs data pool due to this issue. http://tracker.ceph.com/issues/22754
Created 3 new SSD OSDs with accompanying crush rules to use class SSD.
Updated the pools cephfs_metadata and cephfs_cache to use the replicated-ssd crush rule.

2 days after making this change, the 3 SSD OSDs all segfaulted at the same time and refused to come back up. I generated a `debug bluestore 20` log for each of these OSDs, but don't know how you would like me to provide them since they're 80MB/each.


Files

ceph-osd.9.log.debug5 (547 KB) ceph-osd.9.log.debug5 David Turner, 01/25/2018 11:09 AM

Related issues 1 (0 open1 closed)

Related to bluestore - Bug #23040: bluestore: statfs available can go negativeResolved02/19/2018

Actions
Actions

Also available in: Atom PDF