Project

General

Profile

Actions

Bug #39103

closed

bitmap allocator osd make %use of osd over 100%

Added by hoan nv about 5 years ago. Updated almost 5 years ago.

Status:
Resolved
Priority:
Normal
Assignee:
-
Target version:
-
% Done:

0%

Source:
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

after pull request https://github.com/ceph/ceph/pull/26983 is merged.
I install package

ceph-common-13.2.5-101.ga1aa89a.el7.x86_64
ceph-mgr-13.2.5-101.ga1aa89a.el7.x86_64
ceph-mds-13.2.5-101.ga1aa89a.el7.x86_64
ceph-13.2.5-101.ga1aa89a.el7.x86_64
libcephfs2-13.2.5-101.ga1aa89a.el7.x86_64
ceph-selinux-13.2.5-101.ga1aa89a.el7.x86_64
ceph-osd-13.2.5-101.ga1aa89a.el7.x86_64
python-cephfs-13.2.5-101.ga1aa89a.el7.x86_64
ceph-base-13.2.5-101.ga1aa89a.el7.x86_64
ceph-mon-13.2.5-101.ga1aa89a.el7.x86_64

and configured

[osd]
bluestore_allocator = bitmap

after restart 1 osd. use of this osd over 100

ceph osd df tree
ID CLASS WEIGHT  REWEIGHT SIZE    USE     AVAIL   %USE          VAR  PGS TYPE NAME
-4       7.19995        - 7.3 TiB  16 EiB 7.3 TiB  230759568.00 1.13   - root ssd
-3       3.59998        - 3.6 TiB 5.3 GiB 3.6 TiB          0.14    0   -     host ssd-ceph-2
 0   ssd 0.89999  1.00000 930 GiB 1.3 GiB 929 GiB          0.14    0 128         osd.0
 1   ssd 0.89999  1.00000 930 GiB 1.3 GiB 929 GiB          0.14    0 139         osd.1
 2   ssd 0.89999  1.00000 930 GiB 1.3 GiB 929 GiB          0.14    0 114         osd.2
 3   ssd 0.89999  1.00000 930 GiB 1.3 GiB 929 GiB          0.14    0 131         osd.3
-5       3.59998        - 3.6 TiB  16 EiB 3.7 TiB  461405824.00 2.25   -     host ssd-ceph-3
 4   ssd 0.89999  1.00000 930 GiB  16 EiB 965 GiB 1846529920.00 9.00  36         osd.4
 5   ssd 0.89999  1.00000 931 GiB 1.4 GiB 930 GiB          0.15    0 152         osd.5
 6   ssd 0.89999  1.00000 931 GiB 1.3 GiB 930 GiB          0.14    0 133         osd.6
 7   ssd 0.89999  1.00000 931 GiB 1.3 GiB 930 GiB          0.14    0 125         osd.7

It make my cluster not working, iops = 0 and ceph health

1 full osd(s)
1 pool(s) full
Degraded data redundancy: 52/924 objects degraded (5.628%), 38 pgs degraded, 38 pgs undersized
Degraded data redundancy (low space): 2 pgs backfill_toofull, 36 pgs recovery_toofull

log start has

2019-04-04 15:44:37.580 7fa86eb29700 -1 log_channel(cluster) log [ERR] : full status failsafe engaged, dropping updates, now 1846530048% full

detail log in attach file.


Files

bitmap-error.log (217 KB) bitmap-error.log hoan nv, 04/04/2019 09:02 AM
ceph-osd.4.log.tar.gz (664 KB) ceph-osd.4.log.tar.gz start osd with debug bluestore 20 log hoan nv, 04/11/2019 09:16 AM
Actions

Also available in: Atom PDF