Project

General

Profile

Actions

Bug #21388

closed

inconsistent pg but repair does nothing reporting head data_digest != data_digest from auth oi / hopefully data seems ok

Added by Laurent GUERBY over 6 years ago. Updated about 5 years ago.

Status:
Duplicate
Priority:
Normal
Assignee:
David Zafman
Category:
-
Target version:
-
% Done:

0%

Source:
Community (user)
Tags:
Backport:
jewel, luminous
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(RADOS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

ceph pg repair is currently not fixing three "inconsistent" objects
on one of our pg on a replica 3 pool.

For all three objects the 3 replica data objets are identical
(we checked them on disk on the 3 OSD),
the error says "head data_digest != data_digest from auth oi",
see below.

The data in question are used on rbd volumes from KVM and we did a read
from /dev/sdX at the right place on the VM and got a good looking
result : text file, uncorrupted according to our user, so data
currently returned by ceph and replicated 3 times seems fine.

Now the question is how to tell ceph that the replica data is correct
so that the inconsistent message disappears?

We're thinking of doing manually rados get/put but may be this is not
a good idea or there is another way.

May be ceph should handle this via a "force-repair-identical-replica" option
or something similar if automatic "background" repair seems risky in this
case.

  1. ceph --version
    ceph version 10.2.9 (2ee413f77150c0f375ff6f10edd6c8f9c7d060d0)
  2. ceph health detail
    HEALTH_ERR 1 pgs inconsistent; 9 scrub errors;
    pg 58.6c1 is active+clean+inconsistent, acting [46,44,19]
    ...
  3. rados list-inconsistent-obj 58.6c1 --format=json-pretty {
    "epoch": 277681,
    "inconsistents": []
    }
  4. ceph pg repair 58.6c1

... osd 46 /var/log :

shard 19: soid 58:83772424:::rbd_data.30fce9e39dad7a6.000000000007f027:head data_digest 0x783cd2c5 != data_digest 0x501f846c from auth oi 58:83772424:::rbd_data.30fce9e39dad7a6.000000000007f027:head(252707'3100507 osd.2.0:710926 dirty|data_digest|omap_digest s 4194304 uv 3481755 dd 501f846c od ffffffff alloc_hint [0 0])
shard 44: soid 58:83772424:::rbd_data.30fce9e39dad7a6.000000000007f027:head data_digest 0x783cd2c5 != data_digest 0x501f846c from auth oi 58:83772424:::rbd_data.30fce9e39dad7a6.000000000007f027:head(252707'3100507 osd.2.0:710926 dirty|data_digest|omap_digest s 4194304 uv 3481755 dd 501f846c od ffffffff alloc_hint [0 0])
shard 46: soid 58:83772424:::rbd_data.30fce9e39dad7a6.000000000007f027:head data_digest 0x783cd2c5 != data_digest 0x501f846c from auth oi 58:83772424:::rbd_data.30fce9e39dad7a6.000000000007f027:head(252707'3100507 osd.2.0:710926 dirty|data_digest|omap_digest s 4194304 uv 3481755 dd 501f846c od ffffffff alloc_hint [0 0])
soid 58:83772424:::rbd_data.30fce9e39dad7a6.000000000007f027:head: failed to pick suitable auth object
shard 19: soid 58:83772d9e:::rbd_data.68cb7f74b0dc51.000000000000181e:head data_digest 0xd8f6895a != data_digest 0x4edc70a3 from auth oi 58:83772d9e:::rbd_data.68cb7f74b0dc51.000000000000181e:head(77394'2047065 osd.16.0:4500125 dirty|data_digest|omap_digest s 4194304 uv 1895034 dd 4edc70a3 od ffffffff alloc_hint [0 0])
shard 44: soid 58:83772d9e:::rbd_data.68cb7f74b0dc51.000000000000181e:head data_digest 0xd8f6895a != data_digest 0x4edc70a3 from auth oi 58:83772d9e:::rbd_data.68cb7f74b0dc51.000000000000181e:head(77394'2047065 osd.16.0:4500125 dirty|data_digest|omap_digest s 4194304 uv 1895034 dd 4edc70a3 od ffffffff alloc_hint [0 0])
shard 46: soid 58:83772d9e:::rbd_data.68cb7f74b0dc51.000000000000181e:head data_digest 0xd8f6895a != data_digest 0x4edc70a3 from auth oi 58:83772d9e:::rbd_data.68cb7f74b0dc51.000000000000181e:head(77394'2047065 osd.16.0:4500125 dirty|data_digest|omap_digest s 4194304 uv 1895034 dd 4edc70a3 od ffffffff alloc_hint [0 0])
soid 58:83772d9e:::rbd_data.68cb7f74b0dc51.000000000000181e:head: failed to pick suitable auth object
shard 19: soid 58:8377bf9a:::rbd_data.2ef7e1a528b30ea.00000000000254f6:head data_digest 0xdf8916bf != data_digest 0x47b79db8 from auth oi 58:8377bf9a:::rbd_data.2ef7e1a528b30ea.00000000000254f6:head(252707'3100535 osd.2.0:710954 dirty|data_digest|omap_digest s 4194304 uv 3298154 dd 47b79db8 od ffffffff alloc_hint [0 0])
shard 44: soid 58:8377bf9a:::rbd_data.2ef7e1a528b30ea.00000000000254f6:head data_digest 0xdf8916bf != data_digest 0x47b79db8 from auth oi 58:8377bf9a:::rbd_data.2ef7e1a528b30ea.00000000000254f6:head(252707'3100535 osd.2.0:710954 dirty|data_digest|omap_digest s 4194304 uv 3298154 dd 47b79db8 od ffffffff alloc_hint [0 0])
shard 46: soid 58:8377bf9a:::rbd_data.2ef7e1a528b30ea.00000000000254f6:head data_digest 0xdf8916bf != data_digest 0x47b79db8 from auth oi 58:8377bf9a:::rbd_data.2ef7e1a528b30ea.00000000000254f6:head(252707'3100535 osd.2.0:710954 dirty|data_digest|omap_digest s 4194304 uv 3298154 dd 47b79db8 od ffffffff alloc_hint [0 0])
soid 58:8377bf9a:::rbd_data.2ef7e1a528b30ea.00000000000254f6:head: failed to pick suitable auth object
repair 9 errors, 0 fixed


Files

osd-170.log (23.7 KB) osd-170.log Yoann Moulin, 03/01/2018 09:02 AM
osd-78.log (52.2 KB) osd-78.log Yoann Moulin, 03/01/2018 09:02 AM
osd-154.log (14.9 KB) osd-154.log Yoann Moulin, 03/01/2018 09:02 AM
osd-154_dump.log (15 KB) osd-154_dump.log Yoann Moulin, 03/01/2018 09:24 AM
osd-78_dump.log (32.2 KB) osd-78_dump.log Yoann Moulin, 03/01/2018 09:24 AM
osd-170_dump.log (15.2 KB) osd-170_dump.log Yoann Moulin, 03/01/2018 09:24 AM

Related issues 1 (0 open1 closed)

Is duplicate of RADOS - Feature #25085: Allow repair of an object with a bad data_digest in object_info on all replicasResolvedDavid Zafman07/24/2018

Actions
Actions

Also available in: Atom PDF