Project

General

Profile

Actions

Bug #63503

open

data corruption after rbd migration

Added by Nikola Ciprich 7 months ago. Updated 5 months ago.

Status:
New
Priority:
Normal
Assignee:
-
Category:
librbd
Target version:
-
% Done:

0%

Source:
Community (user)
Tags:
Backport:
Regression:
No
Severity:
2 - major
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

we're struggling with strange issue which I think might be a bug
causing snapshot data corruption while migrating RBD image

we've tracked it to minimal set of steps to reproduce using VM
with one 32G drive:

rbd create --size 32768 sata/D2
virsh create xml_orig.xml
rbd snap create ssd/D1@snap1
rbd export-diff ssd/D1@snap1 - | rbd import-diff - sata/D2
rbd export --export-format 1 --no-progress ssd/D1@snap1 - | xxh64sum
505dde3c49775773
rbd export --export-format 1 --no-progress sata/D2@snap1 - | xxh64sum
505dde3c49775773 # <- checksums match - OK

virsh shutdown VM
rbd migration prepare ssd/D1 sata/D1Z
virsh create xml_new.xml
rbd snap create sata/D1Z@snap2
rbd export-diff --from-snap snap1 sata/D1Z@snap2 - | rbd import-diff - sata/D2
rbd migration execute sata/D1Z
rbd migration commit sata/D1Z
rbd export --export-format 1 --no-progress sata/D1Z@snap2 - | xxh64sum
19892545c742c1de
rbd export --export-format 1 --no-progress sata/D2@snap2 - | xxh64sum
cc045975baf74ba8 # <- snapshosts differ

OS is alma 9 based, kernel 5.15.105, CEPH 17.2.6, qemu-8.0.3
we tried disabling VM disk caches as well as discard, to no avail.

Actions #1

Updated by Ilya Dryomov 5 months ago

  • Target version deleted (v17.2.7)
Actions

Also available in: Atom PDF