Project

General

Profile

Actions

Support #54315

open

1 fsck error per osd during nautilus -> octopus upgrade (S3 cluster)

Added by Dan van der Ster about 2 years ago. Updated about 2 years ago.

Status:
New
Priority:
Normal
Target version:
-
% Done:

0%

Tags:
Reviewed:
Affected Versions:
Pull request ID:

Description

At the end of the conversion to per-pool omap, around half of our OSDs had 1 error, but the log didn't show the error.

2022-02-15T16:02:16.554+0100 7fdfde8d8f00  0 bluestore(/var/lib/ceph/osd/ceph-1247) _fsck_check_objects partial offload, done myself 7925084 of 7942492objects, threads 2
2022-02-15T16:02:16.678+0100 7fdfde8d8f00  1 bluestore(/var/lib/ceph/osd/ceph-1247) _fsck_on_open checking shared_blobs
2022-02-15T16:02:16.693+0100 7fdfde8d8f00  1 bluestore(/var/lib/ceph/osd/ceph-1247) _fsck_on_open checking pool_statfs
2022-02-15T16:17:37.407+0100 7fdfde8d8f00  1 bluestore(/var/lib/ceph/osd/ceph-1247) _fsck_on_open <<<FINISH>>> with 1 errors, 318 warnings, 319 repaired, 0 remaining in 1672.130946 seconds

Full log is posted: ceph-post-file: 82f661a7-b10f-4a80-acaf-37f1268f275e

Actions

Also available in: Atom PDF