Reef¶
On-call Schedule¶
- Feb: Venky
- Mar: Patrick
- Apr: Jos
- May: Xiubo
- Jun: Rishabh
- Jul: Kotresh
- Aug: Milind
- Sep: Leonid
- Oct: Dhairya
- Nov: Chris
2024-04-17¶
https://tracker.ceph.com/issues/65393#note-1
https://pulpito.ceph.com/yuriw-2024-04-09_01:14:24-fs-reef-release-distro-default-smithi/
- selinux denials with centos9.stream
- pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main
- qa: ‘Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)’
- Test failure: test_add_ancestor_and_child_directory tasks.cephfs.test_mirroring.TestMirroring
- QA failure: test_fscrypt_dummy_encryption_with_quick_group
- error during scrub thrashing: rank damage found: {'backtrace'}
- dbench: write failed on handle 10009 Resource temporarily unavailable
- qa: fs/snaps/snaptest-git-ceph.sh failed when reseting to tag 'v0.1’
- AttributeError: 'RemoteProcess' object has no attribute 'read’
12 April 2024¶
https://tracker.ceph.com/issues/65264
https://pulpito.ceph.com/rishabh-2024-04-08_08:23:45-fs-wip-rishabh-testing-20240407.092921-reef-testing-default-smithi/
- https://tracker.ceph.com/issues/64711
Test failure: test_cephfs_mirror_cancel_mirroring_and_readd (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/62067
ffsb.sh failure "Resource temporarily unavailable" - https://tracker.ceph.com/issues/53859
Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm) - https://tracker.ceph.com/issues/61820
valgrind issues with ceph-mon - https://tracker.ceph.com/issues/64502
pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main - https://tracker.ceph.com/issues/64616
selinux denials with centos9.stream - https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/65136
QA failure: test_fscrypt_dummy_encryption_with_quick_group - https://tracker.ceph.com/issues/65261
qa/cephfs: cephadm related failure on fs/upgrade job - https://tracker.ceph.com/issues/63265
qa: fs/snaps/snaptest-git-ceph.sh failed when reseting to tag 'v0.1'
2024-04-08¶
https://tracker.ceph.com/issues/65336
- selinux denials with centos9.stream
- pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main
- Test failure: test_fscrypt_dummy_encryption_with_quick_group tasks.cephfs.test_fscrypt.TestFscrypt
- leak in mds.c detected by valgrind during CephFS QA run
- qa: ‘Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)’
- dbench: write failed on handle 10009 Resource temporarily unavailable
- Test failure: test_add_ancestor_and_child_directory tasks.cephfs.test_mirroring.TestMirroring
- AttributeError: 'RemoteProcess' object has no attribute 'read’
- qa: The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported’}
- mds: failed to store backtrace and force file system read-only
2024-04-03¶
https://tracker.ceph.com/issues/65148
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20240326.105515-reef
- error during scrub thrashing: rank damage found: {'backtrace'}
- reef: qa: AttributeError: 'TestSnapSchedulesSubvolAndGroupArguments' object has no attribute 'get_ceph_cmd_stdout'
- selinux denials with centos9.stream
- pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main
- qa: scrub - object missing on disk; some files may be lost
- Test failure: test_fscrypt_dummy_encryption_with_quick_group tasks.cephfs.test_fscrypt.TestFscrypt
- qa: ‘Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)’
- Test failure: test_cephfs_mirror_cancel_mirroring_and_readd
- dbench: write failed on handle 10009 Resource temporarily unavailable
- leak in mds.c detected by valgrind during CephFS QA run
- Test failure: test_pool_perm
3 APR 2024¶
https://pulpito.ceph.com/rishabh-2024-03-29_18:05:24-fs-wip-rishabh-testing-20240327.051042-reef-testing-default-smithi/
https://pulpito.ceph.com/rishabh-2024-03-29_18:05:24-fs-wip-rishabh-testing-20240327.051042-reef-testing-default-smithi/
- https://tracker.ceph.com/issues/53859
Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm) - https://tracker.ceph.com/issues/65136
QA failure: test_fscrypt_dummy_encryption_with_quick_group - https://tracker.ceph.com/issues/47292
Test failure: test_df_for_valid_file (tasks.cephfs.test_cephfs_shell.TestDF)
- https://tracker.ceph.com/issues/48562
qa: scrub - object missing on disk; some files may be lost - https://tracker.ceph.com/issues/57656
[testing] dbench: write failed on handle 10009 (Resource temporarily unavailable) - https://tracker.ceph.com/issues/64502
pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main - https://tracker.ceph.com/issues/64616
selinux denials with centos9.stream - https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/62067
ffsb.sh failure "Resource temporarily unavailable" - https://tracker.ceph.com/issues/57206
ceph_test_libcephfs_reclaim crashes during test - https://tracker.ceph.com/issues/65261
qa/cephfs: cephadm related failure on fs/upgrade job - https://tracker.ceph.com/issues/65262
qa/cephfs: kernel_untar_build.sh failed due to build error
2024-03-28¶
https://tracker.ceph.com/issues/65202
https://pulpito.ceph.com/yuriw-2024-03-29_19:10:12-fs-wip-yuri11-testing-2024-03-28-0753-reef-distro-default-smithi/
- error during scrub thrashing: rank damage found: {'backtrace'}
- qa: Health check failed: Reduced data availability: 1 pg peering
- selinux denials with centos9.stream
- reef: qa: AttributeError: 'TestSnapSchedulesSubvolAndGroupArguments' object has no attribute 'get_ceph_cmd_stdout'
- pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main
- Test failure: test_fscrypt_dummy_encryption_with_quick_group tasks.cephfs.test_fscrypt.TestFscrypt
- qa: scrub - object missing on disk; some files may be lost
- Test failure: test_add_ancestor_and_child_directory tasks.cephfs.test_mirroring.TestMirroring
- dbench: write failed on handle 10009 Resource temporarily unavailable
- ceph_test_libcephfs_reclaim crashes during test
2024-03-26¶
https://trello.com/c/xEeVJoco/1978-wip-yuri-testing-2024-03-12-1240-reef
https://pulpito.ceph.com/yuriw-2024-03-13_19:31:43-fs-wip-yuri-testing-2024-03-12-1240-reef-distro-default-smithi/
- dbench: write failed on handle 10009 Resource temporarily unavailable
- error during scrub thrashing: rank damage found: {'backtrace'}
- pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main
- selinux denials with centos9.stream
- Test failure: test_add_ancestor_and_child_directory tasks.cephfs.test_mirroring.TestMirroring
- Test failure: test_fscrypt_dummy_encryption_with_quick_group tasks.cephfs.test_fscrypt.TestFscrypt
- Test failure: test_multiple_path_r tasks.cephfs.test_admin.TestFsAuthorize
- reef: qa: AttributeError: 'TestSnapSchedulesSubvolAndGroupArguments' object has no attribute 'get_ceph_cmd_stdout'
Test failures caused by PR:
- https://pulpito.ceph.com/yuriw-2024-03-13_19:31:43-fs-wip-yuri-testing-2024-03-12-1240-reef-distro-default-smithi/7598790
- https://pulpito.ceph.com/yuriw-2024-03-13_19:31:43-fs-wip-yuri-testing-2024-03-12-1240-reef-distro-default-smithi/7598726
- https://pulpito.ceph.com/yuriw-2024-03-13_19:31:43-fs-wip-yuri-testing-2024-03-12-1240-reef-distro-default-smithi/7598661
20th March 2024¶
https://trello.com/c/bCURjoHF/1981-wip-yuri8-testing-2024-03-15-0740-reef-old-wip-yuri8-testing-2024-03-14-1513-reef
https://pulpito.ceph.com/yuriw-2024-03-15_20:10:26-fs-wip-yuri8-testing-2024-03-15-0740-reef-distro-default-smithi/
- https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/64502
pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main - https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/64616
selinux denials with centos9.stream - https://tracker.ceph.com/issues/61820
valgrind issues with ceph-mon - https://tracker.ceph.com/issues/59684
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/64937
reef: qa: AttributeError: 'TestSnapSchedulesSubvolAndGroupArguments' object has no attribute 'get_ceph_cmd_stdout' - https://tracker.ceph.com/issues/62221
Test failure: test_add_ancestor_and_child_directory (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/57656
[testing] dbench: write failed on handle 10009 (Resource temporarily unavailable) - https://tracker.ceph.com/issues/59343 (see https://tracker.ceph.com/issues/59343#note-25)
qa: fs/snaps/snaptest-multiple-capsnaps.sh failed (awaiting kernel backport) - https://tracker.ceph.com/issues/65001
mds: ceph-mds might silently ignore client_session(request_close, ...) message
19th March 2024¶
https://trello.com/c/mF32rpDn/1982-wip-yuri10-testing-2024-03-15-1653-reef
https://pulpito.ceph.com/yuriw-2024-03-16_15:03:17-fs-wip-yuri10-testing-2024-03-15-1653-reef-distro-default-smithi/
- https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/64502
pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main - https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/59684
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/61820
valgrind issues with ceph-mon - https://tracker.ceph.com/issues/64616
selinux denials with centos9.stream - https://tracker.ceph.com/issues/64937
reef: qa: AttributeError: 'TestSnapSchedulesSubvolAndGroupArguments' object has no attribute 'get_ceph_cmd_stdout' - https://tracker.ceph.com/issues/48562
qa: scrub - object missing on disk; some files may be lost - https://tracker.ceph.com/issues/57656
[testing] dbench: write failed on handle 10009 (Resource temporarily unavailable) - https://tracker.ceph.com/issues/63265
qa: fs/snaps/snaptest-git-ceph.sh failed when reseting to tag 'v0.1' - https://tracker.ceph.com/issues/59343
qa: fs/snaps/snaptest-multiple-capsnaps.sh failed (awaiting kernel backport)
11th March 2024¶
https://trello.com/c/TLeKDODy/1973-wip-yuri11-testing-2024-03-11-0838-reef-old-wip-yuri11-testing-2024-03-06-0821-reef
https://pulpito.ceph.com/yuriw-2024-03-12_14:59:27-fs-wip-yuri11-testing-2024-03-11-0838-reef-distro-default-smithi/
- https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/64502
pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main - https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/59684
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/61820
valgrind issues with ceph-mon - https://tracker.ceph.com/issues/64616
selinux denials with centos9.stream - https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure - https://tracker.ceph.com/issues/64937
reef: qa: AttributeError: 'TestSnapSchedulesSubvolAndGroupArguments' object has no attribute 'get_ceph_cmd_stdout' - https://tracker.ceph.com/issues/62221
Test failure: test_add_ancestor_and_child_directory (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/48562
qa: scrub - object missing on disk; some files may be lost
PRs failing QA:
- https://github.com/ceph/ceph/pull/54729#pullrequestreview-1937730202
- https://github.com/ceph/ceph/pull/55692#pullrequestreview-1937932954
5th March 2024¶
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-reef-2024-02-28-1133
- https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/59684
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/64616
selinux denials with centos9.stream - https://tracker.ceph.com/issues/57656
[testing] dbench: write failed on handle 10009 (Resource temporarily unavailable) - https://tracker.ceph.com/issues/59684
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/51964
qa: test_cephfs_mirror_restart_sync_on_blocklist failure - https://tracker.ceph.com/issues/59531
"OSD bench result of 228617.361065 IOPS exceeded the threshold limit of 500.000000 IOPS for osd.7. IOPS capacity is unchanged at 315.000000 IOPS. The recommendation is to establish the osd's - https://tracker.ceph.com/issues/64747
postgresql pkg install failure - https://tracker.ceph.com/issues/57206
ceph_test_libcephfs_reclaim crashes during test - https://tracker.ceph.com/issues/64748
reef: snaptest-git-ceph.sh failure
05 Feb 2024¶
- https://tracker.ceph.com/issues/59684
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/59531
"OSD bench result of 228617.361065 IOPS exceeded the threshold limit of 500.000000 IOPS for osd.7. IOPS capacity is unchanged at 315.000000 IOPS. The recommendation is to establish the osd's - https://tracker.ceph.com/issues/59534
Command failed (workunit test suites/dbench.sh) on smithi077 with status 1 - https://tracker.ceph.com/issues/61820
valgrind issues with ceph-mon - https://tracker.ceph.com/issues/63522
No module named ... (PYTHONPATH issues)
29 Jan 2024¶
- https://tracker.ceph.com/issues/59684
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/61831
qa: test_mirroring_init_failure_with_recovery failure - https://tracker.ceph.com/issues/64060
qa: AttributeError: 'TestSubvolumeGroups' object has no attribute '_generate_random_group_name' - https://tracker.ceph.com/issues/59531
"OSD bench result of 228617.361065 IOPS exceeded the threshold limit of 500.000000 IOPS for osd.7. IOPS capacity is unchanged at 315.000000 IOPS. The recommendation is to establish the osd's - https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/51282
cluster [WRN] Health check failed: Degraded data redundancy: 9/46 objects degraded (19.565%), 4 pgs degraded (PG_DEGRADED)" in cluster log - https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/59534
Command failed (workunit test suites/dbench.sh) on smithi077 with status 1
04 Dec 2023¶
https://pulpito.ceph.com/?sha1=a6f82997ac57b70a895455dfed6256360a1e4c32- https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/62658
error during scrub thrashing: reached maximum tries (31) after waiting for 900 seconds - https://tracker.ceph.com/issues/63233
mon|client|mds: valgrind reports possible leaks in the MDS - https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure - https://tracker.ceph.com/issues/63339
reef: mds: warning `clients failing to advance oldest client/flush tid` seen with some workloads (missing backport) - https://tracker.ceph.com/issues/62221
Test failure: test_add_ancestor_and_child_directory (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/62482
qa: "cluster [WRN] Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED)" (backport for cephfs-mirror ignorelist missing) - https://tracker.ceph.com/issues/59684
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/62287
reef: ceph_test_libcephfs_reclaim crashes during test - https://tracker.ceph.com/issues/62081
tasks/fscrypt-common does not finish, timesout
21 Nov 2023¶
https://pulpito.ceph.com/?branch=wip-vshankar-testing2-2023-11-15-1128-reef
- https://tracker.ceph.com/issues/62658
error during scrub thrashing: reached maximum tries (31) after waiting for 900 seconds - https://tracker.ceph.com/issues/51964
qa: test_cephfs_mirror_restart_sync_on_blocklist failure - https://tracker.ceph.com/issues/63339
Test failure: test_client_blocklisted_oldest_tid (tasks.cephfs.test_client_limits.TestClientLimits)
- https://tracker.ceph.com/issues/53859
Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm) - https://tracker.ceph.com/issues/62580
Test failure: test_snapshot_remove (tasks.cephfs.test_strays.TestStrays) - https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1
- https://tracker.ceph.com/issues/57656
dbench: write failed on handle 10010 (Resource temporarily unavailable) - https://tracker.ceph.com/issues/50223
client.xxxx isn't responding to mclientcaps(revoke)
07 Nov 2023¶
https://pulpito.ceph.com/?branch=reef-release (18.2.1)
- https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/63233
mon|client|mds: valgrind reports possible leaks in the MDS - https://tracker.ceph.com/issues/62658
error during scrub thrashing: reached maximum tries (31) after waiting for 900 seconds - https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/62221
Test failure: test_add_ancestor_and_child_directory (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/63488
smoke test fails from "NameError: name 'DEBUGFS_META_DIR' is not defined" (missing kclient patches in ubuntu 20.04) - https://tracker.ceph.com/issues/62081
tasks/fscrypt-common does not finish, timesout - https://tracker.ceph.com/issues/63339
reef: mds: warning `clients failing to advance oldest client/flush tid` seen with some workloads (missing backport)
27 Oct 2023¶
- https://tracker.ceph.com/issues/58126
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/63339
Test failure: test_client_blocklisted_oldest_tid (tasks.cephfs.test_client_limits.TestClientLimits) - https://tracker.ceph.com/issues/48873
Command failed on smithi118 with status 32: 'sudo mount -t nfs -o port=2049 172.21.15.118:/ceph /mnt' - https://tracker.ceph.com/issues/62221
Test failure: test_add_ancestor_and_child_directory (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/55805
Command failed (workunit test kernel_untar_build.sh) on smithi167 with status 2 - https://tracker.ceph.com/issues/57676
error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/62658
error during scrub thrashing: reached maximum tries (31) after waiting for 900 seconds - https://tracker.ceph.com/issues/62580
Test failure: test_snapshot_remove (tasks.cephfs.test_strays.TestStrays) - https://tracker.ceph.com/issues/63340
saw valgrind issues
18 Oct 2023¶
https://pulpito.ceph.com/?branch=wip-vshankar-testing-reef-20231018.163145
- https://tracker.ceph.com/issues/52624
qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" - https://tracker.ceph.com/issues/63233
mon|client|mds: valgrind reports possible leaks in the MDS - https://tracker.ceph.com/issues/59531
"OSD bench result of 228617.361065 IOPS exceeded the threshold limit of 500.000000 IOPS for osd.7. IOPS capacity is unchanged at 315.000000 IOPS. The recommendation is to establish the osd's IOPS capacity using other benchmark tools (e.g. Fio)" - https://tracker.ceph.com/issues/63281
src/mds/MDLog.h: 100: FAILED ceph_assert(!segments.empty()) - https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure - https://tracker.ceph.com/issues/61831
qa: test_mirroring_init_failure_with_recovery failure - https://tracker.ceph.com/issues/62810
Failure in snaptest-git-ceph.sh (it's an async unlink/create bug) -- Need to fix again - https://tracker.ceph.com/issues/57206
ceph_test_libcephfs_reclaim crashes during test - https://tracker.ceph.com/issues/62081
tasks/fscrypt-common does not finish, timesout - https://tracker.ceph.com/issues/59343
qa: fs/snaps/snaptest-multiple-capsnaps.sh failed
11 Oct 2023¶
https://pulpito.ceph.com/?branch=wip-yuri4-testing-2023-10-11-0735-reef
- https://patchwork.kernel.org/project/ceph-devel/patch/20220523160951.8781-1-lhenriques@suse.de/
hit max job timeout, kclient incorrectly decode the session message. And have been backported to downstream kernel-4.18.0-500.el8. - https://tracker.ceph.com/issues/63259
"hit max job timeout", failed to store the backtrace and fs was forced to readonly. - https://tracker.ceph.com/issues/63105
"smithi060 with status 139: 'mkdirp -...qa/workunits/libcephfs/test.sh'" - https://tracker.ceph.com/issues/61820
"saw valgrind issues" - https://tracker.ceph.com/issues/62937
"stderr:logrotate does not support parallel execution on the same set of logfiles" - https://tracker.ceph.com/issues/57676
"error during scrub thrashing: rank damage found: {'backtrace'}" - https://tracker.ceph.com/issues/63212
Failed to download "ior.tbz2". - https://tracker.ceph.com/issues/62277
"Error: Unable to find a match: python2" - https://tracker.ceph.com/issues/61574
qa: build failure for mdtest project - https://tracker.ceph.com/issues/62658
"error during scrub thrashing: reached maximum tries (31) after waiting for 900 seconds" - https://tracker.ceph.com/issues/63265
"Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi062 with status 128: .../fs/snaps/snaptest-git-ceph.sh'" - https://tracker.ceph.com/issues/62221
"Test failure: test_add_ancestor_and_child_directory (tasks.cephfs.test_mirroring.TestMirroring)" - https://tracker.ceph.com/issues/43863
"mkdir: cannot create directory ‘/home/ubuntu/cephtest/archive/audit’: File exists"
09 Oct 2023¶
https://pulpito.ceph.com/?branch=wip-vshankar-testing-reef-20231009.131610
- https://tracker.ceph.com/issues/63211
error during scrub thrashing: reached maximum tries (31) after waiting for 900 seconds - https://tracker.ceph.com/issues/63212
Failed to download "ior.tbz2" - https://tracker.ceph.com/issues/61831
Test failure: test_mirroring_init_failure_with_recovery (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/18845
saw valgrind issues - https://tracker.ceph.com/issues/53859
Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm) - https://tracker.ceph.com/issues/61574
qa: build failure for mdtest project - https://tracker.ceph.com/issues/57655
Failed to build the kernel source - https://tracker.ceph.com/issues/62580
Test failure: test_snapshot_remove (tasks.cephfs.test_strays.TestStrays) - https://tracker.ceph.com/issues/57087
Test failure: test_fragmented_injection (tasks.cephfs.test_data_scan.TestDataScan)
27 Sep 2023¶
https://pulpito.ceph.com/?branch=wip-vshankar-testing-reef-20230927.021134
- https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log - https://tracker.ceph.com/issues/59531
quincy: "OSD bench result of 228617.361065 IOPS exceeded the threshold limit of 500.000000 IOPS for osd.7. IOPS capacity is unchanged at 315.000000 IOPS. The recommendation is to establish the osd's IOPS capacity using other benchmark tools (e.g. Fio)" - https://tracker.ceph.com/issues/61574
qa: "[Makefile:59: mdtest] Error 1" (undefined reference to `fi_strerror') - https://tracker.ceph.com/issues/61399
qa: build failure for ior - https://tracker.ceph.com/issues/62653
qa: unimplemented fcntl command: 1036 with fsstress * https://tracker.ceph.com/issues/62847
mds: blogbench requests stuck (5mds+scrub+snaps-flush) - https://tracker.ceph.com/issues/62081
tasks/fscrypt-common does not finish, timesout - https://tracker.ceph.com/issues/63089
qa: tasks/mirror times out - https://tracker.ceph.com/issues/58244
Test failure: test_rebuild_inotable (tasks.cephfs.test_data_scan.TestDataScan)
26 Sep 2023¶
https://pulpito.ceph.com/?branch=wip-vshankar-testing-reef-20230926.161455
- https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc - https://tracker.ceph.com/issues/59531
quincy: "OSD bench result of 228617.361065 IOPS exceeded the threshold limit of 500.000000 IOPS for osd.7. IOPS capacity is unchanged at 315.000000 IOPS. The recommendation is to establish the osd's IOPS capacity using other benchmark tools (e.g. Fio)" - https://tracker.ceph.com/issues/58126
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/62658
error during scrub thrashing: reached maximum tries (31) after waiting for 900 seconds - https://tracker.ceph.com/issues/62067
ffsb.sh failure "Resource temporarily unavailable"
2 August 2023¶
https://pulpito.ceph.com/yuriw-2023-07-28_23:11:59-fs-reef-release-distro-default-smithi/
https://pulpito.ceph.com/yuriw-2023-07-29_14:02:18-fs-reef-release-distro-default-smithi/
- https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log - https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1 - https://tracker.ceph.com/issues/62187 <<< covered dbench failures
iozone: command not found - https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc - https://tracker.ceph.com/issues/61574
qa: "[Makefile:59: mdtest] Error 1" (undefined reference to `fi_strerror') - https://tracker.ceph.com/issues/61399
qa: build failure for ior - https://tracker.ceph.com/issues/62277
Error: Unable to find a match: python2 with fscrypt tests - https://tracker.ceph.com/issues/62221
Test failure: test_add_ancestor_and_child_directory (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/61182
qa: workloads/cephfs-mirror-ha-workunit - stopping mirror daemon after the test finishes timesout. - https://tracker.ceph.com/issues/57676
qa: error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure
31 July 2023¶
https://pulpito.ceph.com/?branch=wip-yuri-testing-2023-07-25-0833-reef
- https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc - https://tracker.ceph.com/issues/61892
Test failure: test_snapshot_remove (tasks.cephfs.test_strays.TestStrays) - https://tracker.ceph.com/issues/61399
qa: build failure for ior - https://tracker.ceph.com/issues/62081
tasks/fscrypt-common does not finish, timesout - https://tracker.ceph.com/issues/62221
Test failure: test_add_ancestor_and_child_directory (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/61182
qa: workloads/cephfs-mirror-ha-workunit - stopping mirror daemon after the test finishes timesout. - https://tracker.ceph.com/issues/57206
ceph_test_libcephfs_reclaim crashes during test - https://tracker.ceph.com/issues/61574
qa: "[Makefile:59: mdtest] Error 1" (undefined reference to `fi_strerror') - https://tracker.ceph.com/issues/62187
iozone: command not found
25 July 2023¶
https://pulpito.ceph.com/?branch=wip-yuri4-testing-2023-07-06-0738-reef
- https://tracker.ceph.com/issues/51964
qa: test_cephfs_mirror_restart_sync_on_blocklist failure - https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc - https://tracker.ceph.com/issues/58340
mds: fsstress.sh hangs with multimds - https://tracker.ceph.com/issues/57676
qa: error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log - https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1 - https://tracker.ceph.com/issues/61892
Test failure: test_snapshot_remove (tasks.cephfs.test_strays.TestStrays) - https://tracker.ceph.com/issues/58220
Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1: - https://tracker.ceph.com/issues/61399
qa: build failure for ior - https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure - https://tracker.ceph.com/issues/62081
tasks/fscrypt-common does not finish, timesout
19 July 2023¶
NOTE: Results from two runs clubbed into one run entry
https://pulpito.ceph.com/?branch=wip-yuri8-testing-2023-07-14-0803-reef
https://pulpito.ceph.com/?branch=wip-yuri7-testing-2023-07-14-0803-reef
- https://tracker.ceph.com/issues/51964
qa: test_cephfs_mirror_restart_sync_on_blocklist failure - https://tracker.ceph.com/issues/58340
mds: fsstress.sh hangs with multimds - https://tracker.ceph.com/issues/57676
qa: error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1 - https://tracker.ceph.com/issues/61892
Test failure: test_snapshot_remove (tasks.cephfs.test_strays.TestStrays) - https://tracker.ceph.com/issues/48773
qa: scrub does not complete - https://tracker.ceph.com/issues/55332 (fixed - needs backport)
Failure in snaptest-git-ceph.sh - see https://tracker.ceph.com/issues/55332#note-41 - https://tracker.ceph.com/issues/61201
qa: test_rebuild_moved_file (tasks/data-scan) fails because mds crashes in pacific - https://tracker.ceph.com/issues/62076
reef: Test failure: test_grow_shrink (tasks.cephfs.test_failover.TestMultiFilesystems) - https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc - https://tracker.ceph.com/issues/58220
Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1: - https://tracker.ceph.com/issues/44565
src/mds/SimpleLock.h: 528: FAILED ceph_assert(state LOCK_XLOCK || state LOCK_XLOCKDONE || state LOCK_XLOCKSNAP || state LOCK_LOCK_XLOCK || state == LOCK_LOCK || is_locallock()) - https://tracker.ceph.com/issues/62081
tasks/fscrypt-common does not finish, timesout
11 July 2023¶
https://pulpito.ceph.com/?branch=wip-yuri8-testing-2023-07-06-0909-reef
- https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure - https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log - https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1 - https://tracker.ceph.com/issues/58340
mds: fsstress.sh hangs with multimds - https://tracker.ceph.com/issues/61892
Test failure: test_snapshot_remove (tasks.cephfs.test_strays.TestStrays) - https://tracker.ceph.com/issues/51964
qa: test_cephfs_mirror_restart_sync_on_blocklist failure - https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc - https://tracker.ceph.com/issues/58220
Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1: - https://tracker.ceph.com/issues/48773
qa: scrub does not complete
27 Jun 2023¶
- https://tracker.ceph.com/issues/61182
workloads/cephfs-mirror-ha-workunit: reached maximum tries (50) after waiting for 300 seconds (mirror daemon stop times out) - https://tracker.ceph.com/issues/61820
During valgrind test: CommandFailedError: Command failed on smithi133 with status 1: 'sudo ceph --cluster ceph osd crush tunables default' - https://tracker.ceph.com/issues/61399 *** [Makefile:299: ior] Error 1
- https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure - https://tracker.ceph.com/issues/61574
[Makefile:59: mdtest] Error 1 (undefined reference to `fi_strerror') - https://tracker.ceph.com/issues/48773
qa: scrub does not complete - https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log - https://tracker.ceph.com/issues/57206
qa/workunits/libcephfs/test.sh - ceph_test_libcephfs_reclaim core dump - segmentation fault - https://tracker.ceph.com/issues/59344
qa: workunit test fs/quota/quota.sh failed with "setfattr: .: Invalid argument"
05 Jun 2023¶
https://tracker.ceph.com/issues/61515#note-1
Re-run: https://pulpito.ceph.com/yuriw-2023-06-02_15:02:25-fs-reef-release-distro-default-smithi/
Known Failures¶
- https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1 - https://tracker.ceph.com/issues/59683
fs/fscrypt.sh - Error: Unable to find a match: userspace-rcu-devel libedit-devel device-mapper-devel - https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure - https://tracker.ceph.com/issues/61399
qa: "[Makefile:299: ior] Error 1" - https://tracker.ceph.com/issues/57676
qa: error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/61574
[Makefile:59: mdtest] Error 1 (undefined reference to `fi_strerror') - https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc - https://tracker.ceph.com/issues/48773
qa: scrub does not complete - https://tracker.ceph.com/issues/55332 (fixed - needs backport)
Failure in snaptest-git-ceph.sh - https://tracker.ceph.com/issues/61394
mds.a (mds.0) 1 : cluster [WRN] evicting unresponsive client smithi152 (4298), after 303.726 seconds" in cluster log - https://tracker.ceph.com/issues/51279
kclient hangs on umount
Packaging/Installation Issues: (smithi 202 issues)¶
- tasks/workunits/snaps
Error reimaging machines: Expected smithi202's OS to be rhel 8.6 but found ubuntu 22.04
https://pulpito.ceph.com/yuriw-2023-06-02_15:02:25-fs-reef-release-distro-default-smithi/7294352
- tasks/snapshots
Error reimaging machines: 'ssh_keyscan smithi202.front.sepia.ceph.com' reached maximum tries (6) after waiting for 5 seconds
https://pulpito.ceph.com/yuriw-2023-06-02_15:02:25-fs-reef-release-distro-default-smithi/7294353
02 Jun 2023¶
https://tracker.ceph.com/issues/61515#note-1
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/
Known Test Failures¶
- https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288861
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288876
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288892
- https://tracker.ceph.com/issues/57655 ---------------------------CHECK----------------------------
qa: fs:mixed-clients kernel_untar_build failure
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288880
Revert PR isn't merged to reef. Should we merge it before release ?
https://github.com/ceph/ceph/pull/51661
https://github.com/ceph/ceph/pull/51500
- https://tracker.ceph.com/issues/57676
qa: error during scrub thrashing: rank damage found: {'backtrace'}
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288884
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288901
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288941
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288973
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7289012
- https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288896
- https://tracker.ceph.com/issues/59346
qa/workunits/fs/test_python.sh failed with "AssertionError: DiskQuotaExceeded not raised by write"
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288903
- https://tracker.ceph.com/issues/55332 (fixed - needs backport)
Failure in snaptest-git-ceph.sh
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288905
- https://tracker.ceph.com/issues/59344
qa: workunit test fs/quota/quota.sh failed with "setfattr: .: Invalid argument"
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288914
- https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288939
- https://tracker.ceph.com/issues/61394
mds.a (mds.0) 1 : cluster [WRN] evicting unresponsive client smithi152 (4298), after 303.726 seconds" in cluster log
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288948
- https://tracker.ceph.com/issues/61182
workloads/cephfs-mirror-ha-workunit: reached maximum tries (50) after waiting for 300 seconds (mirror daemon stop times out)
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288953
- https://tracker.ceph.com/issues/51279 ---------------------------------CHECK-------------------------------------------
kclient hangs on umount
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288964
Known Make Failures¶
- https://tracker.ceph.com/issues/59683
fs/fscrypt.sh - Error: Unable to find a match: userspace-rcu-devel libedit-devel device-mapper-devel
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288877
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288912
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288995
- https://tracker.ceph.com/issues/61399
qa: "[Makefile:299: ior] Error 1"
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288890
- https://tracker.ceph.com/issues/61574
[Makefile:59: mdtest] Error 1 (undefined reference to `fi_strerror')
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288920
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7289015
Package/Installation Issues¶
- fs/upgrade/mds_upgrade_sequence : Command failed on smithi119 with status 1: 'sudo yum install -y kernel'
2023-05-28T18:58:30.824 INFO:teuthology.orchestra.run.smithi119.stderr:FATAL ERROR: python callback ??? failed, aborting!
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288908
- Command failed on smithi119 with status 100: 'sudo apt-get clean'
2023-05-28T19:06:41.566 INFO:teuthology.orchestra.run.smithi119.stderr:E: Could not get lock /var/cache/apt/archives/lock. It is held by process 2127 (apt-get)
2023-05-28T19:06:41.566 INFO:teuthology.orchestra.run.smithi119.stderr:E: Unable to lock directory /var/cache/apt/archives/
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288928
- Command failed on smithi119 with status 2: 'sudo dpkg -i /tmp/linux-image.deb'
2023-05-28T19:18:49.306 DEBUG:teuthology.orchestra.run.smithi119:> sudo dpkg -i /tmp/linux-image.deb
2023-05-28T19:18:49.315 INFO:teuthology.orchestra.run.smithi119.stderr:dpkg: error: dpkg frontend lock was locked by another process with pid 2205
2023-05-28T19:18:49.315 INFO:teuthology.orchestra.run.smithi119.stderr:Note: removing the lock file is always wrong, and can end up damaging the
2023-05-28T19:18:49.315 INFO:teuthology.orchestra.run.smithi119.stderr:locked area and the entire system. See <https://wiki.debian.org/Teams/Dpkg/FAQ>.
2023-05-28T19:18:49.316 DEBUG:teuthology.orchestra.run:got remote process result: 2
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288942
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288970
- Command failed on smithi119 with status 100: 'sudo apt-get clean'
2023-05-28T19:41:50.171 DEBUG:teuthology.orchestra.run.smithi119:> sudo apt-get clean
...
2023-05-28T19:41:50.398 INFO:teuthology.orchestra.run.smithi119.stderr:E: Could not get lock /var/cache/apt/archives/lock. It is held by process 2120 (apt-get)
2023-05-28T19:41:50.399 INFO:teuthology.orchestra.run.smithi119.stderr:E: Unable to lock directory /var/cache/apt/archives/
2023-05-28T19:41:50.400 DEBUG:teuthology.orchestra.run:got remote process result: 100
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288987
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7289009
- tasks/mds-flush Command failed on smithi119 with status 1: 'sudo yum install -y kernel'
2023-05-28T19:49:47.762 INFO:teuthology.orchestra.run.smithi119.stderr:[Errno 2] No such file or directory: '/var/cache/dnf/baseos-055ffcb2ec25a27f/packages/kernel-4.18.0-492.el8.x86_64.rpm'
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288998
ssh connection to smithi119 was lost¶
FAILED:
- Failed to connect to the host via ssh: ssh: connect to host smithi119.front.sepia.ceph.com port 22: No route to host
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7289021
DEAD:
- SSH connection to smithi119 was lost: 'sudo grub2-mkconfig -o /boot/grub2/grub.cfg'
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288909
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288996
- SSH connection to smithi119 was lost: 'sudo DEBIAN_FRONTEND=noninteractive apt-get -y install linux-image-generic'
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288926
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288938
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288958
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7289008
- SSH connection to smithi119 was lost: 'sudo update-grub'
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7288982
Others¶
- https://tracker.ceph.com/issues/61575
task/mds_thrash.py - assert manager.is_clean()
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7289030/
- https://tracker.ceph.com/issues/61576
Failed to manage policy for boolean nagios_run_sudo: [Errno 11] Resource temporarily unavailable
https://pulpito.ceph.com/yuriw-2023-05-28_14:46:14-fs-reef-release-distro-default-smithi/7289016
23 May 2023¶
- https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1 - https://tracker.ceph.com/issues/59683
fs/fscrypt.sh - Error: Unable to find a match: userspace-rcu-devel libedit-devel device-mapper-devel - https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure - https://tracker.ceph.com/issues/57676
qa: error during scrub thrashing: rank damage found: {'backtrace'} - https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log - https://tracker.ceph.com/issues/56695 (fixed - backport not yet merged in reef)
Command failed (workunit test suites/pjd.sh) - https://tracker.ceph.com/issues/58340
mds: fsstress.sh hangs with multimds - https://tracker.ceph.com/issues/55332 (fixed - needs backport)
Failure in snaptest-git-ceph.sh - https://tracker.ceph.com/issues/61358
osd.2 (osd.2) 3 : cluster [WRN] 1 slow requests (by type [ 'delayed' : 1 ] most affected pool [ 'cephfs_data' : 1 ])" in cluster log - https://tracker.ceph.com/issues/61182
workloads/cephfs-mirror-ha-workunit: reached maximum tries (50) after waiting for 300 seconds (mirror daemon stop times out) - https://tracker.ceph.com/issues/59297
qa: test_join_fs_unset failure - https://tracker.ceph.com/issues/51282
cluster [WRN] Health check failed: Degraded data redundancy: 9/46 objects degraded (19.565%), 4 pgs degraded (PG_DEGRADED)" in cluster log
19 May 2023¶
https://pulpito.ceph.com/yuriw-2023-05-10_18:53:39-fs-wip-yuri3-testing-2023-05-10-0851-reef-distro-default-smithi/
https://pulpito.ceph.com/yuriw-2023-05-15_15:47:06-fs-wip-yuri3-testing-2023-05-10-0851-reef-distro-default-smithi/
- https://tracker.ceph.com/issues/57655
qa: fs:mixed-clients kernel_untar_build failure - https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1 - https://tracker.ceph.com/issues/61265 (NEW - not related)
tasks.cephfs.fuse_mount:process failed to terminate after unmount - https://tracker.ceph.com/issues/56446 (fixed - reef backport yet to be merged)
Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits) - https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log - https://tracker.ceph.com/issues/55332 (fixed - needs backport)
Failure in snaptest-git-ceph.sh - https://tracker.ceph.com/issues/59667
cluster [WRN] client could not reconnect as file system flag refuse_client_session is set" in cluster log - https://tracker.ceph.com/issues/56506
Test failure: test_rebuild_backtraceless (tasks.cephfs.test_data_scan.TestDataScan) - https://tracker.ceph.com/issues/61278 (NEW - teuthology bug)
Config file not found: "/home/teuthworker/src/github.com_ceph_ceph-c_ef2746e02c24971e8f0b792fa43f63ba4ae04ec2/qa/tasks/cephadm.conf". - https://tracker.ceph.com/issues/51282
cluster [WRN] Health check failed: Degraded data redundancy: 9/46 objects degraded (19.565%), 4 pgs degraded (PG_DEGRADED)" in cluster log - https://tracker.ceph.com/issues/59683
fs/fscrypt.sh - Error: Unable to find a match: userspace-rcu-devel libedit-devel device-mapper-devel - https://tracker.ceph.com/issues/48773
workunit/postgres - error during scrub thrashing: reached maximum tries (30) after waiting for 900 seconds - https://tracker.ceph.com/issues/61279
qa: test_dirfrag_limit (tasks.cephfs.test_strays.TestStrays) failed
17 May 2023¶
- https://tracker.ceph.com/issues/59683
fs/fscrypt.sh - Error: Unable to find a match: userspace-rcu-devel libedit-devel device-mapper-devel - https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi002 with status 1 - https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc - https://tracker.ceph.com/issues/56446 (fixed - reef backport yet to be merged)
Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits) - https://tracker.ceph.com/issues/61182
workloads/cephfs-mirror-ha-workunit: reached maximum tries (50) after waiting for 300 seconds (mirror daemon stop times out) - https://tracker.ceph.com/issues/51964
Test failure: test_cephfs_mirror_restart_sync_on_blocklist (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/55332 (fixed - needs backport)
Failure in snaptest-git-ceph.sh - https://tracker.ceph.com/issues/56695 (fixed - backport not yet merged in reef)
Command failed (workunit test suites/pjd.sh) - https://tracker.ceph.com/issues/58340
mds: fsstress.sh hangs with multimds - https://tracker.ceph.com/issues/48773
workunit/postgres - error during scrub thrashing: reached maximum tries (30) after waiting for 900 seconds - https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log - https://tracker.ceph.com/issues/57656
[testing] dbench: write failed on handle 10009 (Resource temporarily unavailable) - https://tracker.ceph.com/issues/53302
tasks/fscrypt-iozone- Command failed on smithi079 with status 3: 'sudo logrotate /etc/logrotate.d/ceph-test.conf'
16 May 2023¶
- https://tracker.ceph.com/issues/58126 [kclient bug]
Test failure: test_fscrypt_dummy_encryption_with_quick_group (tasks.cephfs.test_fscrypt.TestFscrypt) - https://tracker.ceph.com/issues/61182
workloads/cephfs-mirror-ha-workunit: reached maximum tries (50) after waiting for 300 seconds (mirror daemon stop times out) - https://tracker.ceph.com/issues/47292
Test failure: test_df_for_valid_file (tasks.cephfs.test_cephfs_shell.TestDF) - https://tracker.ceph.com/issues/51964
Test failure: test_cephfs_mirror_restart_sync_on_blocklist (tasks.cephfs.test_mirroring.TestMirroring) - https://tracker.ceph.com/issues/59667
cluster [WRN] client could not reconnect as file system flag refuse_client_session is set" in cluster log - https://tracker.ceph.com/issues/52624
cluster [WRN] Health check failed: Reduced data availability: 1 pg inactive, 1 pg peering (PG_AVAILABILITY)" in cluster log - https://tracker.ceph.com/issues/61400
valgrind+ceph-mon: segmentation fault in rocksdb+tcmalloc - https://tracker.ceph.com/issues/48773
workunit/postgres - error during scrub thrashing: reached maximum tries (30) after waiting for 900 seconds - https://tracker.ceph.com/issues/57206
qa/workunits/libcephfs/test.sh - ceph_test_libcephfs_reclaim core dump - segmentation fault - https://tracker.ceph.com/issues/38704 (Very old one resolved - ignore in testcase)
tasks/cephfs_misc_tests - cluster [WRN] Health check failed: 1/3 mons down, quorum a,b (MON_DOWN)" in cluster log' - https://tracker.ceph.com/issues/54460
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi099 with status 1 - https://tracker.ceph.com/issues/53302
tasks/fscrypt-iozone- Command failed on smithi079 with status 3: 'sudo logrotate /etc/logrotate.d/ceph-test.conf'
tasks/cfuse_workunit_suites_blogbench traceless/50pc - Command failed on smithi133 with status 3: 'sudo logrotate /etc/logrotate.d/ceph-test.conf' - https://tracker.ceph.com/issues/58220
Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1 - https://tracker.ceph.com/issues/58340
mds: fsstress.sh hangs with multimds - https://tracker.ceph.com/issues/59683
fs/fscrypt.sh - Error: Unable to find a match: userspace-rcu-devel libedit-devel device-mapper-devel
Updated by Venky Shankar 11 days ago · 86 revisions