Project

General

Profile

Main » History » Version 101

Rishabh Dave, 01/30/2023 05:55 AM

1 79 Venky Shankar
h1. MAIN
2
3 101 Rishabh Dave
h3. 30 JAN 2023
4
5
run: http://pulpito.front.sepia.ceph.com/rishabh-2022-11-28_08:04:11-fs-wip-rishabh-testing-2022Nov24-1818-testing-default-smithi/
6
re-run: http://pulpito.front.sepia.ceph.com/rishabh-2023-01-13_12:08:33-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/
7
re-run of re-run: http://pulpito.front.sepia.ceph.com/rishabh-2023-01-23_18:53:32-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/
8
9
* https://tracker.ceph.com/issues/56695
10
  [RHEL stock] pjd test failures
11
* https://tracker.ceph.com/issues/57676
12
  qa: error during scrub thrashing: rank damage found: {'backtrace'}
13
* https://tracker.ceph.com/issues/55332
14
  Failure in snaptest-git-ceph.sh
15
* https://tracker.ceph.com/issues/51964
16
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
17
* https://tracker.ceph.com/issues/56446
18
  Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
19
* https://tracker.ceph.com/issues/57655 
20
  qa: fs:mixed-clients kernel_untar_build failure
21
* https://tracker.ceph.com/issues/54460
22
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
23
* https://tracker.ceph.com/issues/58340
24
  mds: fsstress.sh hangs with multimds
25
26 100 Venky Shankar
h3. 15 Dec 2022
27
28
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221215.112736
29
30
* https://tracker.ceph.com/issues/52624
31
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
32
* https://tracker.ceph.com/issues/56695
33
    [RHEL stock] pjd test failures
34
* https://tracker.ceph.com/issues/58219
35
* https://tracker.ceph.com/issues/57655
36
* qa: fs:mixed-clients kernel_untar_build failure
37
    Test failure: test_journal_migration (tasks.cephfs.test_journal_migration.TestJournalMigration)
38
* https://tracker.ceph.com/issues/57676
39
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
40
* https://tracker.ceph.com/issues/58340
41
    mds: fsstress.sh hangs with multimds
42
43 96 Venky Shankar
h3. 08 Dec 2022
44
45
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221130.043104
46 99 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221209.043803
47 96 Venky Shankar
48
(lots of transient git.ceph.com failures)
49
50
* https://tracker.ceph.com/issues/52624
51
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
52
* https://tracker.ceph.com/issues/56695
53
    [RHEL stock] pjd test failures
54
* https://tracker.ceph.com/issues/57655
55
    qa: fs:mixed-clients kernel_untar_build failure
56
* https://tracker.ceph.com/issues/58219
57
    Test failure: test_journal_migration (tasks.cephfs.test_journal_migration.TestJournalMigration)
58
* https://tracker.ceph.com/issues/58220
59
    Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1:
60
* https://tracker.ceph.com/issues/57676
61
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
62 97 Venky Shankar
* https://tracker.ceph.com/issues/53859
63
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
64 98 Venky Shankar
* https://tracker.ceph.com/issues/54460
65
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
66
* https://tracker.ceph.com/issues/58244
67
    Test failure: test_rebuild_inotable (tasks.cephfs.test_data_scan.TestDataScan)
68 96 Venky Shankar
69 95 Venky Shankar
h3. 14 Oct 2022
70
71
https://pulpito.ceph.com/vshankar-2022-10-12_04:56:59-fs-wip-vshankar-testing-20221011-145847-testing-default-smithi/
72
https://pulpito.ceph.com/vshankar-2022-10-14_04:04:57-fs-wip-vshankar-testing-20221014-072608-testing-default-smithi/
73
74
* https://tracker.ceph.com/issues/52624
75
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
76
* https://tracker.ceph.com/issues/55804
77
    Command failed (workunit test suites/pjd.sh)
78
* https://tracker.ceph.com/issues/51964
79
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
80
* https://tracker.ceph.com/issues/57682
81
    client: ERROR: test_reconnect_after_blocklisted
82
* https://tracker.ceph.com/issues/54460
83
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
84 90 Rishabh Dave
85 91 Rishabh Dave
h3. 10 Oct 2022
86
87
http://pulpito.front.sepia.ceph.com/rishabh-2022-09-30_19:45:21-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
88 92 Rishabh Dave
89 91 Rishabh Dave
reruns
90
* fs-thrash, passed: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-04_13:19:47-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
91
* fs-verify, passed: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-05_12:25:37-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
92
* cephadm failures also passed after many re-runs: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-06_13:50:51-fs-wip-rishabh-testing-30Sep2022-2-testing-default-smithi/
93 94 Rishabh Dave
    ** needed this PR to be merged in ceph-ci branch - https://github.com/ceph/ceph/pull/47458
94 91 Rishabh Dave
95 93 Rishabh Dave
known bugs
96 91 Rishabh Dave
* https://tracker.ceph.com/issues/52624
97
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
98
* https://tracker.ceph.com/issues/50223
99
  client.xxxx isn't responding to mclientcaps(revoke
100
* https://tracker.ceph.com/issues/57299
101
  qa: test_dump_loads fails with JSONDecodeError
102
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
103
  qa: fs:mixed-clients kernel_untar_build failure
104
* https://tracker.ceph.com/issues/57206
105
  libcephfs/test.sh: ceph_test_libcephfs_reclaim
106
107 90 Rishabh Dave
h3. 2022 Sep 29
108
109
http://pulpito.front.sepia.ceph.com/rishabh-2022-09-14_12:48:43-fs-wip-rishabh-testing-2022Sep9-1708-testing-default-smithi/
110
111
* https://tracker.ceph.com/issues/55804
112
  Command failed (workunit test suites/pjd.sh)
113
* https://tracker.ceph.com/issues/36593
114
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
115
* https://tracker.ceph.com/issues/52624
116
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
117
* https://tracker.ceph.com/issues/51964
118
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
119
* https://tracker.ceph.com/issues/56632
120
  Test failure: test_subvolume_snapshot_clone_quota_exceeded
121
* https://tracker.ceph.com/issues/50821
122
  qa: untar_snap_rm failure during mds thrashing
123
124 88 Patrick Donnelly
h3. 2022 Sep 26
125
126
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220923.171109
127
128
* https://tracker.ceph.com/issues/55804
129
    qa failure: pjd link tests failed
130
* https://tracker.ceph.com/issues/57676
131
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
132
* https://tracker.ceph.com/issues/52624
133
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
134
* https://tracker.ceph.com/issues/57580
135
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
136
* https://tracker.ceph.com/issues/48773
137
    qa: scrub does not complete
138
* https://tracker.ceph.com/issues/57299
139
    qa: test_dump_loads fails with JSONDecodeError
140
* https://tracker.ceph.com/issues/57280
141
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
142
* https://tracker.ceph.com/issues/57205
143
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
144
* https://tracker.ceph.com/issues/57656
145
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
146
* https://tracker.ceph.com/issues/57677
147
    qa: "1 MDSs behind on trimming (MDS_TRIM)"
148
* https://tracker.ceph.com/issues/57206
149
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
150
* https://tracker.ceph.com/issues/57446
151
    qa: test_subvolume_snapshot_info_if_orphan_clone fails
152
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
153
    qa: fs:mixed-clients kernel_untar_build failure
154 89 Patrick Donnelly
* https://tracker.ceph.com/issues/57682
155
    client: ERROR: test_reconnect_after_blocklisted
156 88 Patrick Donnelly
157
158 87 Patrick Donnelly
h3. 2022 Sep 22
159
160
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220920.234701
161
162
* https://tracker.ceph.com/issues/57299
163
    qa: test_dump_loads fails with JSONDecodeError
164
* https://tracker.ceph.com/issues/57205
165
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
166
* https://tracker.ceph.com/issues/52624
167
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
168
* https://tracker.ceph.com/issues/57580
169
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
170
* https://tracker.ceph.com/issues/57280
171
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
172
* https://tracker.ceph.com/issues/48773
173
    qa: scrub does not complete
174
* https://tracker.ceph.com/issues/56446
175
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
176
* https://tracker.ceph.com/issues/57206
177
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
178
* https://tracker.ceph.com/issues/51267
179
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
180
181
NEW:
182
183
* https://tracker.ceph.com/issues/57656
184
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
185
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
186
    qa: fs:mixed-clients kernel_untar_build failure
187
* https://tracker.ceph.com/issues/57657
188
    mds: scrub locates mismatch between child accounted_rstats and self rstats
189
190
Segfault probably caused by: https://github.com/ceph/ceph/pull/47795#issuecomment-1255724799
191
192
193 80 Venky Shankar
h3. 2022 Sep 16
194 79 Venky Shankar
195
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220905-132828
196
197
* https://tracker.ceph.com/issues/57446
198
    qa: test_subvolume_snapshot_info_if_orphan_clone fails
199
* https://tracker.ceph.com/issues/57299
200
    qa: test_dump_loads fails with JSONDecodeError
201
* https://tracker.ceph.com/issues/50223
202
    client.xxxx isn't responding to mclientcaps(revoke)
203
* https://tracker.ceph.com/issues/52624
204
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
205
* https://tracker.ceph.com/issues/57205
206
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
207
* https://tracker.ceph.com/issues/57280
208
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
209
* https://tracker.ceph.com/issues/51282
210
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
211
* https://tracker.ceph.com/issues/48203
212
  https://tracker.ceph.com/issues/36593
213
    qa: quota failure
214
    qa: quota failure caused by clients stepping on each other
215
* https://tracker.ceph.com/issues/57580
216
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
217
218 77 Rishabh Dave
219
h3. 2022 Aug 26
220 76 Rishabh Dave
221
http://pulpito.front.sepia.ceph.com/rishabh-2022-08-22_17:49:59-fs-wip-rishabh-testing-2022Aug19-testing-default-smithi/
222
http://pulpito.front.sepia.ceph.com/rishabh-2022-08-24_11:56:51-fs-wip-rishabh-testing-2022Aug19-testing-default-smithi/
223
224
* https://tracker.ceph.com/issues/57206
225
  libcephfs/test.sh: ceph_test_libcephfs_reclaim
226
* https://tracker.ceph.com/issues/56632
227
  Test failure: test_subvolume_snapshot_clone_quota_exceeded (tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones)
228
* https://tracker.ceph.com/issues/56446
229
  Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
230
* https://tracker.ceph.com/issues/51964
231
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
232
* https://tracker.ceph.com/issues/53859
233
  qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
234
235
* https://tracker.ceph.com/issues/54460
236
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
237
* https://tracker.ceph.com/issues/54462
238
  Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi055 with status 128
239
* https://tracker.ceph.com/issues/54460
240
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
241
* https://tracker.ceph.com/issues/36593
242
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
243
244
* https://tracker.ceph.com/issues/52624
245
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
246
* https://tracker.ceph.com/issues/55804
247
  Command failed (workunit test suites/pjd.sh)
248
* https://tracker.ceph.com/issues/50223
249
  client.xxxx isn't responding to mclientcaps(revoke)
250
251
252 75 Venky Shankar
h3. 2022 Aug 22
253
254
https://pulpito.ceph.com/vshankar-2022-08-12_09:34:24-fs-wip-vshankar-testing1-20220812-072441-testing-default-smithi/
255
https://pulpito.ceph.com/vshankar-2022-08-18_04:30:42-fs-wip-vshankar-testing1-20220818-082047-testing-default-smithi/ (drop problematic PR and re-run)
256
257
* https://tracker.ceph.com/issues/52624
258
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
259
* https://tracker.ceph.com/issues/56446
260
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
261
* https://tracker.ceph.com/issues/55804
262
    Command failed (workunit test suites/pjd.sh)
263
* https://tracker.ceph.com/issues/51278
264
    mds: "FAILED ceph_assert(!segments.empty())"
265
* https://tracker.ceph.com/issues/54460
266
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
267
* https://tracker.ceph.com/issues/57205
268
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
269
* https://tracker.ceph.com/issues/57206
270
    ceph_test_libcephfs_reclaim crashes during test
271
* https://tracker.ceph.com/issues/53859
272
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
273
* https://tracker.ceph.com/issues/50223
274
    client.xxxx isn't responding to mclientcaps(revoke)
275
276 72 Venky Shankar
h3. 2022 Aug 12
277
278
https://pulpito.ceph.com/vshankar-2022-08-10_04:06:00-fs-wip-vshankar-testing-20220805-190751-testing-default-smithi/
279
https://pulpito.ceph.com/vshankar-2022-08-11_12:16:58-fs-wip-vshankar-testing-20220811-145809-testing-default-smithi/ (drop problematic PR and re-run)
280
281
* https://tracker.ceph.com/issues/52624
282
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
283
* https://tracker.ceph.com/issues/56446
284
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
285
* https://tracker.ceph.com/issues/51964
286
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
287
* https://tracker.ceph.com/issues/55804
288
    Command failed (workunit test suites/pjd.sh)
289
* https://tracker.ceph.com/issues/50223
290
    client.xxxx isn't responding to mclientcaps(revoke)
291
* https://tracker.ceph.com/issues/50821
292
    qa: untar_snap_rm failure during mds thrashing
293
* https://tracker.ceph.com/issues/54460
294 73 Venky Shankar
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
295 72 Venky Shankar
296 71 Venky Shankar
h3. 2022 Aug 04
297
298
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220804-123835 (only mgr/volumes, mgr/stats)
299
300
Unrealted teuthology failure on rhel
301
302 69 Rishabh Dave
h3. 2022 Jul 25
303 68 Rishabh Dave
304
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-22_11:34:20-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
305
306
1st re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_03:51:19-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi
307
2nd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
308 74 Rishabh Dave
3rd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
309
4th (final) re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-28_03:59:01-fs-wip-rishabh-testing-2022Jul28-0143-testing-default-smithi/
310 68 Rishabh Dave
311
* https://tracker.ceph.com/issues/55804
312
  Command failed (workunit test suites/pjd.sh)
313
* https://tracker.ceph.com/issues/50223
314
  client.xxxx isn't responding to mclientcaps(revoke)
315
316
* https://tracker.ceph.com/issues/54460
317
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
318
* https://tracker.ceph.com/issues/36593
319
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
320 1 Patrick Donnelly
* https://tracker.ceph.com/issues/54462
321 74 Rishabh Dave
  Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi055 with status 128~
322 68 Rishabh Dave
323 67 Patrick Donnelly
h3. 2022 July 22
324
325
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220721.235756
326
327
MDS_HEALTH_DUMMY error in log fixed by followup commit.
328
transient selinux ping failure
329
330
* https://tracker.ceph.com/issues/56694
331
    qa: avoid blocking forever on hung umount
332
* https://tracker.ceph.com/issues/56695
333
    [RHEL stock] pjd test failures
334
* https://tracker.ceph.com/issues/56696
335
    admin keyring disappears during qa run
336
* https://tracker.ceph.com/issues/56697
337
    qa: fs/snaps fails for fuse
338
* https://tracker.ceph.com/issues/50222
339
    osd: 5.2s0 deep-scrub : stat mismatch
340
* https://tracker.ceph.com/issues/56698
341
    client: FAILED ceph_assert(_size == 0)
342
* https://tracker.ceph.com/issues/50223
343
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
344
345
346 66 Rishabh Dave
h3. 2022 Jul 15
347 65 Rishabh Dave
348
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-08_23:53:34-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
349
350
re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-15_06:42:04-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
351
352
* https://tracker.ceph.com/issues/53859
353
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
354
* https://tracker.ceph.com/issues/55804
355
  Command failed (workunit test suites/pjd.sh)
356
* https://tracker.ceph.com/issues/50223
357
  client.xxxx isn't responding to mclientcaps(revoke)
358
* https://tracker.ceph.com/issues/50222
359
  osd: deep-scrub : stat mismatch
360
361
* https://tracker.ceph.com/issues/56632
362
  Test failure: test_subvolume_snapshot_clone_quota_exceeded (tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones)
363
* https://tracker.ceph.com/issues/56634
364
  workunit test fs/snaps/snaptest-intodir.sh
365
* https://tracker.ceph.com/issues/56644
366
  Test failure: test_rapid_creation (tasks.cephfs.test_fragment.TestFragmentation)
367
368
369
370 61 Rishabh Dave
h3. 2022 July 05
371
372
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-02_14:14:52-fs-wip-rishabh-testing-20220702-1631-testing-default-smithi/
373 62 Rishabh Dave
374 64 Rishabh Dave
On 1st re-run some jobs passed - http://pulpito.front.sepia.ceph.com/rishabh-2022-07-03_15:10:28-fs-wip-rishabh-testing-20220702-1631-distro-default-smithi/
375
376
On 2nd re-run only few jobs failed -
377
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
378
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
379 62 Rishabh Dave
380
* https://tracker.ceph.com/issues/56446
381
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
382
* https://tracker.ceph.com/issues/55804
383
    Command failed (workunit test suites/pjd.sh) on smithi047 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/
384
385
* https://tracker.ceph.com/issues/56445
386
    Command failed on smithi080 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
387
* https://tracker.ceph.com/issues/51267
388 63 Rishabh Dave
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi098 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1
389
* https://tracker.ceph.com/issues/50224
390
    Test failure: test_mirroring_init_failure_with_recovery (tasks.cephfs.test_mirroring.TestMirroring)
391 62 Rishabh Dave
392
393 61 Rishabh Dave
394 58 Venky Shankar
h3. 2022 July 04
395
396
https://pulpito.ceph.com/vshankar-2022-06-29_09:19:00-fs-wip-vshankar-testing-20220627-100931-testing-default-smithi/
397
(rhel runs were borked due to: https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/JSZQFUKVLDND4W33PXDGCABPHNSPT6SS/, tests ran with --filter-out=rhel)
398
399
* https://tracker.ceph.com/issues/56445
400
    Command failed on smithi162 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
401
* https://tracker.ceph.com/issues/56446
402 59 Rishabh Dave
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
403
* https://tracker.ceph.com/issues/51964
404
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
405
* https://tracker.ceph.com/issues/52624
406 60 Rishabh Dave
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
407 59 Rishabh Dave
408 57 Venky Shankar
h3. 2022 June 20
409
410
https://pulpito.ceph.com/vshankar-2022-06-15_04:03:39-fs-wip-vshankar-testing1-20220615-072516-testing-default-smithi/
411
https://pulpito.ceph.com/vshankar-2022-06-19_08:22:46-fs-wip-vshankar-testing1-20220619-102531-testing-default-smithi/
412
413
* https://tracker.ceph.com/issues/52624
414
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
415
* https://tracker.ceph.com/issues/55804
416
    qa failure: pjd link tests failed
417
* https://tracker.ceph.com/issues/54108
418
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
419
* https://tracker.ceph.com/issues/55332
420
    Failure in snaptest-git-ceph.sh (it's an async unlink/create bug)
421
422 56 Patrick Donnelly
h3. 2022 June 13
423
424
https://pulpito.ceph.com/pdonnell-2022-06-12_05:08:12-fs:workload-wip-pdonnell-testing-20220612.004943-distro-default-smithi/
425
426
* https://tracker.ceph.com/issues/56024
427
    cephadm: removes ceph.conf during qa run causing command failure
428
* https://tracker.ceph.com/issues/48773
429
    qa: scrub does not complete
430
* https://tracker.ceph.com/issues/56012
431
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
432
433
434 55 Venky Shankar
h3. 2022 Jun 13
435 54 Venky Shankar
436
https://pulpito.ceph.com/vshankar-2022-06-07_00:25:50-fs-wip-vshankar-testing-20220606-223254-testing-default-smithi/
437
https://pulpito.ceph.com/vshankar-2022-06-10_01:04:46-fs-wip-vshankar-testing-20220609-175550-testing-default-smithi/
438
439
* https://tracker.ceph.com/issues/52624
440
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
441
* https://tracker.ceph.com/issues/51964
442
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
443
* https://tracker.ceph.com/issues/53859
444
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
445
* https://tracker.ceph.com/issues/55804
446
    qa failure: pjd link tests failed
447
* https://tracker.ceph.com/issues/56003
448
    client: src/include/xlist.h: 81: FAILED ceph_assert(_size == 0)
449
* https://tracker.ceph.com/issues/56011
450
    fs/thrash: snaptest-snap-rm-cmp.sh fails in mds5sum comparison
451
* https://tracker.ceph.com/issues/56012
452
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
453
454 53 Venky Shankar
h3. 2022 Jun 07
455
456
https://pulpito.ceph.com/vshankar-2022-06-06_21:25:41-fs-wip-vshankar-testing1-20220606-230129-testing-default-smithi/
457
https://pulpito.ceph.com/vshankar-2022-06-07_10:53:31-fs-wip-vshankar-testing1-20220607-104134-testing-default-smithi/ (rerun after dropping a problematic PR)
458
459
* https://tracker.ceph.com/issues/52624
460
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
461
* https://tracker.ceph.com/issues/50223
462
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
463
* https://tracker.ceph.com/issues/50224
464
    qa: test_mirroring_init_failure_with_recovery failure
465
466 51 Venky Shankar
h3. 2022 May 12
467
468
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220509-125847
469 52 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-13_17:09:16-fs-wip-vshankar-testing-20220513-120051-testing-default-smithi/ (drop prs + rerun)
470 51 Venky Shankar
471
* https://tracker.ceph.com/issues/52624
472
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
473
* https://tracker.ceph.com/issues/50223
474
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
475
* https://tracker.ceph.com/issues/55332
476
    Failure in snaptest-git-ceph.sh
477
* https://tracker.ceph.com/issues/53859
478
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
479
* https://tracker.ceph.com/issues/55538
480 1 Patrick Donnelly
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
481 52 Venky Shankar
* https://tracker.ceph.com/issues/55258
482
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs (cropss up again, though very infrequent)
483 51 Venky Shankar
484 49 Venky Shankar
h3. 2022 May 04
485
486 50 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-01_13:18:44-fs-wip-vshankar-testing1-20220428-204527-testing-default-smithi/
487
https://pulpito.ceph.com/vshankar-2022-05-02_16:58:59-fs-wip-vshankar-testing1-20220502-201957-testing-default-smithi/ (after dropping PRs)
488
489 49 Venky Shankar
* https://tracker.ceph.com/issues/52624
490
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
491
* https://tracker.ceph.com/issues/50223
492
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
493
* https://tracker.ceph.com/issues/55332
494
    Failure in snaptest-git-ceph.sh
495
* https://tracker.ceph.com/issues/53859
496
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
497
* https://tracker.ceph.com/issues/55516
498
    qa: fs suite tests failing with "json.decoder.JSONDecodeError: Extra data: line 2 column 82 (char 82)"
499
* https://tracker.ceph.com/issues/55537
500
    mds: crash during fs:upgrade test
501
* https://tracker.ceph.com/issues/55538
502
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
503
504 48 Venky Shankar
h3. 2022 Apr 25
505
506
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220420-113951 (owner vshankar)
507
508
* https://tracker.ceph.com/issues/52624
509
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
510
* https://tracker.ceph.com/issues/50223
511
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
512
* https://tracker.ceph.com/issues/55258
513
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
514
* https://tracker.ceph.com/issues/55377
515
    kclient: mds revoke Fwb caps stuck after the kclient tries writebcak once
516
517 47 Venky Shankar
h3. 2022 Apr 14
518
519
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220411-144044
520
521
* https://tracker.ceph.com/issues/52624
522
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
523
* https://tracker.ceph.com/issues/50223
524
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
525
* https://tracker.ceph.com/issues/52438
526
    qa: ffsb timeout
527
* https://tracker.ceph.com/issues/55170
528
    mds: crash during rejoin (CDir::fetch_keys)
529
* https://tracker.ceph.com/issues/55331
530
    pjd failure
531
* https://tracker.ceph.com/issues/48773
532
    qa: scrub does not complete
533
* https://tracker.ceph.com/issues/55332
534
    Failure in snaptest-git-ceph.sh
535
* https://tracker.ceph.com/issues/55258
536
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
537
538 45 Venky Shankar
h3. 2022 Apr 11
539
540 46 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-55110-20220408-203242
541 45 Venky Shankar
542
* https://tracker.ceph.com/issues/48773
543
    qa: scrub does not complete
544
* https://tracker.ceph.com/issues/52624
545
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
546
* https://tracker.ceph.com/issues/52438
547
    qa: ffsb timeout
548
* https://tracker.ceph.com/issues/48680
549
    mds: scrubbing stuck "scrub active (0 inodes in the stack)"
550
* https://tracker.ceph.com/issues/55236
551
    qa: fs/snaps tests fails with "hit max job timeout"
552
* https://tracker.ceph.com/issues/54108
553
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
554
* https://tracker.ceph.com/issues/54971
555
    Test failure: test_perf_stats_stale_metrics (tasks.cephfs.test_mds_metrics.TestMDSMetrics)
556
* https://tracker.ceph.com/issues/50223
557
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
558
* https://tracker.ceph.com/issues/55258
559
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
560
561 44 Venky Shankar
h3. 2022 Mar 21
562 42 Venky Shankar
563 43 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-03-20_02:16:37-fs-wip-vshankar-testing-20220319-163539-testing-default-smithi/
564
565
Run didn't go well, lots of failures - debugging by dropping PRs and running against master branch. Only merging unrelated PRs that pass tests.
566
567
568
h3. 2022 Mar 08
569
570 42 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-02-28_04:32:15-fs-wip-vshankar-testing-20220226-211550-testing-default-smithi/
571
572
rerun with
573
- (drop) https://github.com/ceph/ceph/pull/44679
574
- (drop) https://github.com/ceph/ceph/pull/44958
575
https://pulpito.ceph.com/vshankar-2022-03-06_14:47:51-fs-wip-vshankar-testing-20220304-132102-testing-default-smithi/
576
577
* https://tracker.ceph.com/issues/54419 (new)
578
    `ceph orch upgrade start` seems to never reach completion
579
* https://tracker.ceph.com/issues/51964
580
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
581
* https://tracker.ceph.com/issues/52624
582
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
583
* https://tracker.ceph.com/issues/50223
584
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
585
* https://tracker.ceph.com/issues/52438
586
    qa: ffsb timeout
587
* https://tracker.ceph.com/issues/50821
588
    qa: untar_snap_rm failure during mds thrashing
589
590
591 41 Venky Shankar
h3. 2022 Feb 09
592
593
https://pulpito.ceph.com/vshankar-2022-02-05_17:27:49-fs-wip-vshankar-testing-20220201-113815-testing-default-smithi/
594
595
rerun with
596
- (drop) https://github.com/ceph/ceph/pull/37938
597
- (drop) https://github.com/ceph/ceph/pull/44335
598
- (drop) https://github.com/ceph/ceph/pull/44491
599
- (drop) https://github.com/ceph/ceph/pull/44501
600
https://pulpito.ceph.com/vshankar-2022-02-08_14:27:29-fs-wip-vshankar-testing-20220208-181241-testing-default-smithi/
601
602
* https://tracker.ceph.com/issues/51964
603
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
604
* https://tracker.ceph.com/issues/54066
605
    test_subvolume_no_upgrade_v1_sanity fails with `AssertionError: 1000 != 0`
606
* https://tracker.ceph.com/issues/48773
607
    qa: scrub does not complete
608
* https://tracker.ceph.com/issues/52624
609
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
610
* https://tracker.ceph.com/issues/50223
611
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
612
* https://tracker.ceph.com/issues/52438
613
    qa: ffsb timeout
614
615 40 Patrick Donnelly
h3. 2022 Feb 01
616
617
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220127.171526
618
619
* https://tracker.ceph.com/issues/54107
620
    kclient: hang during umount
621
* https://tracker.ceph.com/issues/54106
622
    kclient: hang during workunit cleanup
623
* https://tracker.ceph.com/issues/54108
624
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
625
* https://tracker.ceph.com/issues/48773
626
    qa: scrub does not complete
627
* https://tracker.ceph.com/issues/52438
628
    qa: ffsb timeout
629
630
631 36 Venky Shankar
h3. 2022 Jan 13
632
633
https://pulpito.ceph.com/vshankar-2022-01-06_13:18:41-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
634 39 Venky Shankar
635 36 Venky Shankar
rerun with:
636 38 Venky Shankar
- (add) https://github.com/ceph/ceph/pull/44570
637
- (drop) https://github.com/ceph/ceph/pull/43184
638 36 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-01-13_04:42:40-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
639
640
* https://tracker.ceph.com/issues/50223
641
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
642
* https://tracker.ceph.com/issues/51282
643
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
644
* https://tracker.ceph.com/issues/48773
645
    qa: scrub does not complete
646
* https://tracker.ceph.com/issues/52624
647
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
648
* https://tracker.ceph.com/issues/53859
649
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
650
651 34 Venky Shankar
h3. 2022 Jan 03
652
653
https://pulpito.ceph.com/vshankar-2021-12-22_07:37:44-fs-wip-vshankar-testing-20211216-114012-testing-default-smithi/
654
https://pulpito.ceph.com/vshankar-2022-01-03_12:27:45-fs-wip-vshankar-testing-20220103-142738-testing-default-smithi/ (rerun)
655
656
* https://tracker.ceph.com/issues/50223
657
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
658
* https://tracker.ceph.com/issues/51964
659
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
660
* https://tracker.ceph.com/issues/51267
661
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
662
* https://tracker.ceph.com/issues/51282
663
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
664
* https://tracker.ceph.com/issues/50821
665
    qa: untar_snap_rm failure during mds thrashing
666
* https://tracker.ceph.com/issues/51278
667
    mds: "FAILED ceph_assert(!segments.empty())"
668 35 Ramana Raja
* https://tracker.ceph.com/issues/52279
669
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
670
671 34 Venky Shankar
672 33 Patrick Donnelly
h3. 2021 Dec 22
673
674
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211222.014316
675
676
* https://tracker.ceph.com/issues/52624
677
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
678
* https://tracker.ceph.com/issues/50223
679
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
680
* https://tracker.ceph.com/issues/52279
681
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
682
* https://tracker.ceph.com/issues/50224
683
    qa: test_mirroring_init_failure_with_recovery failure
684
* https://tracker.ceph.com/issues/48773
685
    qa: scrub does not complete
686
687
688 32 Venky Shankar
h3. 2021 Nov 30
689
690
https://pulpito.ceph.com/vshankar-2021-11-24_07:14:27-fs-wip-vshankar-testing-20211124-094330-testing-default-smithi/
691
https://pulpito.ceph.com/vshankar-2021-11-30_06:23:32-fs-wip-vshankar-testing-20211124-094330-distro-default-smithi/ (rerun w/ QA fixes)
692
693
* https://tracker.ceph.com/issues/53436
694
    mds, mon: mds beacon messages get dropped? (mds never reaches up:active state)
695
* https://tracker.ceph.com/issues/51964
696
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
697
* https://tracker.ceph.com/issues/48812
698
    qa: test_scrub_pause_and_resume_with_abort failure
699
* https://tracker.ceph.com/issues/51076
700
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
701
* https://tracker.ceph.com/issues/50223
702
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
703
* https://tracker.ceph.com/issues/52624
704
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
705
* https://tracker.ceph.com/issues/50250
706
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
707
708
709 31 Patrick Donnelly
h3. 2021 November 9
710
711
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211109.180315
712
713
* https://tracker.ceph.com/issues/53214
714
    qa: "dd: error reading '/sys/kernel/debug/ceph/2a934501-6731-4052-a836-f42229a869be.client4874/metrics': Is a directory"
715
* https://tracker.ceph.com/issues/48773
716
    qa: scrub does not complete
717
* https://tracker.ceph.com/issues/50223
718
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
719
* https://tracker.ceph.com/issues/51282
720
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
721
* https://tracker.ceph.com/issues/52624
722
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
723
* https://tracker.ceph.com/issues/53216
724
    qa: "RuntimeError: value of attributes should be either str or None. client_id"
725
* https://tracker.ceph.com/issues/50250
726
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
727
728
729
730 30 Patrick Donnelly
h3. 2021 November 03
731
732
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211103.023355
733
734
* https://tracker.ceph.com/issues/51964
735
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
736
* https://tracker.ceph.com/issues/51282
737
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
738
* https://tracker.ceph.com/issues/52436
739
    fs/ceph: "corrupt mdsmap"
740
* https://tracker.ceph.com/issues/53074
741
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
742
* https://tracker.ceph.com/issues/53150
743
    pybind/mgr/cephadm/upgrade: tolerate MDS failures during upgrade straddling v16.2.5
744
* https://tracker.ceph.com/issues/53155
745
    MDSMonitor: assertion during upgrade to v16.2.5+
746
747
748 29 Patrick Donnelly
h3. 2021 October 26
749
750
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211025.000447
751
752
* https://tracker.ceph.com/issues/53074
753
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
754
* https://tracker.ceph.com/issues/52997
755
    testing: hang ing umount
756
* https://tracker.ceph.com/issues/50824
757
    qa: snaptest-git-ceph bus error
758
* https://tracker.ceph.com/issues/52436
759
    fs/ceph: "corrupt mdsmap"
760
* https://tracker.ceph.com/issues/48773
761
    qa: scrub does not complete
762
* https://tracker.ceph.com/issues/53082
763
    ceph-fuse: segmenetation fault in Client::handle_mds_map
764
* https://tracker.ceph.com/issues/50223
765
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
766
* https://tracker.ceph.com/issues/52624
767
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
768
* https://tracker.ceph.com/issues/50224
769
    qa: test_mirroring_init_failure_with_recovery failure
770
* https://tracker.ceph.com/issues/50821
771
    qa: untar_snap_rm failure during mds thrashing
772
* https://tracker.ceph.com/issues/50250
773
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
774
775
776
777 27 Patrick Donnelly
h3. 2021 October 19
778
779 28 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211019.013028
780 27 Patrick Donnelly
781
* https://tracker.ceph.com/issues/52995
782
    qa: test_standby_count_wanted failure
783
* https://tracker.ceph.com/issues/52948
784
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
785
* https://tracker.ceph.com/issues/52996
786
    qa: test_perf_counters via test_openfiletable
787
* https://tracker.ceph.com/issues/48772
788
    qa: pjd: not ok 9, 44, 80
789
* https://tracker.ceph.com/issues/52997
790
    testing: hang ing umount
791
* https://tracker.ceph.com/issues/50250
792
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
793
* https://tracker.ceph.com/issues/52624
794
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
795
* https://tracker.ceph.com/issues/50223
796
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
797
* https://tracker.ceph.com/issues/50821
798
    qa: untar_snap_rm failure during mds thrashing
799
* https://tracker.ceph.com/issues/48773
800
    qa: scrub does not complete
801
802
803 26 Patrick Donnelly
h3. 2021 October 12
804
805
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211012.192211
806
807
Some failures caused by teuthology bug: https://tracker.ceph.com/issues/52944
808
809
New test caused failure: https://github.com/ceph/ceph/pull/43297#discussion_r729883167
810
811
812
* https://tracker.ceph.com/issues/51282
813
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
814
* https://tracker.ceph.com/issues/52948
815
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
816
* https://tracker.ceph.com/issues/48773
817
    qa: scrub does not complete
818
* https://tracker.ceph.com/issues/50224
819
    qa: test_mirroring_init_failure_with_recovery failure
820
* https://tracker.ceph.com/issues/52949
821
    RuntimeError: The following counters failed to be set on mds daemons: {'mds.dir_split'}
822
823
824 25 Patrick Donnelly
h3. 2021 October 02
825 23 Patrick Donnelly
826 24 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211002.163337
827
828
Some failures caused by cephadm upgrade test. Fixed in follow-up qa commit.
829
830
test_simple failures caused by PR in this set.
831
832
A few reruns because of QA infra noise.
833
834
* https://tracker.ceph.com/issues/52822
835
    qa: failed pacific install on fs:upgrade
836
* https://tracker.ceph.com/issues/52624
837
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
838
* https://tracker.ceph.com/issues/50223
839
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
840
* https://tracker.ceph.com/issues/48773
841
    qa: scrub does not complete
842
843
844
h3. 2021 September 20
845
846 23 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210917.174826
847
848
* https://tracker.ceph.com/issues/52677
849
    qa: test_simple failure
850
* https://tracker.ceph.com/issues/51279
851
    kclient hangs on umount (testing branch)
852
* https://tracker.ceph.com/issues/50223
853
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
854
* https://tracker.ceph.com/issues/50250
855
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
856
* https://tracker.ceph.com/issues/52624
857
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
858
* https://tracker.ceph.com/issues/52438
859
    qa: ffsb timeout
860
861
862 22 Patrick Donnelly
h3. 2021 September 10
863
864
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210910.181451
865
866
* https://tracker.ceph.com/issues/50223
867
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
868
* https://tracker.ceph.com/issues/50250
869
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
870
* https://tracker.ceph.com/issues/52624
871
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
872
* https://tracker.ceph.com/issues/52625
873
    qa: test_kill_mdstable (tasks.cephfs.test_snapshots.TestSnapshots)
874
* https://tracker.ceph.com/issues/52439
875
    qa: acls does not compile on centos stream
876
* https://tracker.ceph.com/issues/50821
877
    qa: untar_snap_rm failure during mds thrashing
878
* https://tracker.ceph.com/issues/48773
879
    qa: scrub does not complete
880
* https://tracker.ceph.com/issues/52626
881
    mds: ScrubStack.cc: 831: FAILED ceph_assert(diri)
882
* https://tracker.ceph.com/issues/51279
883
    kclient hangs on umount (testing branch)
884
885
886 21 Patrick Donnelly
h3. 2021 August 27
887
888
Several jobs died because of device failures.
889
890
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210827.024746
891
892
* https://tracker.ceph.com/issues/52430
893
    mds: fast async create client mount breaks racy test
894
* https://tracker.ceph.com/issues/52436
895
    fs/ceph: "corrupt mdsmap"
896
* https://tracker.ceph.com/issues/52437
897
    mds: InoTable::replay_release_ids abort via test_inotable_sync
898
* https://tracker.ceph.com/issues/51282
899
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
900
* https://tracker.ceph.com/issues/52438
901
    qa: ffsb timeout
902
* https://tracker.ceph.com/issues/52439
903
    qa: acls does not compile on centos stream
904
905
906 20 Patrick Donnelly
h3. 2021 July 30
907
908
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210729.214022
909
910
* https://tracker.ceph.com/issues/50250
911
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
912
* https://tracker.ceph.com/issues/51282
913
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
914
* https://tracker.ceph.com/issues/48773
915
    qa: scrub does not complete
916
* https://tracker.ceph.com/issues/51975
917
    pybind/mgr/stats: KeyError
918
919
920 19 Patrick Donnelly
h3. 2021 July 28
921
922
https://pulpito.ceph.com/pdonnell-2021-07-28_00:39:45-fs-wip-pdonnell-testing-20210727.213757-distro-basic-smithi/
923
924
with qa fix: https://pulpito.ceph.com/pdonnell-2021-07-28_16:20:28-fs-wip-pdonnell-testing-20210728.141004-distro-basic-smithi/
925
926
* https://tracker.ceph.com/issues/51905
927
    qa: "error reading sessionmap 'mds1_sessionmap'"
928
* https://tracker.ceph.com/issues/48773
929
    qa: scrub does not complete
930
* https://tracker.ceph.com/issues/50250
931
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
932
* https://tracker.ceph.com/issues/51267
933
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
934
* https://tracker.ceph.com/issues/51279
935
    kclient hangs on umount (testing branch)
936
937
938 18 Patrick Donnelly
h3. 2021 July 16
939
940
https://pulpito.ceph.com/pdonnell-2021-07-16_05:50:11-fs-wip-pdonnell-testing-20210716.022804-distro-basic-smithi/
941
942
* https://tracker.ceph.com/issues/48773
943
    qa: scrub does not complete
944
* https://tracker.ceph.com/issues/48772
945
    qa: pjd: not ok 9, 44, 80
946
* https://tracker.ceph.com/issues/45434
947
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
948
* https://tracker.ceph.com/issues/51279
949
    kclient hangs on umount (testing branch)
950
* https://tracker.ceph.com/issues/50824
951
    qa: snaptest-git-ceph bus error
952
953
954 17 Patrick Donnelly
h3. 2021 July 04
955
956
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210703.052904
957
958
* https://tracker.ceph.com/issues/48773
959
    qa: scrub does not complete
960
* https://tracker.ceph.com/issues/39150
961
    mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
962
* https://tracker.ceph.com/issues/45434
963
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
964
* https://tracker.ceph.com/issues/51282
965
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
966
* https://tracker.ceph.com/issues/48771
967
    qa: iogen: workload fails to cause balancing
968
* https://tracker.ceph.com/issues/51279
969
    kclient hangs on umount (testing branch)
970
* https://tracker.ceph.com/issues/50250
971
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
972
973
974 16 Patrick Donnelly
h3. 2021 July 01
975
976
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210701.192056
977
978
* https://tracker.ceph.com/issues/51197
979
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
980
* https://tracker.ceph.com/issues/50866
981
    osd: stat mismatch on objects
982
* https://tracker.ceph.com/issues/48773
983
    qa: scrub does not complete
984
985
986 15 Patrick Donnelly
h3. 2021 June 26
987
988
https://pulpito.ceph.com/pdonnell-2021-06-26_00:57:00-fs-wip-pdonnell-testing-20210625.225421-distro-basic-smithi/
989
990
* https://tracker.ceph.com/issues/51183
991
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
992
* https://tracker.ceph.com/issues/51410
993
    kclient: fails to finish reconnect during MDS thrashing (testing branch)
994
* https://tracker.ceph.com/issues/48773
995
    qa: scrub does not complete
996
* https://tracker.ceph.com/issues/51282
997
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
998
* https://tracker.ceph.com/issues/51169
999
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1000
* https://tracker.ceph.com/issues/48772
1001
    qa: pjd: not ok 9, 44, 80
1002
1003
1004 14 Patrick Donnelly
h3. 2021 June 21
1005
1006
https://pulpito.ceph.com/pdonnell-2021-06-22_00:27:21-fs-wip-pdonnell-testing-20210621.231646-distro-basic-smithi/
1007
1008
One failure caused by PR: https://github.com/ceph/ceph/pull/41935#issuecomment-866472599
1009
1010
* https://tracker.ceph.com/issues/51282
1011
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1012
* https://tracker.ceph.com/issues/51183
1013
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1014
* https://tracker.ceph.com/issues/48773
1015
    qa: scrub does not complete
1016
* https://tracker.ceph.com/issues/48771
1017
    qa: iogen: workload fails to cause balancing
1018
* https://tracker.ceph.com/issues/51169
1019
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1020
* https://tracker.ceph.com/issues/50495
1021
    libcephfs: shutdown race fails with status 141
1022
* https://tracker.ceph.com/issues/45434
1023
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1024
* https://tracker.ceph.com/issues/50824
1025
    qa: snaptest-git-ceph bus error
1026
* https://tracker.ceph.com/issues/50223
1027
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1028
1029
1030 13 Patrick Donnelly
h3. 2021 June 16
1031
1032
https://pulpito.ceph.com/pdonnell-2021-06-16_21:26:55-fs-wip-pdonnell-testing-20210616.191804-distro-basic-smithi/
1033
1034
MDS abort class of failures caused by PR: https://github.com/ceph/ceph/pull/41667
1035
1036
* https://tracker.ceph.com/issues/45434
1037
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1038
* https://tracker.ceph.com/issues/51169
1039
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1040
* https://tracker.ceph.com/issues/43216
1041
    MDSMonitor: removes MDS coming out of quorum election
1042
* https://tracker.ceph.com/issues/51278
1043
    mds: "FAILED ceph_assert(!segments.empty())"
1044
* https://tracker.ceph.com/issues/51279
1045
    kclient hangs on umount (testing branch)
1046
* https://tracker.ceph.com/issues/51280
1047
    mds: "FAILED ceph_assert(r == 0 || r == -2)"
1048
* https://tracker.ceph.com/issues/51183
1049
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1050
* https://tracker.ceph.com/issues/51281
1051
    qa: snaptest-snap-rm-cmp.sh: "echo 'FAIL: bad match, /tmp/a 4637e766853d1ad16a7b17079e2c6f03 != real c3883760b18d50e8d78819c54d579b00'"
1052
* https://tracker.ceph.com/issues/48773
1053
    qa: scrub does not complete
1054
* https://tracker.ceph.com/issues/51076
1055
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
1056
* https://tracker.ceph.com/issues/51228
1057
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
1058
* https://tracker.ceph.com/issues/51282
1059
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1060
1061
1062 12 Patrick Donnelly
h3. 2021 June 14
1063
1064
https://pulpito.ceph.com/pdonnell-2021-06-14_20:53:05-fs-wip-pdonnell-testing-20210614.173325-distro-basic-smithi/
1065
1066
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1067
1068
* https://tracker.ceph.com/issues/51169
1069
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1070
* https://tracker.ceph.com/issues/51228
1071
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
1072
* https://tracker.ceph.com/issues/48773
1073
    qa: scrub does not complete
1074
* https://tracker.ceph.com/issues/51183
1075
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1076
* https://tracker.ceph.com/issues/45434
1077
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1078
* https://tracker.ceph.com/issues/51182
1079
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1080
* https://tracker.ceph.com/issues/51229
1081
    qa: test_multi_snap_schedule list difference failure
1082
* https://tracker.ceph.com/issues/50821
1083
    qa: untar_snap_rm failure during mds thrashing
1084
1085
1086 11 Patrick Donnelly
h3. 2021 June 13
1087
1088
https://pulpito.ceph.com/pdonnell-2021-06-12_02:45:35-fs-wip-pdonnell-testing-20210612.002809-distro-basic-smithi/
1089
1090
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1091
1092
* https://tracker.ceph.com/issues/51169
1093
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1094
* https://tracker.ceph.com/issues/48773
1095
    qa: scrub does not complete
1096
* https://tracker.ceph.com/issues/51182
1097
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1098
* https://tracker.ceph.com/issues/51183
1099
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1100
* https://tracker.ceph.com/issues/51197
1101
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
1102
* https://tracker.ceph.com/issues/45434
1103
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1104
1105 10 Patrick Donnelly
h3. 2021 June 11
1106
1107
https://pulpito.ceph.com/pdonnell-2021-06-11_18:02:10-fs-wip-pdonnell-testing-20210611.162716-distro-basic-smithi/
1108
1109
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1110
1111
* https://tracker.ceph.com/issues/51169
1112
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1113
* https://tracker.ceph.com/issues/45434
1114
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1115
* https://tracker.ceph.com/issues/48771
1116
    qa: iogen: workload fails to cause balancing
1117
* https://tracker.ceph.com/issues/43216
1118
    MDSMonitor: removes MDS coming out of quorum election
1119
* https://tracker.ceph.com/issues/51182
1120
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1121
* https://tracker.ceph.com/issues/50223
1122
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1123
* https://tracker.ceph.com/issues/48773
1124
    qa: scrub does not complete
1125
* https://tracker.ceph.com/issues/51183
1126
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1127
* https://tracker.ceph.com/issues/51184
1128
    qa: fs:bugs does not specify distro
1129
1130
1131 9 Patrick Donnelly
h3. 2021 June 03
1132
1133
https://pulpito.ceph.com/pdonnell-2021-06-03_03:40:33-fs-wip-pdonnell-testing-20210603.020013-distro-basic-smithi/
1134
1135
* https://tracker.ceph.com/issues/45434
1136
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1137
* https://tracker.ceph.com/issues/50016
1138
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1139
* https://tracker.ceph.com/issues/50821
1140
    qa: untar_snap_rm failure during mds thrashing
1141
* https://tracker.ceph.com/issues/50622 (regression)
1142
    msg: active_connections regression
1143
* https://tracker.ceph.com/issues/49845#note-2 (regression)
1144
    qa: failed umount in test_volumes
1145
* https://tracker.ceph.com/issues/48773
1146
    qa: scrub does not complete
1147
* https://tracker.ceph.com/issues/43216
1148
    MDSMonitor: removes MDS coming out of quorum election
1149
1150
1151 7 Patrick Donnelly
h3. 2021 May 18
1152
1153 8 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.214114
1154
1155
Regression in testing kernel caused some failures. Ilya fixed those and rerun
1156
looked better. Some odd new noise in the rerun relating to packaging and "No
1157
module named 'tasks.ceph'".
1158
1159
* https://tracker.ceph.com/issues/50824
1160
    qa: snaptest-git-ceph bus error
1161
* https://tracker.ceph.com/issues/50622 (regression)
1162
    msg: active_connections regression
1163
* https://tracker.ceph.com/issues/49845#note-2 (regression)
1164
    qa: failed umount in test_volumes
1165
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1166
    qa: quota failure
1167
1168
1169
h3. 2021 May 18
1170
1171 7 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.025642
1172
1173
* https://tracker.ceph.com/issues/50821
1174
    qa: untar_snap_rm failure during mds thrashing
1175
* https://tracker.ceph.com/issues/48773
1176
    qa: scrub does not complete
1177
* https://tracker.ceph.com/issues/45591
1178
    mgr: FAILED ceph_assert(daemon != nullptr)
1179
* https://tracker.ceph.com/issues/50866
1180
    osd: stat mismatch on objects
1181
* https://tracker.ceph.com/issues/50016
1182
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1183
* https://tracker.ceph.com/issues/50867
1184
    qa: fs:mirror: reduced data availability
1185
* https://tracker.ceph.com/issues/50821
1186
    qa: untar_snap_rm failure during mds thrashing
1187
* https://tracker.ceph.com/issues/50622 (regression)
1188
    msg: active_connections regression
1189
* https://tracker.ceph.com/issues/50223
1190
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1191
* https://tracker.ceph.com/issues/50868
1192
    qa: "kern.log.gz already exists; not overwritten"
1193
* https://tracker.ceph.com/issues/50870
1194
    qa: test_full: "rm: cannot remove 'large_file_a': Permission denied"
1195
1196
1197 6 Patrick Donnelly
h3. 2021 May 11
1198
1199
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210511.232042
1200
1201
* one class of failures caused by PR
1202
* https://tracker.ceph.com/issues/48812
1203
    qa: test_scrub_pause_and_resume_with_abort failure
1204
* https://tracker.ceph.com/issues/50390
1205
    mds: monclient: wait_auth_rotating timed out after 30
1206
* https://tracker.ceph.com/issues/48773
1207
    qa: scrub does not complete
1208
* https://tracker.ceph.com/issues/50821
1209
    qa: untar_snap_rm failure during mds thrashing
1210
* https://tracker.ceph.com/issues/50224
1211
    qa: test_mirroring_init_failure_with_recovery failure
1212
* https://tracker.ceph.com/issues/50622 (regression)
1213
    msg: active_connections regression
1214
* https://tracker.ceph.com/issues/50825
1215
    qa: snaptest-git-ceph hang during mon thrashing v2
1216
* https://tracker.ceph.com/issues/50821
1217
    qa: untar_snap_rm failure during mds thrashing
1218
* https://tracker.ceph.com/issues/50823
1219
    qa: RuntimeError: timeout waiting for cluster to stabilize
1220
1221
1222 5 Patrick Donnelly
h3. 2021 May 14
1223
1224
https://pulpito.ceph.com/pdonnell-2021-05-14_21:45:42-fs-master-distro-basic-smithi/
1225
1226
* https://tracker.ceph.com/issues/48812
1227
    qa: test_scrub_pause_and_resume_with_abort failure
1228
* https://tracker.ceph.com/issues/50821
1229
    qa: untar_snap_rm failure during mds thrashing
1230
* https://tracker.ceph.com/issues/50622 (regression)
1231
    msg: active_connections regression
1232
* https://tracker.ceph.com/issues/50822
1233
    qa: testing kernel patch for client metrics causes mds abort
1234
* https://tracker.ceph.com/issues/48773
1235
    qa: scrub does not complete
1236
* https://tracker.ceph.com/issues/50823
1237
    qa: RuntimeError: timeout waiting for cluster to stabilize
1238
* https://tracker.ceph.com/issues/50824
1239
    qa: snaptest-git-ceph bus error
1240
* https://tracker.ceph.com/issues/50825
1241
    qa: snaptest-git-ceph hang during mon thrashing v2
1242
* https://tracker.ceph.com/issues/50826
1243
    kceph: stock RHEL kernel hangs on snaptests with mon|osd thrashers
1244
1245
1246 4 Patrick Donnelly
h3. 2021 May 01
1247
1248
https://pulpito.ceph.com/pdonnell-2021-05-01_09:07:09-fs-wip-pdonnell-testing-20210501.040415-distro-basic-smithi/
1249
1250
* https://tracker.ceph.com/issues/45434
1251
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1252
* https://tracker.ceph.com/issues/50281
1253
    qa: untar_snap_rm timeout
1254
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1255
    qa: quota failure
1256
* https://tracker.ceph.com/issues/48773
1257
    qa: scrub does not complete
1258
* https://tracker.ceph.com/issues/50390
1259
    mds: monclient: wait_auth_rotating timed out after 30
1260
* https://tracker.ceph.com/issues/50250
1261
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1262
* https://tracker.ceph.com/issues/50622 (regression)
1263
    msg: active_connections regression
1264
* https://tracker.ceph.com/issues/45591
1265
    mgr: FAILED ceph_assert(daemon != nullptr)
1266
* https://tracker.ceph.com/issues/50221
1267
    qa: snaptest-git-ceph failure in git diff
1268
* https://tracker.ceph.com/issues/50016
1269
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1270
1271
1272 3 Patrick Donnelly
h3. 2021 Apr 15
1273
1274
https://pulpito.ceph.com/pdonnell-2021-04-15_01:35:57-fs-wip-pdonnell-testing-20210414.230315-distro-basic-smithi/
1275
1276
* https://tracker.ceph.com/issues/50281
1277
    qa: untar_snap_rm timeout
1278
* https://tracker.ceph.com/issues/50220
1279
    qa: dbench workload timeout
1280
* https://tracker.ceph.com/issues/50246
1281
    mds: failure replaying journal (EMetaBlob)
1282
* https://tracker.ceph.com/issues/50250
1283
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1284
* https://tracker.ceph.com/issues/50016
1285
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1286
* https://tracker.ceph.com/issues/50222
1287
    osd: 5.2s0 deep-scrub : stat mismatch
1288
* https://tracker.ceph.com/issues/45434
1289
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1290
* https://tracker.ceph.com/issues/49845
1291
    qa: failed umount in test_volumes
1292
* https://tracker.ceph.com/issues/37808
1293
    osd: osdmap cache weak_refs assert during shutdown
1294
* https://tracker.ceph.com/issues/50387
1295
    client: fs/snaps failure
1296
* https://tracker.ceph.com/issues/50389
1297
    mds: "cluster [ERR] Error recovering journal 0x203: (2) No such file or directory" in cluster log"
1298
* https://tracker.ceph.com/issues/50216
1299
    qa: "ls: cannot access 'lost+found': No such file or directory"
1300
* https://tracker.ceph.com/issues/50390
1301
    mds: monclient: wait_auth_rotating timed out after 30
1302
1303
1304
1305 1 Patrick Donnelly
h3. 2021 Apr 08
1306
1307 2 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_22:42:24-fs-wip-pdonnell-testing-20210408.192301-distro-basic-smithi/
1308
1309
* https://tracker.ceph.com/issues/45434
1310
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1311
* https://tracker.ceph.com/issues/50016
1312
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1313
* https://tracker.ceph.com/issues/48773
1314
    qa: scrub does not complete
1315
* https://tracker.ceph.com/issues/50279
1316
    qa: "Replacing daemon mds.b as rank 0 with standby daemon mds.c"
1317
* https://tracker.ceph.com/issues/50246
1318
    mds: failure replaying journal (EMetaBlob)
1319
* https://tracker.ceph.com/issues/48365
1320
    qa: ffsb build failure on CentOS 8.2
1321
* https://tracker.ceph.com/issues/50216
1322
    qa: "ls: cannot access 'lost+found': No such file or directory"
1323
* https://tracker.ceph.com/issues/50223
1324
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1325
* https://tracker.ceph.com/issues/50280
1326
    cephadm: RuntimeError: uid/gid not found
1327
* https://tracker.ceph.com/issues/50281
1328
    qa: untar_snap_rm timeout
1329
1330
h3. 2021 Apr 08
1331
1332 1 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_04:31:36-fs-wip-pdonnell-testing-20210408.024225-distro-basic-smithi/
1333
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210408.142238 (with logic inversion / QA fix)
1334
1335
* https://tracker.ceph.com/issues/50246
1336
    mds: failure replaying journal (EMetaBlob)
1337
* https://tracker.ceph.com/issues/50250
1338
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1339
1340
1341
h3. 2021 Apr 07
1342
1343
https://pulpito.ceph.com/pdonnell-2021-04-07_02:12:41-fs-wip-pdonnell-testing-20210406.213012-distro-basic-smithi/
1344
1345
* https://tracker.ceph.com/issues/50215
1346
    qa: "log [ERR] : error reading sessionmap 'mds2_sessionmap'"
1347
* https://tracker.ceph.com/issues/49466
1348
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1349
* https://tracker.ceph.com/issues/50216
1350
    qa: "ls: cannot access 'lost+found': No such file or directory"
1351
* https://tracker.ceph.com/issues/48773
1352
    qa: scrub does not complete
1353
* https://tracker.ceph.com/issues/49845
1354
    qa: failed umount in test_volumes
1355
* https://tracker.ceph.com/issues/50220
1356
    qa: dbench workload timeout
1357
* https://tracker.ceph.com/issues/50221
1358
    qa: snaptest-git-ceph failure in git diff
1359
* https://tracker.ceph.com/issues/50222
1360
    osd: 5.2s0 deep-scrub : stat mismatch
1361
* https://tracker.ceph.com/issues/50223
1362
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1363
* https://tracker.ceph.com/issues/50224
1364
    qa: test_mirroring_init_failure_with_recovery failure
1365
1366
h3. 2021 Apr 01
1367
1368
https://pulpito.ceph.com/pdonnell-2021-04-01_00:45:34-fs-wip-pdonnell-testing-20210331.222326-distro-basic-smithi/
1369
1370
* https://tracker.ceph.com/issues/48772
1371
    qa: pjd: not ok 9, 44, 80
1372
* https://tracker.ceph.com/issues/50177
1373
    osd: "stalled aio... buggy kernel or bad device?"
1374
* https://tracker.ceph.com/issues/48771
1375
    qa: iogen: workload fails to cause balancing
1376
* https://tracker.ceph.com/issues/49845
1377
    qa: failed umount in test_volumes
1378
* https://tracker.ceph.com/issues/48773
1379
    qa: scrub does not complete
1380
* https://tracker.ceph.com/issues/48805
1381
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1382
* https://tracker.ceph.com/issues/50178
1383
    qa: "TypeError: run() got an unexpected keyword argument 'shell'"
1384
* https://tracker.ceph.com/issues/45434
1385
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1386
1387
h3. 2021 Mar 24
1388
1389
https://pulpito.ceph.com/pdonnell-2021-03-24_23:26:35-fs-wip-pdonnell-testing-20210324.190252-distro-basic-smithi/
1390
1391
* https://tracker.ceph.com/issues/49500
1392
    qa: "Assertion `cb_done' failed."
1393
* https://tracker.ceph.com/issues/50019
1394
    qa: mount failure with cephadm "probably no MDS server is up?"
1395
* https://tracker.ceph.com/issues/50020
1396
    qa: "RADOS object not found (Failed to operate read op for oid cephfs_mirror)"
1397
* https://tracker.ceph.com/issues/48773
1398
    qa: scrub does not complete
1399
* https://tracker.ceph.com/issues/45434
1400
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1401
* https://tracker.ceph.com/issues/48805
1402
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1403
* https://tracker.ceph.com/issues/48772
1404
    qa: pjd: not ok 9, 44, 80
1405
* https://tracker.ceph.com/issues/50021
1406
    qa: snaptest-git-ceph failure during mon thrashing
1407
* https://tracker.ceph.com/issues/48771
1408
    qa: iogen: workload fails to cause balancing
1409
* https://tracker.ceph.com/issues/50016
1410
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1411
* https://tracker.ceph.com/issues/49466
1412
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1413
1414
1415
h3. 2021 Mar 18
1416
1417
https://pulpito.ceph.com/pdonnell-2021-03-18_13:46:31-fs-wip-pdonnell-testing-20210318.024145-distro-basic-smithi/
1418
1419
* https://tracker.ceph.com/issues/49466
1420
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1421
* https://tracker.ceph.com/issues/48773
1422
    qa: scrub does not complete
1423
* https://tracker.ceph.com/issues/48805
1424
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1425
* https://tracker.ceph.com/issues/45434
1426
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1427
* https://tracker.ceph.com/issues/49845
1428
    qa: failed umount in test_volumes
1429
* https://tracker.ceph.com/issues/49605
1430
    mgr: drops command on the floor
1431
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1432
    qa: quota failure
1433
* https://tracker.ceph.com/issues/49928
1434
    client: items pinned in cache preventing unmount x2
1435
1436
h3. 2021 Mar 15
1437
1438
https://pulpito.ceph.com/pdonnell-2021-03-15_22:16:56-fs-wip-pdonnell-testing-20210315.182203-distro-basic-smithi/
1439
1440
* https://tracker.ceph.com/issues/49842
1441
    qa: stuck pkg install
1442
* https://tracker.ceph.com/issues/49466
1443
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1444
* https://tracker.ceph.com/issues/49822
1445
    test: test_mirroring_command_idempotency (tasks.cephfs.test_admin.TestMirroringCommands) failure
1446
* https://tracker.ceph.com/issues/49240
1447
    terminate called after throwing an instance of 'std::bad_alloc'
1448
* https://tracker.ceph.com/issues/48773
1449
    qa: scrub does not complete
1450
* https://tracker.ceph.com/issues/45434
1451
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1452
* https://tracker.ceph.com/issues/49500
1453
    qa: "Assertion `cb_done' failed."
1454
* https://tracker.ceph.com/issues/49843
1455
    qa: fs/snaps/snaptest-upchildrealms.sh failure
1456
* https://tracker.ceph.com/issues/49845
1457
    qa: failed umount in test_volumes
1458
* https://tracker.ceph.com/issues/48805
1459
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1460
* https://tracker.ceph.com/issues/49605
1461
    mgr: drops command on the floor
1462
1463
and failure caused by PR: https://github.com/ceph/ceph/pull/39969
1464
1465
1466
h3. 2021 Mar 09
1467
1468
https://pulpito.ceph.com/pdonnell-2021-03-09_03:27:39-fs-wip-pdonnell-testing-20210308.214827-distro-basic-smithi/
1469
1470
* https://tracker.ceph.com/issues/49500
1471
    qa: "Assertion `cb_done' failed."
1472
* https://tracker.ceph.com/issues/48805
1473
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1474
* https://tracker.ceph.com/issues/48773
1475
    qa: scrub does not complete
1476
* https://tracker.ceph.com/issues/45434
1477
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1478
* https://tracker.ceph.com/issues/49240
1479
    terminate called after throwing an instance of 'std::bad_alloc'
1480
* https://tracker.ceph.com/issues/49466
1481
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1482
* https://tracker.ceph.com/issues/49684
1483
    qa: fs:cephadm mount does not wait for mds to be created
1484
* https://tracker.ceph.com/issues/48771
1485
    qa: iogen: workload fails to cause balancing