Project

General

Profile

Main » History » Version 113

Venky Shankar, 03/13/2023 01:49 PM

1 79 Venky Shankar
h1. MAIN
2
3 113 Venky Shankar
h3. 13 Mar 2023
4
5
* https://tracker.ceph.com/issues/56695
6
    [RHEL stock] pjd test failures
7
* https://tracker.ceph.com/issues/57676
8
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
9
* https://tracker.ceph.com/issues/51964
10
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
11
* https://tracker.ceph.com/issues/54460
12
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
13
* https://tracker.ceph.com/issues/57656
14
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
15
16 112 Venky Shankar
h3. 09 Mar 2023
17
18
https://pulpito.ceph.com/vshankar-2023-03-03_04:39:14-fs-wip-vshankar-testing-20230303.023823-testing-default-smithi/
19
https://pulpito.ceph.com/vshankar-2023-03-08_15:12:36-fs-wip-vshankar-testing-20230308.112059-testing-default-smithi/
20
21
* https://tracker.ceph.com/issues/56695
22
    [RHEL stock] pjd test failures
23
* https://tracker.ceph.com/issues/57676
24
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
25
* https://tracker.ceph.com/issues/51964
26
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
27
* https://tracker.ceph.com/issues/54460
28
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
29
* https://tracker.ceph.com/issues/58340
30
    mds: fsstress.sh hangs with multimds
31
* https://tracker.ceph.com/issues/57087
32
    qa: test_fragmented_injection (tasks.cephfs.test_data_scan.TestDataScan) failure
33
34 111 Venky Shankar
h3. 07 Mar 2023
35
36
https://pulpito.ceph.com/vshankar-2023-03-02_09:21:58-fs-wip-vshankar-testing-20230222.044949-testing-default-smithi/
37
https://pulpito.ceph.com/vshankar-2023-03-07_05:15:12-fs-wip-vshankar-testing-20230307.030510-testing-default-smithi/
38
39
* https://tracker.ceph.com/issues/56695
40
    [RHEL stock] pjd test failures
41
* https://tracker.ceph.com/issues/57676
42
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
43
* https://tracker.ceph.com/issues/51964
44
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
45
* https://tracker.ceph.com/issues/57656
46
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
47
* https://tracker.ceph.com/issues/57655
48
    qa: fs:mixed-clients kernel_untar_build failure
49
* https://tracker.ceph.com/issues/58220
50
    Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1:
51
* https://tracker.ceph.com/issues/54460
52
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
53
* https://tracker.ceph.com/issues/58934
54
    snaptest-git-ceph.sh failure with ceph-fuse
55
56 109 Venky Shankar
h3. 28 Feb 2023
57
58
https://pulpito.ceph.com/vshankar-2023-02-24_02:11:45-fs-wip-vshankar-testing-20230222.025426-testing-default-smithi/
59
60
* https://tracker.ceph.com/issues/56695
61
    [RHEL stock] pjd test failures
62
* https://tracker.ceph.com/issues/57676
63
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
64
* https://tracker.ceph.com/issues/56446
65
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
66 110 Venky Shankar
67 109 Venky Shankar
(teuthology infra issues causing testing delays - merging PRs which have tests passing)
68
69 107 Venky Shankar
h3. 25 Jan 2023
70
71
https://pulpito.ceph.com/vshankar-2023-01-25_07:57:32-fs-wip-vshankar-testing-20230125.055346-testing-default-smithi/
72
73
* https://tracker.ceph.com/issues/52624
74
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" 
75
* https://tracker.ceph.com/issues/56695
76
    [RHEL stock] pjd test failures
77
* https://tracker.ceph.com/issues/57676
78
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
79
* https://tracker.ceph.com/issues/56446
80
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
81
* https://tracker.ceph.com/issues/57206
82
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
83
* https://tracker.ceph.com/issues/58220
84
    Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1:
85
* https://tracker.ceph.com/issues/58340
86
  mds: fsstress.sh hangs with multimds
87
* https://tracker.ceph.com/issues/56011
88
    fs/thrash: snaptest-snap-rm-cmp.sh fails in mds5sum comparison
89
* https://tracker.ceph.com/issues/54460
90
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
91
92 101 Rishabh Dave
h3. 30 JAN 2023
93
94
run: http://pulpito.front.sepia.ceph.com/rishabh-2022-11-28_08:04:11-fs-wip-rishabh-testing-2022Nov24-1818-testing-default-smithi/
95
re-run: http://pulpito.front.sepia.ceph.com/rishabh-2023-01-13_12:08:33-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/
96
re-run of re-run: http://pulpito.front.sepia.ceph.com/rishabh-2023-01-23_18:53:32-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/
97
98 105 Rishabh Dave
* https://tracker.ceph.com/issues/52624
99
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" 
100 101 Rishabh Dave
* https://tracker.ceph.com/issues/56695
101
  [RHEL stock] pjd test failures
102
* https://tracker.ceph.com/issues/57676
103
  qa: error during scrub thrashing: rank damage found: {'backtrace'}
104
* https://tracker.ceph.com/issues/55332
105
  Failure in snaptest-git-ceph.sh
106
* https://tracker.ceph.com/issues/51964
107
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
108
* https://tracker.ceph.com/issues/56446
109
  Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
110
* https://tracker.ceph.com/issues/57655 
111
  qa: fs:mixed-clients kernel_untar_build failure
112
* https://tracker.ceph.com/issues/54460
113
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
114
* https://tracker.ceph.com/issues/58340
115
  mds: fsstress.sh hangs with multimds
116 103 Rishabh Dave
* https://tracker.ceph.com/issues/58219
117
  Command crashed: 'ceph-dencoder type inode_backtrace_t import - decode dump_json'
118 101 Rishabh Dave
119 102 Rishabh Dave
* "Failed to load ceph-mgr modules: prometheus" in cluster log"
120
  http://pulpito.front.sepia.ceph.com/rishabh-2023-01-23_18:53:32-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/7134086
121
  Acc to Venky this was fixed in https://github.com/ceph/ceph/commit/cf6089200d96fc56b08ee17a4e31f19823370dc8
122 106 Rishabh Dave
* Created https://tracker.ceph.com/issues/58564
123
  workunit test suites/dbench.sh failed error code 1
124 102 Rishabh Dave
125 100 Venky Shankar
h3. 15 Dec 2022
126
127
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221215.112736
128
129
* https://tracker.ceph.com/issues/52624
130
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
131
* https://tracker.ceph.com/issues/56695
132
    [RHEL stock] pjd test failures
133
* https://tracker.ceph.com/issues/58219
134
* https://tracker.ceph.com/issues/57655
135
* qa: fs:mixed-clients kernel_untar_build failure
136
    Test failure: test_journal_migration (tasks.cephfs.test_journal_migration.TestJournalMigration)
137
* https://tracker.ceph.com/issues/57676
138
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
139
* https://tracker.ceph.com/issues/58340
140
    mds: fsstress.sh hangs with multimds
141
142 96 Venky Shankar
h3. 08 Dec 2022
143
144
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221130.043104
145 99 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221209.043803
146 96 Venky Shankar
147
(lots of transient git.ceph.com failures)
148
149
* https://tracker.ceph.com/issues/52624
150
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
151
* https://tracker.ceph.com/issues/56695
152
    [RHEL stock] pjd test failures
153
* https://tracker.ceph.com/issues/57655
154
    qa: fs:mixed-clients kernel_untar_build failure
155
* https://tracker.ceph.com/issues/58219
156
    Test failure: test_journal_migration (tasks.cephfs.test_journal_migration.TestJournalMigration)
157
* https://tracker.ceph.com/issues/58220
158
    Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1:
159
* https://tracker.ceph.com/issues/57676
160
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
161 97 Venky Shankar
* https://tracker.ceph.com/issues/53859
162
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
163 98 Venky Shankar
* https://tracker.ceph.com/issues/54460
164
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
165
* https://tracker.ceph.com/issues/58244
166
    Test failure: test_rebuild_inotable (tasks.cephfs.test_data_scan.TestDataScan)
167 96 Venky Shankar
168 95 Venky Shankar
h3. 14 Oct 2022
169
170
https://pulpito.ceph.com/vshankar-2022-10-12_04:56:59-fs-wip-vshankar-testing-20221011-145847-testing-default-smithi/
171
https://pulpito.ceph.com/vshankar-2022-10-14_04:04:57-fs-wip-vshankar-testing-20221014-072608-testing-default-smithi/
172
173
* https://tracker.ceph.com/issues/52624
174
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
175
* https://tracker.ceph.com/issues/55804
176
    Command failed (workunit test suites/pjd.sh)
177
* https://tracker.ceph.com/issues/51964
178
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
179
* https://tracker.ceph.com/issues/57682
180
    client: ERROR: test_reconnect_after_blocklisted
181
* https://tracker.ceph.com/issues/54460
182
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
183 90 Rishabh Dave
184 91 Rishabh Dave
h3. 10 Oct 2022
185
186
http://pulpito.front.sepia.ceph.com/rishabh-2022-09-30_19:45:21-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
187 92 Rishabh Dave
188 91 Rishabh Dave
reruns
189
* fs-thrash, passed: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-04_13:19:47-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
190
* fs-verify, passed: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-05_12:25:37-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
191
* cephadm failures also passed after many re-runs: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-06_13:50:51-fs-wip-rishabh-testing-30Sep2022-2-testing-default-smithi/
192 94 Rishabh Dave
    ** needed this PR to be merged in ceph-ci branch - https://github.com/ceph/ceph/pull/47458
193 91 Rishabh Dave
194 93 Rishabh Dave
known bugs
195 91 Rishabh Dave
* https://tracker.ceph.com/issues/52624
196
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
197
* https://tracker.ceph.com/issues/50223
198
  client.xxxx isn't responding to mclientcaps(revoke
199
* https://tracker.ceph.com/issues/57299
200
  qa: test_dump_loads fails with JSONDecodeError
201
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
202
  qa: fs:mixed-clients kernel_untar_build failure
203
* https://tracker.ceph.com/issues/57206
204
  libcephfs/test.sh: ceph_test_libcephfs_reclaim
205
206 90 Rishabh Dave
h3. 2022 Sep 29
207
208
http://pulpito.front.sepia.ceph.com/rishabh-2022-09-14_12:48:43-fs-wip-rishabh-testing-2022Sep9-1708-testing-default-smithi/
209
210
* https://tracker.ceph.com/issues/55804
211
  Command failed (workunit test suites/pjd.sh)
212
* https://tracker.ceph.com/issues/36593
213
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
214
* https://tracker.ceph.com/issues/52624
215
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
216
* https://tracker.ceph.com/issues/51964
217
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
218
* https://tracker.ceph.com/issues/56632
219
  Test failure: test_subvolume_snapshot_clone_quota_exceeded
220
* https://tracker.ceph.com/issues/50821
221
  qa: untar_snap_rm failure during mds thrashing
222
223 88 Patrick Donnelly
h3. 2022 Sep 26
224
225
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220923.171109
226
227
* https://tracker.ceph.com/issues/55804
228
    qa failure: pjd link tests failed
229
* https://tracker.ceph.com/issues/57676
230
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
231
* https://tracker.ceph.com/issues/52624
232
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
233
* https://tracker.ceph.com/issues/57580
234
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
235
* https://tracker.ceph.com/issues/48773
236
    qa: scrub does not complete
237
* https://tracker.ceph.com/issues/57299
238
    qa: test_dump_loads fails with JSONDecodeError
239
* https://tracker.ceph.com/issues/57280
240
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
241
* https://tracker.ceph.com/issues/57205
242
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
243
* https://tracker.ceph.com/issues/57656
244
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
245
* https://tracker.ceph.com/issues/57677
246
    qa: "1 MDSs behind on trimming (MDS_TRIM)"
247
* https://tracker.ceph.com/issues/57206
248
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
249
* https://tracker.ceph.com/issues/57446
250
    qa: test_subvolume_snapshot_info_if_orphan_clone fails
251
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
252
    qa: fs:mixed-clients kernel_untar_build failure
253 89 Patrick Donnelly
* https://tracker.ceph.com/issues/57682
254
    client: ERROR: test_reconnect_after_blocklisted
255 88 Patrick Donnelly
256
257 87 Patrick Donnelly
h3. 2022 Sep 22
258
259
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220920.234701
260
261
* https://tracker.ceph.com/issues/57299
262
    qa: test_dump_loads fails with JSONDecodeError
263
* https://tracker.ceph.com/issues/57205
264
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
265
* https://tracker.ceph.com/issues/52624
266
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
267
* https://tracker.ceph.com/issues/57580
268
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
269
* https://tracker.ceph.com/issues/57280
270
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
271
* https://tracker.ceph.com/issues/48773
272
    qa: scrub does not complete
273
* https://tracker.ceph.com/issues/56446
274
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
275
* https://tracker.ceph.com/issues/57206
276
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
277
* https://tracker.ceph.com/issues/51267
278
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
279
280
NEW:
281
282
* https://tracker.ceph.com/issues/57656
283
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
284
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
285
    qa: fs:mixed-clients kernel_untar_build failure
286
* https://tracker.ceph.com/issues/57657
287
    mds: scrub locates mismatch between child accounted_rstats and self rstats
288
289
Segfault probably caused by: https://github.com/ceph/ceph/pull/47795#issuecomment-1255724799
290
291
292 80 Venky Shankar
h3. 2022 Sep 16
293 79 Venky Shankar
294
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220905-132828
295
296
* https://tracker.ceph.com/issues/57446
297
    qa: test_subvolume_snapshot_info_if_orphan_clone fails
298
* https://tracker.ceph.com/issues/57299
299
    qa: test_dump_loads fails with JSONDecodeError
300
* https://tracker.ceph.com/issues/50223
301
    client.xxxx isn't responding to mclientcaps(revoke)
302
* https://tracker.ceph.com/issues/52624
303
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
304
* https://tracker.ceph.com/issues/57205
305
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
306
* https://tracker.ceph.com/issues/57280
307
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
308
* https://tracker.ceph.com/issues/51282
309
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
310
* https://tracker.ceph.com/issues/48203
311
  https://tracker.ceph.com/issues/36593
312
    qa: quota failure
313
    qa: quota failure caused by clients stepping on each other
314
* https://tracker.ceph.com/issues/57580
315
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
316
317 77 Rishabh Dave
318
h3. 2022 Aug 26
319 76 Rishabh Dave
320
http://pulpito.front.sepia.ceph.com/rishabh-2022-08-22_17:49:59-fs-wip-rishabh-testing-2022Aug19-testing-default-smithi/
321
http://pulpito.front.sepia.ceph.com/rishabh-2022-08-24_11:56:51-fs-wip-rishabh-testing-2022Aug19-testing-default-smithi/
322
323
* https://tracker.ceph.com/issues/57206
324
  libcephfs/test.sh: ceph_test_libcephfs_reclaim
325
* https://tracker.ceph.com/issues/56632
326
  Test failure: test_subvolume_snapshot_clone_quota_exceeded (tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones)
327
* https://tracker.ceph.com/issues/56446
328
  Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
329
* https://tracker.ceph.com/issues/51964
330
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
331
* https://tracker.ceph.com/issues/53859
332
  qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
333
334
* https://tracker.ceph.com/issues/54460
335
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
336
* https://tracker.ceph.com/issues/54462
337
  Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi055 with status 128
338
* https://tracker.ceph.com/issues/54460
339
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
340
* https://tracker.ceph.com/issues/36593
341
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
342
343
* https://tracker.ceph.com/issues/52624
344
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
345
* https://tracker.ceph.com/issues/55804
346
  Command failed (workunit test suites/pjd.sh)
347
* https://tracker.ceph.com/issues/50223
348
  client.xxxx isn't responding to mclientcaps(revoke)
349
350
351 75 Venky Shankar
h3. 2022 Aug 22
352
353
https://pulpito.ceph.com/vshankar-2022-08-12_09:34:24-fs-wip-vshankar-testing1-20220812-072441-testing-default-smithi/
354
https://pulpito.ceph.com/vshankar-2022-08-18_04:30:42-fs-wip-vshankar-testing1-20220818-082047-testing-default-smithi/ (drop problematic PR and re-run)
355
356
* https://tracker.ceph.com/issues/52624
357
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
358
* https://tracker.ceph.com/issues/56446
359
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
360
* https://tracker.ceph.com/issues/55804
361
    Command failed (workunit test suites/pjd.sh)
362
* https://tracker.ceph.com/issues/51278
363
    mds: "FAILED ceph_assert(!segments.empty())"
364
* https://tracker.ceph.com/issues/54460
365
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
366
* https://tracker.ceph.com/issues/57205
367
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
368
* https://tracker.ceph.com/issues/57206
369
    ceph_test_libcephfs_reclaim crashes during test
370
* https://tracker.ceph.com/issues/53859
371
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
372
* https://tracker.ceph.com/issues/50223
373
    client.xxxx isn't responding to mclientcaps(revoke)
374
375 72 Venky Shankar
h3. 2022 Aug 12
376
377
https://pulpito.ceph.com/vshankar-2022-08-10_04:06:00-fs-wip-vshankar-testing-20220805-190751-testing-default-smithi/
378
https://pulpito.ceph.com/vshankar-2022-08-11_12:16:58-fs-wip-vshankar-testing-20220811-145809-testing-default-smithi/ (drop problematic PR and re-run)
379
380
* https://tracker.ceph.com/issues/52624
381
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
382
* https://tracker.ceph.com/issues/56446
383
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
384
* https://tracker.ceph.com/issues/51964
385
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
386
* https://tracker.ceph.com/issues/55804
387
    Command failed (workunit test suites/pjd.sh)
388
* https://tracker.ceph.com/issues/50223
389
    client.xxxx isn't responding to mclientcaps(revoke)
390
* https://tracker.ceph.com/issues/50821
391
    qa: untar_snap_rm failure during mds thrashing
392
* https://tracker.ceph.com/issues/54460
393 73 Venky Shankar
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
394 72 Venky Shankar
395 71 Venky Shankar
h3. 2022 Aug 04
396
397
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220804-123835 (only mgr/volumes, mgr/stats)
398
399
Unrealted teuthology failure on rhel
400
401 69 Rishabh Dave
h3. 2022 Jul 25
402 68 Rishabh Dave
403
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-22_11:34:20-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
404
405
1st re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_03:51:19-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi
406
2nd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
407 74 Rishabh Dave
3rd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
408
4th (final) re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-28_03:59:01-fs-wip-rishabh-testing-2022Jul28-0143-testing-default-smithi/
409 68 Rishabh Dave
410
* https://tracker.ceph.com/issues/55804
411
  Command failed (workunit test suites/pjd.sh)
412
* https://tracker.ceph.com/issues/50223
413
  client.xxxx isn't responding to mclientcaps(revoke)
414
415
* https://tracker.ceph.com/issues/54460
416
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
417
* https://tracker.ceph.com/issues/36593
418
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
419 1 Patrick Donnelly
* https://tracker.ceph.com/issues/54462
420 74 Rishabh Dave
  Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi055 with status 128~
421 68 Rishabh Dave
422 67 Patrick Donnelly
h3. 2022 July 22
423
424
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220721.235756
425
426
MDS_HEALTH_DUMMY error in log fixed by followup commit.
427
transient selinux ping failure
428
429
* https://tracker.ceph.com/issues/56694
430
    qa: avoid blocking forever on hung umount
431
* https://tracker.ceph.com/issues/56695
432
    [RHEL stock] pjd test failures
433
* https://tracker.ceph.com/issues/56696
434
    admin keyring disappears during qa run
435
* https://tracker.ceph.com/issues/56697
436
    qa: fs/snaps fails for fuse
437
* https://tracker.ceph.com/issues/50222
438
    osd: 5.2s0 deep-scrub : stat mismatch
439
* https://tracker.ceph.com/issues/56698
440
    client: FAILED ceph_assert(_size == 0)
441
* https://tracker.ceph.com/issues/50223
442
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
443
444
445 66 Rishabh Dave
h3. 2022 Jul 15
446 65 Rishabh Dave
447
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-08_23:53:34-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
448
449
re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-15_06:42:04-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
450
451
* https://tracker.ceph.com/issues/53859
452
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
453
* https://tracker.ceph.com/issues/55804
454
  Command failed (workunit test suites/pjd.sh)
455
* https://tracker.ceph.com/issues/50223
456
  client.xxxx isn't responding to mclientcaps(revoke)
457
* https://tracker.ceph.com/issues/50222
458
  osd: deep-scrub : stat mismatch
459
460
* https://tracker.ceph.com/issues/56632
461
  Test failure: test_subvolume_snapshot_clone_quota_exceeded (tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones)
462
* https://tracker.ceph.com/issues/56634
463
  workunit test fs/snaps/snaptest-intodir.sh
464
* https://tracker.ceph.com/issues/56644
465
  Test failure: test_rapid_creation (tasks.cephfs.test_fragment.TestFragmentation)
466
467
468
469 61 Rishabh Dave
h3. 2022 July 05
470
471
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-02_14:14:52-fs-wip-rishabh-testing-20220702-1631-testing-default-smithi/
472 62 Rishabh Dave
473 64 Rishabh Dave
On 1st re-run some jobs passed - http://pulpito.front.sepia.ceph.com/rishabh-2022-07-03_15:10:28-fs-wip-rishabh-testing-20220702-1631-distro-default-smithi/
474
475
On 2nd re-run only few jobs failed -
476
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
477
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
478 62 Rishabh Dave
479
* https://tracker.ceph.com/issues/56446
480
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
481
* https://tracker.ceph.com/issues/55804
482
    Command failed (workunit test suites/pjd.sh) on smithi047 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/
483
484
* https://tracker.ceph.com/issues/56445
485
    Command failed on smithi080 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
486
* https://tracker.ceph.com/issues/51267
487 63 Rishabh Dave
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi098 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1
488
* https://tracker.ceph.com/issues/50224
489
    Test failure: test_mirroring_init_failure_with_recovery (tasks.cephfs.test_mirroring.TestMirroring)
490 62 Rishabh Dave
491
492 61 Rishabh Dave
493 58 Venky Shankar
h3. 2022 July 04
494
495
https://pulpito.ceph.com/vshankar-2022-06-29_09:19:00-fs-wip-vshankar-testing-20220627-100931-testing-default-smithi/
496
(rhel runs were borked due to: https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/JSZQFUKVLDND4W33PXDGCABPHNSPT6SS/, tests ran with --filter-out=rhel)
497
498
* https://tracker.ceph.com/issues/56445
499
    Command failed on smithi162 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
500
* https://tracker.ceph.com/issues/56446
501 59 Rishabh Dave
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
502
* https://tracker.ceph.com/issues/51964
503
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
504
* https://tracker.ceph.com/issues/52624
505 60 Rishabh Dave
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
506 59 Rishabh Dave
507 57 Venky Shankar
h3. 2022 June 20
508
509
https://pulpito.ceph.com/vshankar-2022-06-15_04:03:39-fs-wip-vshankar-testing1-20220615-072516-testing-default-smithi/
510
https://pulpito.ceph.com/vshankar-2022-06-19_08:22:46-fs-wip-vshankar-testing1-20220619-102531-testing-default-smithi/
511
512
* https://tracker.ceph.com/issues/52624
513
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
514
* https://tracker.ceph.com/issues/55804
515
    qa failure: pjd link tests failed
516
* https://tracker.ceph.com/issues/54108
517
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
518
* https://tracker.ceph.com/issues/55332
519
    Failure in snaptest-git-ceph.sh (it's an async unlink/create bug)
520
521 56 Patrick Donnelly
h3. 2022 June 13
522
523
https://pulpito.ceph.com/pdonnell-2022-06-12_05:08:12-fs:workload-wip-pdonnell-testing-20220612.004943-distro-default-smithi/
524
525
* https://tracker.ceph.com/issues/56024
526
    cephadm: removes ceph.conf during qa run causing command failure
527
* https://tracker.ceph.com/issues/48773
528
    qa: scrub does not complete
529
* https://tracker.ceph.com/issues/56012
530
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
531
532
533 55 Venky Shankar
h3. 2022 Jun 13
534 54 Venky Shankar
535
https://pulpito.ceph.com/vshankar-2022-06-07_00:25:50-fs-wip-vshankar-testing-20220606-223254-testing-default-smithi/
536
https://pulpito.ceph.com/vshankar-2022-06-10_01:04:46-fs-wip-vshankar-testing-20220609-175550-testing-default-smithi/
537
538
* https://tracker.ceph.com/issues/52624
539
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
540
* https://tracker.ceph.com/issues/51964
541
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
542
* https://tracker.ceph.com/issues/53859
543
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
544
* https://tracker.ceph.com/issues/55804
545
    qa failure: pjd link tests failed
546
* https://tracker.ceph.com/issues/56003
547
    client: src/include/xlist.h: 81: FAILED ceph_assert(_size == 0)
548
* https://tracker.ceph.com/issues/56011
549
    fs/thrash: snaptest-snap-rm-cmp.sh fails in mds5sum comparison
550
* https://tracker.ceph.com/issues/56012
551
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
552
553 53 Venky Shankar
h3. 2022 Jun 07
554
555
https://pulpito.ceph.com/vshankar-2022-06-06_21:25:41-fs-wip-vshankar-testing1-20220606-230129-testing-default-smithi/
556
https://pulpito.ceph.com/vshankar-2022-06-07_10:53:31-fs-wip-vshankar-testing1-20220607-104134-testing-default-smithi/ (rerun after dropping a problematic PR)
557
558
* https://tracker.ceph.com/issues/52624
559
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
560
* https://tracker.ceph.com/issues/50223
561
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
562
* https://tracker.ceph.com/issues/50224
563
    qa: test_mirroring_init_failure_with_recovery failure
564
565 51 Venky Shankar
h3. 2022 May 12
566
567
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220509-125847
568 52 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-13_17:09:16-fs-wip-vshankar-testing-20220513-120051-testing-default-smithi/ (drop prs + rerun)
569 51 Venky Shankar
570
* https://tracker.ceph.com/issues/52624
571
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
572
* https://tracker.ceph.com/issues/50223
573
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
574
* https://tracker.ceph.com/issues/55332
575
    Failure in snaptest-git-ceph.sh
576
* https://tracker.ceph.com/issues/53859
577
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
578
* https://tracker.ceph.com/issues/55538
579 1 Patrick Donnelly
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
580 52 Venky Shankar
* https://tracker.ceph.com/issues/55258
581
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs (cropss up again, though very infrequent)
582 51 Venky Shankar
583 49 Venky Shankar
h3. 2022 May 04
584
585 50 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-01_13:18:44-fs-wip-vshankar-testing1-20220428-204527-testing-default-smithi/
586
https://pulpito.ceph.com/vshankar-2022-05-02_16:58:59-fs-wip-vshankar-testing1-20220502-201957-testing-default-smithi/ (after dropping PRs)
587
588 49 Venky Shankar
* https://tracker.ceph.com/issues/52624
589
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
590
* https://tracker.ceph.com/issues/50223
591
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
592
* https://tracker.ceph.com/issues/55332
593
    Failure in snaptest-git-ceph.sh
594
* https://tracker.ceph.com/issues/53859
595
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
596
* https://tracker.ceph.com/issues/55516
597
    qa: fs suite tests failing with "json.decoder.JSONDecodeError: Extra data: line 2 column 82 (char 82)"
598
* https://tracker.ceph.com/issues/55537
599
    mds: crash during fs:upgrade test
600
* https://tracker.ceph.com/issues/55538
601
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
602
603 48 Venky Shankar
h3. 2022 Apr 25
604
605
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220420-113951 (owner vshankar)
606
607
* https://tracker.ceph.com/issues/52624
608
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
609
* https://tracker.ceph.com/issues/50223
610
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
611
* https://tracker.ceph.com/issues/55258
612
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
613
* https://tracker.ceph.com/issues/55377
614
    kclient: mds revoke Fwb caps stuck after the kclient tries writebcak once
615
616 47 Venky Shankar
h3. 2022 Apr 14
617
618
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220411-144044
619
620
* https://tracker.ceph.com/issues/52624
621
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
622
* https://tracker.ceph.com/issues/50223
623
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
624
* https://tracker.ceph.com/issues/52438
625
    qa: ffsb timeout
626
* https://tracker.ceph.com/issues/55170
627
    mds: crash during rejoin (CDir::fetch_keys)
628
* https://tracker.ceph.com/issues/55331
629
    pjd failure
630
* https://tracker.ceph.com/issues/48773
631
    qa: scrub does not complete
632
* https://tracker.ceph.com/issues/55332
633
    Failure in snaptest-git-ceph.sh
634
* https://tracker.ceph.com/issues/55258
635
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
636
637 45 Venky Shankar
h3. 2022 Apr 11
638
639 46 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-55110-20220408-203242
640 45 Venky Shankar
641
* https://tracker.ceph.com/issues/48773
642
    qa: scrub does not complete
643
* https://tracker.ceph.com/issues/52624
644
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
645
* https://tracker.ceph.com/issues/52438
646
    qa: ffsb timeout
647
* https://tracker.ceph.com/issues/48680
648
    mds: scrubbing stuck "scrub active (0 inodes in the stack)"
649
* https://tracker.ceph.com/issues/55236
650
    qa: fs/snaps tests fails with "hit max job timeout"
651
* https://tracker.ceph.com/issues/54108
652
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
653
* https://tracker.ceph.com/issues/54971
654
    Test failure: test_perf_stats_stale_metrics (tasks.cephfs.test_mds_metrics.TestMDSMetrics)
655
* https://tracker.ceph.com/issues/50223
656
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
657
* https://tracker.ceph.com/issues/55258
658
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
659
660 44 Venky Shankar
h3. 2022 Mar 21
661 42 Venky Shankar
662 43 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-03-20_02:16:37-fs-wip-vshankar-testing-20220319-163539-testing-default-smithi/
663
664
Run didn't go well, lots of failures - debugging by dropping PRs and running against master branch. Only merging unrelated PRs that pass tests.
665
666
667
h3. 2022 Mar 08
668
669 42 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-02-28_04:32:15-fs-wip-vshankar-testing-20220226-211550-testing-default-smithi/
670
671
rerun with
672
- (drop) https://github.com/ceph/ceph/pull/44679
673
- (drop) https://github.com/ceph/ceph/pull/44958
674
https://pulpito.ceph.com/vshankar-2022-03-06_14:47:51-fs-wip-vshankar-testing-20220304-132102-testing-default-smithi/
675
676
* https://tracker.ceph.com/issues/54419 (new)
677
    `ceph orch upgrade start` seems to never reach completion
678
* https://tracker.ceph.com/issues/51964
679
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
680
* https://tracker.ceph.com/issues/52624
681
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
682
* https://tracker.ceph.com/issues/50223
683
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
684
* https://tracker.ceph.com/issues/52438
685
    qa: ffsb timeout
686
* https://tracker.ceph.com/issues/50821
687
    qa: untar_snap_rm failure during mds thrashing
688
689
690 41 Venky Shankar
h3. 2022 Feb 09
691
692
https://pulpito.ceph.com/vshankar-2022-02-05_17:27:49-fs-wip-vshankar-testing-20220201-113815-testing-default-smithi/
693
694
rerun with
695
- (drop) https://github.com/ceph/ceph/pull/37938
696
- (drop) https://github.com/ceph/ceph/pull/44335
697
- (drop) https://github.com/ceph/ceph/pull/44491
698
- (drop) https://github.com/ceph/ceph/pull/44501
699
https://pulpito.ceph.com/vshankar-2022-02-08_14:27:29-fs-wip-vshankar-testing-20220208-181241-testing-default-smithi/
700
701
* https://tracker.ceph.com/issues/51964
702
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
703
* https://tracker.ceph.com/issues/54066
704
    test_subvolume_no_upgrade_v1_sanity fails with `AssertionError: 1000 != 0`
705
* https://tracker.ceph.com/issues/48773
706
    qa: scrub does not complete
707
* https://tracker.ceph.com/issues/52624
708
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
709
* https://tracker.ceph.com/issues/50223
710
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
711
* https://tracker.ceph.com/issues/52438
712
    qa: ffsb timeout
713
714 40 Patrick Donnelly
h3. 2022 Feb 01
715
716
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220127.171526
717
718
* https://tracker.ceph.com/issues/54107
719
    kclient: hang during umount
720
* https://tracker.ceph.com/issues/54106
721
    kclient: hang during workunit cleanup
722
* https://tracker.ceph.com/issues/54108
723
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
724
* https://tracker.ceph.com/issues/48773
725
    qa: scrub does not complete
726
* https://tracker.ceph.com/issues/52438
727
    qa: ffsb timeout
728
729
730 36 Venky Shankar
h3. 2022 Jan 13
731
732
https://pulpito.ceph.com/vshankar-2022-01-06_13:18:41-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
733 39 Venky Shankar
734 36 Venky Shankar
rerun with:
735 38 Venky Shankar
- (add) https://github.com/ceph/ceph/pull/44570
736
- (drop) https://github.com/ceph/ceph/pull/43184
737 36 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-01-13_04:42:40-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
738
739
* https://tracker.ceph.com/issues/50223
740
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
741
* https://tracker.ceph.com/issues/51282
742
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
743
* https://tracker.ceph.com/issues/48773
744
    qa: scrub does not complete
745
* https://tracker.ceph.com/issues/52624
746
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
747
* https://tracker.ceph.com/issues/53859
748
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
749
750 34 Venky Shankar
h3. 2022 Jan 03
751
752
https://pulpito.ceph.com/vshankar-2021-12-22_07:37:44-fs-wip-vshankar-testing-20211216-114012-testing-default-smithi/
753
https://pulpito.ceph.com/vshankar-2022-01-03_12:27:45-fs-wip-vshankar-testing-20220103-142738-testing-default-smithi/ (rerun)
754
755
* https://tracker.ceph.com/issues/50223
756
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
757
* https://tracker.ceph.com/issues/51964
758
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
759
* https://tracker.ceph.com/issues/51267
760
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
761
* https://tracker.ceph.com/issues/51282
762
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
763
* https://tracker.ceph.com/issues/50821
764
    qa: untar_snap_rm failure during mds thrashing
765
* https://tracker.ceph.com/issues/51278
766
    mds: "FAILED ceph_assert(!segments.empty())"
767 35 Ramana Raja
* https://tracker.ceph.com/issues/52279
768
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
769
770 34 Venky Shankar
771 33 Patrick Donnelly
h3. 2021 Dec 22
772
773
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211222.014316
774
775
* https://tracker.ceph.com/issues/52624
776
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
777
* https://tracker.ceph.com/issues/50223
778
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
779
* https://tracker.ceph.com/issues/52279
780
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
781
* https://tracker.ceph.com/issues/50224
782
    qa: test_mirroring_init_failure_with_recovery failure
783
* https://tracker.ceph.com/issues/48773
784
    qa: scrub does not complete
785
786
787 32 Venky Shankar
h3. 2021 Nov 30
788
789
https://pulpito.ceph.com/vshankar-2021-11-24_07:14:27-fs-wip-vshankar-testing-20211124-094330-testing-default-smithi/
790
https://pulpito.ceph.com/vshankar-2021-11-30_06:23:32-fs-wip-vshankar-testing-20211124-094330-distro-default-smithi/ (rerun w/ QA fixes)
791
792
* https://tracker.ceph.com/issues/53436
793
    mds, mon: mds beacon messages get dropped? (mds never reaches up:active state)
794
* https://tracker.ceph.com/issues/51964
795
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
796
* https://tracker.ceph.com/issues/48812
797
    qa: test_scrub_pause_and_resume_with_abort failure
798
* https://tracker.ceph.com/issues/51076
799
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
800
* https://tracker.ceph.com/issues/50223
801
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
802
* https://tracker.ceph.com/issues/52624
803
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
804
* https://tracker.ceph.com/issues/50250
805
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
806
807
808 31 Patrick Donnelly
h3. 2021 November 9
809
810
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211109.180315
811
812
* https://tracker.ceph.com/issues/53214
813
    qa: "dd: error reading '/sys/kernel/debug/ceph/2a934501-6731-4052-a836-f42229a869be.client4874/metrics': Is a directory"
814
* https://tracker.ceph.com/issues/48773
815
    qa: scrub does not complete
816
* https://tracker.ceph.com/issues/50223
817
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
818
* https://tracker.ceph.com/issues/51282
819
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
820
* https://tracker.ceph.com/issues/52624
821
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
822
* https://tracker.ceph.com/issues/53216
823
    qa: "RuntimeError: value of attributes should be either str or None. client_id"
824
* https://tracker.ceph.com/issues/50250
825
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
826
827
828
829 30 Patrick Donnelly
h3. 2021 November 03
830
831
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211103.023355
832
833
* https://tracker.ceph.com/issues/51964
834
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
835
* https://tracker.ceph.com/issues/51282
836
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
837
* https://tracker.ceph.com/issues/52436
838
    fs/ceph: "corrupt mdsmap"
839
* https://tracker.ceph.com/issues/53074
840
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
841
* https://tracker.ceph.com/issues/53150
842
    pybind/mgr/cephadm/upgrade: tolerate MDS failures during upgrade straddling v16.2.5
843
* https://tracker.ceph.com/issues/53155
844
    MDSMonitor: assertion during upgrade to v16.2.5+
845
846
847 29 Patrick Donnelly
h3. 2021 October 26
848
849
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211025.000447
850
851
* https://tracker.ceph.com/issues/53074
852
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
853
* https://tracker.ceph.com/issues/52997
854
    testing: hang ing umount
855
* https://tracker.ceph.com/issues/50824
856
    qa: snaptest-git-ceph bus error
857
* https://tracker.ceph.com/issues/52436
858
    fs/ceph: "corrupt mdsmap"
859
* https://tracker.ceph.com/issues/48773
860
    qa: scrub does not complete
861
* https://tracker.ceph.com/issues/53082
862
    ceph-fuse: segmenetation fault in Client::handle_mds_map
863
* https://tracker.ceph.com/issues/50223
864
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
865
* https://tracker.ceph.com/issues/52624
866
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
867
* https://tracker.ceph.com/issues/50224
868
    qa: test_mirroring_init_failure_with_recovery failure
869
* https://tracker.ceph.com/issues/50821
870
    qa: untar_snap_rm failure during mds thrashing
871
* https://tracker.ceph.com/issues/50250
872
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
873
874
875
876 27 Patrick Donnelly
h3. 2021 October 19
877
878 28 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211019.013028
879 27 Patrick Donnelly
880
* https://tracker.ceph.com/issues/52995
881
    qa: test_standby_count_wanted failure
882
* https://tracker.ceph.com/issues/52948
883
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
884
* https://tracker.ceph.com/issues/52996
885
    qa: test_perf_counters via test_openfiletable
886
* https://tracker.ceph.com/issues/48772
887
    qa: pjd: not ok 9, 44, 80
888
* https://tracker.ceph.com/issues/52997
889
    testing: hang ing umount
890
* https://tracker.ceph.com/issues/50250
891
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
892
* https://tracker.ceph.com/issues/52624
893
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
894
* https://tracker.ceph.com/issues/50223
895
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
896
* https://tracker.ceph.com/issues/50821
897
    qa: untar_snap_rm failure during mds thrashing
898
* https://tracker.ceph.com/issues/48773
899
    qa: scrub does not complete
900
901
902 26 Patrick Donnelly
h3. 2021 October 12
903
904
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211012.192211
905
906
Some failures caused by teuthology bug: https://tracker.ceph.com/issues/52944
907
908
New test caused failure: https://github.com/ceph/ceph/pull/43297#discussion_r729883167
909
910
911
* https://tracker.ceph.com/issues/51282
912
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
913
* https://tracker.ceph.com/issues/52948
914
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
915
* https://tracker.ceph.com/issues/48773
916
    qa: scrub does not complete
917
* https://tracker.ceph.com/issues/50224
918
    qa: test_mirroring_init_failure_with_recovery failure
919
* https://tracker.ceph.com/issues/52949
920
    RuntimeError: The following counters failed to be set on mds daemons: {'mds.dir_split'}
921
922
923 25 Patrick Donnelly
h3. 2021 October 02
924 23 Patrick Donnelly
925 24 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211002.163337
926
927
Some failures caused by cephadm upgrade test. Fixed in follow-up qa commit.
928
929
test_simple failures caused by PR in this set.
930
931
A few reruns because of QA infra noise.
932
933
* https://tracker.ceph.com/issues/52822
934
    qa: failed pacific install on fs:upgrade
935
* https://tracker.ceph.com/issues/52624
936
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
937
* https://tracker.ceph.com/issues/50223
938
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
939
* https://tracker.ceph.com/issues/48773
940
    qa: scrub does not complete
941
942
943
h3. 2021 September 20
944
945 23 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210917.174826
946
947
* https://tracker.ceph.com/issues/52677
948
    qa: test_simple failure
949
* https://tracker.ceph.com/issues/51279
950
    kclient hangs on umount (testing branch)
951
* https://tracker.ceph.com/issues/50223
952
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
953
* https://tracker.ceph.com/issues/50250
954
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
955
* https://tracker.ceph.com/issues/52624
956
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
957
* https://tracker.ceph.com/issues/52438
958
    qa: ffsb timeout
959
960
961 22 Patrick Donnelly
h3. 2021 September 10
962
963
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210910.181451
964
965
* https://tracker.ceph.com/issues/50223
966
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
967
* https://tracker.ceph.com/issues/50250
968
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
969
* https://tracker.ceph.com/issues/52624
970
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
971
* https://tracker.ceph.com/issues/52625
972
    qa: test_kill_mdstable (tasks.cephfs.test_snapshots.TestSnapshots)
973
* https://tracker.ceph.com/issues/52439
974
    qa: acls does not compile on centos stream
975
* https://tracker.ceph.com/issues/50821
976
    qa: untar_snap_rm failure during mds thrashing
977
* https://tracker.ceph.com/issues/48773
978
    qa: scrub does not complete
979
* https://tracker.ceph.com/issues/52626
980
    mds: ScrubStack.cc: 831: FAILED ceph_assert(diri)
981
* https://tracker.ceph.com/issues/51279
982
    kclient hangs on umount (testing branch)
983
984
985 21 Patrick Donnelly
h3. 2021 August 27
986
987
Several jobs died because of device failures.
988
989
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210827.024746
990
991
* https://tracker.ceph.com/issues/52430
992
    mds: fast async create client mount breaks racy test
993
* https://tracker.ceph.com/issues/52436
994
    fs/ceph: "corrupt mdsmap"
995
* https://tracker.ceph.com/issues/52437
996
    mds: InoTable::replay_release_ids abort via test_inotable_sync
997
* https://tracker.ceph.com/issues/51282
998
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
999
* https://tracker.ceph.com/issues/52438
1000
    qa: ffsb timeout
1001
* https://tracker.ceph.com/issues/52439
1002
    qa: acls does not compile on centos stream
1003
1004
1005 20 Patrick Donnelly
h3. 2021 July 30
1006
1007
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210729.214022
1008
1009
* https://tracker.ceph.com/issues/50250
1010
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
1011
* https://tracker.ceph.com/issues/51282
1012
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1013
* https://tracker.ceph.com/issues/48773
1014
    qa: scrub does not complete
1015
* https://tracker.ceph.com/issues/51975
1016
    pybind/mgr/stats: KeyError
1017
1018
1019 19 Patrick Donnelly
h3. 2021 July 28
1020
1021
https://pulpito.ceph.com/pdonnell-2021-07-28_00:39:45-fs-wip-pdonnell-testing-20210727.213757-distro-basic-smithi/
1022
1023
with qa fix: https://pulpito.ceph.com/pdonnell-2021-07-28_16:20:28-fs-wip-pdonnell-testing-20210728.141004-distro-basic-smithi/
1024
1025
* https://tracker.ceph.com/issues/51905
1026
    qa: "error reading sessionmap 'mds1_sessionmap'"
1027
* https://tracker.ceph.com/issues/48773
1028
    qa: scrub does not complete
1029
* https://tracker.ceph.com/issues/50250
1030
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
1031
* https://tracker.ceph.com/issues/51267
1032
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
1033
* https://tracker.ceph.com/issues/51279
1034
    kclient hangs on umount (testing branch)
1035
1036
1037 18 Patrick Donnelly
h3. 2021 July 16
1038
1039
https://pulpito.ceph.com/pdonnell-2021-07-16_05:50:11-fs-wip-pdonnell-testing-20210716.022804-distro-basic-smithi/
1040
1041
* https://tracker.ceph.com/issues/48773
1042
    qa: scrub does not complete
1043
* https://tracker.ceph.com/issues/48772
1044
    qa: pjd: not ok 9, 44, 80
1045
* https://tracker.ceph.com/issues/45434
1046
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1047
* https://tracker.ceph.com/issues/51279
1048
    kclient hangs on umount (testing branch)
1049
* https://tracker.ceph.com/issues/50824
1050
    qa: snaptest-git-ceph bus error
1051
1052
1053 17 Patrick Donnelly
h3. 2021 July 04
1054
1055
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210703.052904
1056
1057
* https://tracker.ceph.com/issues/48773
1058
    qa: scrub does not complete
1059
* https://tracker.ceph.com/issues/39150
1060
    mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
1061
* https://tracker.ceph.com/issues/45434
1062
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1063
* https://tracker.ceph.com/issues/51282
1064
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1065
* https://tracker.ceph.com/issues/48771
1066
    qa: iogen: workload fails to cause balancing
1067
* https://tracker.ceph.com/issues/51279
1068
    kclient hangs on umount (testing branch)
1069
* https://tracker.ceph.com/issues/50250
1070
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
1071
1072
1073 16 Patrick Donnelly
h3. 2021 July 01
1074
1075
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210701.192056
1076
1077
* https://tracker.ceph.com/issues/51197
1078
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
1079
* https://tracker.ceph.com/issues/50866
1080
    osd: stat mismatch on objects
1081
* https://tracker.ceph.com/issues/48773
1082
    qa: scrub does not complete
1083
1084
1085 15 Patrick Donnelly
h3. 2021 June 26
1086
1087
https://pulpito.ceph.com/pdonnell-2021-06-26_00:57:00-fs-wip-pdonnell-testing-20210625.225421-distro-basic-smithi/
1088
1089
* https://tracker.ceph.com/issues/51183
1090
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1091
* https://tracker.ceph.com/issues/51410
1092
    kclient: fails to finish reconnect during MDS thrashing (testing branch)
1093
* https://tracker.ceph.com/issues/48773
1094
    qa: scrub does not complete
1095
* https://tracker.ceph.com/issues/51282
1096
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1097
* https://tracker.ceph.com/issues/51169
1098
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1099
* https://tracker.ceph.com/issues/48772
1100
    qa: pjd: not ok 9, 44, 80
1101
1102
1103 14 Patrick Donnelly
h3. 2021 June 21
1104
1105
https://pulpito.ceph.com/pdonnell-2021-06-22_00:27:21-fs-wip-pdonnell-testing-20210621.231646-distro-basic-smithi/
1106
1107
One failure caused by PR: https://github.com/ceph/ceph/pull/41935#issuecomment-866472599
1108
1109
* https://tracker.ceph.com/issues/51282
1110
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1111
* https://tracker.ceph.com/issues/51183
1112
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1113
* https://tracker.ceph.com/issues/48773
1114
    qa: scrub does not complete
1115
* https://tracker.ceph.com/issues/48771
1116
    qa: iogen: workload fails to cause balancing
1117
* https://tracker.ceph.com/issues/51169
1118
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1119
* https://tracker.ceph.com/issues/50495
1120
    libcephfs: shutdown race fails with status 141
1121
* https://tracker.ceph.com/issues/45434
1122
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1123
* https://tracker.ceph.com/issues/50824
1124
    qa: snaptest-git-ceph bus error
1125
* https://tracker.ceph.com/issues/50223
1126
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1127
1128
1129 13 Patrick Donnelly
h3. 2021 June 16
1130
1131
https://pulpito.ceph.com/pdonnell-2021-06-16_21:26:55-fs-wip-pdonnell-testing-20210616.191804-distro-basic-smithi/
1132
1133
MDS abort class of failures caused by PR: https://github.com/ceph/ceph/pull/41667
1134
1135
* https://tracker.ceph.com/issues/45434
1136
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1137
* https://tracker.ceph.com/issues/51169
1138
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1139
* https://tracker.ceph.com/issues/43216
1140
    MDSMonitor: removes MDS coming out of quorum election
1141
* https://tracker.ceph.com/issues/51278
1142
    mds: "FAILED ceph_assert(!segments.empty())"
1143
* https://tracker.ceph.com/issues/51279
1144
    kclient hangs on umount (testing branch)
1145
* https://tracker.ceph.com/issues/51280
1146
    mds: "FAILED ceph_assert(r == 0 || r == -2)"
1147
* https://tracker.ceph.com/issues/51183
1148
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1149
* https://tracker.ceph.com/issues/51281
1150
    qa: snaptest-snap-rm-cmp.sh: "echo 'FAIL: bad match, /tmp/a 4637e766853d1ad16a7b17079e2c6f03 != real c3883760b18d50e8d78819c54d579b00'"
1151
* https://tracker.ceph.com/issues/48773
1152
    qa: scrub does not complete
1153
* https://tracker.ceph.com/issues/51076
1154
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
1155
* https://tracker.ceph.com/issues/51228
1156
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
1157
* https://tracker.ceph.com/issues/51282
1158
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1159
1160
1161 12 Patrick Donnelly
h3. 2021 June 14
1162
1163
https://pulpito.ceph.com/pdonnell-2021-06-14_20:53:05-fs-wip-pdonnell-testing-20210614.173325-distro-basic-smithi/
1164
1165
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1166
1167
* https://tracker.ceph.com/issues/51169
1168
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1169
* https://tracker.ceph.com/issues/51228
1170
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
1171
* https://tracker.ceph.com/issues/48773
1172
    qa: scrub does not complete
1173
* https://tracker.ceph.com/issues/51183
1174
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1175
* https://tracker.ceph.com/issues/45434
1176
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1177
* https://tracker.ceph.com/issues/51182
1178
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1179
* https://tracker.ceph.com/issues/51229
1180
    qa: test_multi_snap_schedule list difference failure
1181
* https://tracker.ceph.com/issues/50821
1182
    qa: untar_snap_rm failure during mds thrashing
1183
1184
1185 11 Patrick Donnelly
h3. 2021 June 13
1186
1187
https://pulpito.ceph.com/pdonnell-2021-06-12_02:45:35-fs-wip-pdonnell-testing-20210612.002809-distro-basic-smithi/
1188
1189
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1190
1191
* https://tracker.ceph.com/issues/51169
1192
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1193
* https://tracker.ceph.com/issues/48773
1194
    qa: scrub does not complete
1195
* https://tracker.ceph.com/issues/51182
1196
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1197
* https://tracker.ceph.com/issues/51183
1198
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1199
* https://tracker.ceph.com/issues/51197
1200
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
1201
* https://tracker.ceph.com/issues/45434
1202
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1203
1204 10 Patrick Donnelly
h3. 2021 June 11
1205
1206
https://pulpito.ceph.com/pdonnell-2021-06-11_18:02:10-fs-wip-pdonnell-testing-20210611.162716-distro-basic-smithi/
1207
1208
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1209
1210
* https://tracker.ceph.com/issues/51169
1211
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1212
* https://tracker.ceph.com/issues/45434
1213
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1214
* https://tracker.ceph.com/issues/48771
1215
    qa: iogen: workload fails to cause balancing
1216
* https://tracker.ceph.com/issues/43216
1217
    MDSMonitor: removes MDS coming out of quorum election
1218
* https://tracker.ceph.com/issues/51182
1219
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1220
* https://tracker.ceph.com/issues/50223
1221
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1222
* https://tracker.ceph.com/issues/48773
1223
    qa: scrub does not complete
1224
* https://tracker.ceph.com/issues/51183
1225
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1226
* https://tracker.ceph.com/issues/51184
1227
    qa: fs:bugs does not specify distro
1228
1229
1230 9 Patrick Donnelly
h3. 2021 June 03
1231
1232
https://pulpito.ceph.com/pdonnell-2021-06-03_03:40:33-fs-wip-pdonnell-testing-20210603.020013-distro-basic-smithi/
1233
1234
* https://tracker.ceph.com/issues/45434
1235
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1236
* https://tracker.ceph.com/issues/50016
1237
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1238
* https://tracker.ceph.com/issues/50821
1239
    qa: untar_snap_rm failure during mds thrashing
1240
* https://tracker.ceph.com/issues/50622 (regression)
1241
    msg: active_connections regression
1242
* https://tracker.ceph.com/issues/49845#note-2 (regression)
1243
    qa: failed umount in test_volumes
1244
* https://tracker.ceph.com/issues/48773
1245
    qa: scrub does not complete
1246
* https://tracker.ceph.com/issues/43216
1247
    MDSMonitor: removes MDS coming out of quorum election
1248
1249
1250 7 Patrick Donnelly
h3. 2021 May 18
1251
1252 8 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.214114
1253
1254
Regression in testing kernel caused some failures. Ilya fixed those and rerun
1255
looked better. Some odd new noise in the rerun relating to packaging and "No
1256
module named 'tasks.ceph'".
1257
1258
* https://tracker.ceph.com/issues/50824
1259
    qa: snaptest-git-ceph bus error
1260
* https://tracker.ceph.com/issues/50622 (regression)
1261
    msg: active_connections regression
1262
* https://tracker.ceph.com/issues/49845#note-2 (regression)
1263
    qa: failed umount in test_volumes
1264
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1265
    qa: quota failure
1266
1267
1268
h3. 2021 May 18
1269
1270 7 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.025642
1271
1272
* https://tracker.ceph.com/issues/50821
1273
    qa: untar_snap_rm failure during mds thrashing
1274
* https://tracker.ceph.com/issues/48773
1275
    qa: scrub does not complete
1276
* https://tracker.ceph.com/issues/45591
1277
    mgr: FAILED ceph_assert(daemon != nullptr)
1278
* https://tracker.ceph.com/issues/50866
1279
    osd: stat mismatch on objects
1280
* https://tracker.ceph.com/issues/50016
1281
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1282
* https://tracker.ceph.com/issues/50867
1283
    qa: fs:mirror: reduced data availability
1284
* https://tracker.ceph.com/issues/50821
1285
    qa: untar_snap_rm failure during mds thrashing
1286
* https://tracker.ceph.com/issues/50622 (regression)
1287
    msg: active_connections regression
1288
* https://tracker.ceph.com/issues/50223
1289
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1290
* https://tracker.ceph.com/issues/50868
1291
    qa: "kern.log.gz already exists; not overwritten"
1292
* https://tracker.ceph.com/issues/50870
1293
    qa: test_full: "rm: cannot remove 'large_file_a': Permission denied"
1294
1295
1296 6 Patrick Donnelly
h3. 2021 May 11
1297
1298
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210511.232042
1299
1300
* one class of failures caused by PR
1301
* https://tracker.ceph.com/issues/48812
1302
    qa: test_scrub_pause_and_resume_with_abort failure
1303
* https://tracker.ceph.com/issues/50390
1304
    mds: monclient: wait_auth_rotating timed out after 30
1305
* https://tracker.ceph.com/issues/48773
1306
    qa: scrub does not complete
1307
* https://tracker.ceph.com/issues/50821
1308
    qa: untar_snap_rm failure during mds thrashing
1309
* https://tracker.ceph.com/issues/50224
1310
    qa: test_mirroring_init_failure_with_recovery failure
1311
* https://tracker.ceph.com/issues/50622 (regression)
1312
    msg: active_connections regression
1313
* https://tracker.ceph.com/issues/50825
1314
    qa: snaptest-git-ceph hang during mon thrashing v2
1315
* https://tracker.ceph.com/issues/50821
1316
    qa: untar_snap_rm failure during mds thrashing
1317
* https://tracker.ceph.com/issues/50823
1318
    qa: RuntimeError: timeout waiting for cluster to stabilize
1319
1320
1321 5 Patrick Donnelly
h3. 2021 May 14
1322
1323
https://pulpito.ceph.com/pdonnell-2021-05-14_21:45:42-fs-master-distro-basic-smithi/
1324
1325
* https://tracker.ceph.com/issues/48812
1326
    qa: test_scrub_pause_and_resume_with_abort failure
1327
* https://tracker.ceph.com/issues/50821
1328
    qa: untar_snap_rm failure during mds thrashing
1329
* https://tracker.ceph.com/issues/50622 (regression)
1330
    msg: active_connections regression
1331
* https://tracker.ceph.com/issues/50822
1332
    qa: testing kernel patch for client metrics causes mds abort
1333
* https://tracker.ceph.com/issues/48773
1334
    qa: scrub does not complete
1335
* https://tracker.ceph.com/issues/50823
1336
    qa: RuntimeError: timeout waiting for cluster to stabilize
1337
* https://tracker.ceph.com/issues/50824
1338
    qa: snaptest-git-ceph bus error
1339
* https://tracker.ceph.com/issues/50825
1340
    qa: snaptest-git-ceph hang during mon thrashing v2
1341
* https://tracker.ceph.com/issues/50826
1342
    kceph: stock RHEL kernel hangs on snaptests with mon|osd thrashers
1343
1344
1345 4 Patrick Donnelly
h3. 2021 May 01
1346
1347
https://pulpito.ceph.com/pdonnell-2021-05-01_09:07:09-fs-wip-pdonnell-testing-20210501.040415-distro-basic-smithi/
1348
1349
* https://tracker.ceph.com/issues/45434
1350
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1351
* https://tracker.ceph.com/issues/50281
1352
    qa: untar_snap_rm timeout
1353
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1354
    qa: quota failure
1355
* https://tracker.ceph.com/issues/48773
1356
    qa: scrub does not complete
1357
* https://tracker.ceph.com/issues/50390
1358
    mds: monclient: wait_auth_rotating timed out after 30
1359
* https://tracker.ceph.com/issues/50250
1360
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1361
* https://tracker.ceph.com/issues/50622 (regression)
1362
    msg: active_connections regression
1363
* https://tracker.ceph.com/issues/45591
1364
    mgr: FAILED ceph_assert(daemon != nullptr)
1365
* https://tracker.ceph.com/issues/50221
1366
    qa: snaptest-git-ceph failure in git diff
1367
* https://tracker.ceph.com/issues/50016
1368
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1369
1370
1371 3 Patrick Donnelly
h3. 2021 Apr 15
1372
1373
https://pulpito.ceph.com/pdonnell-2021-04-15_01:35:57-fs-wip-pdonnell-testing-20210414.230315-distro-basic-smithi/
1374
1375
* https://tracker.ceph.com/issues/50281
1376
    qa: untar_snap_rm timeout
1377
* https://tracker.ceph.com/issues/50220
1378
    qa: dbench workload timeout
1379
* https://tracker.ceph.com/issues/50246
1380
    mds: failure replaying journal (EMetaBlob)
1381
* https://tracker.ceph.com/issues/50250
1382
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1383
* https://tracker.ceph.com/issues/50016
1384
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1385
* https://tracker.ceph.com/issues/50222
1386
    osd: 5.2s0 deep-scrub : stat mismatch
1387
* https://tracker.ceph.com/issues/45434
1388
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1389
* https://tracker.ceph.com/issues/49845
1390
    qa: failed umount in test_volumes
1391
* https://tracker.ceph.com/issues/37808
1392
    osd: osdmap cache weak_refs assert during shutdown
1393
* https://tracker.ceph.com/issues/50387
1394
    client: fs/snaps failure
1395
* https://tracker.ceph.com/issues/50389
1396
    mds: "cluster [ERR] Error recovering journal 0x203: (2) No such file or directory" in cluster log"
1397
* https://tracker.ceph.com/issues/50216
1398
    qa: "ls: cannot access 'lost+found': No such file or directory"
1399
* https://tracker.ceph.com/issues/50390
1400
    mds: monclient: wait_auth_rotating timed out after 30
1401
1402
1403
1404 1 Patrick Donnelly
h3. 2021 Apr 08
1405
1406 2 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_22:42:24-fs-wip-pdonnell-testing-20210408.192301-distro-basic-smithi/
1407
1408
* https://tracker.ceph.com/issues/45434
1409
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1410
* https://tracker.ceph.com/issues/50016
1411
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1412
* https://tracker.ceph.com/issues/48773
1413
    qa: scrub does not complete
1414
* https://tracker.ceph.com/issues/50279
1415
    qa: "Replacing daemon mds.b as rank 0 with standby daemon mds.c"
1416
* https://tracker.ceph.com/issues/50246
1417
    mds: failure replaying journal (EMetaBlob)
1418
* https://tracker.ceph.com/issues/48365
1419
    qa: ffsb build failure on CentOS 8.2
1420
* https://tracker.ceph.com/issues/50216
1421
    qa: "ls: cannot access 'lost+found': No such file or directory"
1422
* https://tracker.ceph.com/issues/50223
1423
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1424
* https://tracker.ceph.com/issues/50280
1425
    cephadm: RuntimeError: uid/gid not found
1426
* https://tracker.ceph.com/issues/50281
1427
    qa: untar_snap_rm timeout
1428
1429
h3. 2021 Apr 08
1430
1431 1 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_04:31:36-fs-wip-pdonnell-testing-20210408.024225-distro-basic-smithi/
1432
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210408.142238 (with logic inversion / QA fix)
1433
1434
* https://tracker.ceph.com/issues/50246
1435
    mds: failure replaying journal (EMetaBlob)
1436
* https://tracker.ceph.com/issues/50250
1437
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1438
1439
1440
h3. 2021 Apr 07
1441
1442
https://pulpito.ceph.com/pdonnell-2021-04-07_02:12:41-fs-wip-pdonnell-testing-20210406.213012-distro-basic-smithi/
1443
1444
* https://tracker.ceph.com/issues/50215
1445
    qa: "log [ERR] : error reading sessionmap 'mds2_sessionmap'"
1446
* https://tracker.ceph.com/issues/49466
1447
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1448
* https://tracker.ceph.com/issues/50216
1449
    qa: "ls: cannot access 'lost+found': No such file or directory"
1450
* https://tracker.ceph.com/issues/48773
1451
    qa: scrub does not complete
1452
* https://tracker.ceph.com/issues/49845
1453
    qa: failed umount in test_volumes
1454
* https://tracker.ceph.com/issues/50220
1455
    qa: dbench workload timeout
1456
* https://tracker.ceph.com/issues/50221
1457
    qa: snaptest-git-ceph failure in git diff
1458
* https://tracker.ceph.com/issues/50222
1459
    osd: 5.2s0 deep-scrub : stat mismatch
1460
* https://tracker.ceph.com/issues/50223
1461
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1462
* https://tracker.ceph.com/issues/50224
1463
    qa: test_mirroring_init_failure_with_recovery failure
1464
1465
h3. 2021 Apr 01
1466
1467
https://pulpito.ceph.com/pdonnell-2021-04-01_00:45:34-fs-wip-pdonnell-testing-20210331.222326-distro-basic-smithi/
1468
1469
* https://tracker.ceph.com/issues/48772
1470
    qa: pjd: not ok 9, 44, 80
1471
* https://tracker.ceph.com/issues/50177
1472
    osd: "stalled aio... buggy kernel or bad device?"
1473
* https://tracker.ceph.com/issues/48771
1474
    qa: iogen: workload fails to cause balancing
1475
* https://tracker.ceph.com/issues/49845
1476
    qa: failed umount in test_volumes
1477
* https://tracker.ceph.com/issues/48773
1478
    qa: scrub does not complete
1479
* https://tracker.ceph.com/issues/48805
1480
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1481
* https://tracker.ceph.com/issues/50178
1482
    qa: "TypeError: run() got an unexpected keyword argument 'shell'"
1483
* https://tracker.ceph.com/issues/45434
1484
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1485
1486
h3. 2021 Mar 24
1487
1488
https://pulpito.ceph.com/pdonnell-2021-03-24_23:26:35-fs-wip-pdonnell-testing-20210324.190252-distro-basic-smithi/
1489
1490
* https://tracker.ceph.com/issues/49500
1491
    qa: "Assertion `cb_done' failed."
1492
* https://tracker.ceph.com/issues/50019
1493
    qa: mount failure with cephadm "probably no MDS server is up?"
1494
* https://tracker.ceph.com/issues/50020
1495
    qa: "RADOS object not found (Failed to operate read op for oid cephfs_mirror)"
1496
* https://tracker.ceph.com/issues/48773
1497
    qa: scrub does not complete
1498
* https://tracker.ceph.com/issues/45434
1499
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1500
* https://tracker.ceph.com/issues/48805
1501
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1502
* https://tracker.ceph.com/issues/48772
1503
    qa: pjd: not ok 9, 44, 80
1504
* https://tracker.ceph.com/issues/50021
1505
    qa: snaptest-git-ceph failure during mon thrashing
1506
* https://tracker.ceph.com/issues/48771
1507
    qa: iogen: workload fails to cause balancing
1508
* https://tracker.ceph.com/issues/50016
1509
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1510
* https://tracker.ceph.com/issues/49466
1511
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1512
1513
1514
h3. 2021 Mar 18
1515
1516
https://pulpito.ceph.com/pdonnell-2021-03-18_13:46:31-fs-wip-pdonnell-testing-20210318.024145-distro-basic-smithi/
1517
1518
* https://tracker.ceph.com/issues/49466
1519
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1520
* https://tracker.ceph.com/issues/48773
1521
    qa: scrub does not complete
1522
* https://tracker.ceph.com/issues/48805
1523
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1524
* https://tracker.ceph.com/issues/45434
1525
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1526
* https://tracker.ceph.com/issues/49845
1527
    qa: failed umount in test_volumes
1528
* https://tracker.ceph.com/issues/49605
1529
    mgr: drops command on the floor
1530
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1531
    qa: quota failure
1532
* https://tracker.ceph.com/issues/49928
1533
    client: items pinned in cache preventing unmount x2
1534
1535
h3. 2021 Mar 15
1536
1537
https://pulpito.ceph.com/pdonnell-2021-03-15_22:16:56-fs-wip-pdonnell-testing-20210315.182203-distro-basic-smithi/
1538
1539
* https://tracker.ceph.com/issues/49842
1540
    qa: stuck pkg install
1541
* https://tracker.ceph.com/issues/49466
1542
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1543
* https://tracker.ceph.com/issues/49822
1544
    test: test_mirroring_command_idempotency (tasks.cephfs.test_admin.TestMirroringCommands) failure
1545
* https://tracker.ceph.com/issues/49240
1546
    terminate called after throwing an instance of 'std::bad_alloc'
1547
* https://tracker.ceph.com/issues/48773
1548
    qa: scrub does not complete
1549
* https://tracker.ceph.com/issues/45434
1550
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1551
* https://tracker.ceph.com/issues/49500
1552
    qa: "Assertion `cb_done' failed."
1553
* https://tracker.ceph.com/issues/49843
1554
    qa: fs/snaps/snaptest-upchildrealms.sh failure
1555
* https://tracker.ceph.com/issues/49845
1556
    qa: failed umount in test_volumes
1557
* https://tracker.ceph.com/issues/48805
1558
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1559
* https://tracker.ceph.com/issues/49605
1560
    mgr: drops command on the floor
1561
1562
and failure caused by PR: https://github.com/ceph/ceph/pull/39969
1563
1564
1565
h3. 2021 Mar 09
1566
1567
https://pulpito.ceph.com/pdonnell-2021-03-09_03:27:39-fs-wip-pdonnell-testing-20210308.214827-distro-basic-smithi/
1568
1569
* https://tracker.ceph.com/issues/49500
1570
    qa: "Assertion `cb_done' failed."
1571
* https://tracker.ceph.com/issues/48805
1572
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1573
* https://tracker.ceph.com/issues/48773
1574
    qa: scrub does not complete
1575
* https://tracker.ceph.com/issues/45434
1576
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1577
* https://tracker.ceph.com/issues/49240
1578
    terminate called after throwing an instance of 'std::bad_alloc'
1579
* https://tracker.ceph.com/issues/49466
1580
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1581
* https://tracker.ceph.com/issues/49684
1582
    qa: fs:cephadm mount does not wait for mds to be created
1583
* https://tracker.ceph.com/issues/48771
1584
    qa: iogen: workload fails to cause balancing