Project

General

Profile

Main » History » Version 108

Venky Shankar, 01/30/2023 04:29 PM

1 79 Venky Shankar
h1. MAIN
2
3 107 Venky Shankar
h3. 25 Jan 2023
4
5
https://pulpito.ceph.com/vshankar-2023-01-25_07:57:32-fs-wip-vshankar-testing-20230125.055346-testing-default-smithi/
6
7
* https://tracker.ceph.com/issues/52624
8
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" 
9
* https://tracker.ceph.com/issues/56695
10
    [RHEL stock] pjd test failures
11
* https://tracker.ceph.com/issues/57676
12
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
13
* https://tracker.ceph.com/issues/56446
14
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
15
* https://tracker.ceph.com/issues/57206
16
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
17
* https://tracker.ceph.com/issues/58220
18
    Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1:
19
* https://tracker.ceph.com/issues/58340
20
  mds: fsstress.sh hangs with multimds
21
* https://tracker.ceph.com/issues/56011
22
    fs/thrash: snaptest-snap-rm-cmp.sh fails in mds5sum comparison
23
* https://tracker.ceph.com/issues/54460
24
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
25
26 101 Rishabh Dave
h3. 30 JAN 2023
27
28
run: http://pulpito.front.sepia.ceph.com/rishabh-2022-11-28_08:04:11-fs-wip-rishabh-testing-2022Nov24-1818-testing-default-smithi/
29
re-run: http://pulpito.front.sepia.ceph.com/rishabh-2023-01-13_12:08:33-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/
30
re-run of re-run: http://pulpito.front.sepia.ceph.com/rishabh-2023-01-23_18:53:32-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/
31
32 105 Rishabh Dave
* https://tracker.ceph.com/issues/52624
33
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" 
34 101 Rishabh Dave
* https://tracker.ceph.com/issues/56695
35
  [RHEL stock] pjd test failures
36
* https://tracker.ceph.com/issues/57676
37
  qa: error during scrub thrashing: rank damage found: {'backtrace'}
38
* https://tracker.ceph.com/issues/55332
39
  Failure in snaptest-git-ceph.sh
40
* https://tracker.ceph.com/issues/51964
41
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
42
* https://tracker.ceph.com/issues/56446
43
  Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
44
* https://tracker.ceph.com/issues/57655 
45
  qa: fs:mixed-clients kernel_untar_build failure
46
* https://tracker.ceph.com/issues/54460
47
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
48
* https://tracker.ceph.com/issues/58340
49
  mds: fsstress.sh hangs with multimds
50 103 Rishabh Dave
* https://tracker.ceph.com/issues/58219
51
  Command crashed: 'ceph-dencoder type inode_backtrace_t import - decode dump_json'
52 101 Rishabh Dave
53 102 Rishabh Dave
* "Failed to load ceph-mgr modules: prometheus" in cluster log"
54
  http://pulpito.front.sepia.ceph.com/rishabh-2023-01-23_18:53:32-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/7134086
55
  Acc to Venky this was fixed in https://github.com/ceph/ceph/commit/cf6089200d96fc56b08ee17a4e31f19823370dc8
56 106 Rishabh Dave
* Created https://tracker.ceph.com/issues/58564
57
  workunit test suites/dbench.sh failed error code 1
58 102 Rishabh Dave
59 100 Venky Shankar
h3. 15 Dec 2022
60
61
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221215.112736
62
63
* https://tracker.ceph.com/issues/52624
64
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
65
* https://tracker.ceph.com/issues/56695
66
    [RHEL stock] pjd test failures
67
* https://tracker.ceph.com/issues/58219
68
* https://tracker.ceph.com/issues/57655
69
* qa: fs:mixed-clients kernel_untar_build failure
70
    Test failure: test_journal_migration (tasks.cephfs.test_journal_migration.TestJournalMigration)
71
* https://tracker.ceph.com/issues/57676
72
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
73
* https://tracker.ceph.com/issues/58340
74
    mds: fsstress.sh hangs with multimds
75
76 96 Venky Shankar
h3. 08 Dec 2022
77
78
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221130.043104
79 99 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221209.043803
80 96 Venky Shankar
81
(lots of transient git.ceph.com failures)
82
83
* https://tracker.ceph.com/issues/52624
84
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
85
* https://tracker.ceph.com/issues/56695
86
    [RHEL stock] pjd test failures
87
* https://tracker.ceph.com/issues/57655
88
    qa: fs:mixed-clients kernel_untar_build failure
89
* https://tracker.ceph.com/issues/58219
90
    Test failure: test_journal_migration (tasks.cephfs.test_journal_migration.TestJournalMigration)
91
* https://tracker.ceph.com/issues/58220
92
    Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1:
93
* https://tracker.ceph.com/issues/57676
94
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
95 97 Venky Shankar
* https://tracker.ceph.com/issues/53859
96
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
97 98 Venky Shankar
* https://tracker.ceph.com/issues/54460
98
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
99
* https://tracker.ceph.com/issues/58244
100
    Test failure: test_rebuild_inotable (tasks.cephfs.test_data_scan.TestDataScan)
101 96 Venky Shankar
102 95 Venky Shankar
h3. 14 Oct 2022
103
104
https://pulpito.ceph.com/vshankar-2022-10-12_04:56:59-fs-wip-vshankar-testing-20221011-145847-testing-default-smithi/
105
https://pulpito.ceph.com/vshankar-2022-10-14_04:04:57-fs-wip-vshankar-testing-20221014-072608-testing-default-smithi/
106
107
* https://tracker.ceph.com/issues/52624
108
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
109
* https://tracker.ceph.com/issues/55804
110
    Command failed (workunit test suites/pjd.sh)
111
* https://tracker.ceph.com/issues/51964
112
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
113
* https://tracker.ceph.com/issues/57682
114
    client: ERROR: test_reconnect_after_blocklisted
115
* https://tracker.ceph.com/issues/54460
116
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
117 90 Rishabh Dave
118 91 Rishabh Dave
h3. 10 Oct 2022
119
120
http://pulpito.front.sepia.ceph.com/rishabh-2022-09-30_19:45:21-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
121 92 Rishabh Dave
122 91 Rishabh Dave
reruns
123
* fs-thrash, passed: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-04_13:19:47-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
124
* fs-verify, passed: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-05_12:25:37-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
125
* cephadm failures also passed after many re-runs: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-06_13:50:51-fs-wip-rishabh-testing-30Sep2022-2-testing-default-smithi/
126 94 Rishabh Dave
    ** needed this PR to be merged in ceph-ci branch - https://github.com/ceph/ceph/pull/47458
127 91 Rishabh Dave
128 93 Rishabh Dave
known bugs
129 91 Rishabh Dave
* https://tracker.ceph.com/issues/52624
130
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
131
* https://tracker.ceph.com/issues/50223
132
  client.xxxx isn't responding to mclientcaps(revoke
133
* https://tracker.ceph.com/issues/57299
134
  qa: test_dump_loads fails with JSONDecodeError
135
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
136
  qa: fs:mixed-clients kernel_untar_build failure
137
* https://tracker.ceph.com/issues/57206
138
  libcephfs/test.sh: ceph_test_libcephfs_reclaim
139
140 90 Rishabh Dave
h3. 2022 Sep 29
141
142
http://pulpito.front.sepia.ceph.com/rishabh-2022-09-14_12:48:43-fs-wip-rishabh-testing-2022Sep9-1708-testing-default-smithi/
143
144
* https://tracker.ceph.com/issues/55804
145
  Command failed (workunit test suites/pjd.sh)
146
* https://tracker.ceph.com/issues/36593
147
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
148
* https://tracker.ceph.com/issues/52624
149
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
150
* https://tracker.ceph.com/issues/51964
151
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
152
* https://tracker.ceph.com/issues/56632
153
  Test failure: test_subvolume_snapshot_clone_quota_exceeded
154
* https://tracker.ceph.com/issues/50821
155
  qa: untar_snap_rm failure during mds thrashing
156
157 88 Patrick Donnelly
h3. 2022 Sep 26
158
159
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220923.171109
160
161
* https://tracker.ceph.com/issues/55804
162
    qa failure: pjd link tests failed
163
* https://tracker.ceph.com/issues/57676
164
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
165
* https://tracker.ceph.com/issues/52624
166
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
167
* https://tracker.ceph.com/issues/57580
168
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
169
* https://tracker.ceph.com/issues/48773
170
    qa: scrub does not complete
171
* https://tracker.ceph.com/issues/57299
172
    qa: test_dump_loads fails with JSONDecodeError
173
* https://tracker.ceph.com/issues/57280
174
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
175
* https://tracker.ceph.com/issues/57205
176
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
177
* https://tracker.ceph.com/issues/57656
178
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
179
* https://tracker.ceph.com/issues/57677
180
    qa: "1 MDSs behind on trimming (MDS_TRIM)"
181
* https://tracker.ceph.com/issues/57206
182
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
183
* https://tracker.ceph.com/issues/57446
184
    qa: test_subvolume_snapshot_info_if_orphan_clone fails
185
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
186
    qa: fs:mixed-clients kernel_untar_build failure
187 89 Patrick Donnelly
* https://tracker.ceph.com/issues/57682
188
    client: ERROR: test_reconnect_after_blocklisted
189 88 Patrick Donnelly
190
191 87 Patrick Donnelly
h3. 2022 Sep 22
192
193
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220920.234701
194
195
* https://tracker.ceph.com/issues/57299
196
    qa: test_dump_loads fails with JSONDecodeError
197
* https://tracker.ceph.com/issues/57205
198
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
199
* https://tracker.ceph.com/issues/52624
200
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
201
* https://tracker.ceph.com/issues/57580
202
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
203
* https://tracker.ceph.com/issues/57280
204
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
205
* https://tracker.ceph.com/issues/48773
206
    qa: scrub does not complete
207
* https://tracker.ceph.com/issues/56446
208
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
209
* https://tracker.ceph.com/issues/57206
210
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
211
* https://tracker.ceph.com/issues/51267
212
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
213
214
NEW:
215
216
* https://tracker.ceph.com/issues/57656
217
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
218
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
219
    qa: fs:mixed-clients kernel_untar_build failure
220
* https://tracker.ceph.com/issues/57657
221
    mds: scrub locates mismatch between child accounted_rstats and self rstats
222
223
Segfault probably caused by: https://github.com/ceph/ceph/pull/47795#issuecomment-1255724799
224
225
226 80 Venky Shankar
h3. 2022 Sep 16
227 79 Venky Shankar
228
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220905-132828
229
230
* https://tracker.ceph.com/issues/57446
231
    qa: test_subvolume_snapshot_info_if_orphan_clone fails
232
* https://tracker.ceph.com/issues/57299
233
    qa: test_dump_loads fails with JSONDecodeError
234
* https://tracker.ceph.com/issues/50223
235
    client.xxxx isn't responding to mclientcaps(revoke)
236
* https://tracker.ceph.com/issues/52624
237
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
238
* https://tracker.ceph.com/issues/57205
239
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
240
* https://tracker.ceph.com/issues/57280
241
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
242
* https://tracker.ceph.com/issues/51282
243
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
244
* https://tracker.ceph.com/issues/48203
245
  https://tracker.ceph.com/issues/36593
246
    qa: quota failure
247
    qa: quota failure caused by clients stepping on each other
248
* https://tracker.ceph.com/issues/57580
249
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
250
251 77 Rishabh Dave
252
h3. 2022 Aug 26
253 76 Rishabh Dave
254
http://pulpito.front.sepia.ceph.com/rishabh-2022-08-22_17:49:59-fs-wip-rishabh-testing-2022Aug19-testing-default-smithi/
255
http://pulpito.front.sepia.ceph.com/rishabh-2022-08-24_11:56:51-fs-wip-rishabh-testing-2022Aug19-testing-default-smithi/
256
257
* https://tracker.ceph.com/issues/57206
258
  libcephfs/test.sh: ceph_test_libcephfs_reclaim
259
* https://tracker.ceph.com/issues/56632
260
  Test failure: test_subvolume_snapshot_clone_quota_exceeded (tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones)
261
* https://tracker.ceph.com/issues/56446
262
  Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
263
* https://tracker.ceph.com/issues/51964
264
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
265
* https://tracker.ceph.com/issues/53859
266
  qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
267
268
* https://tracker.ceph.com/issues/54460
269
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
270
* https://tracker.ceph.com/issues/54462
271
  Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi055 with status 128
272
* https://tracker.ceph.com/issues/54460
273
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
274
* https://tracker.ceph.com/issues/36593
275
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
276
277
* https://tracker.ceph.com/issues/52624
278
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
279
* https://tracker.ceph.com/issues/55804
280
  Command failed (workunit test suites/pjd.sh)
281
* https://tracker.ceph.com/issues/50223
282
  client.xxxx isn't responding to mclientcaps(revoke)
283
284
285 75 Venky Shankar
h3. 2022 Aug 22
286
287
https://pulpito.ceph.com/vshankar-2022-08-12_09:34:24-fs-wip-vshankar-testing1-20220812-072441-testing-default-smithi/
288
https://pulpito.ceph.com/vshankar-2022-08-18_04:30:42-fs-wip-vshankar-testing1-20220818-082047-testing-default-smithi/ (drop problematic PR and re-run)
289
290
* https://tracker.ceph.com/issues/52624
291
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
292
* https://tracker.ceph.com/issues/56446
293
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
294
* https://tracker.ceph.com/issues/55804
295
    Command failed (workunit test suites/pjd.sh)
296
* https://tracker.ceph.com/issues/51278
297
    mds: "FAILED ceph_assert(!segments.empty())"
298
* https://tracker.ceph.com/issues/54460
299
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
300
* https://tracker.ceph.com/issues/57205
301
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
302
* https://tracker.ceph.com/issues/57206
303
    ceph_test_libcephfs_reclaim crashes during test
304
* https://tracker.ceph.com/issues/53859
305
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
306
* https://tracker.ceph.com/issues/50223
307
    client.xxxx isn't responding to mclientcaps(revoke)
308
309 72 Venky Shankar
h3. 2022 Aug 12
310
311
https://pulpito.ceph.com/vshankar-2022-08-10_04:06:00-fs-wip-vshankar-testing-20220805-190751-testing-default-smithi/
312
https://pulpito.ceph.com/vshankar-2022-08-11_12:16:58-fs-wip-vshankar-testing-20220811-145809-testing-default-smithi/ (drop problematic PR and re-run)
313
314
* https://tracker.ceph.com/issues/52624
315
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
316
* https://tracker.ceph.com/issues/56446
317
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
318
* https://tracker.ceph.com/issues/51964
319
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
320
* https://tracker.ceph.com/issues/55804
321
    Command failed (workunit test suites/pjd.sh)
322
* https://tracker.ceph.com/issues/50223
323
    client.xxxx isn't responding to mclientcaps(revoke)
324
* https://tracker.ceph.com/issues/50821
325
    qa: untar_snap_rm failure during mds thrashing
326
* https://tracker.ceph.com/issues/54460
327 73 Venky Shankar
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
328 72 Venky Shankar
329 71 Venky Shankar
h3. 2022 Aug 04
330
331
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220804-123835 (only mgr/volumes, mgr/stats)
332
333
Unrealted teuthology failure on rhel
334
335 69 Rishabh Dave
h3. 2022 Jul 25
336 68 Rishabh Dave
337
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-22_11:34:20-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
338
339
1st re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_03:51:19-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi
340
2nd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
341 74 Rishabh Dave
3rd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
342
4th (final) re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-28_03:59:01-fs-wip-rishabh-testing-2022Jul28-0143-testing-default-smithi/
343 68 Rishabh Dave
344
* https://tracker.ceph.com/issues/55804
345
  Command failed (workunit test suites/pjd.sh)
346
* https://tracker.ceph.com/issues/50223
347
  client.xxxx isn't responding to mclientcaps(revoke)
348
349
* https://tracker.ceph.com/issues/54460
350
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
351
* https://tracker.ceph.com/issues/36593
352
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
353 1 Patrick Donnelly
* https://tracker.ceph.com/issues/54462
354 74 Rishabh Dave
  Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi055 with status 128~
355 68 Rishabh Dave
356 67 Patrick Donnelly
h3. 2022 July 22
357
358
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220721.235756
359
360
MDS_HEALTH_DUMMY error in log fixed by followup commit.
361
transient selinux ping failure
362
363
* https://tracker.ceph.com/issues/56694
364
    qa: avoid blocking forever on hung umount
365
* https://tracker.ceph.com/issues/56695
366
    [RHEL stock] pjd test failures
367
* https://tracker.ceph.com/issues/56696
368
    admin keyring disappears during qa run
369
* https://tracker.ceph.com/issues/56697
370
    qa: fs/snaps fails for fuse
371
* https://tracker.ceph.com/issues/50222
372
    osd: 5.2s0 deep-scrub : stat mismatch
373
* https://tracker.ceph.com/issues/56698
374
    client: FAILED ceph_assert(_size == 0)
375
* https://tracker.ceph.com/issues/50223
376
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
377
378
379 66 Rishabh Dave
h3. 2022 Jul 15
380 65 Rishabh Dave
381
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-08_23:53:34-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
382
383
re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-15_06:42:04-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
384
385
* https://tracker.ceph.com/issues/53859
386
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
387
* https://tracker.ceph.com/issues/55804
388
  Command failed (workunit test suites/pjd.sh)
389
* https://tracker.ceph.com/issues/50223
390
  client.xxxx isn't responding to mclientcaps(revoke)
391
* https://tracker.ceph.com/issues/50222
392
  osd: deep-scrub : stat mismatch
393
394
* https://tracker.ceph.com/issues/56632
395
  Test failure: test_subvolume_snapshot_clone_quota_exceeded (tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones)
396
* https://tracker.ceph.com/issues/56634
397
  workunit test fs/snaps/snaptest-intodir.sh
398
* https://tracker.ceph.com/issues/56644
399
  Test failure: test_rapid_creation (tasks.cephfs.test_fragment.TestFragmentation)
400
401
402
403 61 Rishabh Dave
h3. 2022 July 05
404
405
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-02_14:14:52-fs-wip-rishabh-testing-20220702-1631-testing-default-smithi/
406 62 Rishabh Dave
407 64 Rishabh Dave
On 1st re-run some jobs passed - http://pulpito.front.sepia.ceph.com/rishabh-2022-07-03_15:10:28-fs-wip-rishabh-testing-20220702-1631-distro-default-smithi/
408
409
On 2nd re-run only few jobs failed -
410
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
411
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
412 62 Rishabh Dave
413
* https://tracker.ceph.com/issues/56446
414
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
415
* https://tracker.ceph.com/issues/55804
416
    Command failed (workunit test suites/pjd.sh) on smithi047 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/
417
418
* https://tracker.ceph.com/issues/56445
419
    Command failed on smithi080 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
420
* https://tracker.ceph.com/issues/51267
421 63 Rishabh Dave
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi098 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1
422
* https://tracker.ceph.com/issues/50224
423
    Test failure: test_mirroring_init_failure_with_recovery (tasks.cephfs.test_mirroring.TestMirroring)
424 62 Rishabh Dave
425
426 61 Rishabh Dave
427 58 Venky Shankar
h3. 2022 July 04
428
429
https://pulpito.ceph.com/vshankar-2022-06-29_09:19:00-fs-wip-vshankar-testing-20220627-100931-testing-default-smithi/
430
(rhel runs were borked due to: https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/JSZQFUKVLDND4W33PXDGCABPHNSPT6SS/, tests ran with --filter-out=rhel)
431
432
* https://tracker.ceph.com/issues/56445
433
    Command failed on smithi162 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
434
* https://tracker.ceph.com/issues/56446
435 59 Rishabh Dave
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
436
* https://tracker.ceph.com/issues/51964
437
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
438
* https://tracker.ceph.com/issues/52624
439 60 Rishabh Dave
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
440 59 Rishabh Dave
441 57 Venky Shankar
h3. 2022 June 20
442
443
https://pulpito.ceph.com/vshankar-2022-06-15_04:03:39-fs-wip-vshankar-testing1-20220615-072516-testing-default-smithi/
444
https://pulpito.ceph.com/vshankar-2022-06-19_08:22:46-fs-wip-vshankar-testing1-20220619-102531-testing-default-smithi/
445
446
* https://tracker.ceph.com/issues/52624
447
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
448
* https://tracker.ceph.com/issues/55804
449
    qa failure: pjd link tests failed
450
* https://tracker.ceph.com/issues/54108
451
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
452
* https://tracker.ceph.com/issues/55332
453
    Failure in snaptest-git-ceph.sh (it's an async unlink/create bug)
454
455 56 Patrick Donnelly
h3. 2022 June 13
456
457
https://pulpito.ceph.com/pdonnell-2022-06-12_05:08:12-fs:workload-wip-pdonnell-testing-20220612.004943-distro-default-smithi/
458
459
* https://tracker.ceph.com/issues/56024
460
    cephadm: removes ceph.conf during qa run causing command failure
461
* https://tracker.ceph.com/issues/48773
462
    qa: scrub does not complete
463
* https://tracker.ceph.com/issues/56012
464
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
465
466
467 55 Venky Shankar
h3. 2022 Jun 13
468 54 Venky Shankar
469
https://pulpito.ceph.com/vshankar-2022-06-07_00:25:50-fs-wip-vshankar-testing-20220606-223254-testing-default-smithi/
470
https://pulpito.ceph.com/vshankar-2022-06-10_01:04:46-fs-wip-vshankar-testing-20220609-175550-testing-default-smithi/
471
472
* https://tracker.ceph.com/issues/52624
473
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
474
* https://tracker.ceph.com/issues/51964
475
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
476
* https://tracker.ceph.com/issues/53859
477
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
478
* https://tracker.ceph.com/issues/55804
479
    qa failure: pjd link tests failed
480
* https://tracker.ceph.com/issues/56003
481
    client: src/include/xlist.h: 81: FAILED ceph_assert(_size == 0)
482
* https://tracker.ceph.com/issues/56011
483
    fs/thrash: snaptest-snap-rm-cmp.sh fails in mds5sum comparison
484
* https://tracker.ceph.com/issues/56012
485
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
486
487 53 Venky Shankar
h3. 2022 Jun 07
488
489
https://pulpito.ceph.com/vshankar-2022-06-06_21:25:41-fs-wip-vshankar-testing1-20220606-230129-testing-default-smithi/
490
https://pulpito.ceph.com/vshankar-2022-06-07_10:53:31-fs-wip-vshankar-testing1-20220607-104134-testing-default-smithi/ (rerun after dropping a problematic PR)
491
492
* https://tracker.ceph.com/issues/52624
493
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
494
* https://tracker.ceph.com/issues/50223
495
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
496
* https://tracker.ceph.com/issues/50224
497
    qa: test_mirroring_init_failure_with_recovery failure
498
499 51 Venky Shankar
h3. 2022 May 12
500
501
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220509-125847
502 52 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-13_17:09:16-fs-wip-vshankar-testing-20220513-120051-testing-default-smithi/ (drop prs + rerun)
503 51 Venky Shankar
504
* https://tracker.ceph.com/issues/52624
505
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
506
* https://tracker.ceph.com/issues/50223
507
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
508
* https://tracker.ceph.com/issues/55332
509
    Failure in snaptest-git-ceph.sh
510
* https://tracker.ceph.com/issues/53859
511
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
512
* https://tracker.ceph.com/issues/55538
513 1 Patrick Donnelly
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
514 52 Venky Shankar
* https://tracker.ceph.com/issues/55258
515
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs (cropss up again, though very infrequent)
516 51 Venky Shankar
517 49 Venky Shankar
h3. 2022 May 04
518
519 50 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-01_13:18:44-fs-wip-vshankar-testing1-20220428-204527-testing-default-smithi/
520
https://pulpito.ceph.com/vshankar-2022-05-02_16:58:59-fs-wip-vshankar-testing1-20220502-201957-testing-default-smithi/ (after dropping PRs)
521
522 49 Venky Shankar
* https://tracker.ceph.com/issues/52624
523
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
524
* https://tracker.ceph.com/issues/50223
525
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
526
* https://tracker.ceph.com/issues/55332
527
    Failure in snaptest-git-ceph.sh
528
* https://tracker.ceph.com/issues/53859
529
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
530
* https://tracker.ceph.com/issues/55516
531
    qa: fs suite tests failing with "json.decoder.JSONDecodeError: Extra data: line 2 column 82 (char 82)"
532
* https://tracker.ceph.com/issues/55537
533
    mds: crash during fs:upgrade test
534
* https://tracker.ceph.com/issues/55538
535
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
536
537 48 Venky Shankar
h3. 2022 Apr 25
538
539
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220420-113951 (owner vshankar)
540
541
* https://tracker.ceph.com/issues/52624
542
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
543
* https://tracker.ceph.com/issues/50223
544
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
545
* https://tracker.ceph.com/issues/55258
546
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
547
* https://tracker.ceph.com/issues/55377
548
    kclient: mds revoke Fwb caps stuck after the kclient tries writebcak once
549
550 47 Venky Shankar
h3. 2022 Apr 14
551
552
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220411-144044
553
554
* https://tracker.ceph.com/issues/52624
555
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
556
* https://tracker.ceph.com/issues/50223
557
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
558
* https://tracker.ceph.com/issues/52438
559
    qa: ffsb timeout
560
* https://tracker.ceph.com/issues/55170
561
    mds: crash during rejoin (CDir::fetch_keys)
562
* https://tracker.ceph.com/issues/55331
563
    pjd failure
564
* https://tracker.ceph.com/issues/48773
565
    qa: scrub does not complete
566
* https://tracker.ceph.com/issues/55332
567
    Failure in snaptest-git-ceph.sh
568
* https://tracker.ceph.com/issues/55258
569
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
570
571 45 Venky Shankar
h3. 2022 Apr 11
572
573 46 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-55110-20220408-203242
574 45 Venky Shankar
575
* https://tracker.ceph.com/issues/48773
576
    qa: scrub does not complete
577
* https://tracker.ceph.com/issues/52624
578
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
579
* https://tracker.ceph.com/issues/52438
580
    qa: ffsb timeout
581
* https://tracker.ceph.com/issues/48680
582
    mds: scrubbing stuck "scrub active (0 inodes in the stack)"
583
* https://tracker.ceph.com/issues/55236
584
    qa: fs/snaps tests fails with "hit max job timeout"
585
* https://tracker.ceph.com/issues/54108
586
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
587
* https://tracker.ceph.com/issues/54971
588
    Test failure: test_perf_stats_stale_metrics (tasks.cephfs.test_mds_metrics.TestMDSMetrics)
589
* https://tracker.ceph.com/issues/50223
590
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
591
* https://tracker.ceph.com/issues/55258
592
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
593
594 44 Venky Shankar
h3. 2022 Mar 21
595 42 Venky Shankar
596 43 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-03-20_02:16:37-fs-wip-vshankar-testing-20220319-163539-testing-default-smithi/
597
598
Run didn't go well, lots of failures - debugging by dropping PRs and running against master branch. Only merging unrelated PRs that pass tests.
599
600
601
h3. 2022 Mar 08
602
603 42 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-02-28_04:32:15-fs-wip-vshankar-testing-20220226-211550-testing-default-smithi/
604
605
rerun with
606
- (drop) https://github.com/ceph/ceph/pull/44679
607
- (drop) https://github.com/ceph/ceph/pull/44958
608
https://pulpito.ceph.com/vshankar-2022-03-06_14:47:51-fs-wip-vshankar-testing-20220304-132102-testing-default-smithi/
609
610
* https://tracker.ceph.com/issues/54419 (new)
611
    `ceph orch upgrade start` seems to never reach completion
612
* https://tracker.ceph.com/issues/51964
613
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
614
* https://tracker.ceph.com/issues/52624
615
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
616
* https://tracker.ceph.com/issues/50223
617
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
618
* https://tracker.ceph.com/issues/52438
619
    qa: ffsb timeout
620
* https://tracker.ceph.com/issues/50821
621
    qa: untar_snap_rm failure during mds thrashing
622
623
624 41 Venky Shankar
h3. 2022 Feb 09
625
626
https://pulpito.ceph.com/vshankar-2022-02-05_17:27:49-fs-wip-vshankar-testing-20220201-113815-testing-default-smithi/
627
628
rerun with
629
- (drop) https://github.com/ceph/ceph/pull/37938
630
- (drop) https://github.com/ceph/ceph/pull/44335
631
- (drop) https://github.com/ceph/ceph/pull/44491
632
- (drop) https://github.com/ceph/ceph/pull/44501
633
https://pulpito.ceph.com/vshankar-2022-02-08_14:27:29-fs-wip-vshankar-testing-20220208-181241-testing-default-smithi/
634
635
* https://tracker.ceph.com/issues/51964
636
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
637
* https://tracker.ceph.com/issues/54066
638
    test_subvolume_no_upgrade_v1_sanity fails with `AssertionError: 1000 != 0`
639
* https://tracker.ceph.com/issues/48773
640
    qa: scrub does not complete
641
* https://tracker.ceph.com/issues/52624
642
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
643
* https://tracker.ceph.com/issues/50223
644
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
645
* https://tracker.ceph.com/issues/52438
646
    qa: ffsb timeout
647
648 40 Patrick Donnelly
h3. 2022 Feb 01
649
650
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220127.171526
651
652
* https://tracker.ceph.com/issues/54107
653
    kclient: hang during umount
654
* https://tracker.ceph.com/issues/54106
655
    kclient: hang during workunit cleanup
656
* https://tracker.ceph.com/issues/54108
657
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
658
* https://tracker.ceph.com/issues/48773
659
    qa: scrub does not complete
660
* https://tracker.ceph.com/issues/52438
661
    qa: ffsb timeout
662
663
664 36 Venky Shankar
h3. 2022 Jan 13
665
666
https://pulpito.ceph.com/vshankar-2022-01-06_13:18:41-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
667 39 Venky Shankar
668 36 Venky Shankar
rerun with:
669 38 Venky Shankar
- (add) https://github.com/ceph/ceph/pull/44570
670
- (drop) https://github.com/ceph/ceph/pull/43184
671 36 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-01-13_04:42:40-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
672
673
* https://tracker.ceph.com/issues/50223
674
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
675
* https://tracker.ceph.com/issues/51282
676
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
677
* https://tracker.ceph.com/issues/48773
678
    qa: scrub does not complete
679
* https://tracker.ceph.com/issues/52624
680
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
681
* https://tracker.ceph.com/issues/53859
682
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
683
684 34 Venky Shankar
h3. 2022 Jan 03
685
686
https://pulpito.ceph.com/vshankar-2021-12-22_07:37:44-fs-wip-vshankar-testing-20211216-114012-testing-default-smithi/
687
https://pulpito.ceph.com/vshankar-2022-01-03_12:27:45-fs-wip-vshankar-testing-20220103-142738-testing-default-smithi/ (rerun)
688
689
* https://tracker.ceph.com/issues/50223
690
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
691
* https://tracker.ceph.com/issues/51964
692
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
693
* https://tracker.ceph.com/issues/51267
694
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
695
* https://tracker.ceph.com/issues/51282
696
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
697
* https://tracker.ceph.com/issues/50821
698
    qa: untar_snap_rm failure during mds thrashing
699
* https://tracker.ceph.com/issues/51278
700
    mds: "FAILED ceph_assert(!segments.empty())"
701 35 Ramana Raja
* https://tracker.ceph.com/issues/52279
702
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
703
704 34 Venky Shankar
705 33 Patrick Donnelly
h3. 2021 Dec 22
706
707
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211222.014316
708
709
* https://tracker.ceph.com/issues/52624
710
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
711
* https://tracker.ceph.com/issues/50223
712
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
713
* https://tracker.ceph.com/issues/52279
714
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
715
* https://tracker.ceph.com/issues/50224
716
    qa: test_mirroring_init_failure_with_recovery failure
717
* https://tracker.ceph.com/issues/48773
718
    qa: scrub does not complete
719
720
721 32 Venky Shankar
h3. 2021 Nov 30
722
723
https://pulpito.ceph.com/vshankar-2021-11-24_07:14:27-fs-wip-vshankar-testing-20211124-094330-testing-default-smithi/
724
https://pulpito.ceph.com/vshankar-2021-11-30_06:23:32-fs-wip-vshankar-testing-20211124-094330-distro-default-smithi/ (rerun w/ QA fixes)
725
726
* https://tracker.ceph.com/issues/53436
727
    mds, mon: mds beacon messages get dropped? (mds never reaches up:active state)
728
* https://tracker.ceph.com/issues/51964
729
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
730
* https://tracker.ceph.com/issues/48812
731
    qa: test_scrub_pause_and_resume_with_abort failure
732
* https://tracker.ceph.com/issues/51076
733
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
734
* https://tracker.ceph.com/issues/50223
735
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
736
* https://tracker.ceph.com/issues/52624
737
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
738
* https://tracker.ceph.com/issues/50250
739
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
740
741
742 31 Patrick Donnelly
h3. 2021 November 9
743
744
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211109.180315
745
746
* https://tracker.ceph.com/issues/53214
747
    qa: "dd: error reading '/sys/kernel/debug/ceph/2a934501-6731-4052-a836-f42229a869be.client4874/metrics': Is a directory"
748
* https://tracker.ceph.com/issues/48773
749
    qa: scrub does not complete
750
* https://tracker.ceph.com/issues/50223
751
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
752
* https://tracker.ceph.com/issues/51282
753
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
754
* https://tracker.ceph.com/issues/52624
755
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
756
* https://tracker.ceph.com/issues/53216
757
    qa: "RuntimeError: value of attributes should be either str or None. client_id"
758
* https://tracker.ceph.com/issues/50250
759
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
760
761
762
763 30 Patrick Donnelly
h3. 2021 November 03
764
765
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211103.023355
766
767
* https://tracker.ceph.com/issues/51964
768
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
769
* https://tracker.ceph.com/issues/51282
770
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
771
* https://tracker.ceph.com/issues/52436
772
    fs/ceph: "corrupt mdsmap"
773
* https://tracker.ceph.com/issues/53074
774
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
775
* https://tracker.ceph.com/issues/53150
776
    pybind/mgr/cephadm/upgrade: tolerate MDS failures during upgrade straddling v16.2.5
777
* https://tracker.ceph.com/issues/53155
778
    MDSMonitor: assertion during upgrade to v16.2.5+
779
780
781 29 Patrick Donnelly
h3. 2021 October 26
782
783
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211025.000447
784
785
* https://tracker.ceph.com/issues/53074
786
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
787
* https://tracker.ceph.com/issues/52997
788
    testing: hang ing umount
789
* https://tracker.ceph.com/issues/50824
790
    qa: snaptest-git-ceph bus error
791
* https://tracker.ceph.com/issues/52436
792
    fs/ceph: "corrupt mdsmap"
793
* https://tracker.ceph.com/issues/48773
794
    qa: scrub does not complete
795
* https://tracker.ceph.com/issues/53082
796
    ceph-fuse: segmenetation fault in Client::handle_mds_map
797
* https://tracker.ceph.com/issues/50223
798
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
799
* https://tracker.ceph.com/issues/52624
800
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
801
* https://tracker.ceph.com/issues/50224
802
    qa: test_mirroring_init_failure_with_recovery failure
803
* https://tracker.ceph.com/issues/50821
804
    qa: untar_snap_rm failure during mds thrashing
805
* https://tracker.ceph.com/issues/50250
806
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
807
808
809
810 27 Patrick Donnelly
h3. 2021 October 19
811
812 28 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211019.013028
813 27 Patrick Donnelly
814
* https://tracker.ceph.com/issues/52995
815
    qa: test_standby_count_wanted failure
816
* https://tracker.ceph.com/issues/52948
817
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
818
* https://tracker.ceph.com/issues/52996
819
    qa: test_perf_counters via test_openfiletable
820
* https://tracker.ceph.com/issues/48772
821
    qa: pjd: not ok 9, 44, 80
822
* https://tracker.ceph.com/issues/52997
823
    testing: hang ing umount
824
* https://tracker.ceph.com/issues/50250
825
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
826
* https://tracker.ceph.com/issues/52624
827
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
828
* https://tracker.ceph.com/issues/50223
829
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
830
* https://tracker.ceph.com/issues/50821
831
    qa: untar_snap_rm failure during mds thrashing
832
* https://tracker.ceph.com/issues/48773
833
    qa: scrub does not complete
834
835
836 26 Patrick Donnelly
h3. 2021 October 12
837
838
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211012.192211
839
840
Some failures caused by teuthology bug: https://tracker.ceph.com/issues/52944
841
842
New test caused failure: https://github.com/ceph/ceph/pull/43297#discussion_r729883167
843
844
845
* https://tracker.ceph.com/issues/51282
846
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
847
* https://tracker.ceph.com/issues/52948
848
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
849
* https://tracker.ceph.com/issues/48773
850
    qa: scrub does not complete
851
* https://tracker.ceph.com/issues/50224
852
    qa: test_mirroring_init_failure_with_recovery failure
853
* https://tracker.ceph.com/issues/52949
854
    RuntimeError: The following counters failed to be set on mds daemons: {'mds.dir_split'}
855
856
857 25 Patrick Donnelly
h3. 2021 October 02
858 23 Patrick Donnelly
859 24 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211002.163337
860
861
Some failures caused by cephadm upgrade test. Fixed in follow-up qa commit.
862
863
test_simple failures caused by PR in this set.
864
865
A few reruns because of QA infra noise.
866
867
* https://tracker.ceph.com/issues/52822
868
    qa: failed pacific install on fs:upgrade
869
* https://tracker.ceph.com/issues/52624
870
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
871
* https://tracker.ceph.com/issues/50223
872
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
873
* https://tracker.ceph.com/issues/48773
874
    qa: scrub does not complete
875
876
877
h3. 2021 September 20
878
879 23 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210917.174826
880
881
* https://tracker.ceph.com/issues/52677
882
    qa: test_simple failure
883
* https://tracker.ceph.com/issues/51279
884
    kclient hangs on umount (testing branch)
885
* https://tracker.ceph.com/issues/50223
886
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
887
* https://tracker.ceph.com/issues/50250
888
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
889
* https://tracker.ceph.com/issues/52624
890
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
891
* https://tracker.ceph.com/issues/52438
892
    qa: ffsb timeout
893
894
895 22 Patrick Donnelly
h3. 2021 September 10
896
897
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210910.181451
898
899
* https://tracker.ceph.com/issues/50223
900
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
901
* https://tracker.ceph.com/issues/50250
902
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
903
* https://tracker.ceph.com/issues/52624
904
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
905
* https://tracker.ceph.com/issues/52625
906
    qa: test_kill_mdstable (tasks.cephfs.test_snapshots.TestSnapshots)
907
* https://tracker.ceph.com/issues/52439
908
    qa: acls does not compile on centos stream
909
* https://tracker.ceph.com/issues/50821
910
    qa: untar_snap_rm failure during mds thrashing
911
* https://tracker.ceph.com/issues/48773
912
    qa: scrub does not complete
913
* https://tracker.ceph.com/issues/52626
914
    mds: ScrubStack.cc: 831: FAILED ceph_assert(diri)
915
* https://tracker.ceph.com/issues/51279
916
    kclient hangs on umount (testing branch)
917
918
919 21 Patrick Donnelly
h3. 2021 August 27
920
921
Several jobs died because of device failures.
922
923
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210827.024746
924
925
* https://tracker.ceph.com/issues/52430
926
    mds: fast async create client mount breaks racy test
927
* https://tracker.ceph.com/issues/52436
928
    fs/ceph: "corrupt mdsmap"
929
* https://tracker.ceph.com/issues/52437
930
    mds: InoTable::replay_release_ids abort via test_inotable_sync
931
* https://tracker.ceph.com/issues/51282
932
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
933
* https://tracker.ceph.com/issues/52438
934
    qa: ffsb timeout
935
* https://tracker.ceph.com/issues/52439
936
    qa: acls does not compile on centos stream
937
938
939 20 Patrick Donnelly
h3. 2021 July 30
940
941
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210729.214022
942
943
* https://tracker.ceph.com/issues/50250
944
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
945
* https://tracker.ceph.com/issues/51282
946
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
947
* https://tracker.ceph.com/issues/48773
948
    qa: scrub does not complete
949
* https://tracker.ceph.com/issues/51975
950
    pybind/mgr/stats: KeyError
951
952
953 19 Patrick Donnelly
h3. 2021 July 28
954
955
https://pulpito.ceph.com/pdonnell-2021-07-28_00:39:45-fs-wip-pdonnell-testing-20210727.213757-distro-basic-smithi/
956
957
with qa fix: https://pulpito.ceph.com/pdonnell-2021-07-28_16:20:28-fs-wip-pdonnell-testing-20210728.141004-distro-basic-smithi/
958
959
* https://tracker.ceph.com/issues/51905
960
    qa: "error reading sessionmap 'mds1_sessionmap'"
961
* https://tracker.ceph.com/issues/48773
962
    qa: scrub does not complete
963
* https://tracker.ceph.com/issues/50250
964
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
965
* https://tracker.ceph.com/issues/51267
966
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
967
* https://tracker.ceph.com/issues/51279
968
    kclient hangs on umount (testing branch)
969
970
971 18 Patrick Donnelly
h3. 2021 July 16
972
973
https://pulpito.ceph.com/pdonnell-2021-07-16_05:50:11-fs-wip-pdonnell-testing-20210716.022804-distro-basic-smithi/
974
975
* https://tracker.ceph.com/issues/48773
976
    qa: scrub does not complete
977
* https://tracker.ceph.com/issues/48772
978
    qa: pjd: not ok 9, 44, 80
979
* https://tracker.ceph.com/issues/45434
980
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
981
* https://tracker.ceph.com/issues/51279
982
    kclient hangs on umount (testing branch)
983
* https://tracker.ceph.com/issues/50824
984
    qa: snaptest-git-ceph bus error
985
986
987 17 Patrick Donnelly
h3. 2021 July 04
988
989
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210703.052904
990
991
* https://tracker.ceph.com/issues/48773
992
    qa: scrub does not complete
993
* https://tracker.ceph.com/issues/39150
994
    mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
995
* https://tracker.ceph.com/issues/45434
996
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
997
* https://tracker.ceph.com/issues/51282
998
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
999
* https://tracker.ceph.com/issues/48771
1000
    qa: iogen: workload fails to cause balancing
1001
* https://tracker.ceph.com/issues/51279
1002
    kclient hangs on umount (testing branch)
1003
* https://tracker.ceph.com/issues/50250
1004
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
1005
1006
1007 16 Patrick Donnelly
h3. 2021 July 01
1008
1009
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210701.192056
1010
1011
* https://tracker.ceph.com/issues/51197
1012
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
1013
* https://tracker.ceph.com/issues/50866
1014
    osd: stat mismatch on objects
1015
* https://tracker.ceph.com/issues/48773
1016
    qa: scrub does not complete
1017
1018
1019 15 Patrick Donnelly
h3. 2021 June 26
1020
1021
https://pulpito.ceph.com/pdonnell-2021-06-26_00:57:00-fs-wip-pdonnell-testing-20210625.225421-distro-basic-smithi/
1022
1023
* https://tracker.ceph.com/issues/51183
1024
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1025
* https://tracker.ceph.com/issues/51410
1026
    kclient: fails to finish reconnect during MDS thrashing (testing branch)
1027
* https://tracker.ceph.com/issues/48773
1028
    qa: scrub does not complete
1029
* https://tracker.ceph.com/issues/51282
1030
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1031
* https://tracker.ceph.com/issues/51169
1032
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1033
* https://tracker.ceph.com/issues/48772
1034
    qa: pjd: not ok 9, 44, 80
1035
1036
1037 14 Patrick Donnelly
h3. 2021 June 21
1038
1039
https://pulpito.ceph.com/pdonnell-2021-06-22_00:27:21-fs-wip-pdonnell-testing-20210621.231646-distro-basic-smithi/
1040
1041
One failure caused by PR: https://github.com/ceph/ceph/pull/41935#issuecomment-866472599
1042
1043
* https://tracker.ceph.com/issues/51282
1044
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1045
* https://tracker.ceph.com/issues/51183
1046
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1047
* https://tracker.ceph.com/issues/48773
1048
    qa: scrub does not complete
1049
* https://tracker.ceph.com/issues/48771
1050
    qa: iogen: workload fails to cause balancing
1051
* https://tracker.ceph.com/issues/51169
1052
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1053
* https://tracker.ceph.com/issues/50495
1054
    libcephfs: shutdown race fails with status 141
1055
* https://tracker.ceph.com/issues/45434
1056
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1057
* https://tracker.ceph.com/issues/50824
1058
    qa: snaptest-git-ceph bus error
1059
* https://tracker.ceph.com/issues/50223
1060
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1061
1062
1063 13 Patrick Donnelly
h3. 2021 June 16
1064
1065
https://pulpito.ceph.com/pdonnell-2021-06-16_21:26:55-fs-wip-pdonnell-testing-20210616.191804-distro-basic-smithi/
1066
1067
MDS abort class of failures caused by PR: https://github.com/ceph/ceph/pull/41667
1068
1069
* https://tracker.ceph.com/issues/45434
1070
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1071
* https://tracker.ceph.com/issues/51169
1072
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1073
* https://tracker.ceph.com/issues/43216
1074
    MDSMonitor: removes MDS coming out of quorum election
1075
* https://tracker.ceph.com/issues/51278
1076
    mds: "FAILED ceph_assert(!segments.empty())"
1077
* https://tracker.ceph.com/issues/51279
1078
    kclient hangs on umount (testing branch)
1079
* https://tracker.ceph.com/issues/51280
1080
    mds: "FAILED ceph_assert(r == 0 || r == -2)"
1081
* https://tracker.ceph.com/issues/51183
1082
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1083
* https://tracker.ceph.com/issues/51281
1084
    qa: snaptest-snap-rm-cmp.sh: "echo 'FAIL: bad match, /tmp/a 4637e766853d1ad16a7b17079e2c6f03 != real c3883760b18d50e8d78819c54d579b00'"
1085
* https://tracker.ceph.com/issues/48773
1086
    qa: scrub does not complete
1087
* https://tracker.ceph.com/issues/51076
1088
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
1089
* https://tracker.ceph.com/issues/51228
1090
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
1091
* https://tracker.ceph.com/issues/51282
1092
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1093
1094
1095 12 Patrick Donnelly
h3. 2021 June 14
1096
1097
https://pulpito.ceph.com/pdonnell-2021-06-14_20:53:05-fs-wip-pdonnell-testing-20210614.173325-distro-basic-smithi/
1098
1099
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1100
1101
* https://tracker.ceph.com/issues/51169
1102
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1103
* https://tracker.ceph.com/issues/51228
1104
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
1105
* https://tracker.ceph.com/issues/48773
1106
    qa: scrub does not complete
1107
* https://tracker.ceph.com/issues/51183
1108
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1109
* https://tracker.ceph.com/issues/45434
1110
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1111
* https://tracker.ceph.com/issues/51182
1112
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1113
* https://tracker.ceph.com/issues/51229
1114
    qa: test_multi_snap_schedule list difference failure
1115
* https://tracker.ceph.com/issues/50821
1116
    qa: untar_snap_rm failure during mds thrashing
1117
1118
1119 11 Patrick Donnelly
h3. 2021 June 13
1120
1121
https://pulpito.ceph.com/pdonnell-2021-06-12_02:45:35-fs-wip-pdonnell-testing-20210612.002809-distro-basic-smithi/
1122
1123
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1124
1125
* https://tracker.ceph.com/issues/51169
1126
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1127
* https://tracker.ceph.com/issues/48773
1128
    qa: scrub does not complete
1129
* https://tracker.ceph.com/issues/51182
1130
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1131
* https://tracker.ceph.com/issues/51183
1132
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1133
* https://tracker.ceph.com/issues/51197
1134
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
1135
* https://tracker.ceph.com/issues/45434
1136
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1137
1138 10 Patrick Donnelly
h3. 2021 June 11
1139
1140
https://pulpito.ceph.com/pdonnell-2021-06-11_18:02:10-fs-wip-pdonnell-testing-20210611.162716-distro-basic-smithi/
1141
1142
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1143
1144
* https://tracker.ceph.com/issues/51169
1145
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1146
* https://tracker.ceph.com/issues/45434
1147
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1148
* https://tracker.ceph.com/issues/48771
1149
    qa: iogen: workload fails to cause balancing
1150
* https://tracker.ceph.com/issues/43216
1151
    MDSMonitor: removes MDS coming out of quorum election
1152
* https://tracker.ceph.com/issues/51182
1153
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1154
* https://tracker.ceph.com/issues/50223
1155
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1156
* https://tracker.ceph.com/issues/48773
1157
    qa: scrub does not complete
1158
* https://tracker.ceph.com/issues/51183
1159
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1160
* https://tracker.ceph.com/issues/51184
1161
    qa: fs:bugs does not specify distro
1162
1163
1164 9 Patrick Donnelly
h3. 2021 June 03
1165
1166
https://pulpito.ceph.com/pdonnell-2021-06-03_03:40:33-fs-wip-pdonnell-testing-20210603.020013-distro-basic-smithi/
1167
1168
* https://tracker.ceph.com/issues/45434
1169
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1170
* https://tracker.ceph.com/issues/50016
1171
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1172
* https://tracker.ceph.com/issues/50821
1173
    qa: untar_snap_rm failure during mds thrashing
1174
* https://tracker.ceph.com/issues/50622 (regression)
1175
    msg: active_connections regression
1176
* https://tracker.ceph.com/issues/49845#note-2 (regression)
1177
    qa: failed umount in test_volumes
1178
* https://tracker.ceph.com/issues/48773
1179
    qa: scrub does not complete
1180
* https://tracker.ceph.com/issues/43216
1181
    MDSMonitor: removes MDS coming out of quorum election
1182
1183
1184 7 Patrick Donnelly
h3. 2021 May 18
1185
1186 8 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.214114
1187
1188
Regression in testing kernel caused some failures. Ilya fixed those and rerun
1189
looked better. Some odd new noise in the rerun relating to packaging and "No
1190
module named 'tasks.ceph'".
1191
1192
* https://tracker.ceph.com/issues/50824
1193
    qa: snaptest-git-ceph bus error
1194
* https://tracker.ceph.com/issues/50622 (regression)
1195
    msg: active_connections regression
1196
* https://tracker.ceph.com/issues/49845#note-2 (regression)
1197
    qa: failed umount in test_volumes
1198
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1199
    qa: quota failure
1200
1201
1202
h3. 2021 May 18
1203
1204 7 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.025642
1205
1206
* https://tracker.ceph.com/issues/50821
1207
    qa: untar_snap_rm failure during mds thrashing
1208
* https://tracker.ceph.com/issues/48773
1209
    qa: scrub does not complete
1210
* https://tracker.ceph.com/issues/45591
1211
    mgr: FAILED ceph_assert(daemon != nullptr)
1212
* https://tracker.ceph.com/issues/50866
1213
    osd: stat mismatch on objects
1214
* https://tracker.ceph.com/issues/50016
1215
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1216
* https://tracker.ceph.com/issues/50867
1217
    qa: fs:mirror: reduced data availability
1218
* https://tracker.ceph.com/issues/50821
1219
    qa: untar_snap_rm failure during mds thrashing
1220
* https://tracker.ceph.com/issues/50622 (regression)
1221
    msg: active_connections regression
1222
* https://tracker.ceph.com/issues/50223
1223
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1224
* https://tracker.ceph.com/issues/50868
1225
    qa: "kern.log.gz already exists; not overwritten"
1226
* https://tracker.ceph.com/issues/50870
1227
    qa: test_full: "rm: cannot remove 'large_file_a': Permission denied"
1228
1229
1230 6 Patrick Donnelly
h3. 2021 May 11
1231
1232
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210511.232042
1233
1234
* one class of failures caused by PR
1235
* https://tracker.ceph.com/issues/48812
1236
    qa: test_scrub_pause_and_resume_with_abort failure
1237
* https://tracker.ceph.com/issues/50390
1238
    mds: monclient: wait_auth_rotating timed out after 30
1239
* https://tracker.ceph.com/issues/48773
1240
    qa: scrub does not complete
1241
* https://tracker.ceph.com/issues/50821
1242
    qa: untar_snap_rm failure during mds thrashing
1243
* https://tracker.ceph.com/issues/50224
1244
    qa: test_mirroring_init_failure_with_recovery failure
1245
* https://tracker.ceph.com/issues/50622 (regression)
1246
    msg: active_connections regression
1247
* https://tracker.ceph.com/issues/50825
1248
    qa: snaptest-git-ceph hang during mon thrashing v2
1249
* https://tracker.ceph.com/issues/50821
1250
    qa: untar_snap_rm failure during mds thrashing
1251
* https://tracker.ceph.com/issues/50823
1252
    qa: RuntimeError: timeout waiting for cluster to stabilize
1253
1254
1255 5 Patrick Donnelly
h3. 2021 May 14
1256
1257
https://pulpito.ceph.com/pdonnell-2021-05-14_21:45:42-fs-master-distro-basic-smithi/
1258
1259
* https://tracker.ceph.com/issues/48812
1260
    qa: test_scrub_pause_and_resume_with_abort failure
1261
* https://tracker.ceph.com/issues/50821
1262
    qa: untar_snap_rm failure during mds thrashing
1263
* https://tracker.ceph.com/issues/50622 (regression)
1264
    msg: active_connections regression
1265
* https://tracker.ceph.com/issues/50822
1266
    qa: testing kernel patch for client metrics causes mds abort
1267
* https://tracker.ceph.com/issues/48773
1268
    qa: scrub does not complete
1269
* https://tracker.ceph.com/issues/50823
1270
    qa: RuntimeError: timeout waiting for cluster to stabilize
1271
* https://tracker.ceph.com/issues/50824
1272
    qa: snaptest-git-ceph bus error
1273
* https://tracker.ceph.com/issues/50825
1274
    qa: snaptest-git-ceph hang during mon thrashing v2
1275
* https://tracker.ceph.com/issues/50826
1276
    kceph: stock RHEL kernel hangs on snaptests with mon|osd thrashers
1277
1278
1279 4 Patrick Donnelly
h3. 2021 May 01
1280
1281
https://pulpito.ceph.com/pdonnell-2021-05-01_09:07:09-fs-wip-pdonnell-testing-20210501.040415-distro-basic-smithi/
1282
1283
* https://tracker.ceph.com/issues/45434
1284
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1285
* https://tracker.ceph.com/issues/50281
1286
    qa: untar_snap_rm timeout
1287
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1288
    qa: quota failure
1289
* https://tracker.ceph.com/issues/48773
1290
    qa: scrub does not complete
1291
* https://tracker.ceph.com/issues/50390
1292
    mds: monclient: wait_auth_rotating timed out after 30
1293
* https://tracker.ceph.com/issues/50250
1294
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1295
* https://tracker.ceph.com/issues/50622 (regression)
1296
    msg: active_connections regression
1297
* https://tracker.ceph.com/issues/45591
1298
    mgr: FAILED ceph_assert(daemon != nullptr)
1299
* https://tracker.ceph.com/issues/50221
1300
    qa: snaptest-git-ceph failure in git diff
1301
* https://tracker.ceph.com/issues/50016
1302
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1303
1304
1305 3 Patrick Donnelly
h3. 2021 Apr 15
1306
1307
https://pulpito.ceph.com/pdonnell-2021-04-15_01:35:57-fs-wip-pdonnell-testing-20210414.230315-distro-basic-smithi/
1308
1309
* https://tracker.ceph.com/issues/50281
1310
    qa: untar_snap_rm timeout
1311
* https://tracker.ceph.com/issues/50220
1312
    qa: dbench workload timeout
1313
* https://tracker.ceph.com/issues/50246
1314
    mds: failure replaying journal (EMetaBlob)
1315
* https://tracker.ceph.com/issues/50250
1316
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1317
* https://tracker.ceph.com/issues/50016
1318
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1319
* https://tracker.ceph.com/issues/50222
1320
    osd: 5.2s0 deep-scrub : stat mismatch
1321
* https://tracker.ceph.com/issues/45434
1322
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1323
* https://tracker.ceph.com/issues/49845
1324
    qa: failed umount in test_volumes
1325
* https://tracker.ceph.com/issues/37808
1326
    osd: osdmap cache weak_refs assert during shutdown
1327
* https://tracker.ceph.com/issues/50387
1328
    client: fs/snaps failure
1329
* https://tracker.ceph.com/issues/50389
1330
    mds: "cluster [ERR] Error recovering journal 0x203: (2) No such file or directory" in cluster log"
1331
* https://tracker.ceph.com/issues/50216
1332
    qa: "ls: cannot access 'lost+found': No such file or directory"
1333
* https://tracker.ceph.com/issues/50390
1334
    mds: monclient: wait_auth_rotating timed out after 30
1335
1336
1337
1338 1 Patrick Donnelly
h3. 2021 Apr 08
1339
1340 2 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_22:42:24-fs-wip-pdonnell-testing-20210408.192301-distro-basic-smithi/
1341
1342
* https://tracker.ceph.com/issues/45434
1343
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1344
* https://tracker.ceph.com/issues/50016
1345
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1346
* https://tracker.ceph.com/issues/48773
1347
    qa: scrub does not complete
1348
* https://tracker.ceph.com/issues/50279
1349
    qa: "Replacing daemon mds.b as rank 0 with standby daemon mds.c"
1350
* https://tracker.ceph.com/issues/50246
1351
    mds: failure replaying journal (EMetaBlob)
1352
* https://tracker.ceph.com/issues/48365
1353
    qa: ffsb build failure on CentOS 8.2
1354
* https://tracker.ceph.com/issues/50216
1355
    qa: "ls: cannot access 'lost+found': No such file or directory"
1356
* https://tracker.ceph.com/issues/50223
1357
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1358
* https://tracker.ceph.com/issues/50280
1359
    cephadm: RuntimeError: uid/gid not found
1360
* https://tracker.ceph.com/issues/50281
1361
    qa: untar_snap_rm timeout
1362
1363
h3. 2021 Apr 08
1364
1365 1 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_04:31:36-fs-wip-pdonnell-testing-20210408.024225-distro-basic-smithi/
1366
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210408.142238 (with logic inversion / QA fix)
1367
1368
* https://tracker.ceph.com/issues/50246
1369
    mds: failure replaying journal (EMetaBlob)
1370
* https://tracker.ceph.com/issues/50250
1371
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1372
1373
1374
h3. 2021 Apr 07
1375
1376
https://pulpito.ceph.com/pdonnell-2021-04-07_02:12:41-fs-wip-pdonnell-testing-20210406.213012-distro-basic-smithi/
1377
1378
* https://tracker.ceph.com/issues/50215
1379
    qa: "log [ERR] : error reading sessionmap 'mds2_sessionmap'"
1380
* https://tracker.ceph.com/issues/49466
1381
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1382
* https://tracker.ceph.com/issues/50216
1383
    qa: "ls: cannot access 'lost+found': No such file or directory"
1384
* https://tracker.ceph.com/issues/48773
1385
    qa: scrub does not complete
1386
* https://tracker.ceph.com/issues/49845
1387
    qa: failed umount in test_volumes
1388
* https://tracker.ceph.com/issues/50220
1389
    qa: dbench workload timeout
1390
* https://tracker.ceph.com/issues/50221
1391
    qa: snaptest-git-ceph failure in git diff
1392
* https://tracker.ceph.com/issues/50222
1393
    osd: 5.2s0 deep-scrub : stat mismatch
1394
* https://tracker.ceph.com/issues/50223
1395
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1396
* https://tracker.ceph.com/issues/50224
1397
    qa: test_mirroring_init_failure_with_recovery failure
1398
1399
h3. 2021 Apr 01
1400
1401
https://pulpito.ceph.com/pdonnell-2021-04-01_00:45:34-fs-wip-pdonnell-testing-20210331.222326-distro-basic-smithi/
1402
1403
* https://tracker.ceph.com/issues/48772
1404
    qa: pjd: not ok 9, 44, 80
1405
* https://tracker.ceph.com/issues/50177
1406
    osd: "stalled aio... buggy kernel or bad device?"
1407
* https://tracker.ceph.com/issues/48771
1408
    qa: iogen: workload fails to cause balancing
1409
* https://tracker.ceph.com/issues/49845
1410
    qa: failed umount in test_volumes
1411
* https://tracker.ceph.com/issues/48773
1412
    qa: scrub does not complete
1413
* https://tracker.ceph.com/issues/48805
1414
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1415
* https://tracker.ceph.com/issues/50178
1416
    qa: "TypeError: run() got an unexpected keyword argument 'shell'"
1417
* https://tracker.ceph.com/issues/45434
1418
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1419
1420
h3. 2021 Mar 24
1421
1422
https://pulpito.ceph.com/pdonnell-2021-03-24_23:26:35-fs-wip-pdonnell-testing-20210324.190252-distro-basic-smithi/
1423
1424
* https://tracker.ceph.com/issues/49500
1425
    qa: "Assertion `cb_done' failed."
1426
* https://tracker.ceph.com/issues/50019
1427
    qa: mount failure with cephadm "probably no MDS server is up?"
1428
* https://tracker.ceph.com/issues/50020
1429
    qa: "RADOS object not found (Failed to operate read op for oid cephfs_mirror)"
1430
* https://tracker.ceph.com/issues/48773
1431
    qa: scrub does not complete
1432
* https://tracker.ceph.com/issues/45434
1433
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1434
* https://tracker.ceph.com/issues/48805
1435
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1436
* https://tracker.ceph.com/issues/48772
1437
    qa: pjd: not ok 9, 44, 80
1438
* https://tracker.ceph.com/issues/50021
1439
    qa: snaptest-git-ceph failure during mon thrashing
1440
* https://tracker.ceph.com/issues/48771
1441
    qa: iogen: workload fails to cause balancing
1442
* https://tracker.ceph.com/issues/50016
1443
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1444
* https://tracker.ceph.com/issues/49466
1445
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1446
1447
1448
h3. 2021 Mar 18
1449
1450
https://pulpito.ceph.com/pdonnell-2021-03-18_13:46:31-fs-wip-pdonnell-testing-20210318.024145-distro-basic-smithi/
1451
1452
* https://tracker.ceph.com/issues/49466
1453
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1454
* https://tracker.ceph.com/issues/48773
1455
    qa: scrub does not complete
1456
* https://tracker.ceph.com/issues/48805
1457
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1458
* https://tracker.ceph.com/issues/45434
1459
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1460
* https://tracker.ceph.com/issues/49845
1461
    qa: failed umount in test_volumes
1462
* https://tracker.ceph.com/issues/49605
1463
    mgr: drops command on the floor
1464
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1465
    qa: quota failure
1466
* https://tracker.ceph.com/issues/49928
1467
    client: items pinned in cache preventing unmount x2
1468
1469
h3. 2021 Mar 15
1470
1471
https://pulpito.ceph.com/pdonnell-2021-03-15_22:16:56-fs-wip-pdonnell-testing-20210315.182203-distro-basic-smithi/
1472
1473
* https://tracker.ceph.com/issues/49842
1474
    qa: stuck pkg install
1475
* https://tracker.ceph.com/issues/49466
1476
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1477
* https://tracker.ceph.com/issues/49822
1478
    test: test_mirroring_command_idempotency (tasks.cephfs.test_admin.TestMirroringCommands) failure
1479
* https://tracker.ceph.com/issues/49240
1480
    terminate called after throwing an instance of 'std::bad_alloc'
1481
* https://tracker.ceph.com/issues/48773
1482
    qa: scrub does not complete
1483
* https://tracker.ceph.com/issues/45434
1484
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1485
* https://tracker.ceph.com/issues/49500
1486
    qa: "Assertion `cb_done' failed."
1487
* https://tracker.ceph.com/issues/49843
1488
    qa: fs/snaps/snaptest-upchildrealms.sh failure
1489
* https://tracker.ceph.com/issues/49845
1490
    qa: failed umount in test_volumes
1491
* https://tracker.ceph.com/issues/48805
1492
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1493
* https://tracker.ceph.com/issues/49605
1494
    mgr: drops command on the floor
1495
1496
and failure caused by PR: https://github.com/ceph/ceph/pull/39969
1497
1498
1499
h3. 2021 Mar 09
1500
1501
https://pulpito.ceph.com/pdonnell-2021-03-09_03:27:39-fs-wip-pdonnell-testing-20210308.214827-distro-basic-smithi/
1502
1503
* https://tracker.ceph.com/issues/49500
1504
    qa: "Assertion `cb_done' failed."
1505
* https://tracker.ceph.com/issues/48805
1506
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1507
* https://tracker.ceph.com/issues/48773
1508
    qa: scrub does not complete
1509
* https://tracker.ceph.com/issues/45434
1510
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1511
* https://tracker.ceph.com/issues/49240
1512
    terminate called after throwing an instance of 'std::bad_alloc'
1513
* https://tracker.ceph.com/issues/49466
1514
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1515
* https://tracker.ceph.com/issues/49684
1516
    qa: fs:cephadm mount does not wait for mds to be created
1517
* https://tracker.ceph.com/issues/48771
1518
    qa: iogen: workload fails to cause balancing