Project

General

Profile

Main » History » Version 103

Rishabh Dave, 01/30/2023 06:01 AM

1 79 Venky Shankar
h1. MAIN
2
3 101 Rishabh Dave
h3. 30 JAN 2023
4
5
run: http://pulpito.front.sepia.ceph.com/rishabh-2022-11-28_08:04:11-fs-wip-rishabh-testing-2022Nov24-1818-testing-default-smithi/
6
re-run: http://pulpito.front.sepia.ceph.com/rishabh-2023-01-13_12:08:33-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/
7
re-run of re-run: http://pulpito.front.sepia.ceph.com/rishabh-2023-01-23_18:53:32-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/
8
9
* https://tracker.ceph.com/issues/56695
10
  [RHEL stock] pjd test failures
11
* https://tracker.ceph.com/issues/57676
12
  qa: error during scrub thrashing: rank damage found: {'backtrace'}
13
* https://tracker.ceph.com/issues/55332
14
  Failure in snaptest-git-ceph.sh
15
* https://tracker.ceph.com/issues/51964
16
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
17
* https://tracker.ceph.com/issues/56446
18
  Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
19
* https://tracker.ceph.com/issues/57655 
20
  qa: fs:mixed-clients kernel_untar_build failure
21
* https://tracker.ceph.com/issues/54460
22
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
23
* https://tracker.ceph.com/issues/58340
24
  mds: fsstress.sh hangs with multimds
25 103 Rishabh Dave
* https://tracker.ceph.com/issues/58219
26
  Command crashed: 'ceph-dencoder type inode_backtrace_t import - decode dump_json'
27 101 Rishabh Dave
28 102 Rishabh Dave
* "Failed to load ceph-mgr modules: prometheus" in cluster log"
29
  http://pulpito.front.sepia.ceph.com/rishabh-2023-01-23_18:53:32-fs-wip-rishabh-testing-2022Nov24-11Jan2023-distro-default-smithi/7134086
30
  Acc to Venky this was fixed in https://github.com/ceph/ceph/commit/cf6089200d96fc56b08ee17a4e31f19823370dc8
31
32 100 Venky Shankar
h3. 15 Dec 2022
33
34
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221215.112736
35
36
* https://tracker.ceph.com/issues/52624
37
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
38
* https://tracker.ceph.com/issues/56695
39
    [RHEL stock] pjd test failures
40
* https://tracker.ceph.com/issues/58219
41
* https://tracker.ceph.com/issues/57655
42
* qa: fs:mixed-clients kernel_untar_build failure
43
    Test failure: test_journal_migration (tasks.cephfs.test_journal_migration.TestJournalMigration)
44
* https://tracker.ceph.com/issues/57676
45
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
46
* https://tracker.ceph.com/issues/58340
47
    mds: fsstress.sh hangs with multimds
48
49 96 Venky Shankar
h3. 08 Dec 2022
50
51
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221130.043104
52 99 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20221209.043803
53 96 Venky Shankar
54
(lots of transient git.ceph.com failures)
55
56
* https://tracker.ceph.com/issues/52624
57
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
58
* https://tracker.ceph.com/issues/56695
59
    [RHEL stock] pjd test failures
60
* https://tracker.ceph.com/issues/57655
61
    qa: fs:mixed-clients kernel_untar_build failure
62
* https://tracker.ceph.com/issues/58219
63
    Test failure: test_journal_migration (tasks.cephfs.test_journal_migration.TestJournalMigration)
64
* https://tracker.ceph.com/issues/58220
65
    Command failed (workunit test fs/quota/quota.sh) on smithi081 with status 1:
66
* https://tracker.ceph.com/issues/57676
67
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
68 97 Venky Shankar
* https://tracker.ceph.com/issues/53859
69
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
70 98 Venky Shankar
* https://tracker.ceph.com/issues/54460
71
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
72
* https://tracker.ceph.com/issues/58244
73
    Test failure: test_rebuild_inotable (tasks.cephfs.test_data_scan.TestDataScan)
74 96 Venky Shankar
75 95 Venky Shankar
h3. 14 Oct 2022
76
77
https://pulpito.ceph.com/vshankar-2022-10-12_04:56:59-fs-wip-vshankar-testing-20221011-145847-testing-default-smithi/
78
https://pulpito.ceph.com/vshankar-2022-10-14_04:04:57-fs-wip-vshankar-testing-20221014-072608-testing-default-smithi/
79
80
* https://tracker.ceph.com/issues/52624
81
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
82
* https://tracker.ceph.com/issues/55804
83
    Command failed (workunit test suites/pjd.sh)
84
* https://tracker.ceph.com/issues/51964
85
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
86
* https://tracker.ceph.com/issues/57682
87
    client: ERROR: test_reconnect_after_blocklisted
88
* https://tracker.ceph.com/issues/54460
89
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
90 90 Rishabh Dave
91 91 Rishabh Dave
h3. 10 Oct 2022
92
93
http://pulpito.front.sepia.ceph.com/rishabh-2022-09-30_19:45:21-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
94 92 Rishabh Dave
95 91 Rishabh Dave
reruns
96
* fs-thrash, passed: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-04_13:19:47-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
97
* fs-verify, passed: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-05_12:25:37-fs-wip-rishabh-testing-30Sep2022-testing-default-smithi/
98
* cephadm failures also passed after many re-runs: http://pulpito.front.sepia.ceph.com/rishabh-2022-10-06_13:50:51-fs-wip-rishabh-testing-30Sep2022-2-testing-default-smithi/
99 94 Rishabh Dave
    ** needed this PR to be merged in ceph-ci branch - https://github.com/ceph/ceph/pull/47458
100 91 Rishabh Dave
101 93 Rishabh Dave
known bugs
102 91 Rishabh Dave
* https://tracker.ceph.com/issues/52624
103
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
104
* https://tracker.ceph.com/issues/50223
105
  client.xxxx isn't responding to mclientcaps(revoke
106
* https://tracker.ceph.com/issues/57299
107
  qa: test_dump_loads fails with JSONDecodeError
108
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
109
  qa: fs:mixed-clients kernel_untar_build failure
110
* https://tracker.ceph.com/issues/57206
111
  libcephfs/test.sh: ceph_test_libcephfs_reclaim
112
113 90 Rishabh Dave
h3. 2022 Sep 29
114
115
http://pulpito.front.sepia.ceph.com/rishabh-2022-09-14_12:48:43-fs-wip-rishabh-testing-2022Sep9-1708-testing-default-smithi/
116
117
* https://tracker.ceph.com/issues/55804
118
  Command failed (workunit test suites/pjd.sh)
119
* https://tracker.ceph.com/issues/36593
120
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
121
* https://tracker.ceph.com/issues/52624
122
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
123
* https://tracker.ceph.com/issues/51964
124
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
125
* https://tracker.ceph.com/issues/56632
126
  Test failure: test_subvolume_snapshot_clone_quota_exceeded
127
* https://tracker.ceph.com/issues/50821
128
  qa: untar_snap_rm failure during mds thrashing
129
130 88 Patrick Donnelly
h3. 2022 Sep 26
131
132
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220923.171109
133
134
* https://tracker.ceph.com/issues/55804
135
    qa failure: pjd link tests failed
136
* https://tracker.ceph.com/issues/57676
137
    qa: error during scrub thrashing: rank damage found: {'backtrace'}
138
* https://tracker.ceph.com/issues/52624
139
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
140
* https://tracker.ceph.com/issues/57580
141
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
142
* https://tracker.ceph.com/issues/48773
143
    qa: scrub does not complete
144
* https://tracker.ceph.com/issues/57299
145
    qa: test_dump_loads fails with JSONDecodeError
146
* https://tracker.ceph.com/issues/57280
147
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
148
* https://tracker.ceph.com/issues/57205
149
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
150
* https://tracker.ceph.com/issues/57656
151
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
152
* https://tracker.ceph.com/issues/57677
153
    qa: "1 MDSs behind on trimming (MDS_TRIM)"
154
* https://tracker.ceph.com/issues/57206
155
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
156
* https://tracker.ceph.com/issues/57446
157
    qa: test_subvolume_snapshot_info_if_orphan_clone fails
158
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
159
    qa: fs:mixed-clients kernel_untar_build failure
160 89 Patrick Donnelly
* https://tracker.ceph.com/issues/57682
161
    client: ERROR: test_reconnect_after_blocklisted
162 88 Patrick Donnelly
163
164 87 Patrick Donnelly
h3. 2022 Sep 22
165
166
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220920.234701
167
168
* https://tracker.ceph.com/issues/57299
169
    qa: test_dump_loads fails with JSONDecodeError
170
* https://tracker.ceph.com/issues/57205
171
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
172
* https://tracker.ceph.com/issues/52624
173
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
174
* https://tracker.ceph.com/issues/57580
175
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
176
* https://tracker.ceph.com/issues/57280
177
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
178
* https://tracker.ceph.com/issues/48773
179
    qa: scrub does not complete
180
* https://tracker.ceph.com/issues/56446
181
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
182
* https://tracker.ceph.com/issues/57206
183
    libcephfs/test.sh: ceph_test_libcephfs_reclaim
184
* https://tracker.ceph.com/issues/51267
185
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
186
187
NEW:
188
189
* https://tracker.ceph.com/issues/57656
190
    [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
191
* https://tracker.ceph.com/issues/57655 [Exist in main as well]
192
    qa: fs:mixed-clients kernel_untar_build failure
193
* https://tracker.ceph.com/issues/57657
194
    mds: scrub locates mismatch between child accounted_rstats and self rstats
195
196
Segfault probably caused by: https://github.com/ceph/ceph/pull/47795#issuecomment-1255724799
197
198
199 80 Venky Shankar
h3. 2022 Sep 16
200 79 Venky Shankar
201
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220905-132828
202
203
* https://tracker.ceph.com/issues/57446
204
    qa: test_subvolume_snapshot_info_if_orphan_clone fails
205
* https://tracker.ceph.com/issues/57299
206
    qa: test_dump_loads fails with JSONDecodeError
207
* https://tracker.ceph.com/issues/50223
208
    client.xxxx isn't responding to mclientcaps(revoke)
209
* https://tracker.ceph.com/issues/52624
210
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
211
* https://tracker.ceph.com/issues/57205
212
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
213
* https://tracker.ceph.com/issues/57280
214
    qa: tasks/kernel_cfuse_workunits_untarbuild_blogbench fails - Failed to fetch package version from shaman
215
* https://tracker.ceph.com/issues/51282
216
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
217
* https://tracker.ceph.com/issues/48203
218
  https://tracker.ceph.com/issues/36593
219
    qa: quota failure
220
    qa: quota failure caused by clients stepping on each other
221
* https://tracker.ceph.com/issues/57580
222
    Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.TestNewOps)
223
224 77 Rishabh Dave
225
h3. 2022 Aug 26
226 76 Rishabh Dave
227
http://pulpito.front.sepia.ceph.com/rishabh-2022-08-22_17:49:59-fs-wip-rishabh-testing-2022Aug19-testing-default-smithi/
228
http://pulpito.front.sepia.ceph.com/rishabh-2022-08-24_11:56:51-fs-wip-rishabh-testing-2022Aug19-testing-default-smithi/
229
230
* https://tracker.ceph.com/issues/57206
231
  libcephfs/test.sh: ceph_test_libcephfs_reclaim
232
* https://tracker.ceph.com/issues/56632
233
  Test failure: test_subvolume_snapshot_clone_quota_exceeded (tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones)
234
* https://tracker.ceph.com/issues/56446
235
  Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
236
* https://tracker.ceph.com/issues/51964
237
  qa: test_cephfs_mirror_restart_sync_on_blocklist failure
238
* https://tracker.ceph.com/issues/53859
239
  qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
240
241
* https://tracker.ceph.com/issues/54460
242
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
243
* https://tracker.ceph.com/issues/54462
244
  Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi055 with status 128
245
* https://tracker.ceph.com/issues/54460
246
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
247
* https://tracker.ceph.com/issues/36593
248
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
249
250
* https://tracker.ceph.com/issues/52624
251
  qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
252
* https://tracker.ceph.com/issues/55804
253
  Command failed (workunit test suites/pjd.sh)
254
* https://tracker.ceph.com/issues/50223
255
  client.xxxx isn't responding to mclientcaps(revoke)
256
257
258 75 Venky Shankar
h3. 2022 Aug 22
259
260
https://pulpito.ceph.com/vshankar-2022-08-12_09:34:24-fs-wip-vshankar-testing1-20220812-072441-testing-default-smithi/
261
https://pulpito.ceph.com/vshankar-2022-08-18_04:30:42-fs-wip-vshankar-testing1-20220818-082047-testing-default-smithi/ (drop problematic PR and re-run)
262
263
* https://tracker.ceph.com/issues/52624
264
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
265
* https://tracker.ceph.com/issues/56446
266
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
267
* https://tracker.ceph.com/issues/55804
268
    Command failed (workunit test suites/pjd.sh)
269
* https://tracker.ceph.com/issues/51278
270
    mds: "FAILED ceph_assert(!segments.empty())"
271
* https://tracker.ceph.com/issues/54460
272
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
273
* https://tracker.ceph.com/issues/57205
274
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
275
* https://tracker.ceph.com/issues/57206
276
    ceph_test_libcephfs_reclaim crashes during test
277
* https://tracker.ceph.com/issues/53859
278
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
279
* https://tracker.ceph.com/issues/50223
280
    client.xxxx isn't responding to mclientcaps(revoke)
281
282 72 Venky Shankar
h3. 2022 Aug 12
283
284
https://pulpito.ceph.com/vshankar-2022-08-10_04:06:00-fs-wip-vshankar-testing-20220805-190751-testing-default-smithi/
285
https://pulpito.ceph.com/vshankar-2022-08-11_12:16:58-fs-wip-vshankar-testing-20220811-145809-testing-default-smithi/ (drop problematic PR and re-run)
286
287
* https://tracker.ceph.com/issues/52624
288
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
289
* https://tracker.ceph.com/issues/56446
290
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
291
* https://tracker.ceph.com/issues/51964
292
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
293
* https://tracker.ceph.com/issues/55804
294
    Command failed (workunit test suites/pjd.sh)
295
* https://tracker.ceph.com/issues/50223
296
    client.xxxx isn't responding to mclientcaps(revoke)
297
* https://tracker.ceph.com/issues/50821
298
    qa: untar_snap_rm failure during mds thrashing
299
* https://tracker.ceph.com/issues/54460
300 73 Venky Shankar
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
301 72 Venky Shankar
302 71 Venky Shankar
h3. 2022 Aug 04
303
304
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220804-123835 (only mgr/volumes, mgr/stats)
305
306
Unrealted teuthology failure on rhel
307
308 69 Rishabh Dave
h3. 2022 Jul 25
309 68 Rishabh Dave
310
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-22_11:34:20-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
311
312
1st re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_03:51:19-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi
313
2nd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
314 74 Rishabh Dave
3rd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
315
4th (final) re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-28_03:59:01-fs-wip-rishabh-testing-2022Jul28-0143-testing-default-smithi/
316 68 Rishabh Dave
317
* https://tracker.ceph.com/issues/55804
318
  Command failed (workunit test suites/pjd.sh)
319
* https://tracker.ceph.com/issues/50223
320
  client.xxxx isn't responding to mclientcaps(revoke)
321
322
* https://tracker.ceph.com/issues/54460
323
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
324
* https://tracker.ceph.com/issues/36593
325
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
326 1 Patrick Donnelly
* https://tracker.ceph.com/issues/54462
327 74 Rishabh Dave
  Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi055 with status 128~
328 68 Rishabh Dave
329 67 Patrick Donnelly
h3. 2022 July 22
330
331
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220721.235756
332
333
MDS_HEALTH_DUMMY error in log fixed by followup commit.
334
transient selinux ping failure
335
336
* https://tracker.ceph.com/issues/56694
337
    qa: avoid blocking forever on hung umount
338
* https://tracker.ceph.com/issues/56695
339
    [RHEL stock] pjd test failures
340
* https://tracker.ceph.com/issues/56696
341
    admin keyring disappears during qa run
342
* https://tracker.ceph.com/issues/56697
343
    qa: fs/snaps fails for fuse
344
* https://tracker.ceph.com/issues/50222
345
    osd: 5.2s0 deep-scrub : stat mismatch
346
* https://tracker.ceph.com/issues/56698
347
    client: FAILED ceph_assert(_size == 0)
348
* https://tracker.ceph.com/issues/50223
349
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
350
351
352 66 Rishabh Dave
h3. 2022 Jul 15
353 65 Rishabh Dave
354
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-08_23:53:34-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
355
356
re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-15_06:42:04-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
357
358
* https://tracker.ceph.com/issues/53859
359
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
360
* https://tracker.ceph.com/issues/55804
361
  Command failed (workunit test suites/pjd.sh)
362
* https://tracker.ceph.com/issues/50223
363
  client.xxxx isn't responding to mclientcaps(revoke)
364
* https://tracker.ceph.com/issues/50222
365
  osd: deep-scrub : stat mismatch
366
367
* https://tracker.ceph.com/issues/56632
368
  Test failure: test_subvolume_snapshot_clone_quota_exceeded (tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones)
369
* https://tracker.ceph.com/issues/56634
370
  workunit test fs/snaps/snaptest-intodir.sh
371
* https://tracker.ceph.com/issues/56644
372
  Test failure: test_rapid_creation (tasks.cephfs.test_fragment.TestFragmentation)
373
374
375
376 61 Rishabh Dave
h3. 2022 July 05
377
378
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-02_14:14:52-fs-wip-rishabh-testing-20220702-1631-testing-default-smithi/
379 62 Rishabh Dave
380 64 Rishabh Dave
On 1st re-run some jobs passed - http://pulpito.front.sepia.ceph.com/rishabh-2022-07-03_15:10:28-fs-wip-rishabh-testing-20220702-1631-distro-default-smithi/
381
382
On 2nd re-run only few jobs failed -
383
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
384
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
385 62 Rishabh Dave
386
* https://tracker.ceph.com/issues/56446
387
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
388
* https://tracker.ceph.com/issues/55804
389
    Command failed (workunit test suites/pjd.sh) on smithi047 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/
390
391
* https://tracker.ceph.com/issues/56445
392
    Command failed on smithi080 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
393
* https://tracker.ceph.com/issues/51267
394 63 Rishabh Dave
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi098 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1
395
* https://tracker.ceph.com/issues/50224
396
    Test failure: test_mirroring_init_failure_with_recovery (tasks.cephfs.test_mirroring.TestMirroring)
397 62 Rishabh Dave
398
399 61 Rishabh Dave
400 58 Venky Shankar
h3. 2022 July 04
401
402
https://pulpito.ceph.com/vshankar-2022-06-29_09:19:00-fs-wip-vshankar-testing-20220627-100931-testing-default-smithi/
403
(rhel runs were borked due to: https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/JSZQFUKVLDND4W33PXDGCABPHNSPT6SS/, tests ran with --filter-out=rhel)
404
405
* https://tracker.ceph.com/issues/56445
406
    Command failed on smithi162 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
407
* https://tracker.ceph.com/issues/56446
408 59 Rishabh Dave
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
409
* https://tracker.ceph.com/issues/51964
410
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
411
* https://tracker.ceph.com/issues/52624
412 60 Rishabh Dave
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
413 59 Rishabh Dave
414 57 Venky Shankar
h3. 2022 June 20
415
416
https://pulpito.ceph.com/vshankar-2022-06-15_04:03:39-fs-wip-vshankar-testing1-20220615-072516-testing-default-smithi/
417
https://pulpito.ceph.com/vshankar-2022-06-19_08:22:46-fs-wip-vshankar-testing1-20220619-102531-testing-default-smithi/
418
419
* https://tracker.ceph.com/issues/52624
420
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
421
* https://tracker.ceph.com/issues/55804
422
    qa failure: pjd link tests failed
423
* https://tracker.ceph.com/issues/54108
424
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
425
* https://tracker.ceph.com/issues/55332
426
    Failure in snaptest-git-ceph.sh (it's an async unlink/create bug)
427
428 56 Patrick Donnelly
h3. 2022 June 13
429
430
https://pulpito.ceph.com/pdonnell-2022-06-12_05:08:12-fs:workload-wip-pdonnell-testing-20220612.004943-distro-default-smithi/
431
432
* https://tracker.ceph.com/issues/56024
433
    cephadm: removes ceph.conf during qa run causing command failure
434
* https://tracker.ceph.com/issues/48773
435
    qa: scrub does not complete
436
* https://tracker.ceph.com/issues/56012
437
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
438
439
440 55 Venky Shankar
h3. 2022 Jun 13
441 54 Venky Shankar
442
https://pulpito.ceph.com/vshankar-2022-06-07_00:25:50-fs-wip-vshankar-testing-20220606-223254-testing-default-smithi/
443
https://pulpito.ceph.com/vshankar-2022-06-10_01:04:46-fs-wip-vshankar-testing-20220609-175550-testing-default-smithi/
444
445
* https://tracker.ceph.com/issues/52624
446
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
447
* https://tracker.ceph.com/issues/51964
448
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
449
* https://tracker.ceph.com/issues/53859
450
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
451
* https://tracker.ceph.com/issues/55804
452
    qa failure: pjd link tests failed
453
* https://tracker.ceph.com/issues/56003
454
    client: src/include/xlist.h: 81: FAILED ceph_assert(_size == 0)
455
* https://tracker.ceph.com/issues/56011
456
    fs/thrash: snaptest-snap-rm-cmp.sh fails in mds5sum comparison
457
* https://tracker.ceph.com/issues/56012
458
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
459
460 53 Venky Shankar
h3. 2022 Jun 07
461
462
https://pulpito.ceph.com/vshankar-2022-06-06_21:25:41-fs-wip-vshankar-testing1-20220606-230129-testing-default-smithi/
463
https://pulpito.ceph.com/vshankar-2022-06-07_10:53:31-fs-wip-vshankar-testing1-20220607-104134-testing-default-smithi/ (rerun after dropping a problematic PR)
464
465
* https://tracker.ceph.com/issues/52624
466
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
467
* https://tracker.ceph.com/issues/50223
468
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
469
* https://tracker.ceph.com/issues/50224
470
    qa: test_mirroring_init_failure_with_recovery failure
471
472 51 Venky Shankar
h3. 2022 May 12
473
474
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220509-125847
475 52 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-13_17:09:16-fs-wip-vshankar-testing-20220513-120051-testing-default-smithi/ (drop prs + rerun)
476 51 Venky Shankar
477
* https://tracker.ceph.com/issues/52624
478
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
479
* https://tracker.ceph.com/issues/50223
480
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
481
* https://tracker.ceph.com/issues/55332
482
    Failure in snaptest-git-ceph.sh
483
* https://tracker.ceph.com/issues/53859
484
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
485
* https://tracker.ceph.com/issues/55538
486 1 Patrick Donnelly
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
487 52 Venky Shankar
* https://tracker.ceph.com/issues/55258
488
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs (cropss up again, though very infrequent)
489 51 Venky Shankar
490 49 Venky Shankar
h3. 2022 May 04
491
492 50 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-01_13:18:44-fs-wip-vshankar-testing1-20220428-204527-testing-default-smithi/
493
https://pulpito.ceph.com/vshankar-2022-05-02_16:58:59-fs-wip-vshankar-testing1-20220502-201957-testing-default-smithi/ (after dropping PRs)
494
495 49 Venky Shankar
* https://tracker.ceph.com/issues/52624
496
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
497
* https://tracker.ceph.com/issues/50223
498
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
499
* https://tracker.ceph.com/issues/55332
500
    Failure in snaptest-git-ceph.sh
501
* https://tracker.ceph.com/issues/53859
502
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
503
* https://tracker.ceph.com/issues/55516
504
    qa: fs suite tests failing with "json.decoder.JSONDecodeError: Extra data: line 2 column 82 (char 82)"
505
* https://tracker.ceph.com/issues/55537
506
    mds: crash during fs:upgrade test
507
* https://tracker.ceph.com/issues/55538
508
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
509
510 48 Venky Shankar
h3. 2022 Apr 25
511
512
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220420-113951 (owner vshankar)
513
514
* https://tracker.ceph.com/issues/52624
515
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
516
* https://tracker.ceph.com/issues/50223
517
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
518
* https://tracker.ceph.com/issues/55258
519
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
520
* https://tracker.ceph.com/issues/55377
521
    kclient: mds revoke Fwb caps stuck after the kclient tries writebcak once
522
523 47 Venky Shankar
h3. 2022 Apr 14
524
525
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220411-144044
526
527
* https://tracker.ceph.com/issues/52624
528
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
529
* https://tracker.ceph.com/issues/50223
530
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
531
* https://tracker.ceph.com/issues/52438
532
    qa: ffsb timeout
533
* https://tracker.ceph.com/issues/55170
534
    mds: crash during rejoin (CDir::fetch_keys)
535
* https://tracker.ceph.com/issues/55331
536
    pjd failure
537
* https://tracker.ceph.com/issues/48773
538
    qa: scrub does not complete
539
* https://tracker.ceph.com/issues/55332
540
    Failure in snaptest-git-ceph.sh
541
* https://tracker.ceph.com/issues/55258
542
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
543
544 45 Venky Shankar
h3. 2022 Apr 11
545
546 46 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-55110-20220408-203242
547 45 Venky Shankar
548
* https://tracker.ceph.com/issues/48773
549
    qa: scrub does not complete
550
* https://tracker.ceph.com/issues/52624
551
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
552
* https://tracker.ceph.com/issues/52438
553
    qa: ffsb timeout
554
* https://tracker.ceph.com/issues/48680
555
    mds: scrubbing stuck "scrub active (0 inodes in the stack)"
556
* https://tracker.ceph.com/issues/55236
557
    qa: fs/snaps tests fails with "hit max job timeout"
558
* https://tracker.ceph.com/issues/54108
559
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
560
* https://tracker.ceph.com/issues/54971
561
    Test failure: test_perf_stats_stale_metrics (tasks.cephfs.test_mds_metrics.TestMDSMetrics)
562
* https://tracker.ceph.com/issues/50223
563
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
564
* https://tracker.ceph.com/issues/55258
565
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
566
567 44 Venky Shankar
h3. 2022 Mar 21
568 42 Venky Shankar
569 43 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-03-20_02:16:37-fs-wip-vshankar-testing-20220319-163539-testing-default-smithi/
570
571
Run didn't go well, lots of failures - debugging by dropping PRs and running against master branch. Only merging unrelated PRs that pass tests.
572
573
574
h3. 2022 Mar 08
575
576 42 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-02-28_04:32:15-fs-wip-vshankar-testing-20220226-211550-testing-default-smithi/
577
578
rerun with
579
- (drop) https://github.com/ceph/ceph/pull/44679
580
- (drop) https://github.com/ceph/ceph/pull/44958
581
https://pulpito.ceph.com/vshankar-2022-03-06_14:47:51-fs-wip-vshankar-testing-20220304-132102-testing-default-smithi/
582
583
* https://tracker.ceph.com/issues/54419 (new)
584
    `ceph orch upgrade start` seems to never reach completion
585
* https://tracker.ceph.com/issues/51964
586
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
587
* https://tracker.ceph.com/issues/52624
588
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
589
* https://tracker.ceph.com/issues/50223
590
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
591
* https://tracker.ceph.com/issues/52438
592
    qa: ffsb timeout
593
* https://tracker.ceph.com/issues/50821
594
    qa: untar_snap_rm failure during mds thrashing
595
596
597 41 Venky Shankar
h3. 2022 Feb 09
598
599
https://pulpito.ceph.com/vshankar-2022-02-05_17:27:49-fs-wip-vshankar-testing-20220201-113815-testing-default-smithi/
600
601
rerun with
602
- (drop) https://github.com/ceph/ceph/pull/37938
603
- (drop) https://github.com/ceph/ceph/pull/44335
604
- (drop) https://github.com/ceph/ceph/pull/44491
605
- (drop) https://github.com/ceph/ceph/pull/44501
606
https://pulpito.ceph.com/vshankar-2022-02-08_14:27:29-fs-wip-vshankar-testing-20220208-181241-testing-default-smithi/
607
608
* https://tracker.ceph.com/issues/51964
609
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
610
* https://tracker.ceph.com/issues/54066
611
    test_subvolume_no_upgrade_v1_sanity fails with `AssertionError: 1000 != 0`
612
* https://tracker.ceph.com/issues/48773
613
    qa: scrub does not complete
614
* https://tracker.ceph.com/issues/52624
615
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
616
* https://tracker.ceph.com/issues/50223
617
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
618
* https://tracker.ceph.com/issues/52438
619
    qa: ffsb timeout
620
621 40 Patrick Donnelly
h3. 2022 Feb 01
622
623
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220127.171526
624
625
* https://tracker.ceph.com/issues/54107
626
    kclient: hang during umount
627
* https://tracker.ceph.com/issues/54106
628
    kclient: hang during workunit cleanup
629
* https://tracker.ceph.com/issues/54108
630
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
631
* https://tracker.ceph.com/issues/48773
632
    qa: scrub does not complete
633
* https://tracker.ceph.com/issues/52438
634
    qa: ffsb timeout
635
636
637 36 Venky Shankar
h3. 2022 Jan 13
638
639
https://pulpito.ceph.com/vshankar-2022-01-06_13:18:41-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
640 39 Venky Shankar
641 36 Venky Shankar
rerun with:
642 38 Venky Shankar
- (add) https://github.com/ceph/ceph/pull/44570
643
- (drop) https://github.com/ceph/ceph/pull/43184
644 36 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-01-13_04:42:40-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
645
646
* https://tracker.ceph.com/issues/50223
647
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
648
* https://tracker.ceph.com/issues/51282
649
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
650
* https://tracker.ceph.com/issues/48773
651
    qa: scrub does not complete
652
* https://tracker.ceph.com/issues/52624
653
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
654
* https://tracker.ceph.com/issues/53859
655
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
656
657 34 Venky Shankar
h3. 2022 Jan 03
658
659
https://pulpito.ceph.com/vshankar-2021-12-22_07:37:44-fs-wip-vshankar-testing-20211216-114012-testing-default-smithi/
660
https://pulpito.ceph.com/vshankar-2022-01-03_12:27:45-fs-wip-vshankar-testing-20220103-142738-testing-default-smithi/ (rerun)
661
662
* https://tracker.ceph.com/issues/50223
663
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
664
* https://tracker.ceph.com/issues/51964
665
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
666
* https://tracker.ceph.com/issues/51267
667
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
668
* https://tracker.ceph.com/issues/51282
669
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
670
* https://tracker.ceph.com/issues/50821
671
    qa: untar_snap_rm failure during mds thrashing
672
* https://tracker.ceph.com/issues/51278
673
    mds: "FAILED ceph_assert(!segments.empty())"
674 35 Ramana Raja
* https://tracker.ceph.com/issues/52279
675
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
676
677 34 Venky Shankar
678 33 Patrick Donnelly
h3. 2021 Dec 22
679
680
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211222.014316
681
682
* https://tracker.ceph.com/issues/52624
683
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
684
* https://tracker.ceph.com/issues/50223
685
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
686
* https://tracker.ceph.com/issues/52279
687
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
688
* https://tracker.ceph.com/issues/50224
689
    qa: test_mirroring_init_failure_with_recovery failure
690
* https://tracker.ceph.com/issues/48773
691
    qa: scrub does not complete
692
693
694 32 Venky Shankar
h3. 2021 Nov 30
695
696
https://pulpito.ceph.com/vshankar-2021-11-24_07:14:27-fs-wip-vshankar-testing-20211124-094330-testing-default-smithi/
697
https://pulpito.ceph.com/vshankar-2021-11-30_06:23:32-fs-wip-vshankar-testing-20211124-094330-distro-default-smithi/ (rerun w/ QA fixes)
698
699
* https://tracker.ceph.com/issues/53436
700
    mds, mon: mds beacon messages get dropped? (mds never reaches up:active state)
701
* https://tracker.ceph.com/issues/51964
702
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
703
* https://tracker.ceph.com/issues/48812
704
    qa: test_scrub_pause_and_resume_with_abort failure
705
* https://tracker.ceph.com/issues/51076
706
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
707
* https://tracker.ceph.com/issues/50223
708
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
709
* https://tracker.ceph.com/issues/52624
710
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
711
* https://tracker.ceph.com/issues/50250
712
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
713
714
715 31 Patrick Donnelly
h3. 2021 November 9
716
717
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211109.180315
718
719
* https://tracker.ceph.com/issues/53214
720
    qa: "dd: error reading '/sys/kernel/debug/ceph/2a934501-6731-4052-a836-f42229a869be.client4874/metrics': Is a directory"
721
* https://tracker.ceph.com/issues/48773
722
    qa: scrub does not complete
723
* https://tracker.ceph.com/issues/50223
724
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
725
* https://tracker.ceph.com/issues/51282
726
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
727
* https://tracker.ceph.com/issues/52624
728
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
729
* https://tracker.ceph.com/issues/53216
730
    qa: "RuntimeError: value of attributes should be either str or None. client_id"
731
* https://tracker.ceph.com/issues/50250
732
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
733
734
735
736 30 Patrick Donnelly
h3. 2021 November 03
737
738
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211103.023355
739
740
* https://tracker.ceph.com/issues/51964
741
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
742
* https://tracker.ceph.com/issues/51282
743
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
744
* https://tracker.ceph.com/issues/52436
745
    fs/ceph: "corrupt mdsmap"
746
* https://tracker.ceph.com/issues/53074
747
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
748
* https://tracker.ceph.com/issues/53150
749
    pybind/mgr/cephadm/upgrade: tolerate MDS failures during upgrade straddling v16.2.5
750
* https://tracker.ceph.com/issues/53155
751
    MDSMonitor: assertion during upgrade to v16.2.5+
752
753
754 29 Patrick Donnelly
h3. 2021 October 26
755
756
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211025.000447
757
758
* https://tracker.ceph.com/issues/53074
759
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
760
* https://tracker.ceph.com/issues/52997
761
    testing: hang ing umount
762
* https://tracker.ceph.com/issues/50824
763
    qa: snaptest-git-ceph bus error
764
* https://tracker.ceph.com/issues/52436
765
    fs/ceph: "corrupt mdsmap"
766
* https://tracker.ceph.com/issues/48773
767
    qa: scrub does not complete
768
* https://tracker.ceph.com/issues/53082
769
    ceph-fuse: segmenetation fault in Client::handle_mds_map
770
* https://tracker.ceph.com/issues/50223
771
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
772
* https://tracker.ceph.com/issues/52624
773
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
774
* https://tracker.ceph.com/issues/50224
775
    qa: test_mirroring_init_failure_with_recovery failure
776
* https://tracker.ceph.com/issues/50821
777
    qa: untar_snap_rm failure during mds thrashing
778
* https://tracker.ceph.com/issues/50250
779
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
780
781
782
783 27 Patrick Donnelly
h3. 2021 October 19
784
785 28 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211019.013028
786 27 Patrick Donnelly
787
* https://tracker.ceph.com/issues/52995
788
    qa: test_standby_count_wanted failure
789
* https://tracker.ceph.com/issues/52948
790
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
791
* https://tracker.ceph.com/issues/52996
792
    qa: test_perf_counters via test_openfiletable
793
* https://tracker.ceph.com/issues/48772
794
    qa: pjd: not ok 9, 44, 80
795
* https://tracker.ceph.com/issues/52997
796
    testing: hang ing umount
797
* https://tracker.ceph.com/issues/50250
798
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
799
* https://tracker.ceph.com/issues/52624
800
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
801
* https://tracker.ceph.com/issues/50223
802
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
803
* https://tracker.ceph.com/issues/50821
804
    qa: untar_snap_rm failure during mds thrashing
805
* https://tracker.ceph.com/issues/48773
806
    qa: scrub does not complete
807
808
809 26 Patrick Donnelly
h3. 2021 October 12
810
811
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211012.192211
812
813
Some failures caused by teuthology bug: https://tracker.ceph.com/issues/52944
814
815
New test caused failure: https://github.com/ceph/ceph/pull/43297#discussion_r729883167
816
817
818
* https://tracker.ceph.com/issues/51282
819
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
820
* https://tracker.ceph.com/issues/52948
821
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
822
* https://tracker.ceph.com/issues/48773
823
    qa: scrub does not complete
824
* https://tracker.ceph.com/issues/50224
825
    qa: test_mirroring_init_failure_with_recovery failure
826
* https://tracker.ceph.com/issues/52949
827
    RuntimeError: The following counters failed to be set on mds daemons: {'mds.dir_split'}
828
829
830 25 Patrick Donnelly
h3. 2021 October 02
831 23 Patrick Donnelly
832 24 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211002.163337
833
834
Some failures caused by cephadm upgrade test. Fixed in follow-up qa commit.
835
836
test_simple failures caused by PR in this set.
837
838
A few reruns because of QA infra noise.
839
840
* https://tracker.ceph.com/issues/52822
841
    qa: failed pacific install on fs:upgrade
842
* https://tracker.ceph.com/issues/52624
843
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
844
* https://tracker.ceph.com/issues/50223
845
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
846
* https://tracker.ceph.com/issues/48773
847
    qa: scrub does not complete
848
849
850
h3. 2021 September 20
851
852 23 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210917.174826
853
854
* https://tracker.ceph.com/issues/52677
855
    qa: test_simple failure
856
* https://tracker.ceph.com/issues/51279
857
    kclient hangs on umount (testing branch)
858
* https://tracker.ceph.com/issues/50223
859
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
860
* https://tracker.ceph.com/issues/50250
861
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
862
* https://tracker.ceph.com/issues/52624
863
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
864
* https://tracker.ceph.com/issues/52438
865
    qa: ffsb timeout
866
867
868 22 Patrick Donnelly
h3. 2021 September 10
869
870
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210910.181451
871
872
* https://tracker.ceph.com/issues/50223
873
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
874
* https://tracker.ceph.com/issues/50250
875
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
876
* https://tracker.ceph.com/issues/52624
877
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
878
* https://tracker.ceph.com/issues/52625
879
    qa: test_kill_mdstable (tasks.cephfs.test_snapshots.TestSnapshots)
880
* https://tracker.ceph.com/issues/52439
881
    qa: acls does not compile on centos stream
882
* https://tracker.ceph.com/issues/50821
883
    qa: untar_snap_rm failure during mds thrashing
884
* https://tracker.ceph.com/issues/48773
885
    qa: scrub does not complete
886
* https://tracker.ceph.com/issues/52626
887
    mds: ScrubStack.cc: 831: FAILED ceph_assert(diri)
888
* https://tracker.ceph.com/issues/51279
889
    kclient hangs on umount (testing branch)
890
891
892 21 Patrick Donnelly
h3. 2021 August 27
893
894
Several jobs died because of device failures.
895
896
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210827.024746
897
898
* https://tracker.ceph.com/issues/52430
899
    mds: fast async create client mount breaks racy test
900
* https://tracker.ceph.com/issues/52436
901
    fs/ceph: "corrupt mdsmap"
902
* https://tracker.ceph.com/issues/52437
903
    mds: InoTable::replay_release_ids abort via test_inotable_sync
904
* https://tracker.ceph.com/issues/51282
905
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
906
* https://tracker.ceph.com/issues/52438
907
    qa: ffsb timeout
908
* https://tracker.ceph.com/issues/52439
909
    qa: acls does not compile on centos stream
910
911
912 20 Patrick Donnelly
h3. 2021 July 30
913
914
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210729.214022
915
916
* https://tracker.ceph.com/issues/50250
917
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
918
* https://tracker.ceph.com/issues/51282
919
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
920
* https://tracker.ceph.com/issues/48773
921
    qa: scrub does not complete
922
* https://tracker.ceph.com/issues/51975
923
    pybind/mgr/stats: KeyError
924
925
926 19 Patrick Donnelly
h3. 2021 July 28
927
928
https://pulpito.ceph.com/pdonnell-2021-07-28_00:39:45-fs-wip-pdonnell-testing-20210727.213757-distro-basic-smithi/
929
930
with qa fix: https://pulpito.ceph.com/pdonnell-2021-07-28_16:20:28-fs-wip-pdonnell-testing-20210728.141004-distro-basic-smithi/
931
932
* https://tracker.ceph.com/issues/51905
933
    qa: "error reading sessionmap 'mds1_sessionmap'"
934
* https://tracker.ceph.com/issues/48773
935
    qa: scrub does not complete
936
* https://tracker.ceph.com/issues/50250
937
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
938
* https://tracker.ceph.com/issues/51267
939
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
940
* https://tracker.ceph.com/issues/51279
941
    kclient hangs on umount (testing branch)
942
943
944 18 Patrick Donnelly
h3. 2021 July 16
945
946
https://pulpito.ceph.com/pdonnell-2021-07-16_05:50:11-fs-wip-pdonnell-testing-20210716.022804-distro-basic-smithi/
947
948
* https://tracker.ceph.com/issues/48773
949
    qa: scrub does not complete
950
* https://tracker.ceph.com/issues/48772
951
    qa: pjd: not ok 9, 44, 80
952
* https://tracker.ceph.com/issues/45434
953
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
954
* https://tracker.ceph.com/issues/51279
955
    kclient hangs on umount (testing branch)
956
* https://tracker.ceph.com/issues/50824
957
    qa: snaptest-git-ceph bus error
958
959
960 17 Patrick Donnelly
h3. 2021 July 04
961
962
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210703.052904
963
964
* https://tracker.ceph.com/issues/48773
965
    qa: scrub does not complete
966
* https://tracker.ceph.com/issues/39150
967
    mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
968
* https://tracker.ceph.com/issues/45434
969
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
970
* https://tracker.ceph.com/issues/51282
971
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
972
* https://tracker.ceph.com/issues/48771
973
    qa: iogen: workload fails to cause balancing
974
* https://tracker.ceph.com/issues/51279
975
    kclient hangs on umount (testing branch)
976
* https://tracker.ceph.com/issues/50250
977
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
978
979
980 16 Patrick Donnelly
h3. 2021 July 01
981
982
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210701.192056
983
984
* https://tracker.ceph.com/issues/51197
985
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
986
* https://tracker.ceph.com/issues/50866
987
    osd: stat mismatch on objects
988
* https://tracker.ceph.com/issues/48773
989
    qa: scrub does not complete
990
991
992 15 Patrick Donnelly
h3. 2021 June 26
993
994
https://pulpito.ceph.com/pdonnell-2021-06-26_00:57:00-fs-wip-pdonnell-testing-20210625.225421-distro-basic-smithi/
995
996
* https://tracker.ceph.com/issues/51183
997
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
998
* https://tracker.ceph.com/issues/51410
999
    kclient: fails to finish reconnect during MDS thrashing (testing branch)
1000
* https://tracker.ceph.com/issues/48773
1001
    qa: scrub does not complete
1002
* https://tracker.ceph.com/issues/51282
1003
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1004
* https://tracker.ceph.com/issues/51169
1005
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1006
* https://tracker.ceph.com/issues/48772
1007
    qa: pjd: not ok 9, 44, 80
1008
1009
1010 14 Patrick Donnelly
h3. 2021 June 21
1011
1012
https://pulpito.ceph.com/pdonnell-2021-06-22_00:27:21-fs-wip-pdonnell-testing-20210621.231646-distro-basic-smithi/
1013
1014
One failure caused by PR: https://github.com/ceph/ceph/pull/41935#issuecomment-866472599
1015
1016
* https://tracker.ceph.com/issues/51282
1017
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1018
* https://tracker.ceph.com/issues/51183
1019
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1020
* https://tracker.ceph.com/issues/48773
1021
    qa: scrub does not complete
1022
* https://tracker.ceph.com/issues/48771
1023
    qa: iogen: workload fails to cause balancing
1024
* https://tracker.ceph.com/issues/51169
1025
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1026
* https://tracker.ceph.com/issues/50495
1027
    libcephfs: shutdown race fails with status 141
1028
* https://tracker.ceph.com/issues/45434
1029
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1030
* https://tracker.ceph.com/issues/50824
1031
    qa: snaptest-git-ceph bus error
1032
* https://tracker.ceph.com/issues/50223
1033
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1034
1035
1036 13 Patrick Donnelly
h3. 2021 June 16
1037
1038
https://pulpito.ceph.com/pdonnell-2021-06-16_21:26:55-fs-wip-pdonnell-testing-20210616.191804-distro-basic-smithi/
1039
1040
MDS abort class of failures caused by PR: https://github.com/ceph/ceph/pull/41667
1041
1042
* https://tracker.ceph.com/issues/45434
1043
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1044
* https://tracker.ceph.com/issues/51169
1045
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1046
* https://tracker.ceph.com/issues/43216
1047
    MDSMonitor: removes MDS coming out of quorum election
1048
* https://tracker.ceph.com/issues/51278
1049
    mds: "FAILED ceph_assert(!segments.empty())"
1050
* https://tracker.ceph.com/issues/51279
1051
    kclient hangs on umount (testing branch)
1052
* https://tracker.ceph.com/issues/51280
1053
    mds: "FAILED ceph_assert(r == 0 || r == -2)"
1054
* https://tracker.ceph.com/issues/51183
1055
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1056
* https://tracker.ceph.com/issues/51281
1057
    qa: snaptest-snap-rm-cmp.sh: "echo 'FAIL: bad match, /tmp/a 4637e766853d1ad16a7b17079e2c6f03 != real c3883760b18d50e8d78819c54d579b00'"
1058
* https://tracker.ceph.com/issues/48773
1059
    qa: scrub does not complete
1060
* https://tracker.ceph.com/issues/51076
1061
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
1062
* https://tracker.ceph.com/issues/51228
1063
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
1064
* https://tracker.ceph.com/issues/51282
1065
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
1066
1067
1068 12 Patrick Donnelly
h3. 2021 June 14
1069
1070
https://pulpito.ceph.com/pdonnell-2021-06-14_20:53:05-fs-wip-pdonnell-testing-20210614.173325-distro-basic-smithi/
1071
1072
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1073
1074
* https://tracker.ceph.com/issues/51169
1075
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1076
* https://tracker.ceph.com/issues/51228
1077
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
1078
* https://tracker.ceph.com/issues/48773
1079
    qa: scrub does not complete
1080
* https://tracker.ceph.com/issues/51183
1081
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1082
* https://tracker.ceph.com/issues/45434
1083
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1084
* https://tracker.ceph.com/issues/51182
1085
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1086
* https://tracker.ceph.com/issues/51229
1087
    qa: test_multi_snap_schedule list difference failure
1088
* https://tracker.ceph.com/issues/50821
1089
    qa: untar_snap_rm failure during mds thrashing
1090
1091
1092 11 Patrick Donnelly
h3. 2021 June 13
1093
1094
https://pulpito.ceph.com/pdonnell-2021-06-12_02:45:35-fs-wip-pdonnell-testing-20210612.002809-distro-basic-smithi/
1095
1096
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1097
1098
* https://tracker.ceph.com/issues/51169
1099
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1100
* https://tracker.ceph.com/issues/48773
1101
    qa: scrub does not complete
1102
* https://tracker.ceph.com/issues/51182
1103
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1104
* https://tracker.ceph.com/issues/51183
1105
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1106
* https://tracker.ceph.com/issues/51197
1107
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
1108
* https://tracker.ceph.com/issues/45434
1109
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1110
1111 10 Patrick Donnelly
h3. 2021 June 11
1112
1113
https://pulpito.ceph.com/pdonnell-2021-06-11_18:02:10-fs-wip-pdonnell-testing-20210611.162716-distro-basic-smithi/
1114
1115
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
1116
1117
* https://tracker.ceph.com/issues/51169
1118
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
1119
* https://tracker.ceph.com/issues/45434
1120
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1121
* https://tracker.ceph.com/issues/48771
1122
    qa: iogen: workload fails to cause balancing
1123
* https://tracker.ceph.com/issues/43216
1124
    MDSMonitor: removes MDS coming out of quorum election
1125
* https://tracker.ceph.com/issues/51182
1126
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
1127
* https://tracker.ceph.com/issues/50223
1128
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1129
* https://tracker.ceph.com/issues/48773
1130
    qa: scrub does not complete
1131
* https://tracker.ceph.com/issues/51183
1132
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
1133
* https://tracker.ceph.com/issues/51184
1134
    qa: fs:bugs does not specify distro
1135
1136
1137 9 Patrick Donnelly
h3. 2021 June 03
1138
1139
https://pulpito.ceph.com/pdonnell-2021-06-03_03:40:33-fs-wip-pdonnell-testing-20210603.020013-distro-basic-smithi/
1140
1141
* https://tracker.ceph.com/issues/45434
1142
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1143
* https://tracker.ceph.com/issues/50016
1144
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1145
* https://tracker.ceph.com/issues/50821
1146
    qa: untar_snap_rm failure during mds thrashing
1147
* https://tracker.ceph.com/issues/50622 (regression)
1148
    msg: active_connections regression
1149
* https://tracker.ceph.com/issues/49845#note-2 (regression)
1150
    qa: failed umount in test_volumes
1151
* https://tracker.ceph.com/issues/48773
1152
    qa: scrub does not complete
1153
* https://tracker.ceph.com/issues/43216
1154
    MDSMonitor: removes MDS coming out of quorum election
1155
1156
1157 7 Patrick Donnelly
h3. 2021 May 18
1158
1159 8 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.214114
1160
1161
Regression in testing kernel caused some failures. Ilya fixed those and rerun
1162
looked better. Some odd new noise in the rerun relating to packaging and "No
1163
module named 'tasks.ceph'".
1164
1165
* https://tracker.ceph.com/issues/50824
1166
    qa: snaptest-git-ceph bus error
1167
* https://tracker.ceph.com/issues/50622 (regression)
1168
    msg: active_connections regression
1169
* https://tracker.ceph.com/issues/49845#note-2 (regression)
1170
    qa: failed umount in test_volumes
1171
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1172
    qa: quota failure
1173
1174
1175
h3. 2021 May 18
1176
1177 7 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.025642
1178
1179
* https://tracker.ceph.com/issues/50821
1180
    qa: untar_snap_rm failure during mds thrashing
1181
* https://tracker.ceph.com/issues/48773
1182
    qa: scrub does not complete
1183
* https://tracker.ceph.com/issues/45591
1184
    mgr: FAILED ceph_assert(daemon != nullptr)
1185
* https://tracker.ceph.com/issues/50866
1186
    osd: stat mismatch on objects
1187
* https://tracker.ceph.com/issues/50016
1188
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1189
* https://tracker.ceph.com/issues/50867
1190
    qa: fs:mirror: reduced data availability
1191
* https://tracker.ceph.com/issues/50821
1192
    qa: untar_snap_rm failure during mds thrashing
1193
* https://tracker.ceph.com/issues/50622 (regression)
1194
    msg: active_connections regression
1195
* https://tracker.ceph.com/issues/50223
1196
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1197
* https://tracker.ceph.com/issues/50868
1198
    qa: "kern.log.gz already exists; not overwritten"
1199
* https://tracker.ceph.com/issues/50870
1200
    qa: test_full: "rm: cannot remove 'large_file_a': Permission denied"
1201
1202
1203 6 Patrick Donnelly
h3. 2021 May 11
1204
1205
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210511.232042
1206
1207
* one class of failures caused by PR
1208
* https://tracker.ceph.com/issues/48812
1209
    qa: test_scrub_pause_and_resume_with_abort failure
1210
* https://tracker.ceph.com/issues/50390
1211
    mds: monclient: wait_auth_rotating timed out after 30
1212
* https://tracker.ceph.com/issues/48773
1213
    qa: scrub does not complete
1214
* https://tracker.ceph.com/issues/50821
1215
    qa: untar_snap_rm failure during mds thrashing
1216
* https://tracker.ceph.com/issues/50224
1217
    qa: test_mirroring_init_failure_with_recovery failure
1218
* https://tracker.ceph.com/issues/50622 (regression)
1219
    msg: active_connections regression
1220
* https://tracker.ceph.com/issues/50825
1221
    qa: snaptest-git-ceph hang during mon thrashing v2
1222
* https://tracker.ceph.com/issues/50821
1223
    qa: untar_snap_rm failure during mds thrashing
1224
* https://tracker.ceph.com/issues/50823
1225
    qa: RuntimeError: timeout waiting for cluster to stabilize
1226
1227
1228 5 Patrick Donnelly
h3. 2021 May 14
1229
1230
https://pulpito.ceph.com/pdonnell-2021-05-14_21:45:42-fs-master-distro-basic-smithi/
1231
1232
* https://tracker.ceph.com/issues/48812
1233
    qa: test_scrub_pause_and_resume_with_abort failure
1234
* https://tracker.ceph.com/issues/50821
1235
    qa: untar_snap_rm failure during mds thrashing
1236
* https://tracker.ceph.com/issues/50622 (regression)
1237
    msg: active_connections regression
1238
* https://tracker.ceph.com/issues/50822
1239
    qa: testing kernel patch for client metrics causes mds abort
1240
* https://tracker.ceph.com/issues/48773
1241
    qa: scrub does not complete
1242
* https://tracker.ceph.com/issues/50823
1243
    qa: RuntimeError: timeout waiting for cluster to stabilize
1244
* https://tracker.ceph.com/issues/50824
1245
    qa: snaptest-git-ceph bus error
1246
* https://tracker.ceph.com/issues/50825
1247
    qa: snaptest-git-ceph hang during mon thrashing v2
1248
* https://tracker.ceph.com/issues/50826
1249
    kceph: stock RHEL kernel hangs on snaptests with mon|osd thrashers
1250
1251
1252 4 Patrick Donnelly
h3. 2021 May 01
1253
1254
https://pulpito.ceph.com/pdonnell-2021-05-01_09:07:09-fs-wip-pdonnell-testing-20210501.040415-distro-basic-smithi/
1255
1256
* https://tracker.ceph.com/issues/45434
1257
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1258
* https://tracker.ceph.com/issues/50281
1259
    qa: untar_snap_rm timeout
1260
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1261
    qa: quota failure
1262
* https://tracker.ceph.com/issues/48773
1263
    qa: scrub does not complete
1264
* https://tracker.ceph.com/issues/50390
1265
    mds: monclient: wait_auth_rotating timed out after 30
1266
* https://tracker.ceph.com/issues/50250
1267
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1268
* https://tracker.ceph.com/issues/50622 (regression)
1269
    msg: active_connections regression
1270
* https://tracker.ceph.com/issues/45591
1271
    mgr: FAILED ceph_assert(daemon != nullptr)
1272
* https://tracker.ceph.com/issues/50221
1273
    qa: snaptest-git-ceph failure in git diff
1274
* https://tracker.ceph.com/issues/50016
1275
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1276
1277
1278 3 Patrick Donnelly
h3. 2021 Apr 15
1279
1280
https://pulpito.ceph.com/pdonnell-2021-04-15_01:35:57-fs-wip-pdonnell-testing-20210414.230315-distro-basic-smithi/
1281
1282
* https://tracker.ceph.com/issues/50281
1283
    qa: untar_snap_rm timeout
1284
* https://tracker.ceph.com/issues/50220
1285
    qa: dbench workload timeout
1286
* https://tracker.ceph.com/issues/50246
1287
    mds: failure replaying journal (EMetaBlob)
1288
* https://tracker.ceph.com/issues/50250
1289
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1290
* https://tracker.ceph.com/issues/50016
1291
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1292
* https://tracker.ceph.com/issues/50222
1293
    osd: 5.2s0 deep-scrub : stat mismatch
1294
* https://tracker.ceph.com/issues/45434
1295
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1296
* https://tracker.ceph.com/issues/49845
1297
    qa: failed umount in test_volumes
1298
* https://tracker.ceph.com/issues/37808
1299
    osd: osdmap cache weak_refs assert during shutdown
1300
* https://tracker.ceph.com/issues/50387
1301
    client: fs/snaps failure
1302
* https://tracker.ceph.com/issues/50389
1303
    mds: "cluster [ERR] Error recovering journal 0x203: (2) No such file or directory" in cluster log"
1304
* https://tracker.ceph.com/issues/50216
1305
    qa: "ls: cannot access 'lost+found': No such file or directory"
1306
* https://tracker.ceph.com/issues/50390
1307
    mds: monclient: wait_auth_rotating timed out after 30
1308
1309
1310
1311 1 Patrick Donnelly
h3. 2021 Apr 08
1312
1313 2 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_22:42:24-fs-wip-pdonnell-testing-20210408.192301-distro-basic-smithi/
1314
1315
* https://tracker.ceph.com/issues/45434
1316
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1317
* https://tracker.ceph.com/issues/50016
1318
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1319
* https://tracker.ceph.com/issues/48773
1320
    qa: scrub does not complete
1321
* https://tracker.ceph.com/issues/50279
1322
    qa: "Replacing daemon mds.b as rank 0 with standby daemon mds.c"
1323
* https://tracker.ceph.com/issues/50246
1324
    mds: failure replaying journal (EMetaBlob)
1325
* https://tracker.ceph.com/issues/48365
1326
    qa: ffsb build failure on CentOS 8.2
1327
* https://tracker.ceph.com/issues/50216
1328
    qa: "ls: cannot access 'lost+found': No such file or directory"
1329
* https://tracker.ceph.com/issues/50223
1330
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1331
* https://tracker.ceph.com/issues/50280
1332
    cephadm: RuntimeError: uid/gid not found
1333
* https://tracker.ceph.com/issues/50281
1334
    qa: untar_snap_rm timeout
1335
1336
h3. 2021 Apr 08
1337
1338 1 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_04:31:36-fs-wip-pdonnell-testing-20210408.024225-distro-basic-smithi/
1339
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210408.142238 (with logic inversion / QA fix)
1340
1341
* https://tracker.ceph.com/issues/50246
1342
    mds: failure replaying journal (EMetaBlob)
1343
* https://tracker.ceph.com/issues/50250
1344
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1345
1346
1347
h3. 2021 Apr 07
1348
1349
https://pulpito.ceph.com/pdonnell-2021-04-07_02:12:41-fs-wip-pdonnell-testing-20210406.213012-distro-basic-smithi/
1350
1351
* https://tracker.ceph.com/issues/50215
1352
    qa: "log [ERR] : error reading sessionmap 'mds2_sessionmap'"
1353
* https://tracker.ceph.com/issues/49466
1354
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1355
* https://tracker.ceph.com/issues/50216
1356
    qa: "ls: cannot access 'lost+found': No such file or directory"
1357
* https://tracker.ceph.com/issues/48773
1358
    qa: scrub does not complete
1359
* https://tracker.ceph.com/issues/49845
1360
    qa: failed umount in test_volumes
1361
* https://tracker.ceph.com/issues/50220
1362
    qa: dbench workload timeout
1363
* https://tracker.ceph.com/issues/50221
1364
    qa: snaptest-git-ceph failure in git diff
1365
* https://tracker.ceph.com/issues/50222
1366
    osd: 5.2s0 deep-scrub : stat mismatch
1367
* https://tracker.ceph.com/issues/50223
1368
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1369
* https://tracker.ceph.com/issues/50224
1370
    qa: test_mirroring_init_failure_with_recovery failure
1371
1372
h3. 2021 Apr 01
1373
1374
https://pulpito.ceph.com/pdonnell-2021-04-01_00:45:34-fs-wip-pdonnell-testing-20210331.222326-distro-basic-smithi/
1375
1376
* https://tracker.ceph.com/issues/48772
1377
    qa: pjd: not ok 9, 44, 80
1378
* https://tracker.ceph.com/issues/50177
1379
    osd: "stalled aio... buggy kernel or bad device?"
1380
* https://tracker.ceph.com/issues/48771
1381
    qa: iogen: workload fails to cause balancing
1382
* https://tracker.ceph.com/issues/49845
1383
    qa: failed umount in test_volumes
1384
* https://tracker.ceph.com/issues/48773
1385
    qa: scrub does not complete
1386
* https://tracker.ceph.com/issues/48805
1387
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1388
* https://tracker.ceph.com/issues/50178
1389
    qa: "TypeError: run() got an unexpected keyword argument 'shell'"
1390
* https://tracker.ceph.com/issues/45434
1391
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1392
1393
h3. 2021 Mar 24
1394
1395
https://pulpito.ceph.com/pdonnell-2021-03-24_23:26:35-fs-wip-pdonnell-testing-20210324.190252-distro-basic-smithi/
1396
1397
* https://tracker.ceph.com/issues/49500
1398
    qa: "Assertion `cb_done' failed."
1399
* https://tracker.ceph.com/issues/50019
1400
    qa: mount failure with cephadm "probably no MDS server is up?"
1401
* https://tracker.ceph.com/issues/50020
1402
    qa: "RADOS object not found (Failed to operate read op for oid cephfs_mirror)"
1403
* https://tracker.ceph.com/issues/48773
1404
    qa: scrub does not complete
1405
* https://tracker.ceph.com/issues/45434
1406
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1407
* https://tracker.ceph.com/issues/48805
1408
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1409
* https://tracker.ceph.com/issues/48772
1410
    qa: pjd: not ok 9, 44, 80
1411
* https://tracker.ceph.com/issues/50021
1412
    qa: snaptest-git-ceph failure during mon thrashing
1413
* https://tracker.ceph.com/issues/48771
1414
    qa: iogen: workload fails to cause balancing
1415
* https://tracker.ceph.com/issues/50016
1416
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1417
* https://tracker.ceph.com/issues/49466
1418
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1419
1420
1421
h3. 2021 Mar 18
1422
1423
https://pulpito.ceph.com/pdonnell-2021-03-18_13:46:31-fs-wip-pdonnell-testing-20210318.024145-distro-basic-smithi/
1424
1425
* https://tracker.ceph.com/issues/49466
1426
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1427
* https://tracker.ceph.com/issues/48773
1428
    qa: scrub does not complete
1429
* https://tracker.ceph.com/issues/48805
1430
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1431
* https://tracker.ceph.com/issues/45434
1432
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1433
* https://tracker.ceph.com/issues/49845
1434
    qa: failed umount in test_volumes
1435
* https://tracker.ceph.com/issues/49605
1436
    mgr: drops command on the floor
1437
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1438
    qa: quota failure
1439
* https://tracker.ceph.com/issues/49928
1440
    client: items pinned in cache preventing unmount x2
1441
1442
h3. 2021 Mar 15
1443
1444
https://pulpito.ceph.com/pdonnell-2021-03-15_22:16:56-fs-wip-pdonnell-testing-20210315.182203-distro-basic-smithi/
1445
1446
* https://tracker.ceph.com/issues/49842
1447
    qa: stuck pkg install
1448
* https://tracker.ceph.com/issues/49466
1449
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1450
* https://tracker.ceph.com/issues/49822
1451
    test: test_mirroring_command_idempotency (tasks.cephfs.test_admin.TestMirroringCommands) failure
1452
* https://tracker.ceph.com/issues/49240
1453
    terminate called after throwing an instance of 'std::bad_alloc'
1454
* https://tracker.ceph.com/issues/48773
1455
    qa: scrub does not complete
1456
* https://tracker.ceph.com/issues/45434
1457
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1458
* https://tracker.ceph.com/issues/49500
1459
    qa: "Assertion `cb_done' failed."
1460
* https://tracker.ceph.com/issues/49843
1461
    qa: fs/snaps/snaptest-upchildrealms.sh failure
1462
* https://tracker.ceph.com/issues/49845
1463
    qa: failed umount in test_volumes
1464
* https://tracker.ceph.com/issues/48805
1465
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1466
* https://tracker.ceph.com/issues/49605
1467
    mgr: drops command on the floor
1468
1469
and failure caused by PR: https://github.com/ceph/ceph/pull/39969
1470
1471
1472
h3. 2021 Mar 09
1473
1474
https://pulpito.ceph.com/pdonnell-2021-03-09_03:27:39-fs-wip-pdonnell-testing-20210308.214827-distro-basic-smithi/
1475
1476
* https://tracker.ceph.com/issues/49500
1477
    qa: "Assertion `cb_done' failed."
1478
* https://tracker.ceph.com/issues/48805
1479
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1480
* https://tracker.ceph.com/issues/48773
1481
    qa: scrub does not complete
1482
* https://tracker.ceph.com/issues/45434
1483
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1484
* https://tracker.ceph.com/issues/49240
1485
    terminate called after throwing an instance of 'std::bad_alloc'
1486
* https://tracker.ceph.com/issues/49466
1487
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1488
* https://tracker.ceph.com/issues/49684
1489
    qa: fs:cephadm mount does not wait for mds to be created
1490
* https://tracker.ceph.com/issues/48771
1491
    qa: iogen: workload fails to cause balancing