Project

General

Profile

Main » History » Version 75

Venky Shankar, 08/23/2022 07:11 AM

1 75 Venky Shankar
h3. 2022 Aug 22
2
3
https://pulpito.ceph.com/vshankar-2022-08-12_09:34:24-fs-wip-vshankar-testing1-20220812-072441-testing-default-smithi/
4
https://pulpito.ceph.com/vshankar-2022-08-18_04:30:42-fs-wip-vshankar-testing1-20220818-082047-testing-default-smithi/ (drop problematic PR and re-run)
5
6
* https://tracker.ceph.com/issues/52624
7
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
8
* https://tracker.ceph.com/issues/56446
9
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
10
* https://tracker.ceph.com/issues/55804
11
    Command failed (workunit test suites/pjd.sh)
12
* https://tracker.ceph.com/issues/51278
13
    mds: "FAILED ceph_assert(!segments.empty())"
14
* https://tracker.ceph.com/issues/54460
15
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
16
* https://tracker.ceph.com/issues/57205
17
    Test failure: test_subvolume_group_ls_filter_internal_directories (tasks.cephfs.test_volumes.TestSubvolumeGroups)
18
* https://tracker.ceph.com/issues/57206
19
    ceph_test_libcephfs_reclaim crashes during test
20
* https://tracker.ceph.com/issues/53859
21
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
22
* https://tracker.ceph.com/issues/50223
23
    client.xxxx isn't responding to mclientcaps(revoke)
24
25 72 Venky Shankar
h3. 2022 Aug 12
26
27
https://pulpito.ceph.com/vshankar-2022-08-10_04:06:00-fs-wip-vshankar-testing-20220805-190751-testing-default-smithi/
28
https://pulpito.ceph.com/vshankar-2022-08-11_12:16:58-fs-wip-vshankar-testing-20220811-145809-testing-default-smithi/ (drop problematic PR and re-run)
29
30
* https://tracker.ceph.com/issues/52624
31
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
32
* https://tracker.ceph.com/issues/56446
33
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
34
* https://tracker.ceph.com/issues/51964
35
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
36
* https://tracker.ceph.com/issues/55804
37
    Command failed (workunit test suites/pjd.sh)
38
* https://tracker.ceph.com/issues/50223
39
    client.xxxx isn't responding to mclientcaps(revoke)
40
* https://tracker.ceph.com/issues/50821
41
    qa: untar_snap_rm failure during mds thrashing
42
* https://tracker.ceph.com/issues/54460
43 73 Venky Shankar
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
44 72 Venky Shankar
45 71 Venky Shankar
h3. 2022 Aug 04
46
47
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220804-123835 (only mgr/volumes, mgr/stats)
48
49
Unrealted teuthology failure on rhel
50
51 69 Rishabh Dave
h3. 2022 Jul 25
52 68 Rishabh Dave
53
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-22_11:34:20-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
54
55
1st re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_03:51:19-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi
56
2nd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
57 74 Rishabh Dave
3rd re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-24_08:53:36-fs-wip-rishabh-testing-2022Jul22-1400-testing-default-smithi/
58
4th (final) re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-28_03:59:01-fs-wip-rishabh-testing-2022Jul28-0143-testing-default-smithi/
59 68 Rishabh Dave
60
* https://tracker.ceph.com/issues/55804
61
  Command failed (workunit test suites/pjd.sh)
62
* https://tracker.ceph.com/issues/50223
63
  client.xxxx isn't responding to mclientcaps(revoke)
64
65
* https://tracker.ceph.com/issues/54460
66
  Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithixxx with status 1
67
* https://tracker.ceph.com/issues/36593
68
  Command failed (workunit test fs/quota/quota.sh) on smithixxx with status 1
69 1 Patrick Donnelly
* https://tracker.ceph.com/issues/54462
70 74 Rishabh Dave
  Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on smithi055 with status 128~
71 68 Rishabh Dave
72 67 Patrick Donnelly
h3. 2022 July 22
73
74
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220721.235756
75
76
MDS_HEALTH_DUMMY error in log fixed by followup commit.
77
transient selinux ping failure
78
79
* https://tracker.ceph.com/issues/56694
80
    qa: avoid blocking forever on hung umount
81
* https://tracker.ceph.com/issues/56695
82
    [RHEL stock] pjd test failures
83
* https://tracker.ceph.com/issues/56696
84
    admin keyring disappears during qa run
85
* https://tracker.ceph.com/issues/56697
86
    qa: fs/snaps fails for fuse
87
* https://tracker.ceph.com/issues/50222
88
    osd: 5.2s0 deep-scrub : stat mismatch
89
* https://tracker.ceph.com/issues/56698
90
    client: FAILED ceph_assert(_size == 0)
91
* https://tracker.ceph.com/issues/50223
92
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
93
94
95 66 Rishabh Dave
h3. 2022 Jul 15
96 65 Rishabh Dave
97
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-08_23:53:34-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
98
99
re-run: http://pulpito.front.sepia.ceph.com/rishabh-2022-07-15_06:42:04-fs-wip-rishabh-testing-2022Jul08-1820-testing-default-smithi/
100
101
* https://tracker.ceph.com/issues/53859
102
  Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
103
* https://tracker.ceph.com/issues/55804
104
  Command failed (workunit test suites/pjd.sh)
105
* https://tracker.ceph.com/issues/50223
106
  client.xxxx isn't responding to mclientcaps(revoke)
107
* https://tracker.ceph.com/issues/50222
108
  osd: deep-scrub : stat mismatch
109
110
* https://tracker.ceph.com/issues/56632
111
  Test failure: test_subvolume_snapshot_clone_quota_exceeded (tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones)
112
* https://tracker.ceph.com/issues/56634
113
  workunit test fs/snaps/snaptest-intodir.sh
114
* https://tracker.ceph.com/issues/56644
115
  Test failure: test_rapid_creation (tasks.cephfs.test_fragment.TestFragmentation)
116
117
118
119 61 Rishabh Dave
h3. 2022 July 05
120
121
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-02_14:14:52-fs-wip-rishabh-testing-20220702-1631-testing-default-smithi/
122 62 Rishabh Dave
123 64 Rishabh Dave
On 1st re-run some jobs passed - http://pulpito.front.sepia.ceph.com/rishabh-2022-07-03_15:10:28-fs-wip-rishabh-testing-20220702-1631-distro-default-smithi/
124
125
On 2nd re-run only few jobs failed -
126
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
127
http://pulpito.front.sepia.ceph.com/rishabh-2022-07-06_05:24:29-fs-wip-rishabh-testing-20220705-2132-distro-default-smithi/
128 62 Rishabh Dave
129
* https://tracker.ceph.com/issues/56446
130
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
131
* https://tracker.ceph.com/issues/55804
132
    Command failed (workunit test suites/pjd.sh) on smithi047 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/
133
134
* https://tracker.ceph.com/issues/56445
135
    Command failed on smithi080 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
136
* https://tracker.ceph.com/issues/51267
137 63 Rishabh Dave
    Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi098 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1
138
* https://tracker.ceph.com/issues/50224
139
    Test failure: test_mirroring_init_failure_with_recovery (tasks.cephfs.test_mirroring.TestMirroring)
140 62 Rishabh Dave
141
142 61 Rishabh Dave
143 58 Venky Shankar
h3. 2022 July 04
144
145
https://pulpito.ceph.com/vshankar-2022-06-29_09:19:00-fs-wip-vshankar-testing-20220627-100931-testing-default-smithi/
146
(rhel runs were borked due to: https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/JSZQFUKVLDND4W33PXDGCABPHNSPT6SS/, tests ran with --filter-out=rhel)
147
148
* https://tracker.ceph.com/issues/56445
149
    Command failed on smithi162 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
150
* https://tracker.ceph.com/issues/56446
151 59 Rishabh Dave
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
152
* https://tracker.ceph.com/issues/51964
153
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
154
* https://tracker.ceph.com/issues/52624
155 60 Rishabh Dave
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
156 59 Rishabh Dave
157 57 Venky Shankar
h3. 2022 June 20
158
159
https://pulpito.ceph.com/vshankar-2022-06-15_04:03:39-fs-wip-vshankar-testing1-20220615-072516-testing-default-smithi/
160
https://pulpito.ceph.com/vshankar-2022-06-19_08:22:46-fs-wip-vshankar-testing1-20220619-102531-testing-default-smithi/
161
162
* https://tracker.ceph.com/issues/52624
163
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
164
* https://tracker.ceph.com/issues/55804
165
    qa failure: pjd link tests failed
166
* https://tracker.ceph.com/issues/54108
167
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
168
* https://tracker.ceph.com/issues/55332
169
    Failure in snaptest-git-ceph.sh (it's an async unlink/create bug)
170
171 56 Patrick Donnelly
h3. 2022 June 13
172
173
https://pulpito.ceph.com/pdonnell-2022-06-12_05:08:12-fs:workload-wip-pdonnell-testing-20220612.004943-distro-default-smithi/
174
175
* https://tracker.ceph.com/issues/56024
176
    cephadm: removes ceph.conf during qa run causing command failure
177
* https://tracker.ceph.com/issues/48773
178
    qa: scrub does not complete
179
* https://tracker.ceph.com/issues/56012
180
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
181
182
183 55 Venky Shankar
h3. 2022 Jun 13
184 54 Venky Shankar
185
https://pulpito.ceph.com/vshankar-2022-06-07_00:25:50-fs-wip-vshankar-testing-20220606-223254-testing-default-smithi/
186
https://pulpito.ceph.com/vshankar-2022-06-10_01:04:46-fs-wip-vshankar-testing-20220609-175550-testing-default-smithi/
187
188
* https://tracker.ceph.com/issues/52624
189
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
190
* https://tracker.ceph.com/issues/51964
191
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
192
* https://tracker.ceph.com/issues/53859
193
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
194
* https://tracker.ceph.com/issues/55804
195
    qa failure: pjd link tests failed
196
* https://tracker.ceph.com/issues/56003
197
    client: src/include/xlist.h: 81: FAILED ceph_assert(_size == 0)
198
* https://tracker.ceph.com/issues/56011
199
    fs/thrash: snaptest-snap-rm-cmp.sh fails in mds5sum comparison
200
* https://tracker.ceph.com/issues/56012
201
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
202
203 53 Venky Shankar
h3. 2022 Jun 07
204
205
https://pulpito.ceph.com/vshankar-2022-06-06_21:25:41-fs-wip-vshankar-testing1-20220606-230129-testing-default-smithi/
206
https://pulpito.ceph.com/vshankar-2022-06-07_10:53:31-fs-wip-vshankar-testing1-20220607-104134-testing-default-smithi/ (rerun after dropping a problematic PR)
207
208
* https://tracker.ceph.com/issues/52624
209
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
210
* https://tracker.ceph.com/issues/50223
211
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
212
* https://tracker.ceph.com/issues/50224
213
    qa: test_mirroring_init_failure_with_recovery failure
214
215 51 Venky Shankar
h3. 2022 May 12
216
217
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220509-125847
218 52 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-13_17:09:16-fs-wip-vshankar-testing-20220513-120051-testing-default-smithi/ (drop prs + rerun)
219 51 Venky Shankar
220
* https://tracker.ceph.com/issues/52624
221
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
222
* https://tracker.ceph.com/issues/50223
223
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
224
* https://tracker.ceph.com/issues/55332
225
    Failure in snaptest-git-ceph.sh
226
* https://tracker.ceph.com/issues/53859
227
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
228
* https://tracker.ceph.com/issues/55538
229 1 Patrick Donnelly
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
230 52 Venky Shankar
* https://tracker.ceph.com/issues/55258
231
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs (cropss up again, though very infrequent)
232 51 Venky Shankar
233 49 Venky Shankar
h3. 2022 May 04
234
235 50 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-01_13:18:44-fs-wip-vshankar-testing1-20220428-204527-testing-default-smithi/
236
https://pulpito.ceph.com/vshankar-2022-05-02_16:58:59-fs-wip-vshankar-testing1-20220502-201957-testing-default-smithi/ (after dropping PRs)
237
238 49 Venky Shankar
* https://tracker.ceph.com/issues/52624
239
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
240
* https://tracker.ceph.com/issues/50223
241
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
242
* https://tracker.ceph.com/issues/55332
243
    Failure in snaptest-git-ceph.sh
244
* https://tracker.ceph.com/issues/53859
245
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
246
* https://tracker.ceph.com/issues/55516
247
    qa: fs suite tests failing with "json.decoder.JSONDecodeError: Extra data: line 2 column 82 (char 82)"
248
* https://tracker.ceph.com/issues/55537
249
    mds: crash during fs:upgrade test
250
* https://tracker.ceph.com/issues/55538
251
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
252
253 48 Venky Shankar
h3. 2022 Apr 25
254
255
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220420-113951 (owner vshankar)
256
257
* https://tracker.ceph.com/issues/52624
258
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
259
* https://tracker.ceph.com/issues/50223
260
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
261
* https://tracker.ceph.com/issues/55258
262
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
263
* https://tracker.ceph.com/issues/55377
264
    kclient: mds revoke Fwb caps stuck after the kclient tries writebcak once
265
266 47 Venky Shankar
h3. 2022 Apr 14
267
268
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220411-144044
269
270
* https://tracker.ceph.com/issues/52624
271
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
272
* https://tracker.ceph.com/issues/50223
273
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
274
* https://tracker.ceph.com/issues/52438
275
    qa: ffsb timeout
276
* https://tracker.ceph.com/issues/55170
277
    mds: crash during rejoin (CDir::fetch_keys)
278
* https://tracker.ceph.com/issues/55331
279
    pjd failure
280
* https://tracker.ceph.com/issues/48773
281
    qa: scrub does not complete
282
* https://tracker.ceph.com/issues/55332
283
    Failure in snaptest-git-ceph.sh
284
* https://tracker.ceph.com/issues/55258
285
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
286
287 45 Venky Shankar
h3. 2022 Apr 11
288
289 46 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-55110-20220408-203242
290 45 Venky Shankar
291
* https://tracker.ceph.com/issues/48773
292
    qa: scrub does not complete
293
* https://tracker.ceph.com/issues/52624
294
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
295
* https://tracker.ceph.com/issues/52438
296
    qa: ffsb timeout
297
* https://tracker.ceph.com/issues/48680
298
    mds: scrubbing stuck "scrub active (0 inodes in the stack)"
299
* https://tracker.ceph.com/issues/55236
300
    qa: fs/snaps tests fails with "hit max job timeout"
301
* https://tracker.ceph.com/issues/54108
302
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
303
* https://tracker.ceph.com/issues/54971
304
    Test failure: test_perf_stats_stale_metrics (tasks.cephfs.test_mds_metrics.TestMDSMetrics)
305
* https://tracker.ceph.com/issues/50223
306
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
307
* https://tracker.ceph.com/issues/55258
308
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
309
310 44 Venky Shankar
h3. 2022 Mar 21
311 42 Venky Shankar
312 43 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-03-20_02:16:37-fs-wip-vshankar-testing-20220319-163539-testing-default-smithi/
313
314
Run didn't go well, lots of failures - debugging by dropping PRs and running against master branch. Only merging unrelated PRs that pass tests.
315
316
317
h3. 2022 Mar 08
318
319 42 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-02-28_04:32:15-fs-wip-vshankar-testing-20220226-211550-testing-default-smithi/
320
321
rerun with
322
- (drop) https://github.com/ceph/ceph/pull/44679
323
- (drop) https://github.com/ceph/ceph/pull/44958
324
https://pulpito.ceph.com/vshankar-2022-03-06_14:47:51-fs-wip-vshankar-testing-20220304-132102-testing-default-smithi/
325
326
* https://tracker.ceph.com/issues/54419 (new)
327
    `ceph orch upgrade start` seems to never reach completion
328
* https://tracker.ceph.com/issues/51964
329
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
330
* https://tracker.ceph.com/issues/52624
331
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
332
* https://tracker.ceph.com/issues/50223
333
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
334
* https://tracker.ceph.com/issues/52438
335
    qa: ffsb timeout
336
* https://tracker.ceph.com/issues/50821
337
    qa: untar_snap_rm failure during mds thrashing
338
339
340 41 Venky Shankar
h3. 2022 Feb 09
341
342
https://pulpito.ceph.com/vshankar-2022-02-05_17:27:49-fs-wip-vshankar-testing-20220201-113815-testing-default-smithi/
343
344
rerun with
345
- (drop) https://github.com/ceph/ceph/pull/37938
346
- (drop) https://github.com/ceph/ceph/pull/44335
347
- (drop) https://github.com/ceph/ceph/pull/44491
348
- (drop) https://github.com/ceph/ceph/pull/44501
349
https://pulpito.ceph.com/vshankar-2022-02-08_14:27:29-fs-wip-vshankar-testing-20220208-181241-testing-default-smithi/
350
351
* https://tracker.ceph.com/issues/51964
352
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
353
* https://tracker.ceph.com/issues/54066
354
    test_subvolume_no_upgrade_v1_sanity fails with `AssertionError: 1000 != 0`
355
* https://tracker.ceph.com/issues/48773
356
    qa: scrub does not complete
357
* https://tracker.ceph.com/issues/52624
358
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
359
* https://tracker.ceph.com/issues/50223
360
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
361
* https://tracker.ceph.com/issues/52438
362
    qa: ffsb timeout
363
364 40 Patrick Donnelly
h3. 2022 Feb 01
365
366
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220127.171526
367
368
* https://tracker.ceph.com/issues/54107
369
    kclient: hang during umount
370
* https://tracker.ceph.com/issues/54106
371
    kclient: hang during workunit cleanup
372
* https://tracker.ceph.com/issues/54108
373
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
374
* https://tracker.ceph.com/issues/48773
375
    qa: scrub does not complete
376
* https://tracker.ceph.com/issues/52438
377
    qa: ffsb timeout
378
379
380 36 Venky Shankar
h3. 2022 Jan 13
381
382
https://pulpito.ceph.com/vshankar-2022-01-06_13:18:41-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
383 39 Venky Shankar
384 36 Venky Shankar
rerun with:
385 38 Venky Shankar
- (add) https://github.com/ceph/ceph/pull/44570
386
- (drop) https://github.com/ceph/ceph/pull/43184
387 36 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-01-13_04:42:40-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
388
389
* https://tracker.ceph.com/issues/50223
390
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
391
* https://tracker.ceph.com/issues/51282
392
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
393
* https://tracker.ceph.com/issues/48773
394
    qa: scrub does not complete
395
* https://tracker.ceph.com/issues/52624
396
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
397
* https://tracker.ceph.com/issues/53859
398
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
399
400 34 Venky Shankar
h3. 2022 Jan 03
401
402
https://pulpito.ceph.com/vshankar-2021-12-22_07:37:44-fs-wip-vshankar-testing-20211216-114012-testing-default-smithi/
403
https://pulpito.ceph.com/vshankar-2022-01-03_12:27:45-fs-wip-vshankar-testing-20220103-142738-testing-default-smithi/ (rerun)
404
405
* https://tracker.ceph.com/issues/50223
406
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
407
* https://tracker.ceph.com/issues/51964
408
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
409
* https://tracker.ceph.com/issues/51267
410
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
411
* https://tracker.ceph.com/issues/51282
412
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
413
* https://tracker.ceph.com/issues/50821
414
    qa: untar_snap_rm failure during mds thrashing
415
* https://tracker.ceph.com/issues/51278
416
    mds: "FAILED ceph_assert(!segments.empty())"
417 35 Ramana Raja
* https://tracker.ceph.com/issues/52279
418
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
419
420 34 Venky Shankar
421 33 Patrick Donnelly
h3. 2021 Dec 22
422
423
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211222.014316
424
425
* https://tracker.ceph.com/issues/52624
426
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
427
* https://tracker.ceph.com/issues/50223
428
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
429
* https://tracker.ceph.com/issues/52279
430
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
431
* https://tracker.ceph.com/issues/50224
432
    qa: test_mirroring_init_failure_with_recovery failure
433
* https://tracker.ceph.com/issues/48773
434
    qa: scrub does not complete
435
436
437 32 Venky Shankar
h3. 2021 Nov 30
438
439
https://pulpito.ceph.com/vshankar-2021-11-24_07:14:27-fs-wip-vshankar-testing-20211124-094330-testing-default-smithi/
440
https://pulpito.ceph.com/vshankar-2021-11-30_06:23:32-fs-wip-vshankar-testing-20211124-094330-distro-default-smithi/ (rerun w/ QA fixes)
441
442
* https://tracker.ceph.com/issues/53436
443
    mds, mon: mds beacon messages get dropped? (mds never reaches up:active state)
444
* https://tracker.ceph.com/issues/51964
445
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
446
* https://tracker.ceph.com/issues/48812
447
    qa: test_scrub_pause_and_resume_with_abort failure
448
* https://tracker.ceph.com/issues/51076
449
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
450
* https://tracker.ceph.com/issues/50223
451
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
452
* https://tracker.ceph.com/issues/52624
453
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
454
* https://tracker.ceph.com/issues/50250
455
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
456
457
458 31 Patrick Donnelly
h3. 2021 November 9
459
460
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211109.180315
461
462
* https://tracker.ceph.com/issues/53214
463
    qa: "dd: error reading '/sys/kernel/debug/ceph/2a934501-6731-4052-a836-f42229a869be.client4874/metrics': Is a directory"
464
* https://tracker.ceph.com/issues/48773
465
    qa: scrub does not complete
466
* https://tracker.ceph.com/issues/50223
467
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
468
* https://tracker.ceph.com/issues/51282
469
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
470
* https://tracker.ceph.com/issues/52624
471
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
472
* https://tracker.ceph.com/issues/53216
473
    qa: "RuntimeError: value of attributes should be either str or None. client_id"
474
* https://tracker.ceph.com/issues/50250
475
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
476
477
478
479 30 Patrick Donnelly
h3. 2021 November 03
480
481
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211103.023355
482
483
* https://tracker.ceph.com/issues/51964
484
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
485
* https://tracker.ceph.com/issues/51282
486
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
487
* https://tracker.ceph.com/issues/52436
488
    fs/ceph: "corrupt mdsmap"
489
* https://tracker.ceph.com/issues/53074
490
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
491
* https://tracker.ceph.com/issues/53150
492
    pybind/mgr/cephadm/upgrade: tolerate MDS failures during upgrade straddling v16.2.5
493
* https://tracker.ceph.com/issues/53155
494
    MDSMonitor: assertion during upgrade to v16.2.5+
495
496
497 29 Patrick Donnelly
h3. 2021 October 26
498
499
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211025.000447
500
501
* https://tracker.ceph.com/issues/53074
502
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
503
* https://tracker.ceph.com/issues/52997
504
    testing: hang ing umount
505
* https://tracker.ceph.com/issues/50824
506
    qa: snaptest-git-ceph bus error
507
* https://tracker.ceph.com/issues/52436
508
    fs/ceph: "corrupt mdsmap"
509
* https://tracker.ceph.com/issues/48773
510
    qa: scrub does not complete
511
* https://tracker.ceph.com/issues/53082
512
    ceph-fuse: segmenetation fault in Client::handle_mds_map
513
* https://tracker.ceph.com/issues/50223
514
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
515
* https://tracker.ceph.com/issues/52624
516
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
517
* https://tracker.ceph.com/issues/50224
518
    qa: test_mirroring_init_failure_with_recovery failure
519
* https://tracker.ceph.com/issues/50821
520
    qa: untar_snap_rm failure during mds thrashing
521
* https://tracker.ceph.com/issues/50250
522
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
523
524
525
526 27 Patrick Donnelly
h3. 2021 October 19
527
528 28 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211019.013028
529 27 Patrick Donnelly
530
* https://tracker.ceph.com/issues/52995
531
    qa: test_standby_count_wanted failure
532
* https://tracker.ceph.com/issues/52948
533
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
534
* https://tracker.ceph.com/issues/52996
535
    qa: test_perf_counters via test_openfiletable
536
* https://tracker.ceph.com/issues/48772
537
    qa: pjd: not ok 9, 44, 80
538
* https://tracker.ceph.com/issues/52997
539
    testing: hang ing umount
540
* https://tracker.ceph.com/issues/50250
541
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
542
* https://tracker.ceph.com/issues/52624
543
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
544
* https://tracker.ceph.com/issues/50223
545
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
546
* https://tracker.ceph.com/issues/50821
547
    qa: untar_snap_rm failure during mds thrashing
548
* https://tracker.ceph.com/issues/48773
549
    qa: scrub does not complete
550
551
552 26 Patrick Donnelly
h3. 2021 October 12
553
554
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211012.192211
555
556
Some failures caused by teuthology bug: https://tracker.ceph.com/issues/52944
557
558
New test caused failure: https://github.com/ceph/ceph/pull/43297#discussion_r729883167
559
560
561
* https://tracker.ceph.com/issues/51282
562
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
563
* https://tracker.ceph.com/issues/52948
564
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
565
* https://tracker.ceph.com/issues/48773
566
    qa: scrub does not complete
567
* https://tracker.ceph.com/issues/50224
568
    qa: test_mirroring_init_failure_with_recovery failure
569
* https://tracker.ceph.com/issues/52949
570
    RuntimeError: The following counters failed to be set on mds daemons: {'mds.dir_split'}
571
572
573 25 Patrick Donnelly
h3. 2021 October 02
574 23 Patrick Donnelly
575 24 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211002.163337
576
577
Some failures caused by cephadm upgrade test. Fixed in follow-up qa commit.
578
579
test_simple failures caused by PR in this set.
580
581
A few reruns because of QA infra noise.
582
583
* https://tracker.ceph.com/issues/52822
584
    qa: failed pacific install on fs:upgrade
585
* https://tracker.ceph.com/issues/52624
586
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
587
* https://tracker.ceph.com/issues/50223
588
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
589
* https://tracker.ceph.com/issues/48773
590
    qa: scrub does not complete
591
592
593
h3. 2021 September 20
594
595 23 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210917.174826
596
597
* https://tracker.ceph.com/issues/52677
598
    qa: test_simple failure
599
* https://tracker.ceph.com/issues/51279
600
    kclient hangs on umount (testing branch)
601
* https://tracker.ceph.com/issues/50223
602
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
603
* https://tracker.ceph.com/issues/50250
604
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
605
* https://tracker.ceph.com/issues/52624
606
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
607
* https://tracker.ceph.com/issues/52438
608
    qa: ffsb timeout
609
610
611 22 Patrick Donnelly
h3. 2021 September 10
612
613
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210910.181451
614
615
* https://tracker.ceph.com/issues/50223
616
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
617
* https://tracker.ceph.com/issues/50250
618
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
619
* https://tracker.ceph.com/issues/52624
620
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
621
* https://tracker.ceph.com/issues/52625
622
    qa: test_kill_mdstable (tasks.cephfs.test_snapshots.TestSnapshots)
623
* https://tracker.ceph.com/issues/52439
624
    qa: acls does not compile on centos stream
625
* https://tracker.ceph.com/issues/50821
626
    qa: untar_snap_rm failure during mds thrashing
627
* https://tracker.ceph.com/issues/48773
628
    qa: scrub does not complete
629
* https://tracker.ceph.com/issues/52626
630
    mds: ScrubStack.cc: 831: FAILED ceph_assert(diri)
631
* https://tracker.ceph.com/issues/51279
632
    kclient hangs on umount (testing branch)
633
634
635 21 Patrick Donnelly
h3. 2021 August 27
636
637
Several jobs died because of device failures.
638
639
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210827.024746
640
641
* https://tracker.ceph.com/issues/52430
642
    mds: fast async create client mount breaks racy test
643
* https://tracker.ceph.com/issues/52436
644
    fs/ceph: "corrupt mdsmap"
645
* https://tracker.ceph.com/issues/52437
646
    mds: InoTable::replay_release_ids abort via test_inotable_sync
647
* https://tracker.ceph.com/issues/51282
648
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
649
* https://tracker.ceph.com/issues/52438
650
    qa: ffsb timeout
651
* https://tracker.ceph.com/issues/52439
652
    qa: acls does not compile on centos stream
653
654
655 20 Patrick Donnelly
h3. 2021 July 30
656
657
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210729.214022
658
659
* https://tracker.ceph.com/issues/50250
660
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
661
* https://tracker.ceph.com/issues/51282
662
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
663
* https://tracker.ceph.com/issues/48773
664
    qa: scrub does not complete
665
* https://tracker.ceph.com/issues/51975
666
    pybind/mgr/stats: KeyError
667
668
669 19 Patrick Donnelly
h3. 2021 July 28
670
671
https://pulpito.ceph.com/pdonnell-2021-07-28_00:39:45-fs-wip-pdonnell-testing-20210727.213757-distro-basic-smithi/
672
673
with qa fix: https://pulpito.ceph.com/pdonnell-2021-07-28_16:20:28-fs-wip-pdonnell-testing-20210728.141004-distro-basic-smithi/
674
675
* https://tracker.ceph.com/issues/51905
676
    qa: "error reading sessionmap 'mds1_sessionmap'"
677
* https://tracker.ceph.com/issues/48773
678
    qa: scrub does not complete
679
* https://tracker.ceph.com/issues/50250
680
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
681
* https://tracker.ceph.com/issues/51267
682
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
683
* https://tracker.ceph.com/issues/51279
684
    kclient hangs on umount (testing branch)
685
686
687 18 Patrick Donnelly
h3. 2021 July 16
688
689
https://pulpito.ceph.com/pdonnell-2021-07-16_05:50:11-fs-wip-pdonnell-testing-20210716.022804-distro-basic-smithi/
690
691
* https://tracker.ceph.com/issues/48773
692
    qa: scrub does not complete
693
* https://tracker.ceph.com/issues/48772
694
    qa: pjd: not ok 9, 44, 80
695
* https://tracker.ceph.com/issues/45434
696
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
697
* https://tracker.ceph.com/issues/51279
698
    kclient hangs on umount (testing branch)
699
* https://tracker.ceph.com/issues/50824
700
    qa: snaptest-git-ceph bus error
701
702
703 17 Patrick Donnelly
h3. 2021 July 04
704
705
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210703.052904
706
707
* https://tracker.ceph.com/issues/48773
708
    qa: scrub does not complete
709
* https://tracker.ceph.com/issues/39150
710
    mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
711
* https://tracker.ceph.com/issues/45434
712
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
713
* https://tracker.ceph.com/issues/51282
714
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
715
* https://tracker.ceph.com/issues/48771
716
    qa: iogen: workload fails to cause balancing
717
* https://tracker.ceph.com/issues/51279
718
    kclient hangs on umount (testing branch)
719
* https://tracker.ceph.com/issues/50250
720
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
721
722
723 16 Patrick Donnelly
h3. 2021 July 01
724
725
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210701.192056
726
727
* https://tracker.ceph.com/issues/51197
728
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
729
* https://tracker.ceph.com/issues/50866
730
    osd: stat mismatch on objects
731
* https://tracker.ceph.com/issues/48773
732
    qa: scrub does not complete
733
734
735 15 Patrick Donnelly
h3. 2021 June 26
736
737
https://pulpito.ceph.com/pdonnell-2021-06-26_00:57:00-fs-wip-pdonnell-testing-20210625.225421-distro-basic-smithi/
738
739
* https://tracker.ceph.com/issues/51183
740
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
741
* https://tracker.ceph.com/issues/51410
742
    kclient: fails to finish reconnect during MDS thrashing (testing branch)
743
* https://tracker.ceph.com/issues/48773
744
    qa: scrub does not complete
745
* https://tracker.ceph.com/issues/51282
746
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
747
* https://tracker.ceph.com/issues/51169
748
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
749
* https://tracker.ceph.com/issues/48772
750
    qa: pjd: not ok 9, 44, 80
751
752
753 14 Patrick Donnelly
h3. 2021 June 21
754
755
https://pulpito.ceph.com/pdonnell-2021-06-22_00:27:21-fs-wip-pdonnell-testing-20210621.231646-distro-basic-smithi/
756
757
One failure caused by PR: https://github.com/ceph/ceph/pull/41935#issuecomment-866472599
758
759
* https://tracker.ceph.com/issues/51282
760
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
761
* https://tracker.ceph.com/issues/51183
762
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
763
* https://tracker.ceph.com/issues/48773
764
    qa: scrub does not complete
765
* https://tracker.ceph.com/issues/48771
766
    qa: iogen: workload fails to cause balancing
767
* https://tracker.ceph.com/issues/51169
768
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
769
* https://tracker.ceph.com/issues/50495
770
    libcephfs: shutdown race fails with status 141
771
* https://tracker.ceph.com/issues/45434
772
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
773
* https://tracker.ceph.com/issues/50824
774
    qa: snaptest-git-ceph bus error
775
* https://tracker.ceph.com/issues/50223
776
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
777
778
779 13 Patrick Donnelly
h3. 2021 June 16
780
781
https://pulpito.ceph.com/pdonnell-2021-06-16_21:26:55-fs-wip-pdonnell-testing-20210616.191804-distro-basic-smithi/
782
783
MDS abort class of failures caused by PR: https://github.com/ceph/ceph/pull/41667
784
785
* https://tracker.ceph.com/issues/45434
786
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
787
* https://tracker.ceph.com/issues/51169
788
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
789
* https://tracker.ceph.com/issues/43216
790
    MDSMonitor: removes MDS coming out of quorum election
791
* https://tracker.ceph.com/issues/51278
792
    mds: "FAILED ceph_assert(!segments.empty())"
793
* https://tracker.ceph.com/issues/51279
794
    kclient hangs on umount (testing branch)
795
* https://tracker.ceph.com/issues/51280
796
    mds: "FAILED ceph_assert(r == 0 || r == -2)"
797
* https://tracker.ceph.com/issues/51183
798
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
799
* https://tracker.ceph.com/issues/51281
800
    qa: snaptest-snap-rm-cmp.sh: "echo 'FAIL: bad match, /tmp/a 4637e766853d1ad16a7b17079e2c6f03 != real c3883760b18d50e8d78819c54d579b00'"
801
* https://tracker.ceph.com/issues/48773
802
    qa: scrub does not complete
803
* https://tracker.ceph.com/issues/51076
804
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
805
* https://tracker.ceph.com/issues/51228
806
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
807
* https://tracker.ceph.com/issues/51282
808
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
809
810
811 12 Patrick Donnelly
h3. 2021 June 14
812
813
https://pulpito.ceph.com/pdonnell-2021-06-14_20:53:05-fs-wip-pdonnell-testing-20210614.173325-distro-basic-smithi/
814
815
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
816
817
* https://tracker.ceph.com/issues/51169
818
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
819
* https://tracker.ceph.com/issues/51228
820
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
821
* https://tracker.ceph.com/issues/48773
822
    qa: scrub does not complete
823
* https://tracker.ceph.com/issues/51183
824
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
825
* https://tracker.ceph.com/issues/45434
826
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
827
* https://tracker.ceph.com/issues/51182
828
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
829
* https://tracker.ceph.com/issues/51229
830
    qa: test_multi_snap_schedule list difference failure
831
* https://tracker.ceph.com/issues/50821
832
    qa: untar_snap_rm failure during mds thrashing
833
834
835 11 Patrick Donnelly
h3. 2021 June 13
836
837
https://pulpito.ceph.com/pdonnell-2021-06-12_02:45:35-fs-wip-pdonnell-testing-20210612.002809-distro-basic-smithi/
838
839
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
840
841
* https://tracker.ceph.com/issues/51169
842
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
843
* https://tracker.ceph.com/issues/48773
844
    qa: scrub does not complete
845
* https://tracker.ceph.com/issues/51182
846
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
847
* https://tracker.ceph.com/issues/51183
848
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
849
* https://tracker.ceph.com/issues/51197
850
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
851
* https://tracker.ceph.com/issues/45434
852
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
853
854 10 Patrick Donnelly
h3. 2021 June 11
855
856
https://pulpito.ceph.com/pdonnell-2021-06-11_18:02:10-fs-wip-pdonnell-testing-20210611.162716-distro-basic-smithi/
857
858
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
859
860
* https://tracker.ceph.com/issues/51169
861
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
862
* https://tracker.ceph.com/issues/45434
863
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
864
* https://tracker.ceph.com/issues/48771
865
    qa: iogen: workload fails to cause balancing
866
* https://tracker.ceph.com/issues/43216
867
    MDSMonitor: removes MDS coming out of quorum election
868
* https://tracker.ceph.com/issues/51182
869
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
870
* https://tracker.ceph.com/issues/50223
871
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
872
* https://tracker.ceph.com/issues/48773
873
    qa: scrub does not complete
874
* https://tracker.ceph.com/issues/51183
875
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
876
* https://tracker.ceph.com/issues/51184
877
    qa: fs:bugs does not specify distro
878
879
880 9 Patrick Donnelly
h3. 2021 June 03
881
882
https://pulpito.ceph.com/pdonnell-2021-06-03_03:40:33-fs-wip-pdonnell-testing-20210603.020013-distro-basic-smithi/
883
884
* https://tracker.ceph.com/issues/45434
885
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
886
* https://tracker.ceph.com/issues/50016
887
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
888
* https://tracker.ceph.com/issues/50821
889
    qa: untar_snap_rm failure during mds thrashing
890
* https://tracker.ceph.com/issues/50622 (regression)
891
    msg: active_connections regression
892
* https://tracker.ceph.com/issues/49845#note-2 (regression)
893
    qa: failed umount in test_volumes
894
* https://tracker.ceph.com/issues/48773
895
    qa: scrub does not complete
896
* https://tracker.ceph.com/issues/43216
897
    MDSMonitor: removes MDS coming out of quorum election
898
899
900 7 Patrick Donnelly
h3. 2021 May 18
901
902 8 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.214114
903
904
Regression in testing kernel caused some failures. Ilya fixed those and rerun
905
looked better. Some odd new noise in the rerun relating to packaging and "No
906
module named 'tasks.ceph'".
907
908
* https://tracker.ceph.com/issues/50824
909
    qa: snaptest-git-ceph bus error
910
* https://tracker.ceph.com/issues/50622 (regression)
911
    msg: active_connections regression
912
* https://tracker.ceph.com/issues/49845#note-2 (regression)
913
    qa: failed umount in test_volumes
914
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
915
    qa: quota failure
916
917
918
h3. 2021 May 18
919
920 7 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.025642
921
922
* https://tracker.ceph.com/issues/50821
923
    qa: untar_snap_rm failure during mds thrashing
924
* https://tracker.ceph.com/issues/48773
925
    qa: scrub does not complete
926
* https://tracker.ceph.com/issues/45591
927
    mgr: FAILED ceph_assert(daemon != nullptr)
928
* https://tracker.ceph.com/issues/50866
929
    osd: stat mismatch on objects
930
* https://tracker.ceph.com/issues/50016
931
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
932
* https://tracker.ceph.com/issues/50867
933
    qa: fs:mirror: reduced data availability
934
* https://tracker.ceph.com/issues/50821
935
    qa: untar_snap_rm failure during mds thrashing
936
* https://tracker.ceph.com/issues/50622 (regression)
937
    msg: active_connections regression
938
* https://tracker.ceph.com/issues/50223
939
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
940
* https://tracker.ceph.com/issues/50868
941
    qa: "kern.log.gz already exists; not overwritten"
942
* https://tracker.ceph.com/issues/50870
943
    qa: test_full: "rm: cannot remove 'large_file_a': Permission denied"
944
945
946 6 Patrick Donnelly
h3. 2021 May 11
947
948
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210511.232042
949
950
* one class of failures caused by PR
951
* https://tracker.ceph.com/issues/48812
952
    qa: test_scrub_pause_and_resume_with_abort failure
953
* https://tracker.ceph.com/issues/50390
954
    mds: monclient: wait_auth_rotating timed out after 30
955
* https://tracker.ceph.com/issues/48773
956
    qa: scrub does not complete
957
* https://tracker.ceph.com/issues/50821
958
    qa: untar_snap_rm failure during mds thrashing
959
* https://tracker.ceph.com/issues/50224
960
    qa: test_mirroring_init_failure_with_recovery failure
961
* https://tracker.ceph.com/issues/50622 (regression)
962
    msg: active_connections regression
963
* https://tracker.ceph.com/issues/50825
964
    qa: snaptest-git-ceph hang during mon thrashing v2
965
* https://tracker.ceph.com/issues/50821
966
    qa: untar_snap_rm failure during mds thrashing
967
* https://tracker.ceph.com/issues/50823
968
    qa: RuntimeError: timeout waiting for cluster to stabilize
969
970
971 5 Patrick Donnelly
h3. 2021 May 14
972
973
https://pulpito.ceph.com/pdonnell-2021-05-14_21:45:42-fs-master-distro-basic-smithi/
974
975
* https://tracker.ceph.com/issues/48812
976
    qa: test_scrub_pause_and_resume_with_abort failure
977
* https://tracker.ceph.com/issues/50821
978
    qa: untar_snap_rm failure during mds thrashing
979
* https://tracker.ceph.com/issues/50622 (regression)
980
    msg: active_connections regression
981
* https://tracker.ceph.com/issues/50822
982
    qa: testing kernel patch for client metrics causes mds abort
983
* https://tracker.ceph.com/issues/48773
984
    qa: scrub does not complete
985
* https://tracker.ceph.com/issues/50823
986
    qa: RuntimeError: timeout waiting for cluster to stabilize
987
* https://tracker.ceph.com/issues/50824
988
    qa: snaptest-git-ceph bus error
989
* https://tracker.ceph.com/issues/50825
990
    qa: snaptest-git-ceph hang during mon thrashing v2
991
* https://tracker.ceph.com/issues/50826
992
    kceph: stock RHEL kernel hangs on snaptests with mon|osd thrashers
993
994
995 4 Patrick Donnelly
h3. 2021 May 01
996
997
https://pulpito.ceph.com/pdonnell-2021-05-01_09:07:09-fs-wip-pdonnell-testing-20210501.040415-distro-basic-smithi/
998
999
* https://tracker.ceph.com/issues/45434
1000
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1001
* https://tracker.ceph.com/issues/50281
1002
    qa: untar_snap_rm timeout
1003
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1004
    qa: quota failure
1005
* https://tracker.ceph.com/issues/48773
1006
    qa: scrub does not complete
1007
* https://tracker.ceph.com/issues/50390
1008
    mds: monclient: wait_auth_rotating timed out after 30
1009
* https://tracker.ceph.com/issues/50250
1010
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1011
* https://tracker.ceph.com/issues/50622 (regression)
1012
    msg: active_connections regression
1013
* https://tracker.ceph.com/issues/45591
1014
    mgr: FAILED ceph_assert(daemon != nullptr)
1015
* https://tracker.ceph.com/issues/50221
1016
    qa: snaptest-git-ceph failure in git diff
1017
* https://tracker.ceph.com/issues/50016
1018
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1019
1020
1021 3 Patrick Donnelly
h3. 2021 Apr 15
1022
1023
https://pulpito.ceph.com/pdonnell-2021-04-15_01:35:57-fs-wip-pdonnell-testing-20210414.230315-distro-basic-smithi/
1024
1025
* https://tracker.ceph.com/issues/50281
1026
    qa: untar_snap_rm timeout
1027
* https://tracker.ceph.com/issues/50220
1028
    qa: dbench workload timeout
1029
* https://tracker.ceph.com/issues/50246
1030
    mds: failure replaying journal (EMetaBlob)
1031
* https://tracker.ceph.com/issues/50250
1032
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1033
* https://tracker.ceph.com/issues/50016
1034
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1035
* https://tracker.ceph.com/issues/50222
1036
    osd: 5.2s0 deep-scrub : stat mismatch
1037
* https://tracker.ceph.com/issues/45434
1038
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1039
* https://tracker.ceph.com/issues/49845
1040
    qa: failed umount in test_volumes
1041
* https://tracker.ceph.com/issues/37808
1042
    osd: osdmap cache weak_refs assert during shutdown
1043
* https://tracker.ceph.com/issues/50387
1044
    client: fs/snaps failure
1045
* https://tracker.ceph.com/issues/50389
1046
    mds: "cluster [ERR] Error recovering journal 0x203: (2) No such file or directory" in cluster log"
1047
* https://tracker.ceph.com/issues/50216
1048
    qa: "ls: cannot access 'lost+found': No such file or directory"
1049
* https://tracker.ceph.com/issues/50390
1050
    mds: monclient: wait_auth_rotating timed out after 30
1051
1052
1053
1054 1 Patrick Donnelly
h3. 2021 Apr 08
1055
1056 2 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_22:42:24-fs-wip-pdonnell-testing-20210408.192301-distro-basic-smithi/
1057
1058
* https://tracker.ceph.com/issues/45434
1059
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1060
* https://tracker.ceph.com/issues/50016
1061
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1062
* https://tracker.ceph.com/issues/48773
1063
    qa: scrub does not complete
1064
* https://tracker.ceph.com/issues/50279
1065
    qa: "Replacing daemon mds.b as rank 0 with standby daemon mds.c"
1066
* https://tracker.ceph.com/issues/50246
1067
    mds: failure replaying journal (EMetaBlob)
1068
* https://tracker.ceph.com/issues/48365
1069
    qa: ffsb build failure on CentOS 8.2
1070
* https://tracker.ceph.com/issues/50216
1071
    qa: "ls: cannot access 'lost+found': No such file or directory"
1072
* https://tracker.ceph.com/issues/50223
1073
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1074
* https://tracker.ceph.com/issues/50280
1075
    cephadm: RuntimeError: uid/gid not found
1076
* https://tracker.ceph.com/issues/50281
1077
    qa: untar_snap_rm timeout
1078
1079
h3. 2021 Apr 08
1080
1081 1 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_04:31:36-fs-wip-pdonnell-testing-20210408.024225-distro-basic-smithi/
1082
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210408.142238 (with logic inversion / QA fix)
1083
1084
* https://tracker.ceph.com/issues/50246
1085
    mds: failure replaying journal (EMetaBlob)
1086
* https://tracker.ceph.com/issues/50250
1087
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
1088
1089
1090
h3. 2021 Apr 07
1091
1092
https://pulpito.ceph.com/pdonnell-2021-04-07_02:12:41-fs-wip-pdonnell-testing-20210406.213012-distro-basic-smithi/
1093
1094
* https://tracker.ceph.com/issues/50215
1095
    qa: "log [ERR] : error reading sessionmap 'mds2_sessionmap'"
1096
* https://tracker.ceph.com/issues/49466
1097
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1098
* https://tracker.ceph.com/issues/50216
1099
    qa: "ls: cannot access 'lost+found': No such file or directory"
1100
* https://tracker.ceph.com/issues/48773
1101
    qa: scrub does not complete
1102
* https://tracker.ceph.com/issues/49845
1103
    qa: failed umount in test_volumes
1104
* https://tracker.ceph.com/issues/50220
1105
    qa: dbench workload timeout
1106
* https://tracker.ceph.com/issues/50221
1107
    qa: snaptest-git-ceph failure in git diff
1108
* https://tracker.ceph.com/issues/50222
1109
    osd: 5.2s0 deep-scrub : stat mismatch
1110
* https://tracker.ceph.com/issues/50223
1111
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
1112
* https://tracker.ceph.com/issues/50224
1113
    qa: test_mirroring_init_failure_with_recovery failure
1114
1115
h3. 2021 Apr 01
1116
1117
https://pulpito.ceph.com/pdonnell-2021-04-01_00:45:34-fs-wip-pdonnell-testing-20210331.222326-distro-basic-smithi/
1118
1119
* https://tracker.ceph.com/issues/48772
1120
    qa: pjd: not ok 9, 44, 80
1121
* https://tracker.ceph.com/issues/50177
1122
    osd: "stalled aio... buggy kernel or bad device?"
1123
* https://tracker.ceph.com/issues/48771
1124
    qa: iogen: workload fails to cause balancing
1125
* https://tracker.ceph.com/issues/49845
1126
    qa: failed umount in test_volumes
1127
* https://tracker.ceph.com/issues/48773
1128
    qa: scrub does not complete
1129
* https://tracker.ceph.com/issues/48805
1130
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1131
* https://tracker.ceph.com/issues/50178
1132
    qa: "TypeError: run() got an unexpected keyword argument 'shell'"
1133
* https://tracker.ceph.com/issues/45434
1134
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1135
1136
h3. 2021 Mar 24
1137
1138
https://pulpito.ceph.com/pdonnell-2021-03-24_23:26:35-fs-wip-pdonnell-testing-20210324.190252-distro-basic-smithi/
1139
1140
* https://tracker.ceph.com/issues/49500
1141
    qa: "Assertion `cb_done' failed."
1142
* https://tracker.ceph.com/issues/50019
1143
    qa: mount failure with cephadm "probably no MDS server is up?"
1144
* https://tracker.ceph.com/issues/50020
1145
    qa: "RADOS object not found (Failed to operate read op for oid cephfs_mirror)"
1146
* https://tracker.ceph.com/issues/48773
1147
    qa: scrub does not complete
1148
* https://tracker.ceph.com/issues/45434
1149
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1150
* https://tracker.ceph.com/issues/48805
1151
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1152
* https://tracker.ceph.com/issues/48772
1153
    qa: pjd: not ok 9, 44, 80
1154
* https://tracker.ceph.com/issues/50021
1155
    qa: snaptest-git-ceph failure during mon thrashing
1156
* https://tracker.ceph.com/issues/48771
1157
    qa: iogen: workload fails to cause balancing
1158
* https://tracker.ceph.com/issues/50016
1159
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1160
* https://tracker.ceph.com/issues/49466
1161
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1162
1163
1164
h3. 2021 Mar 18
1165
1166
https://pulpito.ceph.com/pdonnell-2021-03-18_13:46:31-fs-wip-pdonnell-testing-20210318.024145-distro-basic-smithi/
1167
1168
* https://tracker.ceph.com/issues/49466
1169
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1170
* https://tracker.ceph.com/issues/48773
1171
    qa: scrub does not complete
1172
* https://tracker.ceph.com/issues/48805
1173
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1174
* https://tracker.ceph.com/issues/45434
1175
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1176
* https://tracker.ceph.com/issues/49845
1177
    qa: failed umount in test_volumes
1178
* https://tracker.ceph.com/issues/49605
1179
    mgr: drops command on the floor
1180
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1181
    qa: quota failure
1182
* https://tracker.ceph.com/issues/49928
1183
    client: items pinned in cache preventing unmount x2
1184
1185
h3. 2021 Mar 15
1186
1187
https://pulpito.ceph.com/pdonnell-2021-03-15_22:16:56-fs-wip-pdonnell-testing-20210315.182203-distro-basic-smithi/
1188
1189
* https://tracker.ceph.com/issues/49842
1190
    qa: stuck pkg install
1191
* https://tracker.ceph.com/issues/49466
1192
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1193
* https://tracker.ceph.com/issues/49822
1194
    test: test_mirroring_command_idempotency (tasks.cephfs.test_admin.TestMirroringCommands) failure
1195
* https://tracker.ceph.com/issues/49240
1196
    terminate called after throwing an instance of 'std::bad_alloc'
1197
* https://tracker.ceph.com/issues/48773
1198
    qa: scrub does not complete
1199
* https://tracker.ceph.com/issues/45434
1200
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1201
* https://tracker.ceph.com/issues/49500
1202
    qa: "Assertion `cb_done' failed."
1203
* https://tracker.ceph.com/issues/49843
1204
    qa: fs/snaps/snaptest-upchildrealms.sh failure
1205
* https://tracker.ceph.com/issues/49845
1206
    qa: failed umount in test_volumes
1207
* https://tracker.ceph.com/issues/48805
1208
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1209
* https://tracker.ceph.com/issues/49605
1210
    mgr: drops command on the floor
1211
1212
and failure caused by PR: https://github.com/ceph/ceph/pull/39969
1213
1214
1215
h3. 2021 Mar 09
1216
1217
https://pulpito.ceph.com/pdonnell-2021-03-09_03:27:39-fs-wip-pdonnell-testing-20210308.214827-distro-basic-smithi/
1218
1219
* https://tracker.ceph.com/issues/49500
1220
    qa: "Assertion `cb_done' failed."
1221
* https://tracker.ceph.com/issues/48805
1222
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1223
* https://tracker.ceph.com/issues/48773
1224
    qa: scrub does not complete
1225
* https://tracker.ceph.com/issues/45434
1226
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1227
* https://tracker.ceph.com/issues/49240
1228
    terminate called after throwing an instance of 'std::bad_alloc'
1229
* https://tracker.ceph.com/issues/49466
1230
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1231
* https://tracker.ceph.com/issues/49684
1232
    qa: fs:cephadm mount does not wait for mds to be created
1233
* https://tracker.ceph.com/issues/48771
1234
    qa: iogen: workload fails to cause balancing