Project

General

Profile

Main » History » Version 58

Venky Shankar, 07/04/2022 02:49 AM

1 58 Venky Shankar
h3. 2022 July 04
2
3
https://pulpito.ceph.com/vshankar-2022-06-29_09:19:00-fs-wip-vshankar-testing-20220627-100931-testing-default-smithi/
4
(rhel runs were borked due to: https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/JSZQFUKVLDND4W33PXDGCABPHNSPT6SS/, tests ran with --filter-out=rhel)
5
6
* https://tracker.ceph.com/issues/56445
7
    Command failed on smithi162 with status 123: "find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --no-run-if-empty -- gzip --"
8
* https://tracker.ceph.com/issues/56446
9
    Test failure: test_client_cache_size (tasks.cephfs.test_client_limits.TestClientLimits)
10
* https://tracker.ceph.com/issues/51964
11
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
12
* https://tracker.ceph.com/issues/52624
13
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
14
15 57 Venky Shankar
h3. 2022 June 20
16
17
https://pulpito.ceph.com/vshankar-2022-06-15_04:03:39-fs-wip-vshankar-testing1-20220615-072516-testing-default-smithi/
18
https://pulpito.ceph.com/vshankar-2022-06-19_08:22:46-fs-wip-vshankar-testing1-20220619-102531-testing-default-smithi/
19
20
* https://tracker.ceph.com/issues/52624
21
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
22
* https://tracker.ceph.com/issues/55804
23
    qa failure: pjd link tests failed
24
* https://tracker.ceph.com/issues/54108
25
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
26
* https://tracker.ceph.com/issues/55332
27
    Failure in snaptest-git-ceph.sh (it's an async unlink/create bug)
28
29 56 Patrick Donnelly
h3. 2022 June 13
30
31
https://pulpito.ceph.com/pdonnell-2022-06-12_05:08:12-fs:workload-wip-pdonnell-testing-20220612.004943-distro-default-smithi/
32
33
* https://tracker.ceph.com/issues/56024
34
    cephadm: removes ceph.conf during qa run causing command failure
35
* https://tracker.ceph.com/issues/48773
36
    qa: scrub does not complete
37
* https://tracker.ceph.com/issues/56012
38
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
39
40
41 55 Venky Shankar
h3. 2022 Jun 13
42 54 Venky Shankar
43
https://pulpito.ceph.com/vshankar-2022-06-07_00:25:50-fs-wip-vshankar-testing-20220606-223254-testing-default-smithi/
44
https://pulpito.ceph.com/vshankar-2022-06-10_01:04:46-fs-wip-vshankar-testing-20220609-175550-testing-default-smithi/
45
46
* https://tracker.ceph.com/issues/52624
47
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
48
* https://tracker.ceph.com/issues/51964
49
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
50
* https://tracker.ceph.com/issues/53859
51
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
52
* https://tracker.ceph.com/issues/55804
53
    qa failure: pjd link tests failed
54
* https://tracker.ceph.com/issues/56003
55
    client: src/include/xlist.h: 81: FAILED ceph_assert(_size == 0)
56
* https://tracker.ceph.com/issues/56011
57
    fs/thrash: snaptest-snap-rm-cmp.sh fails in mds5sum comparison
58
* https://tracker.ceph.com/issues/56012
59
    mds: src/mds/MDLog.cc: 283: FAILED ceph_assert(!mds->is_ any_replay())
60
61 53 Venky Shankar
h3. 2022 Jun 07
62
63
https://pulpito.ceph.com/vshankar-2022-06-06_21:25:41-fs-wip-vshankar-testing1-20220606-230129-testing-default-smithi/
64
https://pulpito.ceph.com/vshankar-2022-06-07_10:53:31-fs-wip-vshankar-testing1-20220607-104134-testing-default-smithi/ (rerun after dropping a problematic PR)
65
66
* https://tracker.ceph.com/issues/52624
67
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
68
* https://tracker.ceph.com/issues/50223
69
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
70
* https://tracker.ceph.com/issues/50224
71
    qa: test_mirroring_init_failure_with_recovery failure
72
73 51 Venky Shankar
h3. 2022 May 12
74
75
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220509-125847
76 52 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-13_17:09:16-fs-wip-vshankar-testing-20220513-120051-testing-default-smithi/ (drop prs + rerun)
77 51 Venky Shankar
78
* https://tracker.ceph.com/issues/52624
79
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
80
* https://tracker.ceph.com/issues/50223
81
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
82
* https://tracker.ceph.com/issues/55332
83
    Failure in snaptest-git-ceph.sh
84
* https://tracker.ceph.com/issues/53859
85
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
86
* https://tracker.ceph.com/issues/55538
87 1 Patrick Donnelly
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
88 52 Venky Shankar
* https://tracker.ceph.com/issues/55258
89
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs (cropss up again, though very infrequent)
90 51 Venky Shankar
91 49 Venky Shankar
h3. 2022 May 04
92
93 50 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-05-01_13:18:44-fs-wip-vshankar-testing1-20220428-204527-testing-default-smithi/
94
https://pulpito.ceph.com/vshankar-2022-05-02_16:58:59-fs-wip-vshankar-testing1-20220502-201957-testing-default-smithi/ (after dropping PRs)
95
96 49 Venky Shankar
* https://tracker.ceph.com/issues/52624
97
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
98
* https://tracker.ceph.com/issues/50223
99
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
100
* https://tracker.ceph.com/issues/55332
101
    Failure in snaptest-git-ceph.sh
102
* https://tracker.ceph.com/issues/53859
103
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
104
* https://tracker.ceph.com/issues/55516
105
    qa: fs suite tests failing with "json.decoder.JSONDecodeError: Extra data: line 2 column 82 (char 82)"
106
* https://tracker.ceph.com/issues/55537
107
    mds: crash during fs:upgrade test
108
* https://tracker.ceph.com/issues/55538
109
    Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
110
111 48 Venky Shankar
h3. 2022 Apr 25
112
113
https://pulpito.ceph.com/?branch=wip-vshankar-testing-20220420-113951 (owner vshankar)
114
115
* https://tracker.ceph.com/issues/52624
116
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
117
* https://tracker.ceph.com/issues/50223
118
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
119
* https://tracker.ceph.com/issues/55258
120
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
121
* https://tracker.ceph.com/issues/55377
122
    kclient: mds revoke Fwb caps stuck after the kclient tries writebcak once
123
124 47 Venky Shankar
h3. 2022 Apr 14
125
126
https://pulpito.ceph.com/?branch=wip-vshankar-testing1-20220411-144044
127
128
* https://tracker.ceph.com/issues/52624
129
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
130
* https://tracker.ceph.com/issues/50223
131
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
132
* https://tracker.ceph.com/issues/52438
133
    qa: ffsb timeout
134
* https://tracker.ceph.com/issues/55170
135
    mds: crash during rejoin (CDir::fetch_keys)
136
* https://tracker.ceph.com/issues/55331
137
    pjd failure
138
* https://tracker.ceph.com/issues/48773
139
    qa: scrub does not complete
140
* https://tracker.ceph.com/issues/55332
141
    Failure in snaptest-git-ceph.sh
142
* https://tracker.ceph.com/issues/55258
143
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
144
145 45 Venky Shankar
h3. 2022 Apr 11
146
147 46 Venky Shankar
https://pulpito.ceph.com/?branch=wip-vshankar-testing-55110-20220408-203242
148 45 Venky Shankar
149
* https://tracker.ceph.com/issues/48773
150
    qa: scrub does not complete
151
* https://tracker.ceph.com/issues/52624
152
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
153
* https://tracker.ceph.com/issues/52438
154
    qa: ffsb timeout
155
* https://tracker.ceph.com/issues/48680
156
    mds: scrubbing stuck "scrub active (0 inodes in the stack)"
157
* https://tracker.ceph.com/issues/55236
158
    qa: fs/snaps tests fails with "hit max job timeout"
159
* https://tracker.ceph.com/issues/54108
160
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
161
* https://tracker.ceph.com/issues/54971
162
    Test failure: test_perf_stats_stale_metrics (tasks.cephfs.test_mds_metrics.TestMDSMetrics)
163
* https://tracker.ceph.com/issues/50223
164
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
165
* https://tracker.ceph.com/issues/55258
166
    lots of "heartbeat_check: no reply from X.X.X.X" in OSD logs
167
168 44 Venky Shankar
h3. 2022 Mar 21
169 42 Venky Shankar
170 43 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-03-20_02:16:37-fs-wip-vshankar-testing-20220319-163539-testing-default-smithi/
171
172
Run didn't go well, lots of failures - debugging by dropping PRs and running against master branch. Only merging unrelated PRs that pass tests.
173
174
175
h3. 2022 Mar 08
176
177 42 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-02-28_04:32:15-fs-wip-vshankar-testing-20220226-211550-testing-default-smithi/
178
179
rerun with
180
- (drop) https://github.com/ceph/ceph/pull/44679
181
- (drop) https://github.com/ceph/ceph/pull/44958
182
https://pulpito.ceph.com/vshankar-2022-03-06_14:47:51-fs-wip-vshankar-testing-20220304-132102-testing-default-smithi/
183
184
* https://tracker.ceph.com/issues/54419 (new)
185
    `ceph orch upgrade start` seems to never reach completion
186
* https://tracker.ceph.com/issues/51964
187
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
188
* https://tracker.ceph.com/issues/52624
189
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
190
* https://tracker.ceph.com/issues/50223
191
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
192
* https://tracker.ceph.com/issues/52438
193
    qa: ffsb timeout
194
* https://tracker.ceph.com/issues/50821
195
    qa: untar_snap_rm failure during mds thrashing
196
197
198 41 Venky Shankar
h3. 2022 Feb 09
199
200
https://pulpito.ceph.com/vshankar-2022-02-05_17:27:49-fs-wip-vshankar-testing-20220201-113815-testing-default-smithi/
201
202
rerun with
203
- (drop) https://github.com/ceph/ceph/pull/37938
204
- (drop) https://github.com/ceph/ceph/pull/44335
205
- (drop) https://github.com/ceph/ceph/pull/44491
206
- (drop) https://github.com/ceph/ceph/pull/44501
207
https://pulpito.ceph.com/vshankar-2022-02-08_14:27:29-fs-wip-vshankar-testing-20220208-181241-testing-default-smithi/
208
209
* https://tracker.ceph.com/issues/51964
210
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
211
* https://tracker.ceph.com/issues/54066
212
    test_subvolume_no_upgrade_v1_sanity fails with `AssertionError: 1000 != 0`
213
* https://tracker.ceph.com/issues/48773
214
    qa: scrub does not complete
215
* https://tracker.ceph.com/issues/52624
216
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
217
* https://tracker.ceph.com/issues/50223
218
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
219
* https://tracker.ceph.com/issues/52438
220
    qa: ffsb timeout
221
222 40 Patrick Donnelly
h3. 2022 Feb 01
223
224
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20220127.171526
225
226
* https://tracker.ceph.com/issues/54107
227
    kclient: hang during umount
228
* https://tracker.ceph.com/issues/54106
229
    kclient: hang during workunit cleanup
230
* https://tracker.ceph.com/issues/54108
231
    qa: iogen workunit: "The following counters failed to be set on mds daemons: {'mds.exported', 'mds.imported'}"
232
* https://tracker.ceph.com/issues/48773
233
    qa: scrub does not complete
234
* https://tracker.ceph.com/issues/52438
235
    qa: ffsb timeout
236
237
238 36 Venky Shankar
h3. 2022 Jan 13
239
240
https://pulpito.ceph.com/vshankar-2022-01-06_13:18:41-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
241 39 Venky Shankar
242 36 Venky Shankar
rerun with:
243 38 Venky Shankar
- (add) https://github.com/ceph/ceph/pull/44570
244
- (drop) https://github.com/ceph/ceph/pull/43184
245 36 Venky Shankar
https://pulpito.ceph.com/vshankar-2022-01-13_04:42:40-fs-wip-vshankar-testing-20220106-145819-testing-default-smithi/
246
247
* https://tracker.ceph.com/issues/50223
248
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
249
* https://tracker.ceph.com/issues/51282
250
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
251
* https://tracker.ceph.com/issues/48773
252
    qa: scrub does not complete
253
* https://tracker.ceph.com/issues/52624
254
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
255
* https://tracker.ceph.com/issues/53859
256
    qa: Test failure: test_pool_perm (tasks.cephfs.test_pool_perm.TestPoolPerm)
257
258 34 Venky Shankar
h3. 2022 Jan 03
259
260
https://pulpito.ceph.com/vshankar-2021-12-22_07:37:44-fs-wip-vshankar-testing-20211216-114012-testing-default-smithi/
261
https://pulpito.ceph.com/vshankar-2022-01-03_12:27:45-fs-wip-vshankar-testing-20220103-142738-testing-default-smithi/ (rerun)
262
263
* https://tracker.ceph.com/issues/50223
264
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
265
* https://tracker.ceph.com/issues/51964
266
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
267
* https://tracker.ceph.com/issues/51267
268
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
269
* https://tracker.ceph.com/issues/51282
270
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
271
* https://tracker.ceph.com/issues/50821
272
    qa: untar_snap_rm failure during mds thrashing
273
* https://tracker.ceph.com/issues/51278
274
    mds: "FAILED ceph_assert(!segments.empty())"
275 35 Ramana Raja
* https://tracker.ceph.com/issues/52279
276
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
277
278 34 Venky Shankar
279 33 Patrick Donnelly
h3. 2021 Dec 22
280
281
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211222.014316
282
283
* https://tracker.ceph.com/issues/52624
284
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
285
* https://tracker.ceph.com/issues/50223
286
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
287
* https://tracker.ceph.com/issues/52279
288
    cephadm tests fail due to: error adding seccomp filter rule for syscall bdflush: requested action matches default action of filter
289
* https://tracker.ceph.com/issues/50224
290
    qa: test_mirroring_init_failure_with_recovery failure
291
* https://tracker.ceph.com/issues/48773
292
    qa: scrub does not complete
293
294
295 32 Venky Shankar
h3. 2021 Nov 30
296
297
https://pulpito.ceph.com/vshankar-2021-11-24_07:14:27-fs-wip-vshankar-testing-20211124-094330-testing-default-smithi/
298
https://pulpito.ceph.com/vshankar-2021-11-30_06:23:32-fs-wip-vshankar-testing-20211124-094330-distro-default-smithi/ (rerun w/ QA fixes)
299
300
* https://tracker.ceph.com/issues/53436
301
    mds, mon: mds beacon messages get dropped? (mds never reaches up:active state)
302
* https://tracker.ceph.com/issues/51964
303
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
304
* https://tracker.ceph.com/issues/48812
305
    qa: test_scrub_pause_and_resume_with_abort failure
306
* https://tracker.ceph.com/issues/51076
307
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
308
* https://tracker.ceph.com/issues/50223
309
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
310
* https://tracker.ceph.com/issues/52624
311
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
312
* https://tracker.ceph.com/issues/50250
313
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
314
315
316 31 Patrick Donnelly
h3. 2021 November 9
317
318
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211109.180315
319
320
* https://tracker.ceph.com/issues/53214
321
    qa: "dd: error reading '/sys/kernel/debug/ceph/2a934501-6731-4052-a836-f42229a869be.client4874/metrics': Is a directory"
322
* https://tracker.ceph.com/issues/48773
323
    qa: scrub does not complete
324
* https://tracker.ceph.com/issues/50223
325
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
326
* https://tracker.ceph.com/issues/51282
327
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
328
* https://tracker.ceph.com/issues/52624
329
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
330
* https://tracker.ceph.com/issues/53216
331
    qa: "RuntimeError: value of attributes should be either str or None. client_id"
332
* https://tracker.ceph.com/issues/50250
333
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
334
335
336
337 30 Patrick Donnelly
h3. 2021 November 03
338
339
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211103.023355
340
341
* https://tracker.ceph.com/issues/51964
342
    qa: test_cephfs_mirror_restart_sync_on_blocklist failure
343
* https://tracker.ceph.com/issues/51282
344
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
345
* https://tracker.ceph.com/issues/52436
346
    fs/ceph: "corrupt mdsmap"
347
* https://tracker.ceph.com/issues/53074
348
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
349
* https://tracker.ceph.com/issues/53150
350
    pybind/mgr/cephadm/upgrade: tolerate MDS failures during upgrade straddling v16.2.5
351
* https://tracker.ceph.com/issues/53155
352
    MDSMonitor: assertion during upgrade to v16.2.5+
353
354
355 29 Patrick Donnelly
h3. 2021 October 26
356
357
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211025.000447
358
359
* https://tracker.ceph.com/issues/53074
360
    pybind/mgr/cephadm: upgrade sequence does not continue if no MDS are active
361
* https://tracker.ceph.com/issues/52997
362
    testing: hang ing umount
363
* https://tracker.ceph.com/issues/50824
364
    qa: snaptest-git-ceph bus error
365
* https://tracker.ceph.com/issues/52436
366
    fs/ceph: "corrupt mdsmap"
367
* https://tracker.ceph.com/issues/48773
368
    qa: scrub does not complete
369
* https://tracker.ceph.com/issues/53082
370
    ceph-fuse: segmenetation fault in Client::handle_mds_map
371
* https://tracker.ceph.com/issues/50223
372
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
373
* https://tracker.ceph.com/issues/52624
374
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
375
* https://tracker.ceph.com/issues/50224
376
    qa: test_mirroring_init_failure_with_recovery failure
377
* https://tracker.ceph.com/issues/50821
378
    qa: untar_snap_rm failure during mds thrashing
379
* https://tracker.ceph.com/issues/50250
380
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
381
382
383
384 27 Patrick Donnelly
h3. 2021 October 19
385
386 28 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211019.013028
387 27 Patrick Donnelly
388
* https://tracker.ceph.com/issues/52995
389
    qa: test_standby_count_wanted failure
390
* https://tracker.ceph.com/issues/52948
391
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
392
* https://tracker.ceph.com/issues/52996
393
    qa: test_perf_counters via test_openfiletable
394
* https://tracker.ceph.com/issues/48772
395
    qa: pjd: not ok 9, 44, 80
396
* https://tracker.ceph.com/issues/52997
397
    testing: hang ing umount
398
* https://tracker.ceph.com/issues/50250
399
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
400
* https://tracker.ceph.com/issues/52624
401
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
402
* https://tracker.ceph.com/issues/50223
403
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
404
* https://tracker.ceph.com/issues/50821
405
    qa: untar_snap_rm failure during mds thrashing
406
* https://tracker.ceph.com/issues/48773
407
    qa: scrub does not complete
408
409
410 26 Patrick Donnelly
h3. 2021 October 12
411
412
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211012.192211
413
414
Some failures caused by teuthology bug: https://tracker.ceph.com/issues/52944
415
416
New test caused failure: https://github.com/ceph/ceph/pull/43297#discussion_r729883167
417
418
419
* https://tracker.ceph.com/issues/51282
420
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
421
* https://tracker.ceph.com/issues/52948
422
    osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
423
* https://tracker.ceph.com/issues/48773
424
    qa: scrub does not complete
425
* https://tracker.ceph.com/issues/50224
426
    qa: test_mirroring_init_failure_with_recovery failure
427
* https://tracker.ceph.com/issues/52949
428
    RuntimeError: The following counters failed to be set on mds daemons: {'mds.dir_split'}
429
430
431 25 Patrick Donnelly
h3. 2021 October 02
432 23 Patrick Donnelly
433 24 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20211002.163337
434
435
Some failures caused by cephadm upgrade test. Fixed in follow-up qa commit.
436
437
test_simple failures caused by PR in this set.
438
439
A few reruns because of QA infra noise.
440
441
* https://tracker.ceph.com/issues/52822
442
    qa: failed pacific install on fs:upgrade
443
* https://tracker.ceph.com/issues/52624
444
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
445
* https://tracker.ceph.com/issues/50223
446
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
447
* https://tracker.ceph.com/issues/48773
448
    qa: scrub does not complete
449
450
451
h3. 2021 September 20
452
453 23 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210917.174826
454
455
* https://tracker.ceph.com/issues/52677
456
    qa: test_simple failure
457
* https://tracker.ceph.com/issues/51279
458
    kclient hangs on umount (testing branch)
459
* https://tracker.ceph.com/issues/50223
460
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
461
* https://tracker.ceph.com/issues/50250
462
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
463
* https://tracker.ceph.com/issues/52624
464
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
465
* https://tracker.ceph.com/issues/52438
466
    qa: ffsb timeout
467
468
469 22 Patrick Donnelly
h3. 2021 September 10
470
471
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210910.181451
472
473
* https://tracker.ceph.com/issues/50223
474
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
475
* https://tracker.ceph.com/issues/50250
476
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
477
* https://tracker.ceph.com/issues/52624
478
    qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
479
* https://tracker.ceph.com/issues/52625
480
    qa: test_kill_mdstable (tasks.cephfs.test_snapshots.TestSnapshots)
481
* https://tracker.ceph.com/issues/52439
482
    qa: acls does not compile on centos stream
483
* https://tracker.ceph.com/issues/50821
484
    qa: untar_snap_rm failure during mds thrashing
485
* https://tracker.ceph.com/issues/48773
486
    qa: scrub does not complete
487
* https://tracker.ceph.com/issues/52626
488
    mds: ScrubStack.cc: 831: FAILED ceph_assert(diri)
489
* https://tracker.ceph.com/issues/51279
490
    kclient hangs on umount (testing branch)
491
492
493 21 Patrick Donnelly
h3. 2021 August 27
494
495
Several jobs died because of device failures.
496
497
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210827.024746
498
499
* https://tracker.ceph.com/issues/52430
500
    mds: fast async create client mount breaks racy test
501
* https://tracker.ceph.com/issues/52436
502
    fs/ceph: "corrupt mdsmap"
503
* https://tracker.ceph.com/issues/52437
504
    mds: InoTable::replay_release_ids abort via test_inotable_sync
505
* https://tracker.ceph.com/issues/51282
506
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
507
* https://tracker.ceph.com/issues/52438
508
    qa: ffsb timeout
509
* https://tracker.ceph.com/issues/52439
510
    qa: acls does not compile on centos stream
511
512
513 20 Patrick Donnelly
h3. 2021 July 30
514
515
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210729.214022
516
517
* https://tracker.ceph.com/issues/50250
518
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
519
* https://tracker.ceph.com/issues/51282
520
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
521
* https://tracker.ceph.com/issues/48773
522
    qa: scrub does not complete
523
* https://tracker.ceph.com/issues/51975
524
    pybind/mgr/stats: KeyError
525
526
527 19 Patrick Donnelly
h3. 2021 July 28
528
529
https://pulpito.ceph.com/pdonnell-2021-07-28_00:39:45-fs-wip-pdonnell-testing-20210727.213757-distro-basic-smithi/
530
531
with qa fix: https://pulpito.ceph.com/pdonnell-2021-07-28_16:20:28-fs-wip-pdonnell-testing-20210728.141004-distro-basic-smithi/
532
533
* https://tracker.ceph.com/issues/51905
534
    qa: "error reading sessionmap 'mds1_sessionmap'"
535
* https://tracker.ceph.com/issues/48773
536
    qa: scrub does not complete
537
* https://tracker.ceph.com/issues/50250
538
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
539
* https://tracker.ceph.com/issues/51267
540
    CommandFailedError: Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi096 with status 1:...
541
* https://tracker.ceph.com/issues/51279
542
    kclient hangs on umount (testing branch)
543
544
545 18 Patrick Donnelly
h3. 2021 July 16
546
547
https://pulpito.ceph.com/pdonnell-2021-07-16_05:50:11-fs-wip-pdonnell-testing-20210716.022804-distro-basic-smithi/
548
549
* https://tracker.ceph.com/issues/48773
550
    qa: scrub does not complete
551
* https://tracker.ceph.com/issues/48772
552
    qa: pjd: not ok 9, 44, 80
553
* https://tracker.ceph.com/issues/45434
554
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
555
* https://tracker.ceph.com/issues/51279
556
    kclient hangs on umount (testing branch)
557
* https://tracker.ceph.com/issues/50824
558
    qa: snaptest-git-ceph bus error
559
560
561 17 Patrick Donnelly
h3. 2021 July 04
562
563
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210703.052904
564
565
* https://tracker.ceph.com/issues/48773
566
    qa: scrub does not complete
567
* https://tracker.ceph.com/issues/39150
568
    mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
569
* https://tracker.ceph.com/issues/45434
570
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
571
* https://tracker.ceph.com/issues/51282
572
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
573
* https://tracker.ceph.com/issues/48771
574
    qa: iogen: workload fails to cause balancing
575
* https://tracker.ceph.com/issues/51279
576
    kclient hangs on umount (testing branch)
577
* https://tracker.ceph.com/issues/50250
578
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details" ("freshly-calculated rstats don't match existing ones")
579
580
581 16 Patrick Donnelly
h3. 2021 July 01
582
583
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210701.192056
584
585
* https://tracker.ceph.com/issues/51197
586
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
587
* https://tracker.ceph.com/issues/50866
588
    osd: stat mismatch on objects
589
* https://tracker.ceph.com/issues/48773
590
    qa: scrub does not complete
591
592
593 15 Patrick Donnelly
h3. 2021 June 26
594
595
https://pulpito.ceph.com/pdonnell-2021-06-26_00:57:00-fs-wip-pdonnell-testing-20210625.225421-distro-basic-smithi/
596
597
* https://tracker.ceph.com/issues/51183
598
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
599
* https://tracker.ceph.com/issues/51410
600
    kclient: fails to finish reconnect during MDS thrashing (testing branch)
601
* https://tracker.ceph.com/issues/48773
602
    qa: scrub does not complete
603
* https://tracker.ceph.com/issues/51282
604
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
605
* https://tracker.ceph.com/issues/51169
606
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
607
* https://tracker.ceph.com/issues/48772
608
    qa: pjd: not ok 9, 44, 80
609
610
611 14 Patrick Donnelly
h3. 2021 June 21
612
613
https://pulpito.ceph.com/pdonnell-2021-06-22_00:27:21-fs-wip-pdonnell-testing-20210621.231646-distro-basic-smithi/
614
615
One failure caused by PR: https://github.com/ceph/ceph/pull/41935#issuecomment-866472599
616
617
* https://tracker.ceph.com/issues/51282
618
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
619
* https://tracker.ceph.com/issues/51183
620
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
621
* https://tracker.ceph.com/issues/48773
622
    qa: scrub does not complete
623
* https://tracker.ceph.com/issues/48771
624
    qa: iogen: workload fails to cause balancing
625
* https://tracker.ceph.com/issues/51169
626
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
627
* https://tracker.ceph.com/issues/50495
628
    libcephfs: shutdown race fails with status 141
629
* https://tracker.ceph.com/issues/45434
630
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
631
* https://tracker.ceph.com/issues/50824
632
    qa: snaptest-git-ceph bus error
633
* https://tracker.ceph.com/issues/50223
634
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
635
636
637 13 Patrick Donnelly
h3. 2021 June 16
638
639
https://pulpito.ceph.com/pdonnell-2021-06-16_21:26:55-fs-wip-pdonnell-testing-20210616.191804-distro-basic-smithi/
640
641
MDS abort class of failures caused by PR: https://github.com/ceph/ceph/pull/41667
642
643
* https://tracker.ceph.com/issues/45434
644
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
645
* https://tracker.ceph.com/issues/51169
646
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
647
* https://tracker.ceph.com/issues/43216
648
    MDSMonitor: removes MDS coming out of quorum election
649
* https://tracker.ceph.com/issues/51278
650
    mds: "FAILED ceph_assert(!segments.empty())"
651
* https://tracker.ceph.com/issues/51279
652
    kclient hangs on umount (testing branch)
653
* https://tracker.ceph.com/issues/51280
654
    mds: "FAILED ceph_assert(r == 0 || r == -2)"
655
* https://tracker.ceph.com/issues/51183
656
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
657
* https://tracker.ceph.com/issues/51281
658
    qa: snaptest-snap-rm-cmp.sh: "echo 'FAIL: bad match, /tmp/a 4637e766853d1ad16a7b17079e2c6f03 != real c3883760b18d50e8d78819c54d579b00'"
659
* https://tracker.ceph.com/issues/48773
660
    qa: scrub does not complete
661
* https://tracker.ceph.com/issues/51076
662
    "wait_for_recovery: failed before timeout expired" during thrashosd test with EC backend.
663
* https://tracker.ceph.com/issues/51228
664
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
665
* https://tracker.ceph.com/issues/51282
666
    pybind/mgr/mgr_util: .mgr pool may be created to early causing spurious PG_DEGRADED warnings
667
668
669 12 Patrick Donnelly
h3. 2021 June 14
670
671
https://pulpito.ceph.com/pdonnell-2021-06-14_20:53:05-fs-wip-pdonnell-testing-20210614.173325-distro-basic-smithi/
672
673
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
674
675
* https://tracker.ceph.com/issues/51169
676
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
677
* https://tracker.ceph.com/issues/51228
678
    qa: rmdir: failed to remove 'a/.snap/*': No such file or directory
679
* https://tracker.ceph.com/issues/48773
680
    qa: scrub does not complete
681
* https://tracker.ceph.com/issues/51183
682
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
683
* https://tracker.ceph.com/issues/45434
684
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
685
* https://tracker.ceph.com/issues/51182
686
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
687
* https://tracker.ceph.com/issues/51229
688
    qa: test_multi_snap_schedule list difference failure
689
* https://tracker.ceph.com/issues/50821
690
    qa: untar_snap_rm failure during mds thrashing
691
692
693 11 Patrick Donnelly
h3. 2021 June 13
694
695
https://pulpito.ceph.com/pdonnell-2021-06-12_02:45:35-fs-wip-pdonnell-testing-20210612.002809-distro-basic-smithi/
696
697
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
698
699
* https://tracker.ceph.com/issues/51169
700
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
701
* https://tracker.ceph.com/issues/48773
702
    qa: scrub does not complete
703
* https://tracker.ceph.com/issues/51182
704
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
705
* https://tracker.ceph.com/issues/51183
706
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
707
* https://tracker.ceph.com/issues/51197
708
    qa: [WRN] Scrub error on inode 0x10000001520 (/client.0/tmp/t/linux-5.4/Documentation/driver-api) see mds.f log and `damage ls` output for details
709
* https://tracker.ceph.com/issues/45434
710
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
711
712 10 Patrick Donnelly
h3. 2021 June 11
713
714
https://pulpito.ceph.com/pdonnell-2021-06-11_18:02:10-fs-wip-pdonnell-testing-20210611.162716-distro-basic-smithi/
715
716
Some Ubuntu 20.04 upgrade fallout. In particular, upgrade tests are failing due to missing packages for 18.04 Pacific.
717
718
* https://tracker.ceph.com/issues/51169
719
    qa: ubuntu 20.04 sys protections prevent multiuser file access in /tmp
720
* https://tracker.ceph.com/issues/45434
721
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
722
* https://tracker.ceph.com/issues/48771
723
    qa: iogen: workload fails to cause balancing
724
* https://tracker.ceph.com/issues/43216
725
    MDSMonitor: removes MDS coming out of quorum election
726
* https://tracker.ceph.com/issues/51182
727
    pybind/mgr/snap_schedule: Invalid command: Unexpected argument 'fs=cephfs'
728
* https://tracker.ceph.com/issues/50223
729
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
730
* https://tracker.ceph.com/issues/48773
731
    qa: scrub does not complete
732
* https://tracker.ceph.com/issues/51183
733
    qa: FileNotFoundError: [Errno 2] No such file or directory: '/sys/kernel/debug/ceph/3fab6bea-f243-47a4-a956-8c03a62b61b5.client4721/mds_sessions'
734
* https://tracker.ceph.com/issues/51184
735
    qa: fs:bugs does not specify distro
736
737
738 9 Patrick Donnelly
h3. 2021 June 03
739
740
https://pulpito.ceph.com/pdonnell-2021-06-03_03:40:33-fs-wip-pdonnell-testing-20210603.020013-distro-basic-smithi/
741
742
* https://tracker.ceph.com/issues/45434
743
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
744
* https://tracker.ceph.com/issues/50016
745
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
746
* https://tracker.ceph.com/issues/50821
747
    qa: untar_snap_rm failure during mds thrashing
748
* https://tracker.ceph.com/issues/50622 (regression)
749
    msg: active_connections regression
750
* https://tracker.ceph.com/issues/49845#note-2 (regression)
751
    qa: failed umount in test_volumes
752
* https://tracker.ceph.com/issues/48773
753
    qa: scrub does not complete
754
* https://tracker.ceph.com/issues/43216
755
    MDSMonitor: removes MDS coming out of quorum election
756
757
758 7 Patrick Donnelly
h3. 2021 May 18
759
760 8 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.214114
761
762
Regression in testing kernel caused some failures. Ilya fixed those and rerun
763
looked better. Some odd new noise in the rerun relating to packaging and "No
764
module named 'tasks.ceph'".
765
766
* https://tracker.ceph.com/issues/50824
767
    qa: snaptest-git-ceph bus error
768
* https://tracker.ceph.com/issues/50622 (regression)
769
    msg: active_connections regression
770
* https://tracker.ceph.com/issues/49845#note-2 (regression)
771
    qa: failed umount in test_volumes
772
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
773
    qa: quota failure
774
775
776
h3. 2021 May 18
777
778 7 Patrick Donnelly
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210518.025642
779
780
* https://tracker.ceph.com/issues/50821
781
    qa: untar_snap_rm failure during mds thrashing
782
* https://tracker.ceph.com/issues/48773
783
    qa: scrub does not complete
784
* https://tracker.ceph.com/issues/45591
785
    mgr: FAILED ceph_assert(daemon != nullptr)
786
* https://tracker.ceph.com/issues/50866
787
    osd: stat mismatch on objects
788
* https://tracker.ceph.com/issues/50016
789
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
790
* https://tracker.ceph.com/issues/50867
791
    qa: fs:mirror: reduced data availability
792
* https://tracker.ceph.com/issues/50821
793
    qa: untar_snap_rm failure during mds thrashing
794
* https://tracker.ceph.com/issues/50622 (regression)
795
    msg: active_connections regression
796
* https://tracker.ceph.com/issues/50223
797
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
798
* https://tracker.ceph.com/issues/50868
799
    qa: "kern.log.gz already exists; not overwritten"
800
* https://tracker.ceph.com/issues/50870
801
    qa: test_full: "rm: cannot remove 'large_file_a': Permission denied"
802
803
804 6 Patrick Donnelly
h3. 2021 May 11
805
806
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210511.232042
807
808
* one class of failures caused by PR
809
* https://tracker.ceph.com/issues/48812
810
    qa: test_scrub_pause_and_resume_with_abort failure
811
* https://tracker.ceph.com/issues/50390
812
    mds: monclient: wait_auth_rotating timed out after 30
813
* https://tracker.ceph.com/issues/48773
814
    qa: scrub does not complete
815
* https://tracker.ceph.com/issues/50821
816
    qa: untar_snap_rm failure during mds thrashing
817
* https://tracker.ceph.com/issues/50224
818
    qa: test_mirroring_init_failure_with_recovery failure
819
* https://tracker.ceph.com/issues/50622 (regression)
820
    msg: active_connections regression
821
* https://tracker.ceph.com/issues/50825
822
    qa: snaptest-git-ceph hang during mon thrashing v2
823
* https://tracker.ceph.com/issues/50821
824
    qa: untar_snap_rm failure during mds thrashing
825
* https://tracker.ceph.com/issues/50823
826
    qa: RuntimeError: timeout waiting for cluster to stabilize
827
828
829 5 Patrick Donnelly
h3. 2021 May 14
830
831
https://pulpito.ceph.com/pdonnell-2021-05-14_21:45:42-fs-master-distro-basic-smithi/
832
833
* https://tracker.ceph.com/issues/48812
834
    qa: test_scrub_pause_and_resume_with_abort failure
835
* https://tracker.ceph.com/issues/50821
836
    qa: untar_snap_rm failure during mds thrashing
837
* https://tracker.ceph.com/issues/50622 (regression)
838
    msg: active_connections regression
839
* https://tracker.ceph.com/issues/50822
840
    qa: testing kernel patch for client metrics causes mds abort
841
* https://tracker.ceph.com/issues/48773
842
    qa: scrub does not complete
843
* https://tracker.ceph.com/issues/50823
844
    qa: RuntimeError: timeout waiting for cluster to stabilize
845
* https://tracker.ceph.com/issues/50824
846
    qa: snaptest-git-ceph bus error
847
* https://tracker.ceph.com/issues/50825
848
    qa: snaptest-git-ceph hang during mon thrashing v2
849
* https://tracker.ceph.com/issues/50826
850
    kceph: stock RHEL kernel hangs on snaptests with mon|osd thrashers
851
852
853 4 Patrick Donnelly
h3. 2021 May 01
854
855
https://pulpito.ceph.com/pdonnell-2021-05-01_09:07:09-fs-wip-pdonnell-testing-20210501.040415-distro-basic-smithi/
856
857
* https://tracker.ceph.com/issues/45434
858
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
859
* https://tracker.ceph.com/issues/50281
860
    qa: untar_snap_rm timeout
861
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
862
    qa: quota failure
863
* https://tracker.ceph.com/issues/48773
864
    qa: scrub does not complete
865
* https://tracker.ceph.com/issues/50390
866
    mds: monclient: wait_auth_rotating timed out after 30
867
* https://tracker.ceph.com/issues/50250
868
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
869
* https://tracker.ceph.com/issues/50622 (regression)
870
    msg: active_connections regression
871
* https://tracker.ceph.com/issues/45591
872
    mgr: FAILED ceph_assert(daemon != nullptr)
873
* https://tracker.ceph.com/issues/50221
874
    qa: snaptest-git-ceph failure in git diff
875
* https://tracker.ceph.com/issues/50016
876
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
877
878
879 3 Patrick Donnelly
h3. 2021 Apr 15
880
881
https://pulpito.ceph.com/pdonnell-2021-04-15_01:35:57-fs-wip-pdonnell-testing-20210414.230315-distro-basic-smithi/
882
883
* https://tracker.ceph.com/issues/50281
884
    qa: untar_snap_rm timeout
885
* https://tracker.ceph.com/issues/50220
886
    qa: dbench workload timeout
887
* https://tracker.ceph.com/issues/50246
888
    mds: failure replaying journal (EMetaBlob)
889
* https://tracker.ceph.com/issues/50250
890
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
891
* https://tracker.ceph.com/issues/50016
892
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
893
* https://tracker.ceph.com/issues/50222
894
    osd: 5.2s0 deep-scrub : stat mismatch
895
* https://tracker.ceph.com/issues/45434
896
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
897
* https://tracker.ceph.com/issues/49845
898
    qa: failed umount in test_volumes
899
* https://tracker.ceph.com/issues/37808
900
    osd: osdmap cache weak_refs assert during shutdown
901
* https://tracker.ceph.com/issues/50387
902
    client: fs/snaps failure
903
* https://tracker.ceph.com/issues/50389
904
    mds: "cluster [ERR] Error recovering journal 0x203: (2) No such file or directory" in cluster log"
905
* https://tracker.ceph.com/issues/50216
906
    qa: "ls: cannot access 'lost+found': No such file or directory"
907
* https://tracker.ceph.com/issues/50390
908
    mds: monclient: wait_auth_rotating timed out after 30
909
910
911
912 1 Patrick Donnelly
h3. 2021 Apr 08
913
914 2 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_22:42:24-fs-wip-pdonnell-testing-20210408.192301-distro-basic-smithi/
915
916
* https://tracker.ceph.com/issues/45434
917
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
918
* https://tracker.ceph.com/issues/50016
919
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
920
* https://tracker.ceph.com/issues/48773
921
    qa: scrub does not complete
922
* https://tracker.ceph.com/issues/50279
923
    qa: "Replacing daemon mds.b as rank 0 with standby daemon mds.c"
924
* https://tracker.ceph.com/issues/50246
925
    mds: failure replaying journal (EMetaBlob)
926
* https://tracker.ceph.com/issues/48365
927
    qa: ffsb build failure on CentOS 8.2
928
* https://tracker.ceph.com/issues/50216
929
    qa: "ls: cannot access 'lost+found': No such file or directory"
930
* https://tracker.ceph.com/issues/50223
931
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
932
* https://tracker.ceph.com/issues/50280
933
    cephadm: RuntimeError: uid/gid not found
934
* https://tracker.ceph.com/issues/50281
935
    qa: untar_snap_rm timeout
936
937
h3. 2021 Apr 08
938
939 1 Patrick Donnelly
https://pulpito.ceph.com/pdonnell-2021-04-08_04:31:36-fs-wip-pdonnell-testing-20210408.024225-distro-basic-smithi/
940
https://pulpito.ceph.com/?branch=wip-pdonnell-testing-20210408.142238 (with logic inversion / QA fix)
941
942
* https://tracker.ceph.com/issues/50246
943
    mds: failure replaying journal (EMetaBlob)
944
* https://tracker.ceph.com/issues/50250
945
    mds: "log [WRN] : Scrub error on inode 0x10000004506 (/client.0/tmp/clients/client3/~dmtmp/COREL) see mds.a log and `damage ls` output for details"
946
947
948
h3. 2021 Apr 07
949
950
https://pulpito.ceph.com/pdonnell-2021-04-07_02:12:41-fs-wip-pdonnell-testing-20210406.213012-distro-basic-smithi/
951
952
* https://tracker.ceph.com/issues/50215
953
    qa: "log [ERR] : error reading sessionmap 'mds2_sessionmap'"
954
* https://tracker.ceph.com/issues/49466
955
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
956
* https://tracker.ceph.com/issues/50216
957
    qa: "ls: cannot access 'lost+found': No such file or directory"
958
* https://tracker.ceph.com/issues/48773
959
    qa: scrub does not complete
960
* https://tracker.ceph.com/issues/49845
961
    qa: failed umount in test_volumes
962
* https://tracker.ceph.com/issues/50220
963
    qa: dbench workload timeout
964
* https://tracker.ceph.com/issues/50221
965
    qa: snaptest-git-ceph failure in git diff
966
* https://tracker.ceph.com/issues/50222
967
    osd: 5.2s0 deep-scrub : stat mismatch
968
* https://tracker.ceph.com/issues/50223
969
    qa: "client.4737 isn't responding to mclientcaps(revoke)"
970
* https://tracker.ceph.com/issues/50224
971
    qa: test_mirroring_init_failure_with_recovery failure
972
973
h3. 2021 Apr 01
974
975
https://pulpito.ceph.com/pdonnell-2021-04-01_00:45:34-fs-wip-pdonnell-testing-20210331.222326-distro-basic-smithi/
976
977
* https://tracker.ceph.com/issues/48772
978
    qa: pjd: not ok 9, 44, 80
979
* https://tracker.ceph.com/issues/50177
980
    osd: "stalled aio... buggy kernel or bad device?"
981
* https://tracker.ceph.com/issues/48771
982
    qa: iogen: workload fails to cause balancing
983
* https://tracker.ceph.com/issues/49845
984
    qa: failed umount in test_volumes
985
* https://tracker.ceph.com/issues/48773
986
    qa: scrub does not complete
987
* https://tracker.ceph.com/issues/48805
988
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
989
* https://tracker.ceph.com/issues/50178
990
    qa: "TypeError: run() got an unexpected keyword argument 'shell'"
991
* https://tracker.ceph.com/issues/45434
992
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
993
994
h3. 2021 Mar 24
995
996
https://pulpito.ceph.com/pdonnell-2021-03-24_23:26:35-fs-wip-pdonnell-testing-20210324.190252-distro-basic-smithi/
997
998
* https://tracker.ceph.com/issues/49500
999
    qa: "Assertion `cb_done' failed."
1000
* https://tracker.ceph.com/issues/50019
1001
    qa: mount failure with cephadm "probably no MDS server is up?"
1002
* https://tracker.ceph.com/issues/50020
1003
    qa: "RADOS object not found (Failed to operate read op for oid cephfs_mirror)"
1004
* https://tracker.ceph.com/issues/48773
1005
    qa: scrub does not complete
1006
* https://tracker.ceph.com/issues/45434
1007
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1008
* https://tracker.ceph.com/issues/48805
1009
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1010
* https://tracker.ceph.com/issues/48772
1011
    qa: pjd: not ok 9, 44, 80
1012
* https://tracker.ceph.com/issues/50021
1013
    qa: snaptest-git-ceph failure during mon thrashing
1014
* https://tracker.ceph.com/issues/48771
1015
    qa: iogen: workload fails to cause balancing
1016
* https://tracker.ceph.com/issues/50016
1017
    qa: test_damage: "RuntimeError: 2 mutations had unexpected outcomes"
1018
* https://tracker.ceph.com/issues/49466
1019
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1020
1021
1022
h3. 2021 Mar 18
1023
1024
https://pulpito.ceph.com/pdonnell-2021-03-18_13:46:31-fs-wip-pdonnell-testing-20210318.024145-distro-basic-smithi/
1025
1026
* https://tracker.ceph.com/issues/49466
1027
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1028
* https://tracker.ceph.com/issues/48773
1029
    qa: scrub does not complete
1030
* https://tracker.ceph.com/issues/48805
1031
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1032
* https://tracker.ceph.com/issues/45434
1033
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1034
* https://tracker.ceph.com/issues/49845
1035
    qa: failed umount in test_volumes
1036
* https://tracker.ceph.com/issues/49605
1037
    mgr: drops command on the floor
1038
* https://tracker.ceph.com/issues/48203 (stock kernel update required)
1039
    qa: quota failure
1040
* https://tracker.ceph.com/issues/49928
1041
    client: items pinned in cache preventing unmount x2
1042
1043
h3. 2021 Mar 15
1044
1045
https://pulpito.ceph.com/pdonnell-2021-03-15_22:16:56-fs-wip-pdonnell-testing-20210315.182203-distro-basic-smithi/
1046
1047
* https://tracker.ceph.com/issues/49842
1048
    qa: stuck pkg install
1049
* https://tracker.ceph.com/issues/49466
1050
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1051
* https://tracker.ceph.com/issues/49822
1052
    test: test_mirroring_command_idempotency (tasks.cephfs.test_admin.TestMirroringCommands) failure
1053
* https://tracker.ceph.com/issues/49240
1054
    terminate called after throwing an instance of 'std::bad_alloc'
1055
* https://tracker.ceph.com/issues/48773
1056
    qa: scrub does not complete
1057
* https://tracker.ceph.com/issues/45434
1058
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1059
* https://tracker.ceph.com/issues/49500
1060
    qa: "Assertion `cb_done' failed."
1061
* https://tracker.ceph.com/issues/49843
1062
    qa: fs/snaps/snaptest-upchildrealms.sh failure
1063
* https://tracker.ceph.com/issues/49845
1064
    qa: failed umount in test_volumes
1065
* https://tracker.ceph.com/issues/48805
1066
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1067
* https://tracker.ceph.com/issues/49605
1068
    mgr: drops command on the floor
1069
1070
and failure caused by PR: https://github.com/ceph/ceph/pull/39969
1071
1072
1073
h3. 2021 Mar 09
1074
1075
https://pulpito.ceph.com/pdonnell-2021-03-09_03:27:39-fs-wip-pdonnell-testing-20210308.214827-distro-basic-smithi/
1076
1077
* https://tracker.ceph.com/issues/49500
1078
    qa: "Assertion `cb_done' failed."
1079
* https://tracker.ceph.com/issues/48805
1080
    mds: "cluster [WRN] Scrub error on inode 0x1000000039d (/client.0/tmp/blogbench-1.0/src/blogtest_in) see mds.a log and `damage ls` output for details"
1081
* https://tracker.ceph.com/issues/48773
1082
    qa: scrub does not complete
1083
* https://tracker.ceph.com/issues/45434
1084
    qa: test_full_fsync (tasks.cephfs.test_full.TestClusterFull) failed
1085
* https://tracker.ceph.com/issues/49240
1086
    terminate called after throwing an instance of 'std::bad_alloc'
1087
* https://tracker.ceph.com/issues/49466
1088
    qa: "Command failed on gibba030 with status 1: 'set -ex\nsudo dd of=/tmp/tmp.ZEeZBasJer'"
1089
* https://tracker.ceph.com/issues/49684
1090
    qa: fs:cephadm mount does not wait for mds to be created
1091
* https://tracker.ceph.com/issues/48771
1092
    qa: iogen: workload fails to cause balancing