Project

General

Profile

Actions

Bug #53807

closed

Dead jobs in rados/cephadm/smoke-roleless{...}: ingress jobs stuck

Added by Laura Flores over 2 years ago. Updated about 2 years ago.

Status:
Resolved
Priority:
Immediate
Category:
-
Target version:
-
% Done:

0%

Source:
Tags:
Backport:
quincy,pacific,octopus
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

Description: rados/cephadm/smoke-roleless/{0-distro/centos_8.3_container_tools_3.0 0-nvme-loop 1-start 2-services/nfs-ingress 3-final}

Failure Reason: hit max job timeout

Jobs:
/a/yuriw-2022-01-06_15:50:38-rados-wip-yuri8-testing-2022-01-05-1411-distro-default-smithi/6598774
/a/yuriw-2022-01-06_15:50:38-rados-wip-yuri8-testing-2022-01-05-1411-distro-default-smithi/6598785
/a/yuriw-2022-01-06_15:50:38-rados-wip-yuri8-testing-2022-01-05-1411-distro-default-smithi/6599316
/a/yuriw-2022-01-06_15:50:38-rados-wip-yuri8-testing-2022-01-05-1411-distro-default-smithi/6599350

Earlier in the log:

2022-01-06T16:33:28.615 INFO:teuthology.task.ansible.out:^M
TASK [common : Check firewalld status] *****************************************^M

2022-01-06T16:33:28.617 INFO:teuthology.task.ansible.out:fatal: [smithi107.front.sepia.ceph.com]: FAILED! => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true}^M
...ignoring^M

2022-01-06T16:33:28.638 INFO:teuthology.task.ansible.out:Thursday 06 January 2022  16:33:28 +0000 (0:00:00.260)       0:02:03.410 ****** ^M

Later in the log:

2022-01-06T16:44:13.943 INFO:journalctl@ceph.mon.smithi107.smithi107.stdout:Jan 06 16:44:13 smithi107 ceph-mon[30485]: from='mgr.14216 172.21.15.107:0/4007838189' entity='mgr.smithi107.tttsho' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]: dispatch
2022-01-06T16:44:13.943 INFO:journalctl@ceph.mon.smithi107.smithi107.stdout:Jan 06 16:44:13 smithi107 ceph-mon[30485]: pgmap v133: 33 pgs: 32 unknown, 1 active+clean; 577 KiB data, 47 MiB used, 715 GiB / 715 GiB avail
2022-01-06T16:44:13.943 INFO:journalctl@ceph.mon.smithi107.smithi107.stdout:Jan 06 16:44:13 smithi107 ceph-mon[30485]: from='mgr.14216 172.21.15.107:0/4007838189' entity='mgr.smithi107.tttsho'
2022-01-06T16:44:13.943 INFO:journalctl@ceph.mon.smithi107.smithi107.stdout:Jan 06 16:44:13 smithi107 conmon[30462]: 2022-01-06T16:44:13.732+0000 7fa865de5700 -1 log_channel(cluster) log [ERR] : Health check failed: 1 filesystem is offline (MDS_ALL_DOWN)
2022-01-06T16:44:14.150 INFO:journalctl@ceph.mon.smithi150.smithi150.stdout:Jan 06 16:44:13 smithi150 ceph-mon[37940]: from='mgr.14216 172.21.15.107:0/4007838189' entity='mgr.smithi107.tttsho' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]': finished
2022-01-06T16:44:14.151 INFO:journalctl@ceph.mon.smithi150.smithi150.stdout:Jan 06 16:44:13 smithi150 ceph-mon[37940]: osdmap e43: 8 total, 8 up, 8 in
2022-01-06T16:44:14.151 INFO:journalctl@ceph.mon.smithi150.smithi150.stdout:Jan 06 16:44:13 smithi150 ceph-mon[37940]: from='mgr.14216 172.21.15.107:0/4007838189' entity='mgr.smithi107.tttsho' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]: dispatch
2022-01-06T16:44:14.151 INFO:journalctl@ceph.mon.smithi150.smithi150.stdout:Jan 06 16:44:13 smithi150 ceph-mon[37940]: pgmap v133: 33 pgs: 32 unknown, 1 active+clean; 577 KiB data, 47 MiB used, 715 GiB / 715 GiB avail
2022-01-06T16:44:14.151 INFO:journalctl@ceph.mon.smithi150.smithi150.stdout:Jan 06 16:44:13 smithi150 ceph-mon[37940]: from='mgr.14216 172.21.15.107:0/4007838189' entity='mgr.smithi107.tttsho'
2022-01-06T16:44:14.535 INFO:teuthology.run_tasks:Running task cephadm.apply...


Related issues 1 (0 open1 closed)

Has duplicate Orchestrator - Bug #53904: cephadm: ingress jobs stuckDuplicateMelissa Li

Actions
Actions

Also available in: Atom PDF