Project

General

Profile

Actions

Bug #53097

closed

"Failed to apply 4 service" in upgrade:octopus-x-master

Added by Yuri Weinstein over 2 years ago. Updated over 2 years ago.

Status:
Resolved
Priority:
Normal
Category:
-
Target version:
-
% Done:

0%

Source:
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

Run: https://pulpito.ceph.com/nojha-2021-10-28_22:29:54-upgrade:octopus-x-master-distro-basic-smithi
Logs: https://pulpito.ceph.com/nojha-2021-10-28_22:29:54-upgrade:octopus-x-master-distro-basic-smithi/6465645/

2021-10-28T23:04:58.542 INFO:journalctl@ceph.mon.b.smithi112.stdout:Oct 28 23:04:58 smithi112 ceph-mon[65235]: Health detail: HEALTH_WARN Failed to apply 4 service(s): alertmanager,grafana,node-exporter,prometheus
2021-10-28T23:04:58.542 INFO:journalctl@ceph.mon.b.smithi112.stdout:Oct 28 23:04:58 smithi112 ceph-mon[65235]: [WRN] CEPHADM_APPLY_SPEC_FAIL: Failed to apply 4 service(s): alertmanager,grafana,node-exporter,prometheus

also:

2021-10-28T23:02:48.525 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]: 2021-10-28T23:02:48.026+0000 7fb3ed83a700 -1 log_channel(cephadm) log [ERR] : Failed to apply alertmanager spec AlertManagerSpec({'placement': PlacementSpec(count=1, hosts=[HostPlacementSpec(hostname='smithi033', network='', name='a')]), 'service_type': 'alertmanager', 'service_id': None, 'unmanaged': False, 'preview_only': False, 'networks': [], 'config': None, 'user_data': {}, 'port': None}): name alertmanager.a already in use
2021-10-28T23:02:48.525 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]: Traceback (most recent call last):
2021-10-28T23:02:48.526 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]:   File "/usr/share/ceph/mgr/cephadm/serve.py", line 545, in _apply_all_services
2021-10-28T23:02:48.526 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]:     if self._apply_service(spec):
2021-10-28T23:02:48.526 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]:   File "/usr/share/ceph/mgr/cephadm/serve.py", line 747, in _apply_service
2021-10-28T23:02:48.526 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]:     rank_generation=slot.rank_generation,
2021-10-28T23:02:48.527 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]:   File "/usr/share/ceph/mgr/cephadm/module.py", line 636, in get_unique_name
2021-10-28T23:02:48.527 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]:     f'name {daemon_type}.{forcename} already in use')
2021-10-28T23:02:48.527 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]: orchestrator._interface.OrchestratorValidationError: name alertmanager.a already in use
2021-10-28T23:02:48.527 INFO:journalctl@ceph.mgr.y.smithi033.stdout:Oct 28 23:02:48 smithi033 conmon[81907]: 2021-10-28T23:02:48.027+0000 7fb3ed83a700 -1 log_channel(cephadm) log [ERR] : Failed to apply grafana spec MonitoringSpec({'placement': PlacementSpec(count=1, hosts=[HostPlacementSpec(hostname='smithi112', network='', name='a')]), 'service_type': 'grafana', 'service_id': None, 'unmanaged': False, 'preview_only': False, 'networks': [], 'config': None, 'port': None}): name grafana.a already in use

Actions #1

Updated by Yuri Weinstein over 2 years ago

  • Tracker changed from Tasks to Bug
  • Project changed from Stable releases to Ceph
  • Regression set to No
  • Severity set to 3 - minor
Actions #2

Updated by Neha Ojha over 2 years ago

  • Project changed from Ceph to Orchestrator
Actions #4

Updated by Daniel Pivonka over 2 years ago

https://github.com/ceph/ceph/pull/43376 seems to have made this issue more visible but the same 'name alertmanager.a already in use' can be seen in a upgrade from octopus to pacific

Actions #5

Updated by Sebastian Wagner over 2 years ago

  • Status changed from New to In Progress
  • Assignee set to Daniel Pivonka
Actions #6

Updated by Neha Ojha over 2 years ago

  • Status changed from In Progress to Fix Under Review
  • Pull request ID set to 43825
Actions #7

Updated by Sebastian Wagner over 2 years ago

  • Status changed from Fix Under Review to Resolved
Actions

Also available in: Atom PDF