Ceph : Issues
https://tracker.ceph.com/
https://tracker.ceph.com/favicon.ico
2022-01-17T16:07:38Z
Ceph
Redmine
Orchestrator - Bug #53904 (Duplicate): cephadm: ingress jobs stuck
https://tracker.ceph.com/issues/53904
2022-01-17T16:07:38Z
Sebastian Wagner
<p><a class="external" href="https://pulpito.ceph.com/swagner-2022-01-17_12:42:04-orch:cephadm-wip-swagner-testing-2022-01-17-1014-distro-default-smithi/">https://pulpito.ceph.com/swagner-2022-01-17_12:42:04-orch:cephadm-wip-swagner-testing-2022-01-17-1014-distro-default-smithi/</a></p>
<pre>
2022-01-17T13:17:17.053 DEBUG:teuthology.orchestra.run.smithi155:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:1cdf02ebbbdd98a055173cbac4d0171328a564dc shell -c /etc/ceph/ceph.conf -k />
2022-01-17T13:17:17.054 DEBUG:teuthology.orchestra.run.smithi155:> for haproxy in `ceph orch ps | grep ^haproxy.nfs.foo. | awk '"'"'{print $1}'"'"'`; do
2022-01-17T13:17:17.054 DEBUG:teuthology.orchestra.run.smithi155:> ceph orch daemon stop $haproxy
2022-01-17T13:17:17.054 DEBUG:teuthology.orchestra.run.smithi155:> while ! ceph orch ps | grep $haproxy | grep stopped; do sleep 1 ; done
2022-01-17T13:17:17.055 DEBUG:teuthology.orchestra.run.smithi155:> cat /mnt/foo/testfile
2022-01-17T13:17:17.055 DEBUG:teuthology.orchestra.run.smithi155:> echo $haproxy > /mnt/foo/testfile
2022-01-17T13:17:17.055 DEBUG:teuthology.orchestra.run.smithi155:> sync
2022-01-17T13:17:17.055 DEBUG:teuthology.orchestra.run.smithi155:> ceph orch daemon start $haproxy
2022-01-17T13:17:17.056 DEBUG:teuthology.orchestra.run.smithi155:> while ! ceph orch ps | grep $haproxy | grep running; do sleep 1 ; done
2022-01-17T13:17:17.056 DEBUG:teuthology.orchestra.run.smithi155:> done
2022-01-17T13:17:17.056 DEBUG:teuthology.orchestra.run.smithi155:> '
</pre><br />...snip...<br /><pre>
2022-01-17T13:17:20.571 INFO:teuthology.orchestra.run.smithi155.stdout:Check with each haproxy down in turn...
2022-01-17T13:17:21.281 INFO:teuthology.orchestra.run.smithi155.stdout:Scheduled to stop haproxy.nfs.foo.smithi155.xhswck on host 'smithi155'
</pre><br />...snip...
<pre>
2022-01-17T13:17:36.893 INFO:teuthology.orchestra.run.smithi155.stdout:haproxy.nfs.foo.smithi155.xhswck smithi155 *:2049,9002 stopped 0s ago 79s - - <unknown> <un>
2022-01-17T13:17:36.898 INFO:teuthology.orchestra.run.smithi155.stdout:test
2022-01-17T13:17:37.528 INFO:teuthology.orchestra.run.smithi155.stdout:Scheduled to start haproxy.nfs.foo.smithi155.xhswck on host 'smithi155'
</pre><br />...snip...<br /><pre>
2022-01-17T13:17:53.182 INFO:teuthology.orchestra.run.smithi155.stdout:haproxy.nfs.foo.smithi155.xhswck smithi155 *:2049,9002 running (5s) 0s ago 95s - - 2.3.17-d1c9119 14b>
2022-01-17T13:17:53.519 INFO:teuthology.orchestra.run.smithi155.stdout:Scheduled to stop haproxy.nfs.foo.smithi162.mahcqs on host 'smithi162'
</pre><br />...snip...<br /><pre>
2022-01-17T13:18:07.810 INFO:teuthology.orchestra.run.smithi155.stdout:haproxy.nfs.foo.smithi162.mahcqs smithi162 *:2049,9002 stopped 0s ago 102s - - <unknown> <unk>
</pre><br />...snip..<br /><pre>
h[14066]: cephadm 2022-01-17T13:17:53.516345+0000 mgr.smithi155.uoijyc (mgr.14206) 339 : cephadm [INF] Schedule stop daemon haproxy.nfs.foo.smithi162.mahcqs
</pre>
<p>But I never see a start of haproxy.nfs.foo.smithi162.mahcqs again.</p>
Orchestrator - Bug #53321 (Duplicate): cephadm tries to use the system disk for osd specs
https://tracker.ceph.com/issues/53321
2021-11-18T15:31:20Z
Sebastian Wagner
<p>Having this spec:</p>
<pre><code class="yaml syntaxhl"><span class="CodeRay"><span class="key">service_type</span>: <span class="string"><span class="content">osd</span></span>
<span class="key">service_id</span>: <span class="string"><span class="content">hybrid</span></span>
<span class="key">service_name</span>: <span class="string"><span class="content">osd.hybrid</span></span>
<span class="key">placement</span>:
<span class="key">host_pattern</span>: <span class="string"><span class="content">host1</span></span>
<span class="key">spec</span>:
<span class="key">data_devices</span>:
<span class="key">rotational</span>: <span class="string"><span class="content">1</span></span>
<span class="key">db_devices</span>:
<span class="key">rotational</span>: <span class="string"><span class="content">0</span></span>
<span class="key">filter_logic</span>: <span class="string"><span class="content">AND</span></span>
<span class="key">objectstore</span>: <span class="string"><span class="content">bluestore</span></span>
</span></code></pre>
<p>And having the system partition being locked:</p>
<pre>
# ceph orch device ls host1
HOST PATH TYPE DEVICE ID SIZE AVAILABLE REJECT REASONS
host1 /dev/nvme0n1 ssd NVMENVMENVMENVMENVMENVMENVMENVME1 1600G Yes
host1 /dev/nvme1n1 ssd NVMENVMENVMENVMENVMENVMENVMENVME4 1600G Yes
host1 /dev/sda hdd IDSDM_012345678901 64.2G Has GPT headers, locked
host1 /dev/sdb hdd AAAAAAAAAAA_0000000000000b0f1 16.0T Yes
host1 /dev/sdc hdd AAAAAAAAAAA_000000000000089cd 16.0T Yes
host1 /dev/sdd hdd AAAAAAAAAAA_0000000000000af6d 16.0T Yes
host1 /dev/sde hdd AAAAAAAAAAA_00000000000008a4d 16.0T Yes
host1 /dev/sdf hdd AAAAAAAAAAA_0000000000000af9d 16.0T Yes
host1 /dev/sdg hdd BBBBBBBBBBBb_00000000000044a5 16.0T Yes
host1 /dev/sdh hdd BBBBBBBBBBBb_000000000000f7f9 16.0T Yes
host1 /dev/sdi hdd AAAAAAAAAAA_000000000000089a1 16.0T Yes
host1 /dev/sdj hdd AAAAAAAAAAA_00000000000008601 16.0T Yes
host1 /dev/sdk hdd AAAAAAAAAAA_00000000000008a71 16.0T Yes
host1 /dev/sdl hdd CCCCCCCCCCCC_000000000000ebdd 16.0T Yes
host1 /dev/sdm hdd AAAAAAAAAAA_000000000000089bd 16.0T Yes
host1 /dev/sdn hdd AAAAAAAAAAA_0000000000000fd31 16.0T Yes
host1 /dev/sdo hdd AAAAAAAAAAA_0000000000000f9a9 16.0T Yes
host1 /dev/sdp hdd AAAAAAAAAAA_00000000000008565 16.0T Yes
host1 /dev/sdq hdd BBBBBBBBBBBb_000000000000f3e5 16.0T Yes
host1 /dev/sdr hdd MG08SCA16TEY_5000039aa858002d 16.0T Yes
host1 /dev/sds hdd AAAAAAAAAAA_0000000000000fa61 16.0T Yes
host1 /dev/sdt hdd BBBBBBBBBBBb_00000000000046a5 16.0T Yes
host1 /dev/sdu hdd BBBBBBBBBBBb_00000000000041a1 16.0T Yes
host1 /dev/sdv hdd BBBBBBBBBBBb_000000000000f46d 16.0T Yes
host1 /dev/sdw hdd BBBBBBBBBBBb_00000000000046a9 16.0T Yes
host1 /dev/sdx hdd CCCCCCCCCCCC_000000000000eec1 16.0T Yes
host1 /dev/sdy hdd BBBBBBBBBBBb_0000000000004509 16.0T Yes
# mount -l | grep sda
/dev/sda2 on / type ext4 (rw,relatime) [root]
/dev/sda1 on /boot/efi type vfat (rw,relatime,fmask=0022,dmask=0022,codepage=437,iocharset=iso8859-1,shortname=mixed,errors=remount-ro) [boot-efi]
</pre>
<p>It fails to apply it:</p>
<pre>
# ceph orch ls --format yaml --service-type osd | python3 -c 'import sys, yaml, json; y=yaml.safe_load_all(sys.stdin.read()); print(json.dumps(list(y)))' | jq -r .[0].events[0]
2021-11-18T14:09:26.296374Z service:osd.hybrid [ERROR] "Failed to apply: cephadm exited with an error code: 1, stderr:Non-zero exit code 2 from /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --net=host --entrypoint /usr/sbin/ceph-volume --privileged --group-add=disk --init -e CONTAINER_IMAGE=quay.ceph.io/ceph-ci/ceph@sha256:sha1 -e NODE_NAME=host1 -e CEPH_USE_RANDOM_NONCE=1 -e CEPH_VOLUME_OSDSPEC_AFFINITY=hybrid -v /var/run/ceph/fsid:/var/run/ceph:z -v /var/log/ceph/fsid:/var/log/ceph:z -v /var/lib/ceph/fsid/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs -v /tmp/ceph-tmp1gohyav7:/etc/ceph/ceph.conf:z -v /tmp/ceph-tmpblgd2lz7:/var/lib/ceph/bootstrap-osd/ceph.keyring:z quay.ceph.io/ceph-ci/ceph@sha256:sha1 lvm batch --no-auto /dev/sda /dev/sdb /dev/sdc /dev/sdd /dev/sde /dev/sdf /dev/sdg /dev/sdh /dev/sdi /dev/sdj /dev/sdk /dev/sdl /dev/sdm /dev/sdn /dev/sdo /dev/sdp /dev/sdq /dev/sdr /dev/sds /dev/sdt /dev/sdu /dev/sdv /dev/sdw /dev/sdx /dev/sdy --db-devices /dev/nvme0n1 /dev/nvme1n1 --yes --no-systemd
/usr/bin/docker: stderr usage: ceph-volume lvm batch [-h] [--db-devices [DB_DEVICES [DB_DEVICES ...]]]
/usr/bin/docker: stderr [--wal-devices [WAL_DEVICES [WAL_DEVICES ...]]]
/usr/bin/docker: stderr [--journal-devices [JOURNAL_DEVICES [JOURNAL_DEVICES ...]]]
/usr/bin/docker: stderr [--auto] [--no-auto] [--bluestore] [--filestore]
/usr/bin/docker: stderr [--report] [--yes]
/usr/bin/docker: stderr [--format {json,json-pretty,pretty}] [--dmcrypt]
/usr/bin/docker: stderr [--crush-device-class CRUSH_DEVICE_CLASS]
/usr/bin/docker: stderr [--no-systemd]
/usr/bin/docker: stderr [--osds-per-device OSDS_PER_DEVICE]
/usr/bin/docker: stderr [--data-slots DATA_SLOTS]
/usr/bin/docker: stderr [--data-allocate-fraction DATA_ALLOCATE_FRACTION]
/usr/bin/docker: stderr [--block-db-size BLOCK_DB_SIZE]
/usr/bin/docker: stderr [--block-db-slots BLOCK_DB_SLOTS]
/usr/bin/docker: stderr [--block-wal-size BLOCK_WAL_SIZE]
/usr/bin/docker: stderr [--block-wal-slots BLOCK_WAL_SLOTS]
/usr/bin/docker: stderr [--journal-size JOURNAL_SIZE]
/usr/bin/docker: stderr [--journal-slots JOURNAL_SLOTS] [--prepare]
/usr/bin/docker: stderr [--osd-ids [OSD_IDS [OSD_IDS ...]]]
/usr/bin/docker: stderr [DEVICES [DEVICES ...]]
/usr/bin/docker: stderr ceph-volume lvm batch: error: GPT headers found, they must be removed on: /dev/sda
Traceback (most recent call last):
File "/var/lib/ceph/fsid/cephadm.hash", line 8331, in <module>
main()
File "/var/lib/ceph/fsid/cephadm.hash", line 8319, in main
r = ctx.func(ctx)
File "/var/lib/ceph/fsid/cephadm.hash", line 1735, in _infer_config
return func(ctx)
File "/var/lib/ceph/fsid/cephadm.hash", line 1676, in _infer_fsid
return func(ctx)
File "/var/lib/ceph/fsid/cephadm.hash", line 1763, in _infer_image
return func(ctx)
File "/var/lib/ceph/fsid/cephadm.hash", line 1663, in _validate_fsid
return func(ctx)
File "/var/lib/ceph/fsid/cephadm.hash", line 5285, in command_ceph_volume
out, err, code = call_throws(ctx, c.run_cmd())
File "/var/lib/ceph/fsid/cephadm.hash", line 1465, in call_throws
raise RuntimeError('Failed command: %s' % ' '.join(command))
RuntimeError: Failed command: /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --net=host --entrypoint /usr/sbin/ceph-volume --privileged --group-add=disk --init -e CONTAINER_IMAGE=quay.ceph.io/ceph-ci/ceph@sha256:sha1 -e NODE_NAME=host1 -e CEPH_USE_RANDOM_NONCE=1 -e CEPH_VOLUME_OSDSPEC_AFFINITY=hybrid -v /var/run/ceph/fsid:/var/run/ceph:z -v /var/log/ceph/fsid:/var/log/ceph:z -v /var/lib/ceph/fsid/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs -v /tmp/ceph-tmp1gohyav7:/etc/ceph/ceph.conf:z -v /tmp/ceph-tmpblgd2lz7:/var/lib/ceph/bootstrap-osd/ceph.keyring:z quay.ceph.io/ceph-ci/ceph@sha256:sha1 lvm batch --no-auto /dev/sda /dev/sdb /dev/sdc /dev/sdd /dev/sde /dev/sdf /dev/sdg /dev/sdh /dev/sdi /dev/sdj /dev/sdk /dev/sdl /dev/sdm /dev/sdn /dev/sdo /dev/sdp /dev/sdq /dev/sdr /dev/sds /dev/sdt /dev/sdu /dev/sdv /dev/sdw /dev/sdx /dev/sdy --db-devices /dev/nvme0n1 /dev/nvme1n1 --yes --no-systemd"
</pre>
mgr - Bug #52603 (Duplicate): pacific ERROR: test_diskprediction_local (tasks.mgr.test_module_sel...
https://tracker.ceph.com/issues/52603
2021-09-14T11:29:31Z
Sebastian Wagner
<pre>
2021-09-12T15:40:17.598 INFO:teuthology.orchestra.run.smithi033.stderr:2021-09-12T15:40:17.603+0000 7f11dac3c700 20 mgrc start_command cmd: [{"prefix": "mgr self-test module", "module": "diskprediction_local", >
2021-09-12T15:40:17.598 INFO:teuthology.orchestra.run.smithi033.stderr:2021-09-12T15:40:17.603+0000 7f11dac3c700 1 -- 172.21.15.33:0/2949193497 --> [v2:172.21.15.200:6824/18725,v1:172.21.15.200:6825/18725] -- >
2021-09-12T15:40:17.599 INFO:tasks.ceph.mgr.z.smithi200.stderr:2021-09-12T15:40:17.607+0000 7f0443f9c700 -1 no module 'diskprediction_local'
2021-09-12T15:40:17.600 INFO:tasks.ceph.mgr.z.smithi200.stderr:2021-09-12T15:40:17.607+0000 7f0443f9c700 -1 mgr handle_command module 'selftest' command handler threw exception: Module not found
2021-09-12T15:40:17.601 INFO:tasks.ceph.mgr.z.smithi200.stderr:2021-09-12T15:40:17.607+0000 7f0443f9c700 -1 mgr.server reply reply (22) Invalid argument Traceback (most recent call last):
2021-09-12T15:40:17.601 INFO:tasks.ceph.mgr.z.smithi200.stderr: File "/usr/share/ceph/mgr/mgr_module.py", line 1353, in _handle_command
2021-09-12T15:40:17.602 INFO:tasks.ceph.mgr.z.smithi200.stderr: return self.handle_command(inbuf, cmd)
2021-09-12T15:40:17.602 INFO:tasks.ceph.mgr.z.smithi200.stderr: File "/usr/share/ceph/mgr/selftest/module.py", line 142, in handle_command
2021-09-12T15:40:17.602 INFO:tasks.ceph.mgr.z.smithi200.stderr: r = self.remote(command['module'], "self_test")
2021-09-12T15:40:17.602 INFO:tasks.ceph.mgr.z.smithi200.stderr: File "/usr/share/ceph/mgr/mgr_module.py", line 1738, in remote
2021-09-12T15:40:17.603 INFO:tasks.ceph.mgr.z.smithi200.stderr: return self._ceph_dispatch_remote(module_name, method_name,
2021-09-12T15:40:17.603 INFO:tasks.ceph.mgr.z.smithi200.stderr:ImportError: Module not found
2021-09-12T15:40:17.603 INFO:tasks.ceph.mgr.z.smithi200.stderr:
</pre>
<p><a class="external" href="https://pulpito.ceph.com/yuriw-2021-09-12_15:18:23-rados-pacific-distro-basic-smithi/6386597">https://pulpito.ceph.com/yuriw-2021-09-12_15:18:23-rados-pacific-distro-basic-smithi/6386597</a></p>
mgr - Bug #51564 (Duplicate): Dashboard URL address is incorrect after bootstrap with IPV6 address
https://tracker.ceph.com/issues/51564
2021-07-07T12:27:06Z
Sebastian Wagner
<p>Description of problem:<br />After cephadm bootstrap with IPV6 address as mon ip gives the dashboard url which is incorrect,</p>
<p>It got missed square brackets</p>
<pre>
[ceph: root@magna081 /]# ceph mgr services
{
"dashboard": "https://2620:52:0:880:225:90ff:fefc:2536:8443/",
"prometheus": "http://2620:52:0:880:225:90ff:fefc:2536:9283/"
}
[root@magna081 ~]# curl -k https://2620:52:0:880:225:90ff:fefc:2536:8443/
curl: (3) Port number ended with ':'
</pre>
<p>working URL</p>
<pre>
[root@magna081 ~]# curl -k https://[2620:52:0:880:225:90ff:fefc:2536]:8443/
<!doctype html>
<html lang="en-US">
<head>
<meta charset="utf-8">
<title>Red Hat Ceph Storage</title>
<script>
document.write('<base href="' + document.location+ '" />');
</script>
<meta name="viewport" content="width=device-width, initial-scale=1, shrink-to-fit=no">
<link rel="icon" type="image/x-icon" id="cdFavicon" href="assets/RedHat_favicon_0319.svg">
<link rel="stylesheet" href="styles.ba0e881c90a69f89b28e.css"></head>
<body>
<noscript>
<div class="noscript container"
ng-if="false">
<div class="jumbotron alert alert-danger">
<h2 i18n>JavaScript required!</h2>
<p i18n>A browser with JavaScript enabled is required in order to use this service.</p>
<p i18n>When using Internet Explorer, please check your security settings and add this address to your trusted sites.</p>
</div>
</div>
</noscript>
<cd-root></cd-root>
<script src="runtime.15da3e1803be577b1f00.js" defer></script><script src="polyfills.b66d1515aae6fe3887b1.js" defer></script><script src="scripts.6bda3fa7e09a87cd4228.js" defer></script><script src="main.b310ff35ff1005ba4a64.js" defer></script></body>
</html>
</pre>
<p>Steps to Reproduce:</p>
<ol>
<li>Bootstrap with IPV6 address as --mon-ip</li>
<li>Get dashboard URL using `ceph mgr services`</li>
<li>try to login to dashboard</li>
</ol>
Orchestrator - Bug #50979 (Duplicate): rook: implement_array_function method already has a docstring
https://tracker.ceph.com/issues/50979
2021-05-26T11:06:05Z
Sebastian Wagner
<pre>
2021-05-24T19:40:05.822+0000 7f56cd302040 -1 mgr[py] Module not found: 'rook'
2021-05-24T19:40:05.826+0000 7f8800ece040 -1 mgr[py] Module not found: 'rook'
2021-05-24T19:40:05.826+0000 7f8800ece040 -1 mgr[py] Traceback (most recent call last):
File "/usr/share/ceph/mgr/rook/__init__.py", line 2, in <module>
from .module import RookOrchestrator
File "/usr/share/ceph/mgr/rook/module.py", line 17, in <module>
from kubernetes import client, config
File "/usr/lib/python3/dist-packages/kubernetes/__init__.py", line 22, in <module>
import kubernetes.stream
File "/usr/lib/python3/dist-packages/kubernetes/stream/__init__.py", line 15, in <module>
from .stream import stream
File "/usr/lib/python3/dist-packages/kubernetes/stream/stream.py", line 13, in <module>
from . import ws_client
File "/usr/lib/python3/dist-packages/kubernetes/stream/ws_client.py", line 19, in <module>
from websocket import WebSocket, ABNF, enableTrace
File "/usr/lib/python3/dist-packages/websocket/__init__.py", line 22, in <module>
from ._abnf import *
File "/usr/lib/python3/dist-packages/websocket/_abnf.py", line 34, in <module>
import numpy
File "/usr/lib/python3/dist-packages/numpy/__init__.py", line 142, in <module>
from . import core
File "/usr/lib/python3/dist-packages/numpy/core/__init__.py", line 17, in <module>
from . import multiarray
File "/usr/lib/python3/dist-packages/numpy/core/multiarray.py", line 14, in <module>
from . import overrides
File "/usr/lib/python3/dist-packages/numpy/core/overrides.py", line 16, in <module>
add_docstring(
RuntimeError: implement_array_function method already has a docstring
</pre>
<p><a class="external" href="https://pulpito.ceph.com/yuriw-2021-05-24_19:17:12-rados-wip-yuri8-testing-2021-05-24-0950-octopus-distro-basic-smithi/6132162">https://pulpito.ceph.com/yuriw-2021-05-24_19:17:12-rados-wip-yuri8-testing-2021-05-24-0950-octopus-distro-basic-smithi/6132162</a></p>
Orchestrator - Documentation #50883 (Duplicate): cephadm: mds_cache_memory_limit
https://tracker.ceph.com/issues/50883
2021-05-19T11:18:57Z
Sebastian Wagner
<p>Users can apply:</p>
<pre><code class="yaml syntaxhl"><span class="CodeRay"><span class="key">service_type</span>: <span class="string"><span class="content">mds</span></span>
<span class="key">service_id</span>: <span class="string"><span class="content">fsname</span></span>
<span class="key">placement</span>:
<span class="key">count</span>: <span class="string"><span class="content">2</span></span>
<span class="key">config</span>:
<span class="key">mds_cache_memory_limit</span>: <span class="string"><span class="content">8Gi</span></span>
</span></code></pre>
<p>And this would set the config option for all daemons of that service.</p>
<p>We should probably add this to the documentation</p>
Orchestrator - Cleanup #50117 (Duplicate): orch apply kind: introduce another layer on top of ser...
https://tracker.ceph.com/issues/50117
2021-04-02T22:32:23Z
Sebastian Wagner
<p><strong>Current situation</strong></p>
<p>Right now, we have already three different type of things that <strong>ceph orch apply</strong> supports:</p>
<ul>
<li>Standard services: <strong>service_type: mon|mgr|rgw...</strong></li>
<li>Host specs: <strong>service_type: host</strong></li>
<li>Generic Deployments: <strong>service_type: container</strong></li>
</ul>
<p><strong>Idea</strong></p>
<p>The idea is to distinguish them better for the user. Make it clear that a host spec is not a service.</p>
<pre><code class="yaml syntaxhl"><span class="CodeRay"><span class="key">kind</span>: <span class="string"><span class="content">host</span></span>
<span class="key">hostname</span>: <span class="string"><span class="content">foobar</span></span>
<span class="key">addr</span>: <span class="string"><span class="content">127.0.0.1</span></span>
</span></code></pre>
<p>is better than</p>
<pre><code class="yaml syntaxhl"><span class="CodeRay"><span class="key">service_type</span>: <span class="string"><span class="content">host</span></span>
<span class="key">hostname</span>: <span class="string"><span class="content">foobar</span></span>
<span class="key">addr</span>: <span class="string"><span class="content">127.0.0.1</span></span>
</span></code></pre>
<p>Thus, I propose to add another layer called <strong>kind: host|service|deployment</strong> on top of service_type.</p>
<p>Like so:</p>
<pre><code class="yaml syntaxhl"><span class="CodeRay"><span class="key">kind</span>: <span class="string"><span class="content">host</span></span>
<span class="key">hostname</span>: <span class="string"><span class="content">myhost1</span></span>
<span class="key">labels</span>:
- <span class="string"><span class="content">rgw</span></span>
<span class="head"><span class="head">---</span></span>
<span class="key">kind</span>: <span class="string"><span class="content">service</span></span>
<span class="key">service_type</span>: <span class="string"><span class="content">rgw</span></span>
<span class="key">service_id</span>: <span class="string"><span class="content">foobar</span></span>
<span class="key">spec</span>:
<span class="key">realm</span>: <span class="string"><span class="content">myrealm</span></span>
<span class="key">zone</span>: <span class="string"><span class="content">myzone</span></span>
<span class="head"><span class="head">---</span></span>
<span class="key">kind</span>: <span class="string"><span class="content">rgw_realm</span></span>
<span class="key">name</span>: <span class="string"><span class="content">realm-a</span></span>
<span class="key">pull_endpoint</span>: <span class="string"><span class="content">http://10.2.105.133:80</span></span>
<span class="head"><span class="head">---</span></span>
<span class="key">kind</span>: <span class="string"><span class="content">rgw_zone</span></span>
<span class="key">name</span>: <span class="string"><span class="content">zone-a</span></span>
<span class="head"><span class="head">---</span></span>
<span class="key">kind</span>: <span class="string"><span class="content">deployment</span></span>
<span class="key">service_name</span>: <span class="string"><span class="content">prometheus-webhook-snmp</span></span>
<span class="key">placement</span>:
<span class="error">ADD_PLACEMENT_HERE</span>
<span class="key">image</span>: <span class="string"><span class="content">hello:latest</span></span>
<span class="key">ports</span>:
- <span class="string"><span class="content">999: 999</span></span>
<span class="key">envs</span>:
- <span class="string"><span class="content">args="--debug --snmp-host=ADD_HOST_GATEWAY_HERE --metrics" </span></span>
</span></code></pre>
<pre>
ceph orch apply -i cluster.yaml
</pre>
<p>or</p>
<pre>
cephadm bootstrap --apply-spec cluster.yaml
</pre>
Orchestrator - Feature #49249 (Duplicate): cephadm: Automatically create OSDs after reinstalling ...
https://tracker.ceph.com/issues/49249
2021-02-11T16:49:37Z
Sebastian Wagner
<p><a class="issue tracker-6 status-1 priority-4 priority-default" title="Documentation: Document manually deploment of OSDs (New)" href="https://tracker.ceph.com/issues/46691">#46691</a> provides the manual process of deploying cephadm OSDs.</p>
<p>we should probably provide an automated way to do this.</p>
Orchestrator - Bug #49191 (Duplicate): cephadm: service_type: osd: Failed to apply: ''NoneType'' ...
https://tracker.ceph.com/issues/49191
2021-02-05T14:29:33Z
Sebastian Wagner
<pre>
ceph orch ls --format yaml
</pre><br /><pre><code class="yaml syntaxhl"><span class="CodeRay"><span class="key">service_type</span>: <span class="string"><span class="content">osd</span></span>
<span class="key">service_id</span>: <span class="string"><span class="content">dashboard-p14040</span></span>
<span class="key">service_name</span>: <span class="string"><span class="content">osd.dashboard-p14040</span></span>
<span class="key">placement</span>:
<span class="key">hosts</span>:
- <span class="string"><span class="content">lnx92252</span></span>
<span class="key">spec</span>:
<span class="key">filter_logic</span>: <span class="string"><span class="content">AND</span></span>
<span class="key">objectstore</span>: <span class="string"><span class="content">bluestore</span></span>
<span class="key">status</span>:
<span class="key">container_image_id</span>: <span class="string"><span class="content">xyz</span></span>
<span class="key">container_image_name</span>: <span class="string"><span class="content">ceph/ceph:latest</span></span>
<span class="key">last_refresh</span>: <span class="string"><span class="content">'2021-02-01T13:53:11.218862'</span></span>
<span class="key">running</span>: <span class="string"><span class="content">1</span></span>
<span class="key">size</span>: <span class="string"><span class="content">0</span></span>
<span class="key">events</span>:
- <span class="string"><span class="content">'2021-01-30T18:40:07.414091 service:osd.dashboard [ERROR] "Failed</span></span>
<span class="key">to apply</span>: <span class="string"><span class="content">''NoneType'' object has no attribute ''paths''"'</span></span>
</span></code></pre>
Orchestrator - Bug #48463 (Duplicate): mon.c: Error: invalid config provided: CapAdd and privileg...
https://tracker.ceph.com/issues/48463
2020-12-04T11:17:07Z
Sebastian Wagner
<p><a class="external" href="https://pulpito.ceph.com/swagner-2020-12-04_10:02:29-rados:cephadm-wip-jmolmo-testing-2020-12-02-1452-distro-basic-smithi/5680473/">https://pulpito.ceph.com/swagner-2020-12-04_10:02:29-rados:cephadm-wip-jmolmo-testing-2020-12-02-1452-distro-basic-smithi/5680473/</a></p>
<pre>
['/bin/podman', 'run', '--rm', '--net=host', '-e', 'CONTAINER_IMAGE=docker.io/ceph/ceph:v15.2.0', '-e', 'NODE_NAME=smithi135', '-v', '/var/log/ceph/c4502caa-3619-11eb-980d-001a4
aab830c:/var/log/ceph:z', '-v', '/tmp/ceph-tmpr_67xdiq:/etc/ceph/ceph.client.admin.keyring:z', '-v', '/tmp/ceph-tmpl0cckyum:/etc/ceph/ceph.conf:z', '-v', '/var/lib/ceph/c4502caa-3619-11eb-980d-001a4aab830c/mon.a:/var/lib/ceph/mon/ceph-a:z', '--entrypoint', '/usr
/bin/ceph', 'docker.io/ceph/ceph:v15.2.0', 'config', 'generate-minimal-conf', '-o', '/var/lib/ceph/mon/ceph-a/config']
</pre>
<p>Turns out, we're now installing podman 2 and then strting the upgrade from 15.2.0, which does not support podman 2.</p>
Orchestrator - Bug #48277 (Duplicate): cephadm infer image: <none>:<none>, despite --filter dangl...
https://tracker.ceph.com/issues/48277
2020-11-18T13:49:38Z
Sebastian Wagner
<pre>
master:~ # podman --version
podman version 2.0.6
master:~ # podman images --filter label=ceph=True --filter dangling=false --format '{{.Repository}} {{.Tag}}'
<none> <none>
registry.suse.com/ses/7/ceph/ceph latest
</pre>
Orchestrator - Feature #46265 (Duplicate): test cephadm MDS deployment
https://tracker.ceph.com/issues/46265
2020-06-29T18:40:07Z
Sebastian Wagner
<p>right now, the test is broken.</p>
<p>workaround is to apply it manuall: <a class="external" href="https://github.com/ceph/ceph/blob/cedf2bbd13daba6059c98046969f646b944042ae/qa/suites/rados/cephadm/workunits/task/test_orch_cli.yaml#L14">https://github.com/ceph/ceph/blob/cedf2bbd13daba6059c98046969f646b944042ae/qa/suites/rados/cephadm/workunits/task/test_orch_cli.yaml#L14</a></p>
Orchestrator - Feature #45905 (Duplicate): cephadm: errors in serve() should create a HEALTH warning
https://tracker.ceph.com/issues/45905
2020-06-05T09:50:18Z
Sebastian Wagner
<p>othwerwise users need to search the mgr log for hints manually.</p>
Orchestrator - Documentation #45564 (Duplicate): cephadm: document workaround for accessing the a...
https://tracker.ceph.com/issues/45564
2020-05-15T09:35:17Z
Sebastian Wagner
<pre>
$ ceph daemon mgr.ceph03 config show
admin_socket: exception getting command descriptions: [Errno 2] No such file or directory
</pre>
<pre>
$ ceph --admin-daemon /var/run/ceph/mgr.ceph03 config show
No such file or directory
</pre>
<p>Probably, users have to run</p>
<pre>
cephadm enter mgr.ceph03
</pre>
<p>to access the admin socket.</p>
Orchestrator - Feature #39057 (Duplicate): orchestrator_cli should check minimum device size
https://tracker.ceph.com/issues/39057
2019-04-01T08:07:47Z
Sebastian Wagner
<p>Otherwise we might get an error like this:</p>
<pre>
2019-04-01 08:01:11.465649 I | --> RuntimeError: Unable to use device 4.00 GB /dev/vdb, LVs would be smaller than 5GB
</pre>
<pre>
[root@kubic-1 /]# k get CephCluster -o yaml
apiVersion: v1
items:
- apiVersion: ceph.rook.io/v1
kind: CephCluster
metadata: ...
spec:
cephVersion:
allowUnsupported: true
image: 192.168.122.1:443/ceph/ceph:latest
dashboard:
enabled: true
dataDirHostPath: /var/lib/rook
mon:
allowMultiplePerNode: true
count: 3
preferredCount: 0
network:
hostNetwork: false
rbdMirroring:
workers: 0
storage:
config:
databaseSizeMB: "1024"
journalSizeMB: "1024"
osdsPerDevice: "1"
directories:
- config: null
path: /var/lib/rook
nodes:
- config: null
devices:
- FullPath: ""
config: null
name: vdb
name: kubic-1
resources: {}
useAllDevices: false
status:
state: Updating
kind: List
metadata:
resourceVersion: ""
selfLink: ""
</pre>
<pre>
[root@kubic-1 /]# ceph orchestrator device ls
Host kubic-1:
Device Path Type Size Rotates Available Model
vdb hdd 5120M False False
vda hdd 24.0G False False
Host kubic-2:
Device Path Type Size Rotates Available Model
vdb hdd 5120M False False
vda hdd 24.0G False False
</pre>