Activity
From 10/04/2022 to 11/02/2022
11/02/2022
- 08:27 PM rgw Feature #57965 (Resolved): Add new zone option to control whether an object's first data stripe i...
- Delete requests are quite slow on clusters that have a data pool backed by HDDs, especially with an EC pool. For exam...
- 06:58 PM rgw Bug #57562: multisite replication issue on Quincy
- Yeah, both those commits are gone, make sure you only have the newest one.
- 06:33 PM rgw Bug #57562: multisite replication issue on Quincy
- Adam Emerson wrote:
> Pushed a new version with what should be a fix for multi-thread and multi-client races.
We ... - 07:22 AM rgw Bug #57562: multisite replication issue on Quincy
- Pushed a new version with what should be a fix for multi-thread and multi-client races.
- 04:09 PM Bug #57964 (New): Cephadm: MONs are not getting back after /var/log filesystem is full
- I had at least two examples where Cephadm controlled cluster had /var/log full in CentOS Linux and it obviously cause...
- 03:53 PM RADOS Fix #57963 (Fix Under Review): osd: Misleading information displayed for the running configuratio...
- With the fix, the following is shown for an OSD with ssd as the underlying device type:...
- 03:26 PM RADOS Fix #57963: osd: Misleading information displayed for the running configuration of osd_mclock_max...
- See BZ: https://bugzilla.redhat.com/show_bug.cgi?id=2111282 for additional information.
- 03:25 PM RADOS Fix #57963 (Resolved): osd: Misleading information displayed for the running configuration of osd...
- For the inactive device type(hdd/ssd) of an OSD, the running configuration option osd_mclock_max_capacity_iops_[hdd|s...
- 03:10 PM ceph-volume Bug #57918 (Fix Under Review): CEPHADM_REFRESH_FAILED: failed to probe daemons or devices
- 01:40 PM ceph-volume Bug #57918 (In Progress): CEPHADM_REFRESH_FAILED: failed to probe daemons or devices
- 01:48 PM Feature #57962 (Pending Backport): ceph-mixin: Add Prometheus Alert for Degraded Bond
- Currently there is no alert for a network interface card to be misconfigured or failed which is part of a network bon...
- 12:51 PM Orchestrator Bug #57897: ceph mgr restart causes restart of all iscsi daemons in a loop
- Attempts to stop or redeploy the daemon don't work as they seem to invoke a dependency check, which then restarts the...
- 10:48 AM Orchestrator Bug #57960: iscsi - rbd-target-api unkillable on container exit, daemon enters error state
- The systemd service shows the following:...
- 10:20 AM Orchestrator Bug #57960 (New): iscsi - rbd-target-api unkillable on container exit, daemon enters error state
- Hi,
We have seen lots of iscsi container restarts due to https://tracker.ceph.com/issues/57897 and during some of ... - 09:04 AM Dashboard Bug #57959 (New): mgr/dashboard: OSD create form without orchestrator error
- h3. Description of problem
If you deploy a ceph-dev cluster on main and you try to click on "create" inside Cluste... - 08:07 AM rgw Bug #57942: rgw leaks rados objects when a part is submitted multiple times in a multipart upload
- FI pull-request https://github.com/ceph/ceph/pull/48704
- 06:40 AM RADOS Bug #57533 (Fix Under Review): Able to modify the mclock reservation, weight and limit parameters...
11/01/2022
- 08:22 PM Orchestrator Bug #57897: ceph mgr restart causes restart of all iscsi daemons in a loop
- We currently have this situation again after an mgr failover for node maintenance/reboots and managed enabled debug l...
- 08:12 PM rgw Bug #57942: rgw leaks rados objects when a part is submitted multiple times in a multipart upload
- FI Working on https://github.com/pgoron/ceph/commits/fix_rgw_rados_leaks_57942 to fix both issues (index entry leaks ...
- 07:25 PM rgw Bug #57562: multisite replication issue on Quincy
- Agree as you mentioned, the other solution could be, secondary not limited to just listening on to orpan part, but co...
- 06:37 PM rgw Bug #57562: multisite replication issue on Quincy
- Ah, I see, I need to update the async lister.
- 06:36 PM rgw Bug #57562: multisite replication issue on Quincy
- That's the point of the commit `rgw/fifo: `part_full` is not a reliable indicator`. There is no 'orphan part' in that...
- 05:39 PM rgw Bug #57562: multisite replication issue on Quincy
- Hey Adam,
Just a heads-up we tested with latest commit and we still see the issue.
The issue is seen when running M... - 02:11 PM rgw Bug #57562: multisite replication issue on Quincy
- Thank you Adam. We'll test with the latest change.
- 06:04 PM Orchestrator Cleanup #57957: cephadm: rename "extra_container_args"
- Additionally, the "args" field specific to Custom Containers. I think that "args" is an attractively named option for...
- 03:32 PM Orchestrator Cleanup #57957 (New): cephadm: rename "extra_container_args"
- we ideally want a different name that makes it more clear that these are arguments for the podman/docker run command ...
- 04:27 PM rgw Bug #44660 (Fix Under Review): Multipart re-uploads cause orphan data
- Actually it looks like there is a simpler solution to this problem, which uses the meta object lock when checking if ...
- 02:33 PM Bug #57956 (New): Ceph monitors in crash loop
Creating a pool causes mon to restart
This is the detailed question:https://github.com/rook/rook/issues/10110
...- 03:43 AM Orchestrator Bug #57954: rook/k8s: nfs cluster creation ends up with no daemons deployment
- I am working on this issue
- 03:42 AM Orchestrator Bug #57954 (Resolved): rook/k8s: nfs cluster creation ends up with no daemons deployment
- steps to reproduce:
1, with rook deploy Ceph 17.2.5 on k8s
2, run the following to enable rook as orchestrator with... - 03:32 AM Support #57953 (New): Ceph runs in the openeuler system, and the pool cannot be created after ini...
- https://github.com/rook/rook/issues/11242
Ceph runs in the openeuler system, and the pool cannot be created after ... - 02:32 AM CephFS Support #57952 (New): Pacific: the buffer_anon_bytes of ceph-mds is too large
- The buffer_anon_bytes will reach 200+GB, then run out of machine memory.It does not seem to be able to effectively fr...
10/31/2022
- 06:56 PM rgw Bug #57562: multisite replication issue on Quincy
- Pushed a new version that should make listing list all the objects reliably.
- 06:33 PM rbd Bug #49947 (Resolved): document supported architectures for PWL cache plugin
- 05:32 PM Backport #57925: pacific: common: use fmt::print for stderr logging
- Conditional on fmt >= 9 ?
https://tracker.ceph.com/issues/57540 - 05:22 PM Orchestrator Feature #57944: add option to allow for setting extra daemon args for containerized services
- We've hit this too.
- 04:40 PM rgw Bug #57951 (Pending Backport): rgw: lc: lc for a single large bucket can run too long
- If this happens, other lc hosts/threads can attempt to process the same bucket, which inflates overhead without any c...
- 04:07 PM Orchestrator Feature #57948 (Fix Under Review): Adding support for a secure ceph monitoring stack
- 10:05 AM Orchestrator Feature #57948 (Resolved): Adding support for a secure ceph monitoring stack
- Right now the communications between the monitoring components (prometheus, alertmanager, node-exporter, ..) is using...
- 03:11 PM Dashboard Backport #57691 (Resolved): pacific: mgr/dashboard: permission denied when creating a NFS export
- 01:58 PM RADOS Bug #53729 (Resolved): ceph-osd takes all memory before oom on boot
- 01:58 PM RADOS Backport #55633 (Rejected): octopus: ceph-osd takes all memory before oom on boot
- Octopus is EOL
- 12:53 PM rbd Bug #57902 (Fix Under Review): [rbd-nbd] add --snap-id option to "rbd device map" to allow mappin...
- 12:41 PM RADOS Bug #57757: ECUtil: terminate called after throwing an instance of 'ceph::buffer::v15_2_0::end_of...
- ...
- 09:50 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- Venky Shankar wrote:
> Hi,
>
> Could the list of PRs that try to address this issue be linked? (so, that we don't... - 09:36 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- Hi,
Could the list of PRs that try to address this issue be linked? (so, that we don't loose track of them).
As... - 04:50 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- zhikuo du wrote:
> > I am afraid this won't work. As I remembered from my test before, the size of ESubtreeMap could... - 02:47 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- > I am afraid this won't work. As I remembered from my test before, the size of ESubtreeMap could reach up to several...
- 02:37 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- zhikuo du wrote:
> > May I ask you a question:
> > What factors decide how many event must have a ESubtreeMap e... - 09:44 AM rgw Feature #57947 (Pending Backport): Improve performance of multi-object delete by handling individ...
- Multi-object deletes are currently quite slow. The handler for this method currently just loops through the list of o...
- 08:10 AM rgw Bug #57942: rgw leaks rados objects when a part is submitted multiple times in a multipart upload
- After digging more on the issue, I think the root cause is linked to following code:
https://github.com/ceph/ceph/... - 05:04 AM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- And suddenly the upgrade is happening!!!
Today I rebooted ceph02, a node that only had the MDS, and suddenly thing... - 04:35 AM CephFS Backport #57946 (In Progress): quincy: cephfs-top: make cephfs-top display scrollable like top
- 04:26 AM CephFS Backport #57946 (Resolved): quincy: cephfs-top: make cephfs-top display scrollable like top
- https://github.com/ceph/ceph/pull/48677
- 04:21 AM CephFS Feature #55197 (Pending Backport): cephfs-top: make cephfs-top display scrollable like top
- 03:57 AM RADOS Bug #52657: MOSDPGLog::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, SERVER_NAUTILUS)'
- Still trying to run a test with added debugging due to the ongoing infra issues but I noticed that Coverity CID 15096...
- 01:30 AM Dashboard Bug #53230: ceph API tests failed
- maybe this is an environmental issue. can close it
10/30/2022
- 05:14 PM rbd Bug #49947 (Fix Under Review): document supported architectures for PWL cache plugin
- 02:18 PM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- > @Xiubo Li @Venky Shankar
>
> I readed the codes about: how the segment is trimmed and how ESubtreeMap/EImportSt... - 01:10 PM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- > May I ask you a question:
> What factors decide how many event must have a ESubtreeMap event? And what is the... - 09:42 AM Bug #57945 (Pending Backport): On Rocky Os mgr could not start do to wrong python version
- Reproduction
Compile and run ceph on Rocky os
run vstart
Result
error in manager process
python module was ...
10/29/2022
- 10:29 PM Documentation #47656 (Closed): Install Guide - Fed 32 installation instructions don't work
- 10:06 PM RADOS Documentation #46126: RGW docs lack an explanation of how permissions management works, especiall...
- You thought that copying this rude exchange verbatim was essential to motivate improving the docs?
Matt - 09:45 PM Documentation #44342 (Resolved): Create a notification and a link to direct people to a particula...
- We have had this versioned-documentation menu in the docs for some time now. I have attached screenshots that confirm...
- 09:32 PM Bug #47063 (Resolved): The RADOS deployment guide refers to ceph-deploy, which is, as of Octopus,...
- Robert Sander removed ceph/deploy.rst in 2020: 1b42759e19b352c22d9e9109ecdf6c3b20feed84
https://github.com/ceph/ceph... - 09:28 PM Bug #47064 (Resolved): rados/deployment is redundant
- Robert Sander removed this in 2020: 1b42759e19b352c22d9e9109ecdf6c3b20feed84
https://github.com/ceph/ceph/commit/1b4... - 05:32 PM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- I am getting ready to add another node to the cluster. Is there anything you can think of I can check, pre or post?
- 02:20 AM CephFS Feature #55197 (Resolved): cephfs-top: make cephfs-top display scrollable like top
- 01:32 AM rgw Bug #57853: multisite sync process block after long time running
- I think something wrong with rgw-coroutine,please check the above PR
- 01:30 AM rgw Bug #57853: multisite sync process block after long time running
- PR https://github.com/ceph/ceph/pull/48626
10/28/2022
- 09:42 PM Bug #57914: centos 8 build failed
- Note from Dan:
It's looking more and more like we're just going to have to rebuild these broken VMs from scratch. ... - 09:10 PM rgw Bug #57770: RGW (pacific) misplaces index entries after dynamically resharding bucket
- J. Eric Ivancich wrote:
> Nick,
>
> I don't know that I have a cluster at my fingertips that might be necessary t... - 07:23 PM rgw Bug #57770: RGW (pacific) misplaces index entries after dynamically resharding bucket
- Nick,
I don't know that I have a cluster at my fingertips that might be necessary to test this potential fix. How ... - 07:21 PM rgw Bug #57770 (Fix Under Review): RGW (pacific) misplaces index entries after dynamically resharding...
- 07:16 PM Orchestrator Feature #57944 (Resolved): add option to allow for setting extra daemon args for containerized se...
- The ceph orchestrator YML specs for service templates has an option for "extra_container_args" which allows the user ...
- 04:25 PM CephFS Bug #53509 (Resolved): quota support for subvolumegroup
- 04:25 PM CephFS Bug #53848 (Resolved): mgr/volumes: Failed to create clones if the source snapshot's quota is exc...
- 02:37 PM ceph-volume Bug #57907: ceph-volume complains about "Insufficient space (<5GB)" on 1.75TB device
- I would need it with hotplug enabled.
anyway, I tried to reproduce... - 01:08 PM ceph-volume Bug #57907: ceph-volume complains about "Insufficient space (<5GB)" on 1.75TB device
- Guillaume Abrioux wrote:
> can you share the output of `ceph-volume inventory --format json` ?
With hotplug disab... - 12:36 PM ceph-volume Bug #57907: ceph-volume complains about "Insufficient space (<5GB)" on 1.75TB device
- can you share the output of `ceph-volume inventory --format json` ?
- 01:35 PM Dashboard Bug #57943 (New): doc/radosgw: "waiting on unpkg.com" for upwards of one minute when http://local...
- h3. Description
*RADOSGW documentation calls unpkg.com when http://localhost:8080/radosgw/multisite/ is loaded in ... - 12:36 PM rgw Bug #57942 (Duplicate): rgw leaks rados objects when a part is submitted multiple times in a mult...
- Hello,
Issue presented below affects all ceph versions at least since 14.2 (reproducer tested on 14.2, 15.2, 16.2,... - 07:11 AM CephFS Backport #57723: pacific: qa: test_subvolume_snapshot_info_if_orphan_clone fails
- Backport of https://github.com/ceph/ceph/pull/48642 is also included with this
- 01:39 AM bluestore Feature #57785: fragmentation score in metrics
- Ultimately, I'd like it in prometheus, so I can setup alerts if it gets too high.
10/27/2022
- 09:21 PM rgw Bug #57562: multisite replication issue on Quincy
- Pushed a newer, newer fix that guards all calls to _prepare_new_head behind check/set of preparing.
- 04:15 PM rgw Bug #57562: multisite replication issue on Quincy
- Pushed a newer fix that does the check in need_new_head()
- 02:01 PM rgw Bug #57562: multisite replication issue on Quincy
- Hi Adam,
We obtained the extra logging with the fix in place.
I think the contention is not within _prepare_ne... - 01:09 AM rgw Bug #57562: multisite replication issue on Quincy
- I expect there are multiple problems with sync in Quincy, so I don't expect this to actually make sync work.
But i... - 12:15 AM rgw Bug #57562: multisite replication issue on Quincy
- Pulled the changes in on top of the commit _9056dbcdeaa7f4350b54a69f669982358ec5448e_ (on main branch). Unfortunately...
- 08:29 PM rbd Bug #57941 (Rejected): Severe performance drop after writing 100 GB of data to RBD volume, depend...
- Write throughput to a mapped RBD volume drops dramatically after the volume reaches a certain usage size. The amount ...
- 06:07 PM RADOS Bug #57940 (Duplicate): ceph osd crashes with FAILED ceph_assert(clone_overlap.count(clone)) when...
- Hi, I have this current crash:
I've experienced a disk failure in my ceph cluster.
I've replaced the disk, but no... - 05:28 PM bluestore Feature #57785: fragmentation score in metrics
- Kevin Fox wrote:
> Currently the bluestore fragmentation score does not seem to be exported in metrics. Due to the i... - 04:50 PM RADOS Bug #57757: ECUtil: terminate called after throwing an instance of 'ceph::buffer::v15_2_0::end_of...
- @Laura, thanks for that! i'll try first with main as you suggested
- 03:32 PM RADOS Bug #57757: ECUtil: terminate called after throwing an instance of 'ceph::buffer::v15_2_0::end_of...
- @Nitzan, here is the branch if you'd like to rebuild it on ci: https://github.com/ljflores/ceph/commits/wip-lflores-t...
- 10:36 AM RADOS Bug #57757: ECUtil: terminate called after throwing an instance of 'ceph::buffer::v15_2_0::end_of...
- The coredump from branch wip-lflores-testing, I was not able to create docker image since this branch is no longer av...
- 02:31 PM rgw Bug #57928 (Duplicate): Octopus:multisite sync process block after long time running
- 02:31 PM rgw Bug #57927 (Duplicate): pacific:multisite sync process block after long time running
- 01:22 PM CephFS Bug #55804 (Duplicate): qa failure: pjd link tests failed
- 01:21 PM CephFS Bug #55804: qa failure: pjd link tests failed
- This issue is probably fixed by PR: https://github.com/ceph/ceph/pull/46331 ("mds: wait unlink to finish to avoid con...
- 12:55 PM CephFS Bug #57446: qa: test_subvolume_snapshot_info_if_orphan_clone fails
- Fixed another possible failure with this test
https://github.com/ceph/ceph/pull/48642 - 12:27 PM CephFS Bug #51278: mds: "FAILED ceph_assert(!segments.empty())"
- Venky Shankar wrote:
> Latest occurrence with similar backtrace - https://pulpito.ceph.com/vshankar-2022-06-03_10:03... - 12:23 PM ceph-volume Bug #57939 (New): Not able to add additional disk sharing common wal/db device
- Ceph version: 17.2.3 (dff484dfc9e19a9819f375586300b3b79d80034d)
quincy (stable)
Each of our nodes has 8x 16T rot... - 12:17 PM RADOS Bug #55141: thrashers/fastread: assertion failure: rollback_info_trimmed_to == head
- Radoslaw Zarzynski wrote:
> Well, just found a new occurance.
Where can i find it?
- 12:13 PM RADOS Bug #50042 (In Progress): rados/test.sh: api_watch_notify failures
- 12:12 PM RADOS Bug #52136 (In Progress): Valgrind reports memory "Leak_DefinitelyLost" errors.
- 11:47 AM RADOS Bug #57751 (In Progress): LibRadosAio.SimpleWritePP hang and pkill
- 10:55 AM RADOS Bug #57751: LibRadosAio.SimpleWritePP hang and pkill
- This is not an issue with the test, not all the osd are up, and we are waiting (valgrind report memory leak from rock...
- 08:24 AM ceph-volume Bug #57907: ceph-volume complains about "Insufficient space (<5GB)" on 1.75TB device
- Actually, this is a major bug for me, as i have to reboot the complete host, to replace one OSD.
- 07:36 AM rgw Cleanup #57938 (Pending Backport): relying on boost flatmap emplace behavior is risky
- see coverity issue: http://folio07.front.sepia.ceph.com/main/ceph-main-98d41855/cov-main-html/3/2253rgw_trim_bilog.cc...
- 06:39 AM Orchestrator Bug #57910: ingress: HAProxy fails to start because keepalived IP address not yet available on ne...
- Happens also (sometimes?) after re-provisioning an ingress server. After OS installed and when cephadm configures the...
- 04:26 AM RADOS Bug #57937 (Rejected): pg autoscaler of rgw pools doesn't work after creating otp pool
- It's about the following my post to ceph-users ML.
https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/threa... - 02:56 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- zhikuo du wrote:
> Xiubo Li wrote:
> > zhikuo du wrote:
> > [...]
> > > 4,I think this problem will seriously aff...
10/26/2022
- 11:25 PM RADOS Bug #57017 (Pending Backport): mon-stretched_cluster: degraded stretched mode lead to Monitor crash
- 10:01 PM rgw Bug #57936 (Pending Backport): 'radosgw-admin bucket chown' doesn't set bucket instance owner or ...
- steps to reproduce:
1. start a vstart cluster and create a bucket as user 'testid'... - 09:29 PM Orchestrator Bug #57755: task/test_orch_cli: test_cephfs_mirror times out
- /a/yuriw-2022-10-19_18:35:19-rados-wip-yuri10-testing-2022-10-19-0810-distro-default-smithi/7074978
- 09:18 PM RADOS Bug #52129: LibRadosWatchNotify.AioWatchDelete failed
- /a/yuriw-2022-10-19_18:35:19-rados-wip-yuri10-testing-2022-10-19-0810-distro-default-smithi/7074802
- 05:08 PM rgw Bug #57562: multisite replication issue on Quincy
- Awesome! Thanks for the quick turn around! Will pull and test.
- 04:49 PM rgw Bug #57562 (Fix Under Review): multisite replication issue on Quincy
- I have a candidate fix at https://github.com/ceph/ceph/pull/48632
- 02:14 PM rgw Bug #57562: multisite replication issue on Quincy
- FYI: We pulled in the 2 PRs Casey posted in the tracker https://tracker.ceph.com/issues/57783, and tested again with ...
- 12:31 PM rgw Bug #57562: multisite replication issue on Quincy
- FWIW, below provides some log snippets with enhanced events. To be specific, some existing log events are added addit...
- 03:47 PM Feature #57109 (Fix Under Review): windows: rbd-wnbd SCSI persistent reservations
- 02:52 PM RADOS Bug #57883 (Resolved): test-erasure-code.sh: TEST_rados_put_get_jerasure fails on "rados_put_get:...
- 02:13 PM CephFS Backport #57717 (Resolved): quincy: libcephfs: incorrectly showing the size for snapdirs when sta...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48414
Merged. - 01:45 PM RADOS Bug #50042: rados/test.sh: api_watch_notify failures
- ...
- 04:58 AM RADOS Bug #50042: rados/test.sh: api_watch_notify failures
- I checked all the list_watchers failures (checking size of watch list), It looks like the watcher timed out and that ...
- 12:52 PM Orchestrator Bug #57931 (Fix Under Review): RGW rgw_frontend_type field is not checked correctly during the sp...
- 08:23 AM Orchestrator Bug #57931 (Resolved): RGW rgw_frontend_type field is not checked correctly during the spec parsing
- A spec with the following type for example:...
- 10:46 AM Bug #57934 (New): Pacific 16.2.10 OSD crashing with tcmalloc
- # ceph crash info 2022-10-25T14:49:09.609527Z_b10565bb-036d-408f-a536-442f7b4c8213
{
"archived": "2022-10-26 04... - 10:34 AM Bug #57933 (New): Do package update on base images before building image to reduce Vulnerability
- *What should the feature do:*
Update the package definition update on base images before building image to install... - 10:09 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- Xiubo Li wrote:
> zhikuo du wrote:
> [...]
> > 4,I think this problem will seriously affect the IOPS of write and ... - 03:49 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- zhikuo du wrote:
> Xiubo Li wrote:
> > zhikuo du wrote:
> > [...]
> > > 4,I think this problem will seriously aff... - 01:42 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- Xiubo Li wrote:
> zhikuo du wrote:
> [...]
> > 4,I think this problem will seriously affect the IOPS of write and ... - 12:42 AM CephFS Bug #57920: mds:ESubtreeMap event size is too large
- zhikuo du wrote:
[...]
> 4,I think this problem will seriously affect the IOPS of write and read.
>
> 5, @Xiubo ... - 10:05 AM CephFS Bug #57856 (Closed): cephfs-top: Skip refresh when the perf stats query shows no metrics
- Closing this, as refreshes are optimised in a better way in https://github.com/ceph/ceph/pull/48090.
- 09:02 AM mgr Bug #57932: Intermittent ceph-mgr segfault MgrStandby::ms_dispatch2()
- Please let me know if I can provide any more detail. If it's helpful I can provide a crash dump
- 09:00 AM mgr Bug #57932 (Need More Info): Intermittent ceph-mgr segfault MgrStandby::ms_dispatch2()
- We're seeing intermittent ceph-mgr segfaults in CI...
- 07:09 AM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- Also, it seems that requests to mds are much slower than writing blocks. When I run the rm command, it sends an avera...
- 06:55 AM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- Xiubo Li wrote:
> Minjong Kim wrote:
> > ceph I used the ceph kernel mount. In fuse-mount it works fine.
> >
> >... - 06:32 AM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- https://gist.github.com/caffeinism/dbfd974374d620911a6c0c3dd1daadfb
I am not good at writing files in a shell scri... - 06:23 AM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- But I haven't checked the testing branch. (I'll check)
- 06:22 AM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- Xiubo Li wrote:
> Minjong Kim wrote:
> > ceph I used the ceph kernel mount. In fuse-mount it works fine.
> >
> >... - 05:59 AM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- Minjong Kim wrote:
> ceph I used the ceph kernel mount. In fuse-mount it works fine.
>
> The test script is nothi... - 05:49 AM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- ceph I used the ceph kernel mount. In fuse-mount it works fine.
The test script is nothing special. I just did the... - 04:55 AM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- Could you upload your test script ?
Do you mean you can also reproduce this by using the ceph-fuse mount, right ? - 04:41 AM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- Hello again
I don't know if anyone is interested, but when tested with an already built kernel (https://kernel.ubunt... - 06:25 AM CephFS Backport #57929 (In Progress): quincy: qa: test_dump_loads fails with JSONDecodeError
- https://github.com/ceph/ceph/pull/54187
- 06:18 AM CephFS Bug #57299 (Pending Backport): qa: test_dump_loads fails with JSONDecodeError
- 06:09 AM RADOS Bug #52657: MOSDPGLog::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, SERVER_NAUTILUS)'
- I was able to gather a coredump and set up a binary compatible environment to debug it from this run Laura started in...
- 04:58 AM RADOS Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
- I wrote up an working explanation of PastIntervals in https://github.com/athanatos/ceph/tree/sjust/wip-49689-past-int...
- 03:48 AM rgw Bug #57853: multisite sync process block after long time running
- Quincy、Pacific、Octopus、 Nautilus has same issue
- 03:31 AM mgr Backport #57077 (Resolved): quincy: "overlapping roots" error message needs documentation
- 03:27 AM rgw Bug #57928 (Duplicate): Octopus:multisite sync process block after long time running
- 1、deploy RADOSGW multisite
2、put lot of objects
3、keep it runing for a long time - 03:25 AM rgw Bug #57927: pacific:multisite sync process block after long time running
- same as https://tracker.ceph.com/issues/57853
- 03:24 AM rgw Bug #57927 (Duplicate): pacific:multisite sync process block after long time running
- 1、deploy RADOSGW multisite
2、put lot of objects
3、keep it runing for a long time - 12:31 AM Backport #57926 (New): quincy: common: use fmt::print for stderr logging
- 12:30 AM Backport #57925 (Rejected): pacific: common: use fmt::print for stderr logging
- 12:17 AM Cleanup #53682 (Pending Backport): common: use fmt::print for stderr logging
- 12:15 AM Bug #57923 (Fix Under Review): log: writes to stderr (pipe) may not be atomic
- 12:07 AM RADOS Bug #57845 (New): MOSDRepOp::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, SERVER_O...
- Notes from rados team meeting:
Seems like the same class of bugs we hit in https://tracker.ceph.com/issues/52657 a...
10/25/2022
- 11:50 PM mgr Backport #57077: quincy: "overlapping roots" error message needs documentation
- FYI, the corresponding PR (#47519) was merged and so it seems to be OK to close this ticket.
- 11:14 PM RADOS Bug #51729: Upmap verification fails for multi-level crush rule
- I put together the following contrived example to
illustrate the problem. Again, this is pacific 16.2.9 on rocky8 li... - 10:20 PM rgw Bug #57562 (In Progress): multisite replication issue on Quincy
- Small reproducer turned out to not be, but fixing that.
- 04:51 PM rgw Bug #57562: multisite replication issue on Quincy
- Thank you.
- 04:34 PM rgw Bug #57562: multisite replication issue on Quincy
- Please see the following FIFO log snippets. And please let me know if you need more.
The creation of data_log.34.n... - 03:53 PM rgw Bug #57562: multisite replication issue on Quincy
- Can we get a more complete log snippet? All the FIFO logging with the relevant TIDs would make tracing what's going o...
- 03:12 PM rgw Bug #57562: multisite replication issue on Quincy
- thanks, that's very interesting
- 02:59 PM rgw Bug #57562: multisite replication issue on Quincy
- We pretty much narrowed down what the problem is: a race condition has been identified in FIFO::_prepare_new_head(..)...
- 07:00 PM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- I rebooted last night, all items report a refreshed time of about 13 hours ago, when I rebooted....
- 12:14 PM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- this seems to imply the cephadm service loop just isn't running at all. Does the REFRESHED column in `ceph orch devic...
- 05:34 PM Orchestrator Bug #57917: cephadm: duplicate log entry for /var/log/ceph/cephadm.log
- As a workaround, I tried a `prerotate` script:...
- 04:08 AM Orchestrator Bug #57917: cephadm: duplicate log entry for /var/log/ceph/cephadm.log
- this is also problematic for users upgrading from Octopus based clusters to Pacific.
- 04:06 AM Orchestrator Bug #57917 (New): cephadm: duplicate log entry for /var/log/ceph/cephadm.log
- Configuration of logrotate for /var/lib/ceph/cephadm.log was added (and backported to Pacific) via this PR:
https://... - 05:19 PM RADOS Bug #50219 (New): qa/standalone/erasure-code/test-erasure-eio.sh fails since pg is not in recover...
- The failure actually reproduced here:
/a/lflores-2022-10-17_18:19:55-rados:standalone-main-distro-default-smithi/7... - 05:06 PM RADOS Bug #57883 (Fix Under Review): test-erasure-code.sh: TEST_rados_put_get_jerasure fails on "rados_...
- 02:21 PM RADOS Bug #57883 (In Progress): test-erasure-code.sh: TEST_rados_put_get_jerasure fails on "rados_put_g...
- 03:39 PM Dashboard Bug #57924: mgr/dashboard: fails with "Module 'dashboard' has failed: key type unsupported" when ...
- the 2nd quoted cert was working. Unfortunately i cannot fix my own bug reports.
- 03:38 PM Dashboard Bug #57924: mgr/dashboard: fails with "Module 'dashboard' has failed: key type unsupported" when ...
- Certificate that did not work:...
- 03:33 PM Dashboard Bug #57924 (New): mgr/dashboard: fails with "Module 'dashboard' has failed: key type unsupported"...
- h3. Description of problem
After generating a recent certificate by letsencrypt and configuring dashboard to use t... - 02:39 PM Documentation #57858 (Resolved): v17.2.4 release does not contain latest cherry-picks
- 17.2.5 was released with all missing commits
- 02:19 PM RADOS Bug #57900 (In Progress): mon/crush_ops.sh: mons out of quorum
- 02:17 PM RADOS Bug #57900: mon/crush_ops.sh: mons out of quorum
- @Radek so the suggestion is to give the mons more time to reboot?
This is the workunit:
https://github.com/ceph/c... - 02:16 PM Bug #57923 (Resolved): log: writes to stderr (pipe) may not be atomic
- This can lead to logging from pods like:...
- 02:09 PM Dashboard Bug #57922 (Resolved): mgr/dashboard: error message displaying when editing journam mirror image
- h3. Description of problem
An error message is being displayed when editing journal mirror image
!image1.png! - 12:52 PM CephFS Feature #57090: MDSMonitor,mds: add MDSMap flag to prevent clients from connecting
- Dhairya, status on this?
- 12:18 PM Orchestrator Cleanup #57921 (Resolved): orchestrator: orch upgrade status help message is wrong
- Currently the `orch upgrade status` help message is just a copy of the `orch upgrade check` help message...
- 09:32 AM CephFS Bug #57920 (New): mds:ESubtreeMap event size is too large
- In production environment, we have a problem: The ESubtreeMap event size is too large.
1,The ESubtreeMap event siz... - 07:28 AM ceph-volume Bug #57918: CEPHADM_REFRESH_FAILED: failed to probe daemons or devices
- Fixed the issue by removing the unused disk, but an empty disk shouldn't be a issue.
- 06:56 AM ceph-volume Bug #57918 (Resolved): CEPHADM_REFRESH_FAILED: failed to probe daemons or devices
- Last friday I upgrade the Ceph cluster successfully from 17.2.3 to 17.2.5 with "ceph orch upgrade start --image local...
- 07:18 AM rgw Bug #57919 (New): bucket can not be resharded after cancelling prior reshard process
- Hi,
we run a multisite setup where only the metadata get synced, but not the actual data.
I wanted to reshard a b... - 05:52 AM rgw Bug #56248: crash: rgw::ARN::ARN(rgw_bucket const&)
- Fixed in https://tracker.ceph.com/issues/55765 and https://github.com/ceph/ceph/pull/47194/commits is waiting for rel...
- 05:47 AM rgw Bug #56248: crash: rgw::ARN::ARN(rgw_bucket const&)
- We had a RGW crash on this as well some hours ago....
- 05:06 AM Backport #57916 (In Progress): quincy: make-dist creates ceph.spec with incorrect Release tag for...
- 03:08 AM Backport #57916 (In Progress): quincy: make-dist creates ceph.spec with incorrect Release tag for...
- https://github.com/ceph/ceph/pull/48613
- 04:00 AM Orchestrator Cleanup #50168 (Resolved): cephadm: move bin/cephadm from the git tree to download.ceph.com
- 03:02 AM Orchestrator Backport #57638 (Resolved): pacific: applying osd service spec with size filter fails if there's ...
- 02:58 AM Bug #57893 (Pending Backport): make-dist creates ceph.spec with incorrect Release tag for SUSE-ba...
10/24/2022
- 06:18 PM RADOS Bug #57852: osd: unhealthy osd cannot be marked down in time
- Not a something we introduced recently but still worth taking a look if nothing urgent is not the plate.
- 06:17 PM RADOS Bug #57852 (New): osd: unhealthy osd cannot be marked down in time
- For the detailed explanation!
- 06:10 PM RADOS Bug #57845: MOSDRepOp::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, SERVER_OCTOPUS...
- Just before the crash time-outs were seen:...
- 06:05 PM RADOS Bug #57915: LibRadosWatchNotify.AioNotify - error callback ceph_assert(ref > 0)
- Yes, this is one of the Notify bugs that i hit during my tests
- 05:14 PM RADOS Bug #57915: LibRadosWatchNotify.AioNotify - error callback ceph_assert(ref > 0)
- Nitzan, I recall you mentioned about some watch-related tests on today's stand-up. Is this one of them?
- 05:57 PM RADOS Bug #57546: rados/thrash-erasure-code: wait_for_recovery timeout due to "active+clean+remapped+la...
- As this is about EC: can be acting's items duplicated?
- 05:55 PM RADOS Bug #57546: rados/thrash-erasure-code: wait_for_recovery timeout due to "active+clean+remapped+la...
- If https://github.com/ceph/ceph/pull/47901/commits/0d07b406dc2f854363f7ae9b970e980400f4f03e is the actual culprit, th...
- 05:42 PM RADOS Bug #57883: test-erasure-code.sh: TEST_rados_put_get_jerasure fails on "rados_put_get: grep '\<5...
- It looks we asked for taking osd.5 down, got a confirmation the command was handled by mon and then @get_osd@ said %5...
- 05:25 PM RADOS Bug #57900: mon/crush_ops.sh: mons out of quorum
- Just **suggestion** from the bug scrub: this is a mon thrashing test. None of mon loga seems to have a trace of crash...
- 05:18 PM RADOS Bug #55141: thrashers/fastread: assertion failure: rollback_info_trimmed_to == head
- Well, just found a new occurance.
- 05:11 PM RADOS Bug #55141: thrashers/fastread: assertion failure: rollback_info_trimmed_to == head
- Lowering the priority as we haven't seen a reoccurence last time.
- 05:17 PM RADOS Bug #57913 (Duplicate): Thrashosd: timeout 120 ceph --cluster ceph osd pool rm unique_pool_2 uniq...
- In the teuthology log:...
- 05:10 PM RADOS Bug #57529 (Fix Under Review): mclock backfill is getting higher priority than WPQ
- 04:41 PM Bug #57914: centos 8 build failed
- All branches failed the same way
- 03:10 PM Bug #57914: centos 8 build failed
- from those logs, i see the builds succeeding. at the end of both, i see:
> Error: authenticating creds for "quay.cep... - 03:52 PM rgw Bug #19988 (Resolved): RGW: can't stack compression and encryption filters
- 11:37 AM rgw Bug #44660: Multipart re-uploads cause orphan data
- Looking at the code. In `MultipartObjectProcessor::process_first_chunk`, if writing the multipart object first chunk ...
- 11:24 AM bluestore Bug #57895: OSD crash in Onode::put()
- OK, thanks Igor for your confirmation, I'm reviewing your patch, we can discuss over there.
- 04:06 AM RADOS Bug #56101: Gibba Cluster: 17.2.0 to 17.2.1 RC upgrade OSD crash in function safe_timer
- Laura Flores wrote:
> Notes from the rados suite review:
>
> We may need to check if we're shutting down while se... - 02:48 AM bluestore Bug #57855: cannot enable level_compaction_dynamic_level_bytes
- db_paths is not compatible with level_compaction_dynamic_level_bytes.
10/23/2022
- 07:05 PM rgw Bug #57899 (Fix Under Review): admin: cannot use tenant with notification topic
- 11:45 AM RADOS Bug #57915 (New): LibRadosWatchNotify.AioNotify - error callback ceph_assert(ref > 0)
- /a//nmordech-2022-10-23_05:26:13-rados:verify-wip-nm-51282-distro-default-smithi/7077932...
- 05:19 AM RADOS Bug #57699: slow osd boot with valgrind (reached maximum tries (50) after waiting for 300 seconds)
- Sridher, yes, those trackers look the same, valgrind make the osd start slower, maybe that's the reason we are seeing...
10/22/2022
- 03:16 AM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- Seems I haven't seen the "host address is empty" error in about 10 days now.... Not sure if that is because of DNS,...
10/21/2022
- 09:16 PM Bug #57914 (Resolved): centos 8 build failed
- I see it on main and pacific
https://jenkins.ceph.com/job/ceph-dev-build/ARCH=x86_64,AVAILABLE_ARCH=x86_64,AVAILAB... - 06:26 PM bluestore Bug #53002: crash BlueStore::Onode::put from BlueStore::TransContext::~TransContext
- Hi Sven,
Thanks for reporting telemetry! The issue you reported is tracked in https://tracker.ceph.com/issues/5620... - 04:41 PM bluestore Bug #53002: crash BlueStore::Onode::put from BlueStore::TransContext::~TransContext
- We have almost daily crashes on our octopus cluster, which are also reported via telemetry, which look like this bug,...
- 05:31 PM Backport #57505 (Resolved): quincy: openSUSE Leap 15.x needs to explicitly specify gcc-11
- 03:26 PM Backport #57505: quincy: openSUSE Leap 15.x needs to explicitly specify gcc-11
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48058
merged - 05:28 PM rbd Bug #52915: rbd du versus rbd diff values wildly different when snapshots are present
- Alex Yarbrough wrote:
> If I _rbd du_ all of the ~200 images that I have, and sum the result, my total is about 24 T... - 03:25 PM rbd Bug #52915: rbd du versus rbd diff values wildly different when snapshots are present
- Ilya, first thank you for the time you put into your messages. I am aware of the issue regarding RBD object size vers...
- 04:52 PM CephFS Backport #57719: quincy: Test failure: test_subvolume_group_ls_filter_internal_directories (tasks...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48327
merged - 04:19 PM RADOS Bug #55809: "Leak_IndirectlyLost" valgrind report on mon.c
- /a/yuriw-2022-10-12_16:24:50-rados-wip-yuri8-testing-2022-10-12-0718-quincy-distro-default-smithi/7063948/
- 04:16 PM RADOS Bug #57913 (Duplicate): Thrashosd: timeout 120 ceph --cluster ceph osd pool rm unique_pool_2 uniq...
- /a/yuriw-2022-10-12_16:24:50-rados-wip-yuri8-testing-2022-10-12-0718-quincy-distro-default-smithi/7063868/
rados/t... - 03:57 PM rbd Backport #57843 (Resolved): quincy: rbd CLI inconsistencies affecting "--namespace" arg
- 03:29 PM rbd Backport #57843: quincy: rbd CLI inconsistencies affecting "--namespace" arg
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48458
merged - 03:55 PM Orchestrator Bug #56951: rook/smoke: Updating cephclusters/rook-ceph is forbidden
- /a/yuriw-2022-10-12_16:24:50-rados-wip-yuri8-testing-2022-10-12-0718-quincy-distro-default-smithi/7063866/
- 03:36 PM Orchestrator Bug #52321: qa/tasks/rook times out: 'check osd count' reached maximum tries (90) after waiting f...
- /a/yuriw-2022-10-12_16:24:50-rados-wip-yuri8-testing-2022-10-12-0718-quincy-distro-default-smithi/7063706/
- 12:41 PM Dashboard Bug #57912 (Fix Under Review): mgr/dashboard: Dashboard creation of NFS exports with RGW backend ...
- 12:12 PM Dashboard Bug #57912 (Fix Under Review): mgr/dashboard: Dashboard creation of NFS exports with RGW backend ...
- When attempting to create a NFS export with RGW as the backend from Dashboard, this fails as per the description.
Ho... - 10:39 AM rgw Bug #57911 (Pending Backport): Segmentation fault when uploading file with bucket policy on Quincy
- RGW crashes when a file is uploaded and a bucket policy has been set up.
The crash has been "reproduced for latest... - 10:28 AM bluestore Bug #57895: OSD crash in Onode::put()
- dongdong tao wrote:
> Yaarit Hatuka wrote:
> > Status changed from "New" to "Duplicate" since this issue duplicates... - 12:20 AM bluestore Bug #57895: OSD crash in Onode::put()
- Yaarit Hatuka wrote:
> Status changed from "New" to "Duplicate" since this issue duplicates https://tracker.ceph.com... - 09:53 AM Orchestrator Bug #57910 (New): ingress: HAProxy fails to start because keepalived IP address not yet available...
- After deploying a new cluster _sometimes_ HAProxy fails to start on ingress nodes:...
- 08:41 AM RADOS Bug #57699: slow osd boot with valgrind (reached maximum tries (50) after waiting for 300 seconds)
- @Nitzan Mordechai this is probably similar to,
https://tracker.ceph.com/issues/52948 and https://tracker.ceph.com/is... - 07:47 AM RADOS Fix #57040 (Resolved): osd: Update osd's IOPS capacity using async Context completion instead of ...
- 07:46 AM RADOS Backport #57443 (Resolved): quincy: osd: Update osd's IOPS capacity using async Context completio...
- 06:03 AM Orchestrator Feature #55490: cephadm: allow passing grafana cert and frontend-api-url in spec
- The OP mentioned @set-grafana-frontend-api-url@ but missed mentioning setting @set-grafana-api-url@ from a spec which...
10/20/2022
- 11:33 PM RADOS Bug #56101: Gibba Cluster: 17.2.0 to 17.2.1 RC upgrade OSD crash in function safe_timer
- Notes from the rados suite review:
We may need to check if we're shutting down while sending pg stats; if so, we d... - 10:47 PM bluestore Feature #57785: fragmentation score in metrics
- I'm just a user so I can't answer some of the questions. I'll fill in what I know though.
1. Not sure
3. No priva... - 10:26 PM bluestore Feature #57785: fragmentation score in metrics
- Hey Kevin (and Vikhyat),
I have a few questions regarding the fragmentation score:
1. Where are all the places ... - 09:25 PM rbd Bug #52915: rbd du versus rbd diff values wildly different when snapshots are present
- Going back to CephRBD_NVMe/vm-101-disk-0 image, your "rbd du" output makes perfect sense to me based on what you said...
- 09:12 PM rbd Bug #52915: rbd du versus rbd diff values wildly different when snapshots are present
- Hi Alex,
"rbd diff CephRBD_NVMe/vm-101-disk-0" reports the allocated areas of the image without taking snapshots i... - 03:40 PM rbd Bug #52915: rbd du versus rbd diff values wildly different when snapshots are present
- Greetings all. I have read through the related issues that are resolved. I do not believe this issue is duplicated or...
- 06:11 PM Orchestrator Feature #57909 (Resolved): cephadm: make logging host refresh data to debug logs configurable
- The amount of data we log in the debug logs when refreshing a host is too verbose, even for debug level. It renders t...
- 04:14 PM Support #57908 (New): rgw common prefix performance on large bucket
- Hi, I'm facing the same issue metioned here:
https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/36P62BOOCJBVVJ... - 04:09 PM ceph-volume Bug #57907: ceph-volume complains about "Insufficient space (<5GB)" on 1.75TB device
- I add a workaround screenshot to disable Hotplug in Bios.
- 04:01 PM ceph-volume Bug #57907: ceph-volume complains about "Insufficient space (<5GB)" on 1.75TB device
- The Problem is that in @util/device.py@ line 582
The call for @int(self.sys_api.get('size', 0))@ is always 0 if s... - 03:15 PM ceph-volume Bug #57907 (Duplicate): ceph-volume complains about "Insufficient space (<5GB)" on 1.75TB device
- On a one week old working cluster 17.2.5, i try to add another host with 2 SSDs and 4 HDDs.
None of them is shown as... - 03:07 PM RADOS Bug #57152 (Resolved): segfault in librados via libcephsqlite
- 03:06 PM RADOS Backport #57373 (Resolved): pacific: segfault in librados via libcephsqlite
- 02:56 PM RADOS Backport #57373: pacific: segfault in librados via libcephsqlite
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48187
merged - 03:01 PM Orchestrator Backport #57638: pacific: applying osd service spec with size filter fails if there's tiny (KB-si...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48243
merged - 02:58 PM Orchestrator Backport #57639: pacific: cephadm: `ceph orch ps` doesn't list container versions in some cases
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48210
merged - 02:55 PM ceph-volume Backport #57566: pacific: inventory a device get_partitions_facts called many times
- Guillaume Abrioux wrote:
> https://github.com/ceph/ceph/pull/48126
merged - 02:53 PM ceph-volume Backport #57564: pacific: functional test lvm-centos8-filestore-create is broken
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48123
merged - 02:45 PM Bug #57906 (New): ceph -s show too many executing tasks
- I got a log of execution tasks with ceph -s but I'm sure there is nothing running. How I can clean this messages? Als...
- 02:24 PM rgw Bug #57770 (Triaged): RGW (pacific) misplaces index entries after dynamically resharding bucket
- 02:24 PM rgw Bug #57770 (New): RGW (pacific) misplaces index entries after dynamically resharding bucket
- 02:21 PM rgw Bug #57783: multisite: data sync reports shards behind after source zone fully trims datalog
- related work in https://github.com/ceph/ceph/pull/47682 and https://github.com/ceph/ceph/pull/48397
- 02:20 PM rgw Bug #57804: Enabling sync on bucket not working
- i can only recommend running the command until it succeeds
- 02:18 PM rgw Bug #57853 (Need More Info): multisite sync process block after long time running
- 02:16 PM rgw Bug #57901 (Fix Under Review): s3:ListBuckets response limited to 1000 buckets (by default) since...
- 02:11 PM rgw Bug #57231 (Resolved): Valgrind: jump on unitialized in s3select
- 01:51 PM bluestore Bug #57895 (Duplicate): OSD crash in Onode::put()
- Status changed from "New" to "Duplicate" since this issue duplicates https://tracker.ceph.com/issues/56382.
- 10:10 AM bluestore Bug #57895: OSD crash in Onode::put()
- Please help to review this one, https://github.com/ceph/ceph/pull/48566
Here is the related log: https://pastebin.... - 01:30 PM rgw Bug #57905 (Pending Backport): multisite: terminate called after throwing an instance of 'ceph::b...
- example from rgw/multisite suite: http://qa-proxy.ceph.com/teuthology/cbodley-2022-10-19_23:28:37-rgw-wip-cbodley-tes...
- 10:54 AM bluestore Bug #56851: crash: int BlueStore::read_allocation_from_onodes(SimpleBitmap*, BlueStore::read_allo...
- @Sudhin - curious if you can reproduce the issue? If so it would be great to get OSD log with debug-bluestore set to ...
- 10:52 AM bluestore Bug #52464: FAILED ceph_assert(current_shard->second->valid())
- IMO this is rather related to DB sharding stuff introduced by https://github.com/ceph/ceph/pull/34006
Hence reassign... - 10:46 AM bluestore Bug #52464: FAILED ceph_assert(current_shard->second->valid())
- Neha Ojha wrote:
> Gabi, I am assigning it to you for now, since this looks related to NCB.
No, apparently this i... - 09:49 AM bluestore Bug #57857 (Fix Under Review): KernelDevice::read doesn't translate error codes correctly
- 09:40 AM bluestore Bug #56382 (Fix Under Review): ONode ref counting is broken
- 09:10 AM bluestore Bug #56382 (Pending Backport): ONode ref counting is broken
- 06:33 AM CephFS Bug #54557 (Fix Under Review): scrub repair does not clear earlier damage health status
- 06:24 AM Dashboard Bug #57284 (Resolved): mgr/dashboard: 500 internal server error seen on ingress service creation ...
- 06:24 AM Dashboard Backport #57485 (Resolved): pacific: mgr/dashboard: 500 internal server error seen on ingress ser...
- 05:57 AM rgw Bug #57562: multisite replication issue on Quincy
- We have an example scenario here where one of the objects in a bucket failed to be synced to the secondary.
* Mdlog... - 05:28 AM CephFS Backport #57716 (Resolved): pacific: libcephfs: incorrectly showing the size for snapdirs when st...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48413
Merged. - 04:54 AM CephFS Backport #57874 (In Progress): quincy: Permissions of the .snap directory do not inherit ACLs
- 04:17 AM CephFS Backport #57723 (Resolved): pacific: qa: test_subvolume_snapshot_info_if_orphan_clone fails
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48417
Merged.
10/19/2022
- 11:31 PM rbd Bug #57902 (Resolved): [rbd-nbd] add --snap-id option to "rbd device map" to allow mapping arbitr...
- As any snapshot in a non-user snapshot namespace, mirror snapshots are inaccessible to most rbd CLI commands. As suc...
- 11:16 PM rbd Bug #57066 (Fix Under Review): rbd snap list not change the last read when more than 64 group snaps
- 09:28 PM rgw Bug #57901 (Resolved): s3:ListBuckets response limited to 1000 buckets (by default) since Octopus
- Since Octopus, s3:ListBuckets is limited to rgw_list_buckets_max_chunk buckets in its response due to loss of truncat...
- 09:21 PM RADOS Backport #52747 (In Progress): pacific: MON_DOWN during mon_join process
- 09:09 PM RADOS Backport #52746 (Rejected): octopus: MON_DOWN during mon_join process
- Octopus is EOL.
- 08:59 PM RADOS Bug #43584: MON_DOWN during mon_join process
- /a/yuriw-2022-10-05_20:44:57-rados-wip-yuri4-testing-2022-10-05-0917-pacific-distro-default-smithi/7055594
- 08:46 PM RADOS Bug #57900 (In Progress): mon/crush_ops.sh: mons out of quorum
- /a/teuthology-2022-10-09_07:01:03-rados-quincy-distro-default-smithi/7059463...
- 05:56 PM Orchestrator Bug #57341: cephadm: failures from tests comparing output strings are difficult to debug
- See attached screenshot for a better colorized example.
- 05:53 PM Orchestrator Bug #57341: cephadm: failures from tests comparing output strings are difficult to debug
- I did a few minutes of research and found two packages that may help:
pytest-mock (https://pytest-mock.readthedocs.i... - 03:38 PM Linux kernel client Bug #57898: ceph client extremely slow kernel version between 5.15 and 6.0
- Even with the ceph-fuse method in the body it gets slow again over time.
- 12:47 PM Linux kernel client Bug #57898 (In Progress): ceph client extremely slow kernel version between 5.15 and 6.0
- hello? I am very new to ceph. Thank you for taking that into consideration and reading.
I recently changed the ker... - 03:20 PM RADOS Bug #57698 (Pending Backport): osd/scrub: "scrub a chunk" requests are sent to the wrong set of r...
- 03:05 PM rgw Bug #16767 (In Progress): RadosGW Multipart Cleanup Failure
- 02:55 PM rgw Bug #16767: RadosGW Multipart Cleanup Failure
- Vicki Good wrote:
> I've encountered this bug in Ceph 14 and 15 and it's a pretty big problem for us for the same re... - 02:16 PM CephFS Backport #57875 (In Progress): pacific: Permissions of the .snap directory do not inherit ACLs
- 01:45 PM bluestore Bug #57855: cannot enable level_compaction_dynamic_level_bytes
- I found that the level_compaction_dynamic_level_bytes option does not apply if opt.db_paths exists when opening rocks...
- 01:26 PM bluestore Bug #55324: rocksdb omap iterators become extremely slow in the presence of large delete range to...
- Benoît Knecht wrote:
> > I see this was backported in: https://github.com/ceph/ceph/pull/45963 but was later reverte... - 12:09 PM bluestore Bug #55324: rocksdb omap iterators become extremely slow in the presence of large delete range to...
- Sven Kieske wrote:
> I assume this was not backported to the last octopus release?
Yes, the octopus is EOL - 12:04 PM bluestore Bug #55324: rocksdb omap iterators become extremely slow in the presence of large delete range to...
- > I see this was backported in: https://github.com/ceph/ceph/pull/45963 but was later reverted in https://github.com/...
- 11:21 AM bluestore Bug #55324: rocksdb omap iterators become extremely slow in the presence of large delete range to...
- Sven Kieske wrote:
> I don't see the PR showing up in any release notes. I assume this was not backported to the las... - 11:16 AM bluestore Bug #55324: rocksdb omap iterators become extremely slow in the presence of large delete range to...
- I don't see the PR showing up in any release notes. I assume this was not backported to the last octopus release? In ...
- 09:06 AM bluestore Bug #55324 (Resolved): rocksdb omap iterators become extremely slow in the presence of large dele...
- 01:20 PM rgw-testing Bug #54104: test_rgw_datacache.py: s3cmd fails with '403 (SignatureDoesNotMatch)' in ubuntu
- ping @Mark, this remains a blocker for enabling ubuntu in the rgw/verify suite. that subsuite contains most of our fu...
- 01:11 PM rgw Bug #57899 (Pending Backport): admin: cannot use tenant with notification topic
- issue was a regression introduced in: 200f71a90c9e77c91452cec128c2c8be0d3d6f1f
topic notification commands should be... - 01:03 PM mgr Bug #55046 (Resolved): mgr: perf counters node exporter
- 12:59 PM mgr Backport #57141 (Resolved): quincy: mgr: perf counters node exporter
- 12:27 PM Orchestrator Bug #57897 (New): ceph mgr restart causes restart of all iscsi daemons in a loop
- We have observed that since v17.2.4, a restart of the active ceph mgr appears to cause all iSCSI daemons to restart a...
- 11:49 AM Dashboard Feature #57896 (New): mgr/dashboard: create per component high level dashboard view
- h3. Description of problem
A great improvemnte to the dashboard would be to have a higher level view of each compo... - 11:49 AM bluestore Bug #57895: OSD crash in Onode::put()
- This is observed from 15.2.16, but I believe the code defect to cause this kind of race condition is still present on...
- 11:42 AM bluestore Bug #57895 (Duplicate): OSD crash in Onode::put()
This issue happens when an Onode is being trimmed right away after it's unpinned. This is possible when the LRU lis...- 11:01 AM Bug #57868: iSCSI: rbd-target-api reports python version and identified 'unsupported version' tri...
- This likely goes for all ceph-container containers... Guillaume, could you please take a look?
- 10:30 AM Orchestrator Feature #57894 (Fix Under Review): Move prometheus spec check to the service_spec module
- 10:18 AM Orchestrator Feature #57894 (Pending Backport): Move prometheus spec check to the service_spec module
- 10:29 AM RADOS Bug #57699: slow osd boot with valgrind (reached maximum tries (50) after waiting for 300 seconds)
- The issue is that we having deadlock on specific condition. When we are trying to update the mClockScheduler config c...
- 09:13 AM CephFS Bug #57882: Kernel Oops, kernel NULL pointer dereference
- Xiubo Li wrote:
> It's a known bug and I will check this today or this week.
Oh my ! I did search for anything pr... - 08:46 AM bluestore Bug #55328 (Closed): OSD crashed due to checksum error
- 08:45 AM Bug #57893 (Fix Under Review): make-dist creates ceph.spec with incorrect Release tag for SUSE-ba...
- 08:04 AM Bug #57893 (Pending Backport): make-dist creates ceph.spec with incorrect Release tag for SUSE-ba...
- @ceph.spec.in@ says:...
- 07:43 AM Dashboard Bug #57805 (Pending Backport): mgr/dashboard: Unable to change subuser permission
- 07:42 AM Dashboard Bug #57805 (Resolved): mgr/dashboard: Unable to change subuser permission
- 07:42 AM Dashboard Backport #57841 (Resolved): quincy: mgr/dashboard: Unable to change subuser permission
- 07:33 AM Dashboard Feature #57826 (Resolved): mgr/dashboard: add server side encryption to rgw/s3
- 07:33 AM Dashboard Backport #57835 (Resolved): quincy: mgr/dashboard: add server side encryption to rgw/s3
- 05:57 AM rbd Bug #57872 (Fix Under Review): [pwl] inconsistent "rbd status" output (clean = true but dirty_byt...
- 05:31 AM RADOS Bug #57546: rados/thrash-erasure-code: wait_for_recovery timeout due to "active+clean+remapped+la...
- I was able to reproduce this using the test Laura mentioned above - http://pulpito.front.sepia.ceph.com/amathuri-2022...
- 05:12 AM Dashboard Bug #39726 (Resolved): mgr/dashboard: "Striping" feature checkbox missing in RBD image dialog
- 05:12 AM Dashboard Backport #56566 (Resolved): pacific: mgr/dashboard: "Striping" feature checkbox missing in RBD im...
- 05:06 AM crimson Bug #57629: crimson: segfault during mkfs
- using GCC 12.2.0 on ubuntu 22.04 facing the same problem.
- 03:26 AM crimson Bug #57549: Crimson: Alienstore not work after ceph enable c++20
- This problem disappeared after update GCC complier to the 12.2.0 version. And I met the Segmentation fault on https:/...
10/18/2022
- 07:16 PM Dashboard Bug #48258: mgr/dashboard: Switch from tslint to eslint
- Thanks Nizam, will get working
- 06:25 PM Documentation #57858: v17.2.4 release does not contain latest cherry-picks
- Bottom line: The quincy-release branch (and future release branches) should be up-to-date on the Ceph repository for ...
- 06:04 PM Orchestrator Bug #57891 (Resolved): [Gibba Cluster] HEALTH_ERR: Upgrade: failed due to an unexpected exception
- - Upgrade paused due to one host not being reachable in the cluster.
- Resumed the upgrade with the resume command
... - 05:29 PM Bug #57890: cmd_getval() throws but many callers don't catch the exception
- For reference, here are crashes with `cmd_getval` in their backtrace:
http://telemetry.front.sepia.ceph.com:4000/d/N... - 05:02 PM Bug #57890 (New): cmd_getval() throws but many callers don't catch the exception
- In https://github.com/ceph/ceph/pull/23557 we switched @cmd_getval()@ to throw on error. This family of functions hav...
- 04:31 PM RADOS Bug #51729: Upmap verification fails for multi-level crush rule
- Chris, can you please provide your osdmap binary?
- 04:13 PM rgw Backport #57889 (Rejected): pacific: amqp: rgw crash when ca location is used for amqp connections
- 04:12 PM rgw Backport #57888 (In Progress): quincy: amqp: rgw crash when ca location is used for amqp connections
- https://github.com/ceph/ceph/pull/54170
- 04:08 PM rgw Bug #57850 (Pending Backport): amqp: rgw crash when ca location is used for amqp connections
- 03:49 PM Orchestrator Backport #57787 (In Progress): quincy: mgr/nfs: Add a sectype field to nfs exports created by nfs...
- 03:39 PM rgw Bug #57881 (Fix Under Review): LDAP invalid password resource leak fix
- 09:56 AM rgw Bug #57881: LDAP invalid password resource leak fix
- I created a pull request for a possible fix:
https://github.com/ceph/ceph/pull/48509 - 01:02 PM rgw Bug #57877 (Fix Under Review): rgw: some operations may not have a valid bucket object
- 09:53 AM mgr Backport #57887 (In Progress): pacific: mgr/prometheus: avoid duplicates and deleted entries for ...
- 09:04 AM mgr Backport #57887 (Resolved): pacific: mgr/prometheus: avoid duplicates and deleted entries for rbd...
- https://github.com/ceph/ceph/pull/48524
- 09:49 AM mgr Backport #57886 (In Progress): quincy: mgr/prometheus: avoid duplicates and deleted entries for r...
- 09:04 AM mgr Backport #57886 (Resolved): quincy: mgr/prometheus: avoid duplicates and deleted entries for rbd_...
- https://github.com/ceph/ceph/pull/48523
- 09:35 AM Linux kernel client Bug #47450 (Resolved): stop parsing the error string in the session reject message
- Fixed in:...
- 09:33 AM Linux kernel client Bug #46904: kclient: cluster [WRN] client.4478 isn't responding to mclientcaps(revoke)
- Fixed it in kernel and the patchwork link: https://patchwork.kernel.org/project/ceph-devel/list/?series=686074
- 09:27 AM Backport #57885 (In Progress): quincy: disable system_pmdk on s390x for SUSE distros
- 08:49 AM Backport #57885 (Resolved): quincy: disable system_pmdk on s390x for SUSE distros
- https://github.com/ceph/ceph/pull/48522
- 09:03 AM RADOS Bug #57845: MOSDRepOp::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, SERVER_OCTOPUS...
- Hi Neha,
the logs from the crash instance that I reported initially are already rotated out on the particular node... - 09:00 AM mgr Bug #57797 (Pending Backport): mgr/prometheus: avoid duplicates and deleted entries for rbd_stats...
- 08:41 AM Bug #57860 (Pending Backport): disable system_pmdk on s390x for SUSE distros
- 08:19 AM Orchestrator Bug #57096: osd not restarting after upgrading to quincy due to podman args --cgroups=split
- I manually created the unit.meta, and it seems to work. thanks again.
- 06:28 AM Orchestrator Bug #57096: osd not restarting after upgrading to quincy due to podman args --cgroups=split
- The unit.meta file is not yet present in Octopus. I'll try to figure something out or wait for the PR release.
Tha... - 02:48 AM RADOS Bug #57852: osd: unhealthy osd cannot be marked down in time
- Radoslaw Zarzynski wrote:
> Could you please clarify a bit? Do you mean there some extra, unnecessary (from the POV ... - 02:19 AM CephFS Backport #57880 (In Progress): pacific: NFS client unable to see newly created files when listing...
- 02:14 AM CephFS Backport #57879 (In Progress): quincy: NFS client unable to see newly created files when listing ...
- 12:52 AM CephFS Bug #57882 (Duplicate): Kernel Oops, kernel NULL pointer dereference
- It's a known bug and I will check this today or this week.
10/17/2022
- 07:29 PM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- alright, looking back at the original traceback...
- 06:55 PM Orchestrator Bug #57884 (Resolved): cephadm: attempting a daemon redeploy of the active mgr with a specified i...
- If I run something like...
- 06:27 PM RADOS Bug #57796: after rebalance of pool via pgupmap balancer, continuous issues in monitor log
- Link to the discussion on ceph-users: https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/AZHAIGY3BIM4SGB...
- 06:20 PM RADOS Bug #57883: test-erasure-code.sh: TEST_rados_put_get_jerasure fails on "rados_put_get: grep '\<5...
- Let's first see if it's easily reproducible:
http://pulpito.front.sepia.ceph.com/lflores-2022-10-17_18:19:55-rados:s... - 06:03 PM RADOS Bug #57883: test-erasure-code.sh: TEST_rados_put_get_jerasure fails on "rados_put_get: grep '\<5...
- The failed function:
qa/standalone/erasure-code/test-erasure-code.sh... - 05:52 PM RADOS Bug #57883 (Resolved): test-erasure-code.sh: TEST_rados_put_get_jerasure fails on "rados_put_get:...
- /a/yuriw-2022-10-13_17:24:48-rados-main-distro-default-smithi/7065580...
- 06:16 PM RADOS Bug #57845 (Need More Info): MOSDRepOp::encode_payload(uint64_t): Assertion `HAVE_FEATURE(feature...
- These reports in telemetry look similar: http://telemetry.front.sepia.ceph.com:4000/d/Nvj6XTaMk/spec-search?orgId=1&v...
- 06:08 PM RADOS Bug #57852 (Need More Info): osd: unhealthy osd cannot be marked down in time
- Could you please clarify a bit? Do you mean there some extra, unnecessary (from the POV of jugging whether an OSD is ...
- 06:01 PM mgr Bug #57460: Json formatted ceph pg dump hangs on large clusters
- Thanks, Radoslow! I'll look into modifying the patch as you suggested, targeting Reef.
- 05:48 PM RADOS Bug #57782: [mon] high cpu usage by fn_monstore thread
- NOT A FIX (extra debugs): https://github.com/ceph/ceph/pull/48513
- 05:45 PM RADOS Bug #57698 (Fix Under Review): osd/scrub: "scrub a chunk" requests are sent to the wrong set of r...
- 05:43 PM RADOS Bug #51729: Upmap verification fails for multi-level crush rule
- A note from bug scrub: this is going to be assigned tomorrow.
- 02:49 PM Bug #57613: Kernel Oops, kernel NULL pointer dereference
- Moved (copied) to cephfs, might have echo from a better spot :) this one can be closed.
- 02:47 PM CephFS Bug #57882 (Duplicate): Kernel Oops, kernel NULL pointer dereference
- (repost from Ceph (#57613), I couldn't find a way to move the bug entry from one project to another)
Hello everyon... - 02:19 PM rbd Backport #57779 (Resolved): quincy: [test] fio 3.16 doesn't build on recent kernels due to remova...
- 02:10 PM rbd Backport #57779: quincy: [test] fio 3.16 doesn't build on recent kernels due to removal of linux/...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48386
merged - 01:45 PM Dashboard Backport #57828 (Resolved): quincy: cephadm/test_dashboard_e2e.sh: Expected to find content: '/^f...
- 01:25 PM rbd Tasks #54312: combine the journal and snapshot test scripts
- Please set the state to Fix Under Review once the lab stuff is sorted out and you have a link to a test run.
- 01:22 PM rbd Bug #57066 (In Progress): rbd snap list not change the last read when more than 64 group snaps
- 12:30 PM rgw Bug #57881 (Pending Backport): LDAP invalid password resource leak fix
- I have noticed that in the case a User tries to log in using LDAP with a wrong password, two new LDAP sessions will b...
- 12:04 PM CephFS Backport #57880 (Resolved): pacific: NFS client unable to see newly created files when listing di...
- https://github.com/ceph/ceph/pull/48521
- 12:04 PM CephFS Backport #57879 (Resolved): quincy: NFS client unable to see newly created files when listing dir...
- https://github.com/ceph/ceph/pull/48520
- 11:57 AM CephFS Bug #57210 (Pending Backport): NFS client unable to see newly created files when listing director...
- 11:53 AM CephFS Backport #57261 (Resolved): pacific: standby-replay mds is removed from MDSMap unexpectedly
- 10:54 AM Orchestrator Feature #57878 (Resolved): Add typing checks for rgw module
- 10:23 AM bluestore Bug #57855: cannot enable level_compaction_dynamic_level_bytes
- I did some more digging on this and found that this PR was the cause.
https://github.com/ceph/ceph/pull/43100
- 09:32 AM Orchestrator Bug #57876 (Fix Under Review): prometheus ERROR failed to collect metrics
- 09:13 AM Orchestrator Bug #57876 (Resolved): prometheus ERROR failed to collect metrics
- ...
- 09:19 AM rgw Bug #57877 (Resolved): rgw: some operations may not have a valid bucket object
- Some codepaths may not always have a valid bucket, so add checks to detect this.
- 08:57 AM Linux kernel client Bug #46904 (Fix Under Review): kclient: cluster [WRN] client.4478 isn't responding to mclientcaps...
- 04:52 AM Linux kernel client Bug #46904: kclient: cluster [WRN] client.4478 isn't responding to mclientcaps(revoke)
- The MDS was waiting for _*Fw*_ caps:...
- 03:43 AM Linux kernel client Bug #56524 (Resolved): xfstest-dev: generic/467 failed with "open_by_handle(/mnt/kcephfs.A/467-di...
- 03:42 AM Linux kernel client Bug #57321 (Resolved): xfstests: ceph/004 setfattr: /mnt/kcephfs.A/test-004/dest: Invalid argument
- 03:41 AM Linux kernel client Bug #57342 (Resolved): kclient: incorrectly showing the size for snapdirs when stating them
10/16/2022
- 02:50 PM CephFS Backport #57875 (Resolved): pacific: Permissions of the .snap directory do not inherit ACLs
- https://github.com/ceph/ceph/pull/48553
- 02:50 PM CephFS Backport #57874 (Resolved): quincy: Permissions of the .snap directory do not inherit ACLs
- https://github.com/ceph/ceph/pull/48563
- 02:49 PM CephFS Bug #57084 (Pending Backport): Permissions of the .snap directory do not inherit ACLs
- 02:46 PM CephFS Bug #57084 (Resolved): Permissions of the .snap directory do not inherit ACLs
10/15/2022
- 08:36 PM crimson Bug #57873 (New): crimson: override overrides.ceph.flavor in crimson_qa_overrides.yaml as well
- overrides.ceph.flavor = default gets set by teuthology/suite/placeholder.py
- 09:19 AM rbd Bug #57872 (Resolved): [pwl] inconsistent "rbd status" output (clean = true but dirty_bytes = 61440)
- This popped up in a quincy integration branch run, but the code in main is exactly the same:...
10/14/2022
- 09:17 PM rgw Bug #52027: XML responses return different order of XML elements
- Hi
I think this is not fully addressed.
I've added a comment to pull request https://github.com/ceph/ceph/pull/42... - 09:13 PM RADOS Bug #51729: Upmap verification fails for multi-level crush rule
- Andras,
Thanks for the extra info. This needs to be addressed. Anyone?
- 08:48 PM RADOS Bug #51729: Upmap verification fails for multi-level crush rule
- Just to clarify - the error "verify_upmap number of buckets X exceeds desired Y" comes from the C++ code in ceph-mon ...
- 06:47 PM RADOS Bug #51729: Upmap verification fails for multi-level crush rule
- I am now seeing this issue on pacific, 16.2.10 on rocky8 linux.
If I have a >2 level rule on an ec pool (6+2), suc... - 06:54 PM rgw Backport #57430: quincy: key is used after move in RGWGetObj_ObjStore_S3::override_range_hdr
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48228
merged - 06:50 PM Orchestrator Bug #57870 (Resolved): cephadm: --apply-spec is trying to do too much and failing as a result
- --apply-spec is intended to do 2 things:
1) distribute ssh keys to hosts with hosts specs in the applied spec
2) ... - 04:15 PM RADOS Bug #57698: osd/scrub: "scrub a chunk" requests are sent to the wrong set of replicas
- Following some discussions: here are excerpts from a run demonstrating this issue.
Test run rfriedma-2022-09-28_15:5... - 04:04 PM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- Oh, by all combinations, I mean I created DNS entries for all hosts, not just ceph02.
- 04:03 PM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- I add DNS entries for all combinations. So both ceph02.oldname.local and ceph02.domain.local are now valid names but...
- 02:05 PM rbd Tasks #54312 (In Progress): combine the journal and snapshot test scripts
- 01:42 PM rgw Bug #44660: Multipart re-uploads cause orphan data
- Writing on behalf of Ulrich Klein <Ulrich.Klein@ulrichklein.de>, he wanted to add some info to this tracker, below is...
- 10:45 AM Orchestrator Bug #57781 (Rejected): Fix prometheus dependencies calculation
- closing as the current behavior is correct. We just need to add some comments to clarify the logic.
- 10:21 AM Orchestrator Bug #57366 (Pending Backport): prometheus is not re-deployed when service-discovery port changes
- 10:20 AM Orchestrator Bug #57816 (Fix Under Review): Add support to configure protocol (http or https) for Grafana url ...
- 09:24 AM Dashboard Bug #48258: mgr/dashboard: Switch from tslint to eslint
- great, thanks Sedrick. you can assign it to you. There are two PRs opened currently. You can go over the discussions ...
- 09:19 AM Dashboard Bug #48258: mgr/dashboard: Switch from tslint to eslint
- Hi, will like to work on this one
- 08:15 AM rgw Bug #57804: Enabling sync on bucket not working
- Hello Casey,
The init command ended after 60 minutes running.
Unfortunately the two errors are returned constan... - 07:46 AM Bug #57868 (New): iSCSI: rbd-target-api reports python version and identified 'unsupported versio...
- When running the cephadm deployed iSCSI container images, the API endpoint exposes python versions. This trigggers vu...
- 04:35 AM Dashboard Cleanup #57867 (Resolved): mgr/dashboard: migrate bootstrap 4 to 5
- h3. Description of problem
_here_
h3. Environment
* @ceph version@ string:
* Platform (OS/distro/release)... - 04:34 AM Dashboard Cleanup #57866 (Resolved): mgr/dashboard: update to angular 13
- 12:14 AM crimson Bug #57549: Crimson: Alienstore not work after ceph enable c++20
- do you mean rados bench works on ubuntu 20.04 in your machine for aliestore?
10/13/2022
- 11:27 PM CephFS Bug #48673 (Fix Under Review): High memory usage on standby replay MDS
- 10:19 PM Documentation #57858: v17.2.4 release does not contain latest cherry-picks
- ...
- 08:02 PM crimson Bug #57791 (Resolved): crimson: zero becomes truncate if region exceeds object bound
- https://github.com/ceph/ceph/pull/48405
- 08:02 PM crimson Bug #57789 (Resolved): crimson: add list_snaps
- https://github.com/ceph/ceph/pull/48405
- 08:02 PM crimson Bug #57773 (Resolved): crimson: TestLibRBD.TestCompareAndWriteStripeUnitSuccessPP fails with EINVAL
- https://github.com/ceph/ceph/pull/48405
- 08:02 PM crimson Bug #57759 (Resolved): crimson: rbdv1 needs TMAP, easier to implement than to skip rbdv1 tests
- https://github.com/ceph/ceph/pull/48405
- 06:40 PM Bug #57864 (In Progress): qa: fail "Checking cluster log for badness" check (and therefore the jo...
- 10:18 AM Bug #57864 (In Progress): qa: fail "Checking cluster log for badness" check (and therefore the jo...
- Discovered in https://github.com/ceph/ceph/pull/48288#discussion_r993883997:
----------
It appears there's a ca... - 04:08 PM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- it's odd that the hostname it reports not having an address for isn't even a hostname it has stored "ceph02.domain.lo...
- 03:10 PM CephFS Bug #54760 (Closed): crash: void CDir::try_remove_dentries_for_stray(): assert(dn->get_linkage()-...
- Venky Shankar wrote:
> I think https://github.com/ceph/ceph/pull/46331 would mitigate this issue, however, the unlin... - 03:07 PM rbd Backport #57780 (Resolved): pacific: [test] fio 3.16 doesn't build on recent kernels due to remov...
- 03:07 PM rbd Backport #57780: pacific: [test] fio 3.16 doesn't build on recent kernels due to removal of linux...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48385
merged - 02:18 PM rgw Bug #57783 (In Progress): multisite: data sync reports shards behind after source zone fully trim...
- 02:17 PM rgw Bug #57804 (Need More Info): Enabling sync on bucket not working
- did the `bucket sync enable` command finish? i imagine it would take a while with 60000 index shards..
- 02:15 PM rgw Bug #57807 (Duplicate): The cloud sync module does not work starting with Pacific
- 02:01 PM rgw Bug #57724 (Triaged): Keys returned by Admin API during user creation on secondary zone not valid
- 12:44 PM rgw Bug #57770: RGW (pacific) misplaces index entries after dynamically resharding bucket
- Nick Janus wrote:
> J. Eric Ivancich wrote:
> > The theory is that the bucket index shard does not exist at this mo... - 09:28 AM Dashboard Tasks #57863 (Resolved): mgr/dashboard: cluster-utilization card
- h3. Description
One of the cards of the new landing page (https://github.com/ceph/ceph/tree/feature-landing-page-r... - 09:18 AM Dashboard Bug #57018: host.containers.internal accessing grafana's performance graphs
- I see similar behaviour here and this started with podman 4.1 where podman is injecting an entry into /etc/hosts insi...
- 09:16 AM Dashboard Tasks #57862 (Resolved): mgr/dashboard: capacity card
- h3. Description
The capacity card is one of the cards for the new landing page. It is currently implemented on the... - 09:08 AM Dashboard Feature #57861 (Pending Backport): mgr/dashboard: Dashboard landing page revamp
- h3. Description
Tasks for the landing page revamp.
- 08:27 AM Linux kernel client Bug #57656 (Need More Info): [testing] dbench: write failed on handle 10009 (Resource temporaril...
- Today I spent more than half day to read the mds, osd side logs, but still couldn't find any suspect logs. Usually if...
- 07:39 AM RADOS Bug #57859 (Fix Under Review): bail from handle_command() if _generate_command_map() fails
- 03:51 AM RADOS Bug #57859 (Resolved): bail from handle_command() if _generate_command_map() fails
- https://tracker.ceph.com/issues/54558 catches an exception from handle_command() to avoid mon termination due to a po...
- 04:34 AM Bug #57860 (Fix Under Review): disable system_pmdk on s390x for SUSE distros
- 04:28 AM Bug #57860 (Pending Backport): disable system_pmdk on s390x for SUSE distros
- Same as https://tracker.ceph.com/issues/56491 which addressed RHEL and Fedora not shipping libpmem on s390x, but for ...
- 04:03 AM RADOS Bug #54558: malformed json in a Ceph RESTful API call can stop all ceph-mon services
- nikhil kshirsagar wrote:
> Ilya Dryomov wrote:
> > I don't think https://github.com/ceph/ceph/pull/45547 is a compl... - 03:21 AM crimson Bug #57549: Crimson: Alienstore not work after ceph enable c++20
- tried the latest version with gcc-12.2.0 and ubuntu 22.04, met the same problem on https://tracker.ceph.com/issues/57...
- 02:59 AM crimson Bug #57693 (Resolved): Messenger test failed against test_messenger_peer.cc
- The fix was merged.
- 02:58 AM crimson Bug #56589 (Resolved): perf-crimson-msgr: segmentation fault happens when shutdown
- The fix was merged.
- 02:56 AM crimson Bug #56520: perf-crimson-msgr: Aborting on shard 0
- The fix was merged.
- 02:55 AM crimson Bug #56520 (Resolved): perf-crimson-msgr: Aborting on shard 0
10/12/2022
- 09:18 PM Documentation #57858: v17.2.4 release does not contain latest cherry-picks
- Here's how I think we should go about this.
We know that the v17.2.4 tag is missing from the Quincy branch. We sho... - 07:35 PM Documentation #57858: v17.2.4 release does not contain latest cherry-picks
- The signed v17.2.4 tag was also not included in https://github.com/ceph/ceph/pull/48290. This seems to have occurred ...
- 06:57 PM Documentation #57858 (Resolved): v17.2.4 release does not contain latest cherry-picks
- Earlier today, I went to check one of the Telemetry commands in the Long Running Cluster, and the command caused a cr...
- 08:59 PM bluestore Bug #56851: crash: int BlueStore::read_allocation_from_onodes(SimpleBitmap*, BlueStore::read_allo...
- Sudhin Bengeri wrote:
> We are running into the same problem in our ceph cluster, we are running ceph v17.2.3.
We... - 08:57 PM bluestore Bug #56851: crash: int BlueStore::read_allocation_from_onodes(SimpleBitmap*, BlueStore::read_allo...
- We are running into the same problem in our ceph cluster, we are running ceph v17.2.3
- 06:41 PM bluestore Bug #57857 (Pending Backport): KernelDevice::read doesn't translate error codes correctly
- "(()+0xf630) [0x7f746eadc630]",
"(gsignal()+0x37) [0x7f746d8cf387]",
"(abort()+0x148) [0x7f... - 06:40 PM CephFS Backport #57848 (In Progress): pacific: mgr/volumes: addition of human-readable flag to volume in...
- 05:57 PM CephFS Backport #57849 (In Progress): quincy: mgr/volumes: addition of human-readable flag to volume inf...
- 05:11 PM crimson Bug #55326 (Resolved): crimson: formatter recursion loop crash
- 05:08 PM RADOS Bug #57782: [mon] high cpu usage by fn_monstore thread
- Hey Radek,
makes sense, I created a debug branch https://github.com/ceph/ceph-ci/pull/new/wip-crush-debug and migh... - 05:00 PM rgw Bug #57770: RGW (pacific) misplaces index entries after dynamically resharding bucket
- J. Eric Ivancich wrote:
> The theory is that the bucket index shard does not exist at this moment, as it was deleted... - 03:45 PM CephFS Bug #57856 (Fix Under Review): cephfs-top: Skip refresh when the perf stats query shows no metrics
- 03:39 PM CephFS Bug #57856 (Closed): cephfs-top: Skip refresh when the perf stats query shows no metrics
- In cephfs-top loading the clients usually takes time. So skip refreshing the main window when there are no metrics.
- 03:15 PM bluestore Feature #57785: fragmentation score in metrics
- Looks like we can get the fragmentation score via an admin socket command:...
- 02:43 PM bluestore Feature #57785: fragmentation score in metrics
- Yaarit/Laura - can we do something in telemetry perf channels?
- 02:54 PM bluestore Bug #57855 (Resolved): cannot enable level_compaction_dynamic_level_bytes
- create an osd with the following options....
- 12:29 PM CephFS Bug #53573 (Fix Under Review): qa: test new clients against older Ceph clusters
- 09:06 AM Linux kernel client Bug #57656: [testing] dbench: write failed on handle 10009 (Resource temporarily unavailable)
- Another failure is :...
- 05:22 AM Linux kernel client Bug #57656 (In Progress): [testing] dbench: write failed on handle 10009 (Resource temporarily u...
- 07:46 AM rbd Backport #57844 (In Progress): pacific: rbd CLI inconsistencies affecting "--namespace" arg
- 07:42 AM rbd Backport #57843 (In Progress): quincy: rbd CLI inconsistencies affecting "--namespace" arg
- 06:19 AM CephFS Bug #57854 (Resolved): mds: make num_fwd and num_retry to __u32
- The num_fwd in MClientRequestForward is int32_t, while the num_fwd
in ceph_mds_request_head is __u8. This is buggy w... - 05:44 AM crimson Bug #57693 (Fix Under Review): Messenger test failed against test_messenger_peer.cc
- Should be fixed by https://github.com/ceph/ceph/pull/48457
- 05:09 AM Dashboard Bug #57166 (Resolved): mgr/dashboard: "Average GET/PUT Latencies" panel lacks details
- 05:09 AM Dashboard Backport #57487 (Resolved): pacific: mgr/dashboard: "Average GET/PUT Latencies" panel lacks details
- 05:08 AM Dashboard Feature #56699 (Resolved): mgr/dashboard: improve dashboard redirect address
- 05:07 AM Dashboard Backport #57661 (Resolved): quincy: mgr/dashboard: improve dashboard redirect address
- 05:07 AM Dashboard Backport #57663 (Resolved): pacific: mgr/dashboard: improve dashboard redirect address
- 04:56 AM CephFS Backport #57836 (In Progress): pacific: Failure in snaptest-git-ceph.sh (it's an async unlink/cre...
- 04:11 AM rgw Backport #57197 (Resolved): pacific: x-amz-date protocol change breaks aws v4 signature logic: wa...
- 04:11 AM rgw Bug #47527 (Resolved): Ceph returns s3 incompatible xml response for listMultipartUploads
- 04:10 AM rgw Backport #53148 (Rejected): octopus: Ceph returns s3 incompatible xml response for listMultipartU...
- Octopus is EOL
- 04:10 AM rgw Backport #53149 (Resolved): pacific: Ceph returns s3 incompatible xml response for listMultipartU...
- 02:52 AM CephFS Backport #57837 (In Progress): quincy: Failure in snaptest-git-ceph.sh (it's an async unlink/crea...
- 02:51 AM rgw Bug #57853 (Pending Backport): multisite sync process block after long time running
- 1、deploy RADOSGW multisite
2、put lot of objects
3、keep it runing for a long time
- 02:39 AM RADOS Bug #57852 (Need More Info): osd: unhealthy osd cannot be marked down in time
- Before an unhealthy osd is marked down by mon, other osd may choose it as
heartbeat peer and then report an incorrec... - 01:03 AM bluestore Bug #55328: OSD crashed due to checksum error
- Hi Igor
> I will start to run the same test senario with a newer Ceph version (v16.2.10) in a few weeks, and run the...
10/11/2022
- 10:25 PM CephFS Backport #57718: pacific: Test failure: test_subvolume_group_ls_filter_internal_directories (task...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48328
merged - 10:25 PM CephFS Backport #57261: pacific: standby-replay mds is removed from MDSMap unexpectedly
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48270
merged - 10:24 PM CephFS Backport #57194: pacific: ceph pacific fails to perform fs/mirror test
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48269
merged - 10:23 PM rgw Backport #57649: pacific: rgw: fix bool/int logic error when calling get_obj_head_ioctx
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48230
merged - 10:22 PM rgw Backport #57429: pacific: key is used after move in RGWGetObj_ObjStore_S3::override_range_hdr
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48229
merged - 10:20 PM rgw Backport #57753: pacific: Log status of individual object deletions for multi-object delete reque...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48348
merged - 10:19 PM rgw Backport #57197: pacific: x-amz-date protocol change breaks aws v4 signature logic: was rfc 2616....
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48313
merged - 10:18 PM rgw Backport #55918: pacific: Bucket sync policy core dumped
- https://github.com/ceph/ceph/pull/47994 merged
- 10:17 PM rgw Backport #57450: pacific: 'radosgw-admin sync flow create' cmd crashes if flow-type omitted
- https://github.com/ceph/ceph/pull/47994 merged
- 10:15 PM rgw Backport #55245: pacific: rgwlc: ordinary expiration can remove delete-markers at end of current...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47231
merged - 10:14 PM rgw Backport #56185: pacific: rgw crash when use swift api
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47230
merged - 10:14 PM rgw Backport #55135: pacific: multisite: data sync only spawns one bucket sync at a time
- Casey Bodley wrote:
> https://github.com/ceph/ceph/pull/45713
merged - 10:13 PM rgw Backport #54144: pacific: bilog trim: segfault in RGWRadosBILogTrimCR::send_request if bucket sha...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44907
merged - 10:11 PM rgw Backport #53149: pacific: Ceph returns s3 incompatible xml response for listMultipartUploads
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44558
merged - 06:04 PM rgw Bug #57770: RGW (pacific) misplaces index entries after dynamically resharding bucket
- Here is the code that does this:...
- 06:02 PM rgw Bug #57770 (Need More Info): RGW (pacific) misplaces index entries after dynamically resharding b...
- So I looked at the code in 16.2.9 to try to understand how this might happen. The final step in adding an object to t...
- 05:44 PM mgr Bug #57851 (Fix Under Review): pybind/mgr/snap_schedule: use temp_store for db
- 05:42 PM mgr Bug #57851 (Resolved): pybind/mgr/snap_schedule: use temp_store for db
- ...
- 04:19 PM Dashboard Backport #57831 (Resolved): quincy: mgr/dashboard: weird data in OSD details
- 04:19 PM Dashboard Backport #57831 (Rejected): quincy: mgr/dashboard: weird data in OSD details
- 06:11 AM Dashboard Backport #57831 (In Progress): quincy: mgr/dashboard: weird data in OSD details
- 05:37 AM Dashboard Backport #57831 (Resolved): quincy: mgr/dashboard: weird data in OSD details
- https://github.com/ceph/ceph/pull/48433
- 04:19 PM Dashboard Backport #57847 (Resolved): quincy: mgr/dashboard: auto-coloring label
- 11:03 AM Dashboard Backport #57847 (Resolved): quincy: mgr/dashboard: auto-coloring label
- https://github.com/ceph/ceph/pull/48433
- 01:50 PM rgw Bug #57850 (Fix Under Review): amqp: rgw crash when ca location is used for amqp connections
- 01:45 PM rgw Bug #57850 (Pending Backport): amqp: rgw crash when ca location is used for amqp connections
- ca location value is stored as a reference, and the original string may already be destroyed when ca location is used
- 01:36 PM Linux kernel client Bug #54044: intermittent hangs waiting for caps
- Hi Xiubo,
here are the answers to the open questions:
* My max_mds value is 1
* My ceph version is 17.2.2
I... - 01:21 PM Dashboard Bug #52811 (Can't reproduce): mgr/dashboard: mgr crashes when viewing unavailable filesystem info...
- Latest main branch looks like it was fixed.
- 11:42 AM CephFS Backport #57821 (In Progress): pacific: cephfs-data-scan: scan_links is not verbose enough
- 11:41 AM CephFS Backport #57820 (In Progress): quincy: cephfs-data-scan: scan_links is not verbose enough
- 11:35 AM CephFS Bug #56162 (Resolved): mgr/stats: add fs_name as field in perf stats command output
- 11:34 AM CephFS Bug #56169 (Resolved): mgr/stats: 'perf stats' command shows incorrect output with non-existing m...
- 11:34 AM CephFS Bug #56483 (Resolved): mgr/stats: missing clients in perf stats command output.
- 11:33 AM CephFS Feature #54978 (Resolved): cephfs-top:addition of filesystem menu(improving GUI)
- 11:33 AM CephFS Bug #55861 (Resolved): Test failure: test_client_metrics_and_metadata (tasks.cephfs.test_mds_metr...
- 11:32 AM CephFS Backport #57283 (Resolved): quincy: cephfs-top:addition of filesystem menu(improving GUI)
- 11:32 AM CephFS Backport #57273 (Resolved): quincy: mgr/stats: missing clients in perf stats command output.
- 11:32 AM CephFS Backport #57330 (Resolved): quincy: Test failure: test_client_metrics_and_metadata (tasks.cephfs....
- 11:31 AM CephFS Backport #57276 (Resolved): quincy: mgr/stats: 'perf stats' command shows incorrect output with n...
- 11:31 AM CephFS Backport #57278 (Resolved): quincy: mgr/stats: add fs_name as field in perf stats command output
- 11:27 AM CephFS Backport #57849 (Resolved): quincy: mgr/volumes: addition of human-readable flag to volume info c...
- https://github.com/ceph/ceph/pull/48466
- 11:26 AM CephFS Backport #57848 (Resolved): pacific: mgr/volumes: addition of human-readable flag to volume info ...
- https://github.com/ceph/ceph/pull/48468
- 11:19 AM CephFS Bug #57620 (Pending Backport): mgr/volumes: addition of human-readable flag to volume info command
- 11:08 AM Dashboard Backport #57838 (Resolved): quincy: mgr/dashboard: prometheus: change name of pg_repaired_objects
- 08:18 AM Dashboard Backport #57838 (In Progress): quincy: mgr/dashboard: prometheus: change name of pg_repaired_objects
- 07:45 AM Dashboard Backport #57838 (Resolved): quincy: mgr/dashboard: prometheus: change name of pg_repaired_objects
- https://github.com/ceph/ceph/pull/48438
- 11:03 AM Dashboard Backport #57846 (Resolved): pacific: mgr/dashboard: auto-coloring label
- https://github.com/ceph/ceph/pull/50121
- 11:00 AM Dashboard Feature #55922 (Pending Backport): mgr/dashboard: auto-coloring label
- 11:00 AM Dashboard Backport #57835 (In Progress): quincy: mgr/dashboard: add server side encryption to rgw/s3
- 07:25 AM Dashboard Backport #57835 (Resolved): quincy: mgr/dashboard: add server side encryption to rgw/s3
- https://github.com/ceph/ceph/pull/48441
- 10:58 AM Dashboard Backport #57841 (In Progress): quincy: mgr/dashboard: Unable to change subuser permission
- 09:41 AM Dashboard Backport #57841 (Resolved): quincy: mgr/dashboard: Unable to change subuser permission
- https://github.com/ceph/ceph/pull/48440
- 10:13 AM RADOS Bug #57845 (New): MOSDRepOp::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, SERVER_O...
- ...
- 10:02 AM rbd Backport #57844 (Resolved): pacific: rbd CLI inconsistencies affecting "--namespace" arg
- https://github.com/ceph/ceph/pull/48459
- 10:02 AM rbd Backport #57843 (Resolved): quincy: rbd CLI inconsistencies affecting "--namespace" arg
- https://github.com/ceph/ceph/pull/48458
- 09:58 AM rbd Bug #57765 (Pending Backport): rbd CLI inconsistencies affecting "--namespace" arg
- 09:43 AM Dashboard Bug #57840 (Triaged): mgr/dashboard: "Add Host" rejects IPv6 addresses
- 09:21 AM Dashboard Bug #57840 (Triaged): mgr/dashboard: "Add Host" rejects IPv6 addresses
- h3. Description of problem
After setup a Cluster with cephadm it is not possible to add Hosts with its IPv6 addess... - 09:41 AM Dashboard Backport #57842 (Rejected): pacific: mgr/dashboard: Unable to change subuser permission
- 09:41 AM Dashboard Bug #57805 (Pending Backport): mgr/dashboard: Unable to change subuser permission
- 08:23 AM Dashboard Backport #57839 (In Progress): pacific: mgr/dashboard: prometheus: change name of pg_repaired_obj...
- 07:46 AM Dashboard Backport #57839 (Resolved): pacific: mgr/dashboard: prometheus: change name of pg_repaired_objects
- https://github.com/ceph/ceph/pull/48439
- 07:55 AM Orchestrator Backport #55991: pacific: Allow setting crush_device_class in OSD service specs
- I just tried to use the feature "crush_device_class" as it's supposed to be available, but it fails (I try to create ...
- 07:33 AM Dashboard Bug #57806 (Pending Backport): mgr/dashboard: prometheus: change name of pg_repaired_objects
- 07:25 AM CephFS Backport #57837 (Resolved): quincy: Failure in snaptest-git-ceph.sh (it's an async unlink/create ...
- https://github.com/ceph/ceph/pull/48452
- 07:25 AM CephFS Backport #57836 (Resolved): pacific: Failure in snaptest-git-ceph.sh (it's an async unlink/create...
- https://github.com/ceph/ceph/pull/48453
- 07:18 AM Dashboard Feature #57826 (Pending Backport): mgr/dashboard: add server side encryption to rgw/s3
- 04:39 AM Dashboard Feature #57826 (Resolved): mgr/dashboard: add server side encryption to rgw/s3
- Add the capability to add serever side encryption to the buckets in rgw
- 07:17 AM CephFS Bug #55332 (Pending Backport): Failure in snaptest-git-ceph.sh (it's an async unlink/create bug)
- 06:53 AM CephFS Documentation #57778: CephFS subvolume metadata not available in pacific
- Thanks for the update and looking into it. :-)
- 05:23 AM CephFS Documentation #57778: CephFS subvolume metadata not available in pacific
- Hi Eugen,
The doc changes are backported when a backport PR gets merged. The 16.2.10 release is a hotfix release w... - 05:37 AM Dashboard Backport #57833 (Resolved): quincy: mgr/dashboard: cephadm dashboard e2e failure "being covered b...
- https://github.com/ceph/ceph/pull/48432
- 05:37 AM Dashboard Backport #57832 (Rejected): pacific: mgr/dashboard: cephadm dashboard e2e failure "being covered ...
- 05:36 AM Dashboard Backport #57830 (Resolved): pacific: mgr/dashboard: weird data in OSD details
- https://github.com/ceph/ceph/pull/50121
- 05:25 AM Dashboard Backport #57828 (In Progress): quincy: cephadm/test_dashboard_e2e.sh: Expected to find content: '...
- 05:02 AM Dashboard Backport #57828 (Resolved): quincy: cephadm/test_dashboard_e2e.sh: Expected to find content: '/^f...
- https://github.com/ceph/ceph/pull/48432
- 05:20 AM Dashboard Bug #57511 (Pending Backport): mgr/dashboard: cephadm dashboard e2e failure "being covered by ano...
- 05:09 AM Dashboard Bug #57803 (Pending Backport): mgr/dashboard: weird data in OSD details
- 05:02 AM Dashboard Backport #57829 (Resolved): pacific: cephadm/test_dashboard_e2e.sh: Expected to find content: '/^...
- https://github.com/ceph/ceph/pull/55415
- 04:43 AM Dashboard Tasks #57827 (New): mgr/dashboard: add e2e tests for cephx user creation
- 04:40 AM Dashboard Bug #57114 (Resolved): mgr/dashboard: Squash is not mandatory field in "Create NFS export" page
- 04:39 AM Dashboard Backport #57435 (Resolved): pacific: mgr/dashboard: Squash is not mandatory field in "Create NFS ...
- 04:39 AM Dashboard Backport #57582 (Resolved): pacific: AssertionError: Expected to find element: `cd-modal .badge:n...
- 04:38 AM Dashboard Backport #57581 (Resolved): quincy: AssertionError: Expected to find element: `cd-modal .badge:no...
- 04:30 AM Dashboard Bug #57386 (Pending Backport): cephadm/test_dashboard_e2e.sh: Expected to find content: '/^foo$/'...
- 04:03 AM CephFS Bug #57299: qa: test_dump_loads fails with JSONDecodeError
- Rishabh, should this change be backported to p/q releases?
- 02:16 AM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- So, I did notice that I had set the domain name on one of the nodes to the "oldname.local" (when I was doing the find...
- 01:19 AM Orchestrator Bug #57800: ceph orch upgrade does not appear to work with FQNDs.
- what does `ceph orch host ls` report for this host? This error should only be raised if we can't find any IP stored f...
- 01:24 AM Orchestrator Documentation #57596: MON Service
- the "networks" parameter inside the service spec is a separate thing from the public/cluster network. The public_netw...
10/10/2022
- 08:54 PM rgw Bug #57807: The cloud sync module does not work starting with Pacific
- Related issue: https://tracker.ceph.com/issues/55310
- 12:48 PM rgw Bug #57807 (Duplicate): The cloud sync module does not work starting with Pacific
- We have a cluster running Ceph Pacific storing objects with S3 and we want to sync the objects with an external endpo...
- 07:19 PM rgw Bug #57562: multisite replication issue on Quincy
- Are there any suggestions/tips on how we can debug this type of multisite/replication issues?
- 06:48 PM CephFS Backport #57825 (Resolved): pacific: qa: mirror tests should cleanup fs during unwind
- https://github.com/ceph/ceph/pull/50765
- 06:47 PM CephFS Backport #57824 (Resolved): quincy: qa: mirror tests should cleanup fs during unwind
- https://github.com/ceph/ceph/pull/50766
- 06:47 PM CephFS Backport #57823 (Rejected): pacific: Test failure: test_newops_getvxattr (tasks.cephfs.test_newop...
- 06:47 PM CephFS Backport #57822 (Rejected): quincy: Test failure: test_newops_getvxattr (tasks.cephfs.test_newops...
- 06:47 PM CephFS Backport #57821 (Resolved): pacific: cephfs-data-scan: scan_links is not verbose enough
- https://github.com/ceph/ceph/pull/48443
- 06:47 PM CephFS Backport #57820 (Resolved): quincy: cephfs-data-scan: scan_links is not verbose enough
- https://github.com/ceph/ceph/pull/48442
- 06:44 PM CephFS Bug #57248 (Pending Backport): qa: mirror tests should cleanup fs during unwind
- 06:37 PM CephFS Bug #57589 (Pending Backport): cephfs-data-scan: scan_links is not verbose enough
- 06:37 PM CephFS Bug #57580 (Pending Backport): Test failure: test_newops_getvxattr (tasks.cephfs.test_newops.Test...
- 06:33 PM RADOS Bug #57796: after rebalance of pool via pgupmap balancer, continuous issues in monitor log
Radoslaw,
Yes, I saw that piece of code too. But i *think* I figured it out just a short time ago. I had the cru...- 06:05 PM RADOS Bug #57796 (Need More Info): after rebalance of pool via pgupmap balancer, continuous issues in m...
- Thanks for the report! The log comes from there:...
- 06:23 PM RADOS Bug #57782 (Need More Info): [mon] high cpu usage by fn_monstore thread
- It looks we're burning CPU in @close(2)@. The single call site I can spot is in @write_data_set_to_csv@. Let's analyz...
- 05:30 PM mgr Bug #57460: Json formatted ceph pg dump hangs on large clusters
- Hi Ponnuvel! Thanks for the analysis. The problem is genuine and the exponential explosion is simply a no-no.
I just... - 05:07 PM mgr Bug #57768 (Fix Under Review): mgr/balancer: check for end_weekday is exclusive, stops balancing ...
- 03:48 PM Dashboard Backport #57819 (New): quincy: mgr/dashboard: update legal links
- 03:48 PM Dashboard Backport #57818 (Rejected): pacific: mgr/dashboard: update legal links
- 03:35 PM Dashboard Bug #57792 (Pending Backport): mgr/dashboard: update legal links
- 02:36 PM Linux kernel client Bug #57686: general protection fault and CephFS kernel client hangs after MDS failover
- I believe that https://tracker.ceph.com/issues/57817 is another instance of this bug, but I wasn't sure so I opened a...
- 02:35 PM Linux kernel client Bug #57817 (Duplicate): general protection fault and CephFS kernel client hangs after MDS failover
- I believe that this is the same bug as https://tracker.ceph.com/issues/57686, but in case I'm wrong, I'm opening this...
- 02:25 PM Orchestrator Bug #57816 (Pending Backport): Add support to configure protocol (http or https) for Grafana url ...
- Right now cephadm always deploy Grafana by using https. In some testing scenarios it would be helpful to configure th...
- 01:45 PM Dashboard Feature #57815 (New): mgr/dashboard: smart automatic capabilities creator
- h3. Description
h1. Capabilities are formed by using known keywords as unknown values are not permitted. This mean... - 01:41 PM Dashboard Feature #57814 (New): mgr/dashboard: add enum fields
- h3. Description
Entities in capabilities are known beforehand therefore we can fill a dropdown and let the user ju... - 01:39 PM Dashboard Feature #57813 (New): mgr/dashboard: include form name in breadcrumbs
- h3. Description
The breadcrumbs should include the current form name (e.g.: Cluster >> Users >> Create), and the ... - 01:39 PM Dashboard Feature #57812 (New): mgr/dashboard: map icons in backend to frontend
- h3. Description
Rather than using literal Font-Awesome icon names in the back-end, we could just use a enum set ... - 01:28 PM Dashboard Feature #57811 (New): mgr/dashboard: infer form path
- h3. Description
f"{obj.action_type} {obj.__class__.__name__.title()}"
h3. Target persona
{{collapse(Example... - 01:25 PM Dashboard Feature #57810 (New): mgr/dashboard: auto generated routing in backend forms
- h3. Description of problem
Backend generated forms should generate the needed routing from the backend too and not... - 01:24 PM Dashboard Bug #57809 (New): mgr/dashboard: disable drag and drop in array forms
- h3. Description of problem
arrays in angular json schema have a drag and drop functionality that must be droped si... - 01:23 PM Dashboard Feature #57808 (New): mgr/dashboard: authx improvements
- h3. Description of problem
This is the epic of followup tasks in the authx feature
h3. Environment
* @ceph v... - 12:52 PM CephFS Bug #57655: qa: fs:mixed-clients kernel_untar_build failure
- Hitting similar issue in my today's run against ubuntu 22.04:
http://qa-proxy.ceph.com/teuthology/dparmar-2022-10-... - 07:53 AM CephFS Bug #57655: qa: fs:mixed-clients kernel_untar_build failure
- Maybe this can help here https://lore.kernel.org/all/alpine.LSU.2.21.2004031057320.25955@pobox.suse.cz/.
- 11:50 AM CephFS Backport #57723 (In Progress): pacific: qa: test_subvolume_snapshot_info_if_orphan_clone fails
- 10:56 AM CephFS Bug #57641: Ceph FS fscrypt clones missing fscrypt metadata
- Thanks Venky, I did not submit this right away as I'm unsure the xattr copy is the right way to do this.
I wonder ... - 08:03 AM CephFS Bug #57641 (Fix Under Review): Ceph FS fscrypt clones missing fscrypt metadata
- Hi Marcel,
I pushed a PR with your commit. We should probably take this opportunity to copy user xattrs during clone. - 10:36 AM Dashboard Bug #57806 (Fix Under Review): mgr/dashboard: prometheus: change name of pg_repaired_objects
- 10:18 AM Dashboard Bug #57806 (Resolved): mgr/dashboard: prometheus: change name of pg_repaired_objects
- h3. Description of problem
pg_repaired_objects > pool_repaired_objects
h3. Environment
* @ceph version@ stri... - 10:36 AM Dashboard Bug #57623 (Resolved): mgr/dashboard: expose num repaired objects metric per pool
- 10:13 AM Dashboard Bug #57805 (Resolved): mgr/dashboard: Unable to change subuser permission
- Tried to edit the permission of a subuser but once changed the permission, edited permission not seen in user info of...
- 09:49 AM CephFS Documentation #57778: CephFS subvolume metadata not available in pacific
- Hi Eugen,
The latest pacific (v16.2.10) only included CVE fixes as per
https://github.com/ceph/ceph/blo... - 09:29 AM CephFS Bug #57610: qa: timeout during unwinding of qa/workunits/suites/fsstress.sh
- Milind, not sure I remember this correctly -- did you RCA this?
- 09:28 AM CephFS Backport #57717 (In Progress): quincy: libcephfs: incorrectly showing the size for snapdirs when ...
- 09:23 AM CephFS Backport #57716 (In Progress): pacific: libcephfs: incorrectly showing the size for snapdirs when...
- 06:47 AM rgw Bug #57804 (Need More Info): Enabling sync on bucket not working
- Hello,
I'm having a problem when trying to enable sync on one of our buckets (multi-site) from master zone.
Her... - 06:33 AM Dashboard Feature #56155 (Resolved): mgr/dashboard: Add daemon logs tab to Cluster -> Logs component
- 06:08 AM RADOS Bug #56101: Gibba Cluster: 17.2.0 to 17.2.1 RC upgrade OSD crash in function safe_timer
- Laura Flores wrote:
> I contacted some Telemetry users. I will report back here with any information.
>
I am on... - 05:52 AM Dashboard Bug #57803 (Fix Under Review): mgr/dashboard: weird data in OSD details
- 05:37 AM Dashboard Bug #57803 (Resolved): mgr/dashboard: weird data in OSD details
- Please see the attached screenshot:
* OSD.3 device shows as "sdcsdc" (rhceph53)
* With daemons OSD.3 and OSD.4
* B... - 05:27 AM crimson Bug #57773 (Fix Under Review): crimson: TestLibRBD.TestCompareAndWriteStripeUnitSuccessPP fails w...
- 02:57 AM Bug #57802 (New): RGW crash when upload file through swift RGWFormPost function
- Hi,
When I use swift RGWFormPost to upload file, I got ** Caught signal (Segmentation fault) ** error. It will case ... - 12:39 AM crimson Bug #57801 (New): crimson: tag pool types as crimson, disallow snapshot, scrub, ec operations
- add mon_pool_default_crimson option to enable it by default, set in vstart,teutholoy
--crimson flag during pool crea...
10/09/2022
- 01:44 AM CephFS Bug #57674: fuse mount crashes the standby MDSes
- Jos Collin wrote:
> This is not a bug, just the limit reached.
>
> Processor -- accept open file descriptions lim... - 01:20 AM Orchestrator Bug #57800 (New): ceph orch upgrade does not appear to work with FQNDs.
- This is purely speculative on my part, but after attempting an upgrade to 17.2.4 from 17.2.3, it just sits there doin...
10/08/2022
- 09:09 PM crimson Bug #57799 (Resolved): crimson: add guard rails to enable crimson on a cluster
- - crimson experimental feature
- ceph osd set-allow-crimson
- disallow crimson-osd booting without that flag
- 05:43 PM crimson Bug #57758: crimson: disable autoscale for crimson in teuthology
- Actually already defaults to this, need to figure out why the test I saw was doing merges.
- 04:42 PM crimson Bug #57798 (Resolved): crimson: actually set CRIMSON_COMPAT for teuthology workunits
- 07:27 AM rgw Bug #56992: rgw_op.cc:Deleting a non-existent object also generates a delete marker
- Because I do not fully understand the PR process, I need to close the old PR, and the new one in https://github.com/c...
- 04:26 AM CephFS Backport #57362 (Resolved): quincy: ffsb.sh test failure
- 04:25 AM CephFS Backport #57240 (Resolved): quincy: ceph-fs crashes on getfattr
10/07/2022
- 09:01 PM rgw Bug #57562: multisite replication issue on Quincy
- Hi,
Here's some extra data from another test which used increased rgw debugging levels by feeding in the options _... - 08:32 PM RADOS Bug #57796: after rebalance of pool via pgupmap balancer, continuous issues in monitor log
- I removed the hosts holding the osds reported by verify_upmap from the default root rule that no one uses, and the lo...
- 05:56 PM RADOS Bug #57796: after rebalance of pool via pgupmap balancer, continuous issues in monitor log
- Note that the balancer balanced a replicated pool, using its own custom crush root too. The hosts in that pool (not i...
- 05:46 PM RADOS Bug #57796: after rebalance of pool via pgupmap balancer, continuous issues in monitor log
- preformatting the crush info so it shows up properly ......
- 05:43 PM RADOS Bug #57796 (Need More Info): after rebalance of pool via pgupmap balancer, continuous issues in m...
The pgupmap balancer was not balancing well, and after setting mgr/balancer/upmap_max_deviation to 1 (ceph config-k...- 08:00 PM mgr Bug #57797 (Resolved): mgr/prometheus: avoid duplicates and deleted entries for rbd_stats_pool
- 04:46 PM RADOS Backport #57795 (In Progress): quincy: intrusive_lru leaking memory when
- https://github.com/ceph/ceph/pull/54557
- 04:46 PM RADOS Backport #57794 (Resolved): pacific: intrusive_lru leaking memory when
- https://github.com/ceph/ceph/pull/54558
- 04:46 PM Orchestrator Backport #57793 (New): quincy: Update monitoring doc to reflect the new location of grafana key/cert
- 04:32 PM Orchestrator Documentation #57769 (Pending Backport): Update monitoring doc to reflect the new location of gra...
- 04:29 PM RADOS Bug #57573 (Pending Backport): intrusive_lru leaking memory when
- 03:10 PM Dashboard Bug #57792 (Fix Under Review): mgr/dashboard: update legal links
- 10:10 AM Dashboard Bug #57792 (Pending Backport): mgr/dashboard: update legal links
- The legal links in the login page are outdated:
* "Help" is broken, and it should probably point to docs.ceph.com
*... - 02:12 PM Feature #57455 (Rejected): msg: change to allow separate port ranges for MDS and OSD
- This has always been possible by way of [osd], [mds], etc sections in ceph.conf file. See discussion in https://gith...
- 02:11 PM rgw Bug #51919 (Duplicate): crash: ceph::common::PerfCounters::inc(int, unsigned long) (in RGWAsyncFe...
- Changed status from Resolved to Duplicate since this issue duplicates https://tracker.ceph.com/issues/49666.
- 12:53 PM CephFS Bug #57594 (In Progress): pacific: Test failure: test_rebuild_moved_dir (tasks.cephfs.test_data_s...
- http://pulpito.front.sepia.ceph.com/jcollin-2022-10-07_11:57:35-fs-wip-jcollin-B57594-main-check-distro-default-smith...
- 12:36 PM mgr Bug #54788: crash: void MonMap::add(const mon_info_t&): assert(addr_mons.count(a) == 0)
- See bug 54744.
- 12:36 PM RADOS Bug #54773: crash: void MonMap::add(const mon_info_t&): assert(addr_mons.count(a) == 0)
- See bug 54744.
- 12:35 PM RADOS Bug #54744: crash: void MonMap::add(const mon_info_t&): assert(addr_mons.count(a) == 0)
- Rook v1.6.5 / Ceph v12.2.9 running on the host network and not inside the Kubernetes SDN caused creating a mon canary...
- 09:40 AM CephFS Bug #57764: Thread md_log_replay is hanged for ever.
- Thanks for the bug report. Seems like you found a subtle race. I haven't gone through the fix yet, but I'll get to it...
- 07:41 AM cleanup Tasks #57569: implement chown admin rest entrypoint
- This item has been repurposed to: *implement chown admin rest entrypoint*.
After a chat with Daniel Gryniewicz we ag... - 07:33 AM rgw Bug #57784: beast frontend crashes on exception from socket.local_endpoint()
- Hey,
here is a full stack trace from the RGW daemon. I removed bucket/file/user names.
The host is:
Ubuntu 20.04...
10/06/2022
- 11:28 PM crimson Bug #57738 (Resolved): crimson: repop ordering bug
- 10:22 PM rbd Bug #56561 (Resolved): rbd perf image iostat/iotop lost the ability to gather data across pools
- 10:11 PM crimson Bug #57791 (Resolved): crimson: zero becomes truncate if region exceeds object bound
- 09:17 PM Bug #56098: api_tier_pp: failure on LibRadosTwoPoolsPP.ManifestRefRead
- /a/yuriw-2022-10-05_21:09:57-rados-main-distro-default-smithi/7056369
- 09:10 PM crimson Bug #57789 (Resolved): crimson: add list_snaps
- Some librbd library functions and tests use list_snaps. Worth adding though snapshot support doesn't really work yet.
- 08:38 PM RADOS Bug #56101: Gibba Cluster: 17.2.0 to 17.2.1 RC upgrade OSD crash in function safe_timer
- I contacted some Telemetry users. I will report back here with any information.
Something to note: The large maj... - 07:49 PM Orchestrator Bug #57303: rados/cephadm: Failed to fetch package version from https://shaman.ceph.com/api/searc...
- Laura Flores wrote:
> /a/yuriw-2022-09-20_17:39:55-rados-wip-yuri5-testing-2022-09-19-1007-pacific-distro-default-sm... - 07:37 PM Bug #57756 (Resolved): upgrade: notify retry canceled due to unrecoverable error after 1 attempts...
- I think this was resolved in 17.2.4, so it shouldn't happen anymore.
- 06:58 PM Orchestrator Backport #57788 (Resolved): pacific: mgr/nfs: Add a sectype field to nfs exports created by nfs m...
- https://github.com/ceph/ceph/pull/49929
- 06:58 PM Orchestrator Backport #57787 (Resolved): quincy: mgr/nfs: Add a sectype field to nfs exports created by nfs mg...
- https://github.com/ceph/ceph/pull/48531
- 06:51 PM Orchestrator Feature #57404 (Pending Backport): mgr/nfs: Add a sectype field to nfs exports created by nfs mgr...
- 06:35 PM Orchestrator Feature #57786 (Resolved): cephadm: open ports in firewall when deploying iscsi
- specifically, 3260 and whatever the user provides for the api_port. We already have logic in the deploy command in th...
- 06:30 PM Orchestrator Bug #57750: cephadm fails to upgrade systems not running sudo
- I can confirm that this was with the root user, no custom user involved.
Upgrades prior to 17.x.x worked like a ch... - 06:24 PM Orchestrator Bug #57750: cephadm fails to upgrade systems not running sudo
- assuming this wasn't with a custom (not-root) ssh user anyway, in which case sudo would be required as cephadm needs ...
- 06:22 PM Orchestrator Bug #57750 (In Progress): cephadm fails to upgrade systems not running sudo
- this is a legit bug. However, I think this should have been fixed by https://github.com/ceph/ceph/pull/47898 which ap...
- 04:10 PM Orchestrator Bug #57750: cephadm fails to upgrade systems not running sudo
- The documentation does not state that one need sudo at all. It's an option. So one cannot make the assumption that ev...
- 03:52 PM Orchestrator Bug #57750 (Need More Info): cephadm fails to upgrade systems not running sudo
- I'd say that the expected behavior. The user you use with cephadm needs passwordless sudo access to all the hosts tha...
- 06:19 PM bluestore Bug #57672: SSD OSD won't start after high framentation score!
- I created an issue to surface the fragmentation score via prom here: https://tracker.ceph.com/issues/57785
Not a 1... - 06:17 PM bluestore Feature #57785 (New): fragmentation score in metrics
- Currently the bluestore fragmentation score does not seem to be exported in metrics. Due to the issue described in ht...
- 05:25 PM rgw Bug #57784 (Fix Under Review): beast frontend crashes on exception from socket.local_endpoint()
- 05:19 PM rgw Bug #57784 (Pending Backport): beast frontend crashes on exception from socket.local_endpoint()
- reported on ceph-users in https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/FSXGU7WVFJTPHW5S5A63IN4AEOV...
- 05:12 PM Backport #50382 (Resolved): pacific: DecayCounter: Expected: (std::abs(total-expected)/expected) ...
- 05:09 PM Backport #50382: pacific: DecayCounter: Expected: (std::abs(total-expected)/expected) < (0.01), a...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48031
merged - 05:11 PM CephFS Backport #57554: quincy: qa: test_subvolume_snapshot_clone_quota_exceeded fails CommandFailedError
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48164
merged - 05:08 PM RADOS Backport #57545: quincy: CommandFailedError: Command failed (workunit test rados/test_python.sh) ...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48113
merged - 05:05 PM RADOS Backport #57496: quincy: Invalid read of size 8 in handle_recovery_delete()
- Nitzan Mordechai wrote:
> https://github.com/ceph/ceph/pull/48039
merged - 05:04 PM RADOS Backport #57443: quincy: osd: Update osd's IOPS capacity using async Context completion instead o...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47983
merged - 05:03 PM RADOS Backport #57346: quincy: expected valgrind issues and found none
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47933
merged - 05:01 PM RADOS Backport #56602: quincy: ceph report missing osdmap_clean_epochs if answered by peon
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47928
merged - 05:00 PM RADOS Backport #55282: quincy: osd: add scrub duration for scrubs after recovery
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47926
merged - 04:58 PM CephFS Backport #57362: quincy: ffsb.sh test failure
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47890
merged - 04:58 PM CephFS Backport #57240: quincy: ceph-fs crashes on getfattr
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47890
merged - 04:56 PM CephFS Backport #57283: quincy: cephfs-top:addition of filesystem menu(improving GUI)
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47820
merged - 04:56 PM CephFS Backport #57273: quincy: mgr/stats: missing clients in perf stats command output.
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47820
merged - 04:56 PM CephFS Backport #57330: quincy: Test failure: test_client_metrics_and_metadata (tasks.cephfs.test_mds_me...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47820
merged - 04:56 PM CephFS Backport #57276: quincy: mgr/stats: 'perf stats' command shows incorrect output with non-existing...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47820
merged - 04:56 PM CephFS Backport #57278: quincy: mgr/stats: add fs_name as field in perf stats command output
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47820
merged - 04:51 PM CephFS Backport #57555: pacific: qa: test_subvolume_snapshot_clone_quota_exceeded fails CommandFailedError
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48165
merged - 04:50 PM Dashboard Backport #57582: pacific: AssertionError: Expected to find element: `cd-modal .badge:not(script,s...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48142
merged - 04:47 PM RADOS Backport #57544: pacific: CommandFailedError: Command failed (workunit test rados/test_python.sh)...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/48112
merged - 04:44 PM rgw Bug #57562: multisite replication issue on Quincy
- The difference between this issue and Bug #57783 is that in our case, the buckets/objects are NOT synced.
I tried a... - 04:44 PM CephFS Backport #56468: pacific: mgr/volumes: display in-progress clones for a snapshot
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/47112
merged - 03:28 PM rgw Bug #57783 (In Progress): multisite: data sync reports shards behind after source zone fully trim...
- workload tests have been producing cases where data sync shows shards behind, although the source zone has fully trim...
- 02:08 PM RADOS Bug #57782 (Fix Under Review): [mon] high cpu usage by fn_monstore thread
- We observed high cpu usage by ms_dispatch and fn_monstore thread (amounting to 100-99% in top) Ceph [ deployment was ...
- 12:45 PM CephFS Bug #57764 (Fix Under Review): Thread md_log_replay is hanged for ever.
- 12:29 PM Bug #56610: FTBFS with fmtlib 9.0.0
- Can confirm adding -DFMT_DEPRECATED_OSTREAM to CXXFLAGS downstream in openSUSE fixes this (as it does for Debian). N...
- 11:49 AM Orchestrator Bug #57695 (Resolved): cephadm: upgrade tests fail with "Upgrade: Paused due to UPGRADE_BAD_TARGE...
- 10:58 AM rbd Backport #57779 (In Progress): quincy: [test] fio 3.16 doesn't build on recent kernels due to rem...
- 10:43 AM rbd Backport #57779 (Resolved): quincy: [test] fio 3.16 doesn't build on recent kernels due to remova...
- https://github.com/ceph/ceph/pull/48386
- 10:57 AM rbd Backport #57780 (In Progress): pacific: [test] fio 3.16 doesn't build on recent kernels due to re...
- 10:43 AM rbd Backport #57780 (Resolved): pacific: [test] fio 3.16 doesn't build on recent kernels due to remov...
- https://github.com/ceph/ceph/pull/48385
- 10:57 AM Orchestrator Bug #57781 (Rejected): Fix prometheus dependencies calculation
- https://github.com/ceph/ceph/pull/46400 introduced a new http service discovery mechanism but we are still including ...
- 10:43 AM rbd Bug #57766 (Pending Backport): [test] fio 3.16 doesn't build on recent kernels due to removal of ...
- 07:18 AM CephFS Bug #54501 (Fix Under Review): libcephfs: client needs to update the mtime and change attr when s...
- 06:07 AM CephFS Documentation #57778 (New): CephFS subvolume metadata not available in pacific
- According to the current Pacific docs [1] it should be possible to set subvolume metadata for a cephfs volume:
<pr... - 05:52 AM Dashboard Cleanup #54356 (Resolved): mgr/dashboard: Grafana e2e tests
- 05:52 AM Dashboard Cleanup #56426 (Resolved): mgr/dashboard: update cypress to 9.7
- 05:51 AM Dashboard Backport #56588 (Resolved): pacific: mgr/dashboard: update cypress to 9.7
- 05:51 AM Dashboard Backport #55468 (Resolved): pacific: mgr/dashboard: Grafana e2e tests
- 05:27 AM CephFS Backport #57777 (In Progress): quincy: Clarify security implications of path-restricted cephx cap...
- https://github.com/ceph/ceph/pull/53559
- 05:26 AM CephFS Backport #57776 (Resolved): pacific: Clarify security implications of path-restricted cephx capab...
- https://github.com/ceph/ceph/pull/53560
- 05:19 AM CephFS Bug #56507: pacific: Test failure: test_rapid_creation (tasks.cephfs.test_fragment.TestFragmentat...
- "New run":http://qa-proxy.ceph.com/teuthology/yuriw-2022-10-03_22:11:49-fs-wip-yuri-testing-2022-10-03-1342-pacific-d...
- 05:11 AM CephFS Documentation #57737 (Pending Backport): Clarify security implications of path-restricted cephx c...
- 01:56 AM crimson Bug #57774 (Closed): crimson: skip snapshot tests for test_librbd
- 01:54 AM crimson Bug #57773 (Resolved): crimson: TestLibRBD.TestCompareAndWriteStripeUnitSuccessPP fails with EINVAL
- ./bin/ceph_test_librbd --gtest_filter=TestLibRBD.TestCompareAndWriteStripeUnitSuccessPP
- 12:33 AM Orchestrator Backport #57772 (New): quincy: cephadm: watch Grafana certificates
- 12:22 AM Orchestrator Feature #44461 (Pending Backport): cephadm: watch Grafana certificates
10/05/2022
- 11:51 PM crimson Bug #57740 (Resolved): crimson: op hang while running ./bin/ceph_test_rados_api_aio_pp and ./bin/...
- https://github.com/ceph/ceph/pull/48352
- 11:48 PM crimson Bug #57617 (Resolved): crimson: need to actually set version/user_version for duplicate ops
- https://github.com/ceph/ceph/pull/48195
- 11:07 PM Orchestrator Bug #51361 (New): KillMode=none is deprecated
- I was wrong about @KillMode=none@ for my use case
- 08:54 PM Orchestrator Bug #57771 (Pending Backport): orch/cephadm suite: 'TESTDIR=/home/ubuntu/cephtest bash -s' fails
- It seems to be failing to install some selinux package...
- 06:49 PM RADOS Bug #57699 (Fix Under Review): slow osd boot with valgrind (reached maximum tries (50) after wait...
- 06:48 PM RADOS Bug #57049 (Duplicate): cluster logging does not adhere to mon_cluster_log_file_level
- 06:46 PM RADOS Bug #50222: osd: 5.2s0 deep-scrub : stat mismatch
- Hi Laura. In luck with verification of the hypothesis from the comment #17?
- 06:43 PM RADOS Bug #57532 (Duplicate): Notice discrepancies in the performance of mclock built-in profiles
- Marked as duplicate per comment #4.
- 06:25 PM RADOS Bug #57757: ECUtil: terminate called after throwing an instance of 'ceph::buffer::v15_2_0::end_of...
- There is a coredump on the teuhtology node (@/ceph/teuthology-archive/yuriw-2022-09-29_16:44:24-rados-wip-lflores-tes...
- 06:19 PM RADOS Bug #57546: rados/thrash-erasure-code: wait_for_recovery timeout due to "active+clean+remapped+la...
- I think this a fix for that got reverted in quincy (https://tracker.ceph.com/issues/53806) but it's still in @main@. ...
- 06:12 PM RADOS Bug #50042: rados/test.sh: api_watch_notify failures
- Assigning to Nitzan just for the sake of testing the hypothesis from https://tracker.ceph.com/issues/50042#note-35.
- 06:06 PM RADOS Cleanup #57587 (Resolved): mon: fix Elector warnings
- Resolved by https://github.com/ceph/ceph/pull/48289.
- 06:05 PM RADOS Bug #56101: Gibba Cluster: 17.2.0 to 17.2.1 RC upgrade OSD crash in function safe_timer
- This won't be easy to reproduce but there are still some options like:
* contacting owners of the external cluster... - 05:58 PM rgw Bug #57770 (Resolved): RGW (pacific) misplaces index entries after dynamically resharding bucket
- When RGW reshards buckets with ~250k index entries*, I've noticed some s3:PutObject requests that return 200 end up w...
- 04:43 PM mgr Bug #57710 (Rejected): Exports cannot be removed with ceph_argparse
- 04:37 PM mgr Bug #57694 (Rejected): Exports not created correctly when using ceph_argparse
- 04:37 PM mgr Bug #57711 (Rejected): Exports not updated correctly when using ceph_argparse
- 03:21 PM crimson Bug #57578 (Fix Under Review): crimson: assertion failure in _do_transaction_step()
- 02:42 PM mgr Bug #57768: mgr/balancer: check for end_weekday is exclusive, stops balancing too early
- *PR*: https://github.com/ceph/ceph/pull/48375
- 01:50 PM mgr Bug #57768: mgr/balancer: check for end_weekday is exclusive, stops balancing too early
- I am working on this issue
- 01:45 PM mgr Bug #57768 (Resolved): mgr/balancer: check for end_weekday is exclusive, stops balancing too early
- According to the "docs":https://docs.ceph.com/en/latest/rados/operations/balancer/ @end_weekday@ restricts automatic ...
- 02:20 PM Orchestrator Documentation #57769 (In Progress): Update monitoring doc to reflect the new location of grafana ...
- 02:20 PM Orchestrator Documentation #57769 (Pending Backport): Update monitoring doc to reflect the new location of gra...
- As part of the PR https://github.com/ceph/ceph/pull/47098 grafane key/cert now are stored per node but doc has not be...
- 01:58 PM Orchestrator Bug #57173 (Resolved): cephadm: bootstrap should return non-zero exit code when applying spec fails
- 01:57 PM Orchestrator Backport #57379 (Resolved): pacific: cephadm: bootstrap should return non-zero exit code when app...
- 10:55 AM rbd Bug #57766 (Fix Under Review): [test] fio 3.16 doesn't build on recent kernels due to removal of ...
- 08:07 AM rbd Bug #57766 (Resolved): [test] fio 3.16 doesn't build on recent kernels due to removal of linux/raw.h
- ...
- 10:23 AM rbd Bug #57765 (Fix Under Review): rbd CLI inconsistencies affecting "--namespace" arg
- Making Lucian the nominal assignee as Stefan doesn't seem to have a tracker account.
- 06:26 AM rbd Bug #57765 (Resolved): rbd CLI inconsistencies affecting "--namespace" arg
- There are a few rbd CLI inconsistencies that affect the "--namespace" parameter:
* unlike "rbd device map", "rbd d... - 08:16 AM ceph-volume Bug #57767 (In Progress): ceph-volume should check if device is locked prior to zapping it
- ceph-volume allows zapping a device although its related `ceph-osd` process is running...
- 08:08 AM Dashboard Bug #57456 (Resolved): mgr/dashboard: Cephfs snapshot creation with same name on UI throws 500 In...
- 08:07 AM Dashboard Backport #57498 (Resolved): pacific: mgr/dashboard: Cephfs snapshot creation with same name on UI...
- 08:00 AM Bug #56610: FTBFS with fmtlib 9.0.0
- We're having the same issue with Pacific on openSUSE (https://bugzilla.opensuse.org/show_bug.cgi?id=1202292).
10/04/2022
- 11:21 PM CephFS Bug #57764 (Resolved): Thread md_log_replay is hanged for ever.
- In production environment, we have a problem: one standby-replay's md_log_replay thread is hanged.
1,The reason:
... - 10:42 PM rgw Bug #57562: multisite replication issue on Quincy
- We are able to consistently reproduce the replication issue now. The following are the environment and the steps to r...
- 07:31 PM Bug #57763: monitor DB grows without bound during rebalance
- edit:
Why do the Monitor DBs continue to grow in size when the rebalance, backfill, balancer, and autoscaler are dis... - 07:17 PM Bug #57763 (New): monitor DB grows without bound during rebalance
- We have a very large cluster of about 680 OSDs across 18 storage servers. The largest and most active pool is our RGW...
- 05:39 PM bluestore Bug #57672: SSD OSD won't start after high framentation score!
- Sure. https://tracker.ceph.com/issues/57762
- 03:51 PM bluestore Bug #57672: SSD OSD won't start after high framentation score!
- Kevin Fox wrote:
> For the record, ssd/ssd or hdd/hdd seems to work fine even though the documentation makes it soun... - 03:30 PM bluestore Bug #57672: SSD OSD won't start after high framentation score!
- For the record, ssd/ssd or hdd/hdd seems to work fine even though the documentation makes it sound like it doesn't.
... - 05:38 PM bluestore Documentation #57762 (New): documentation about same hardware class wrong
- The documentation in at least one place:
https://docs.ceph.com/en/pacific/man/8/ceph-bluestore-tool/ bluefs-bdev-mig... - 05:25 PM RADOS Bug #50042: rados/test.sh: api_watch_notify failures
- /a/yuriw-2022-09-29_16:40:30-rados-wip-all-kickoff-r-distro-default-smithi/7047940...
- 05:17 PM rgw Bug #51574: Segfault when uploading file
- Here is the stacktrace from running the test script with ceph version 17.2.3 (dff484dfc9e19a9819f375586300b3b79d80034...
- 04:02 PM rgw Bug #51574: Segfault when uploading file
- Hello,
As Ceph Quincy 17.2.3 is still segfaulting using the same test script as for Pacific before the fix, we sti... - 01:53 PM Linux kernel client Bug #57703: unable to handle page fault for address and system lockup after MDS failover
- Minor correction: "@ceph: update_snap_trace error -5@" is still seen in dmesg after MDS failovers when mounting the w...
- 01:44 PM CephFS Bug #57674 (Closed): fuse mount crashes the standby MDSes
- This is not a bug, just the limit reached.
Processor -- accept open file descriptions limit reached sd = 20 errno ... - 10:26 AM CephFS Bug #57674 (In Progress): fuse mount crashes the standby MDSes
- 12:56 PM CephFS Backport #57748 (Rejected): pacific: doc: Fix disaster recovery documentation
- Not required for Pacific
- 12:56 PM CephFS Backport #57743 (Rejected): pacific: qa: test_recovery_pool uses wrong recovery procedure
- Not required for pacific.
- 12:49 PM CephFS Bug #57676 (Triaged): qa: error during scrub thrashing: rank damage found: {'backtrace'}
- 12:48 PM CephFS Bug #57682 (Triaged): client: ERROR: test_reconnect_after_blocklisted
- 07:43 AM rbd Backport #57388: quincy: [test] iscsi rest_api_create.t and rest_api_delete.t need formatting adj...
- Hi Guillaume,
Could you please take a look at this? It seemed like a ceph-container issue to me. - 06:26 AM CephFS Backport #57761 (Resolved): pacific: qa: test_scrub_pause_and_resume_with_abort failure
- https://github.com/ceph/ceph/pull/49458
- 06:26 AM CephFS Backport #57760 (Resolved): quincy: qa: test_scrub_pause_and_resume_with_abort failure
- https://github.com/ceph/ceph/pull/49459
- 06:15 AM CephFS Bug #48812 (Pending Backport): qa: test_scrub_pause_and_resume_with_abort failure
- 05:35 AM CephFS Bug #57411: mutiple mds crash seen while running db workloads with regular snapshots and journal ...
- Patrick Donnelly wrote:
> Apparently this one is known.
yeh. and its only seen when running database workloads on... - 12:44 AM crimson Bug #57759: crimson: rbdv1 needs TMAP, easier to implement than to skip rbdv1 tests
- src/test/librbd/test_librbd.cc...
- 12:19 AM crimson Bug #57759 (Resolved): crimson: rbdv1 needs TMAP, easier to implement than to skip rbdv1 tests
- ...
Also available in: Atom