Activity
From 09/20/2020 to 10/19/2020
10/19/2020
- 07:33 PM Bug #45100: qa: Test failure: test_damaged_dentry (tasks.cephfs.test_damage.TestDamage)
- /ceph/teuthology-archive/pdonnell-2020-10-13_22:14:10-kcephfs-wip-pdonnell-testing-20201013.174240-distro-basic-smith...
- 05:21 PM Bug #47786: mds: log [ERR] : failed to commit dir 0x100000005f1.1010* object, errno -2
- $ grep 'failed to commit dir' pdonnell-2020-10-*/*/teu*
Binary file pdonnell-2020-10-07_03:30:19-multimds-wip-pdonne... - 02:02 AM Bug #47786: mds: log [ERR] : failed to commit dir 0x100000005f1.1010* object, errno -2
- For /ceph/teuthology-archive/pdonnell-2020-10-08_01:40:56-multimds-wip-pdonnell-testing-20201007.214100-distro-basic-...
- 04:54 PM Bug #46434: osdc: FAILED ceph_assert(bh->waitfor_read.empty())
- Multiple failures in https://pulpito.ceph.com/yuriw-2020-10-12_15:45:53-powercycle-nautilus-distro-basic-smithi/
- 04:11 PM Bug #47843 (Fix Under Review): mds: stuck in resolve when restarting MDS and reducing max_mds
- 01:41 PM Bug #47842 (Triaged): qa: "fsstress.sh: line 16: 28870 Bus error (core dumped) "$BI...
- 01:40 PM Bug #47881 (Need More Info): mon/MDSMonitor: stop all MDS processes in the cluster at the same ti...
- Would `ceph fs fail <fs_name>` not be the command you want?
- 06:43 AM Bug #47881: mon/MDSMonitor: stop all MDS processes in the cluster at the same time. Some MDS cann...
- Zheng Yan wrote:
> this is by design. monitor never marks laggy mds failed if there is no replacement
Pull Requet... - 12:21 AM Bug #47881: mon/MDSMonitor: stop all MDS processes in the cluster at the same time. Some MDS cann...
- this is by design. monitor never marks laggy mds failed if there is no replacement
- 01:21 PM Fix #15134 (In Progress): multifs: test case exercising mds_thrash for multiple filesystems
- 11:45 AM Documentation #46884 (In Progress): pybind/mgr/mds_autoscaler: add documentation
- 08:33 AM Backport #47891 (Resolved): octopus: mgr/nfs: Pseudo path prints wrong error message
- https://github.com/ceph/ceph/pull/37855
10/17/2020
- 07:25 AM Bug #47881 (Resolved): mon/MDSMonitor: stop all MDS processes in the cluster at the same time. So...
- Stop all MDS processes in the cluster at the same time, After all MDS processes exits, some MDS are still in the "act...
10/16/2020
- 09:35 AM Documentation #45730 (Resolved): MDS config reference lists mds log max expiring
- 09:34 AM Backport #45826 (Rejected): mimic: MDS config reference lists mds log max expiring
- mimic EOL
- 09:12 AM Backport #47877 (In Progress): octopus: Create NFS Ganesha Cluster instructions are misleading
- 09:12 AM Backport #47877 (Resolved): octopus: Create NFS Ganesha Cluster instructions are misleading
- https://github.com/ceph/ceph/pull/37691
- 09:11 AM Bug #46559 (Pending Backport): Create NFS Ganesha Cluster instructions are misleading
10/15/2020
- 12:38 PM Bug #36171: mds: ctime should not use client provided ctime/mtime
- IMHO ctime should always be `ceph_clock_now()` rather than any time from the client.
Here's an XFS demo. Note how ... - 08:41 AM Backport #47824 (In Progress): octopus: pybind/mgr/volumes: Make number of cloner threads configu...
- 08:27 AM Bug #46883: kclient: ghost kernel mount
- From: /ceph/teuthology-archive/pdonnell-2020-08-08_02:19:19-kcephfs-wip-pdonnell-testing-20200808.001303-distro-basic...
10/14/2020
- 07:06 PM Bug #46883: kclient: ghost kernel mount
- I'm not a fan of this noshare option. That seems like a hacky workaround for a problem that I'm not sure any of us fu...
- 02:22 PM Bug #47854 (Fix Under Review): some clients may return failure in the scenario where multiple cli...
- 03:13 AM Bug #47854 (Resolved): some clients may return failure in the scenario where multiple clients cre...
- The issue can be reproduced by the following steps:
(1)ceph version: 14.2.10, multimds , multiple clients mount the ...
10/13/2020
- 05:27 PM Bug #47783 (Pending Backport): mgr/nfs: Pseudo path prints wrong error message
- 05:26 PM Documentation #47784 (Resolved): nfs: Remove doc on creating cephfs exports using rook
- 04:49 PM Bug #42365 (Resolved): client: FAILED assert(dir->readdir_cache[dirp->cache_index] == dn)
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 04:46 PM Bug #47011 (Resolved): client: Client::open() pass wrong cap mask to path_walk
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 04:46 PM Bug #47125 (Resolved): mds: fix possible crash when the MDS is stopping
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 04:45 PM Bug #47224 (Resolved): various quota failures
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 04:45 PM Bug #47353 (Resolved): mds: purge_queue's _calculate_ops is inaccurate
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 04:45 PM Bug #47512 (Resolved): mgr/nfs: Cluster creation throws 'NoneType' object has no attribute 'repla...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 04:30 PM Fix #46696 (Resolved): mds: pre-fragment distributed ephemeral pin directories to distribute the ...
- 04:15 PM Backport #47608 (Resolved): octopus: mds: OpenFileTable::prefetch_inodes during rejoin can cause ...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37383
m... - 04:15 PM Backport #47604 (Resolved): octopus: mds: purge_queue's _calculate_ops is inaccurate
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37372
m... - 04:15 PM Backport #47601 (Resolved): octopus: mgr/nfs: Cluster creation throws 'NoneType' object has no at...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37371
m... - 04:14 PM Backport #47260 (Resolved): octopus: client: FAILED assert(dir->readdir_cache[dirp->cache_index] ...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37370
m... - 04:14 PM Backport #47623 (Resolved): octopus: various quota failures
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37369
m... - 04:14 PM Backport #47255 (Resolved): octopus: client: Client::open() pass wrong cap mask to path_walk
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37369
m... - 04:13 PM Backport #47253 (Resolved): octopus: mds: fix possible crash when the MDS is stopping
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37368
m... - 03:55 PM Bug #47787: mgr/nfs: exercise host-level HA of NFS-Ganesha by killing the process
- Varsha Rao wrote:
> Patrick Donnelly wrote:
> > In my own testing, the process is not respawned and the NFS client ... - 06:06 AM Bug #47787: mgr/nfs: exercise host-level HA of NFS-Ganesha by killing the process
- Patrick Donnelly wrote:
> In my own testing, the process is not respawned and the NFS client hangs. I suspect there'... - 03:20 PM Bug #46883 (Fix Under Review): kclient: ghost kernel mount
- 03:11 PM Bug #47833 (Fix Under Review): mds FAILED ceph_assert(sessions != 0) in function 'void SessionMap...
- 12:16 PM Bug #47833: mds FAILED ceph_assert(sessions != 0) in function 'void SessionMap::hit_session(Sessi...
- I tested the fix in the same env, with the stale fh clients, and now the mds does not crash while stopping.
- 01:53 PM Bug #47849 (Resolved): qa/vstart_runner: LocalRemote.run can't take multiple commands
- The issue is caused by this commit - https://github.com/ceph/ceph/pull/36457/commits/a177b470aa48a84e5346b310efa4fd62...
- 11:43 AM Bug #47844 (Fix Under Review): mds: only update the requesting metrics
- 11:40 AM Bug #47844 (In Progress): mds: only update the requesting metrics
- 11:40 AM Bug #47844 (Resolved): mds: only update the requesting metrics
- Currently for the MDSs without global metrics needed to be refreshed,
the global metric counters will be zero and th... - 10:17 AM Bug #47798 (Duplicate): pybind/mgr/volumes: TypeError: bad operand type for unary -: 'str' for er...
- Closing this as duplicate of https://tracker.ceph.com/issues/46360
- 10:16 AM Bug #47798: pybind/mgr/volumes: TypeError: bad operand type for unary -: 'str' for errno ETIMEDOUT
- The PR https://github.com/ceph/ceph/pull/35934 has already fixed this issue. The issue is tracked by tracked by https...
- 08:01 AM Bug #47843 (Fix Under Review): mds: stuck in resolve when restarting MDS and reducing max_mds
- In multi MDS ceph cluster, first reduce max_mds,before this step is completed, restart one or more MDS immediately. T...
- 03:34 AM Bug #47652 (Resolved): teuthology's misc.sudo_write_file is incompatible with vstart_runner
- Wasn't aware that Ramana too was working on the same issue. The fix was merged in commit https://github.com/ceph/ceph...
- 01:57 AM Bug #47842 (Resolved): qa: "fsstress.sh: line 16: 28870 Bus error (core dumped) "$B...
- ...
- 01:51 AM Feature #24285 (Resolved): mgr: add module which displays current usage of file system (`fs top`)
10/12/2020
- 07:42 PM Bug #47833: mds FAILED ceph_assert(sessions != 0) in function 'void SessionMap::hit_session(Sessi...
- That patch looks correct. Would you like to post the PR Dan?
- 06:36 PM Bug #47833: mds FAILED ceph_assert(sessions != 0) in function 'void SessionMap::hit_session(Sessi...
- I have the coredump so we can debug further. In the hit_session frame, we see the session clearly:...
- 05:53 PM Bug #47833: mds FAILED ceph_assert(sessions != 0) in function 'void SessionMap::hit_session(Sessi...
- Patrick Donnelly wrote:
> Dan van der Ster wrote:
> > Indeed, I evicted the weird clients spinning on Stale file ha... - 05:24 PM Bug #47833: mds FAILED ceph_assert(sessions != 0) in function 'void SessionMap::hit_session(Sessi...
- Dan van der Ster wrote:
> Indeed, I evicted the weird clients spinning on Stale file handles, and then the mds stopp... - 02:50 PM Bug #47833: mds FAILED ceph_assert(sessions != 0) in function 'void SessionMap::hit_session(Sessi...
- Indeed, I evicted the weird clients spinning on Stale file handles, and then the mds stopping procedure finished with...
- 02:35 PM Bug #47833: mds FAILED ceph_assert(sessions != 0) in function 'void SessionMap::hit_session(Sessi...
- Here is a log with debug_mds=10. ceph-post-file: f4f87969-d492-4e1d-8e8e-5c9e81e45d2f
From what I can gather, (... - 02:10 PM Bug #47833 (Resolved): mds FAILED ceph_assert(sessions != 0) in function 'void SessionMap::hit_se...
- We are not able to decrease from max_mds=2 to 1 on our cephfs cluster.
As soon as we decrease max_mds, the mds goe... - 06:39 PM Backport #47608: octopus: mds: OpenFileTable::prefetch_inodes during rejoin can cause out-of-memory
- Zheng Yan wrote:
> https://github.com/ceph/ceph/pull/37383
merged - 06:39 PM Backport #47604: octopus: mds: purge_queue's _calculate_ops is inaccurate
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37372
merged - 06:38 PM Backport #47601: octopus: mgr/nfs: Cluster creation throws 'NoneType' object has no attribute 're...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37371
merged - 06:38 PM Backport #47260: octopus: client: FAILED assert(dir->readdir_cache[dirp->cache_index] == dn)
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37370
merged - 06:37 PM Backport #47623: octopus: various quota failures
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37369
merged - 06:37 PM Backport #47255: octopus: client: Client::open() pass wrong cap mask to path_walk
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37369
merged - 06:37 PM Backport #47253: octopus: mds: fix possible crash when the MDS is stopping
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37368
merged - 02:43 PM Cleanup #47160 (Resolved): qa/tasks/cephfs: Break up test_volumes.py
- I don't think this will be feasible to backport without significant effort. Ramana, do you think it's worth it?
- 01:39 PM Bug #47798 (Triaged): pybind/mgr/volumes: TypeError: bad operand type for unary -: 'str' for errn...
- 09:02 AM Bug #46883: kclient: ghost kernel mount
- Patrick Donnelly wrote:
> So there are two issues here:
[...]
>
> * Use separate auth credentials for each mount... - 06:53 AM Bug #46883: kclient: ghost kernel mount
- Will work on it.
- 02:46 AM Bug #47565 (Fix Under Review): qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x20...
10/10/2020
- 07:05 PM Bug #36389: untar encounters unexpected EPERM on kclient/multimds cluster with thrashing
- Patrick Donnelly wrote:
> I think this might be a dup of #47723
Yes, it probably is. There is no evidence in /ce... - 09:03 AM Bug #40864 (Resolved): cephfs-shell: rmdir doesn't complain when directory is not empty
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 08:58 AM Backport #47824 (Resolved): octopus: pybind/mgr/volumes: Make number of cloner threads configurable
- https://github.com/ceph/ceph/pull/37671
- 08:58 AM Backport #47823 (Resolved): nautilus: pybind/mgr/volumes: Make number of cloner threads configurable
- https://github.com/ceph/ceph/pull/37936
- 08:43 AM Backport #47259: nautilus: client: FAILED assert(dir->readdir_cache[dirp->cache_index] == dn)
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37232
m... - 06:22 AM Backport #47259 (Resolved): nautilus: client: FAILED assert(dir->readdir_cache[dirp->cache_index]...
- 08:43 AM Backport #47252: nautilus: mds: fix possible crash when the MDS is stopping
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37229
m... - 06:22 AM Backport #47252 (Resolved): nautilus: mds: fix possible crash when the MDS is stopping
- 08:43 AM Backport #47246: nautilus: qa: Replacing daemon mds.a as rank 0 with standby daemon mds.b" in clu...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37228
m... - 06:21 AM Backport #47246 (Resolved): nautilus: qa: Replacing daemon mds.a as rank 0 with standby daemon md...
- 08:43 AM Backport #47088 (Resolved): nautilus: mds: recover files after normal session close
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37178
m... - 08:42 AM Backport #47605 (Resolved): nautilus: mds: purge_queue's _calculate_ops is inaccurate
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37481
m...
10/09/2020
- 06:59 PM Bug #47678: mgr: include/interval_set.h: 466: ceph_abort_msg("abort() called")
- Putting this in the fs project for now.
- 06:55 PM Bug #36389: untar encounters unexpected EPERM on kclient/multimds cluster with thrashing
- I think this might be a dup of #47723
- 06:51 PM Bug #47563 (Fix Under Review): qa: kernel client closes session improperly causing eviction due t...
- 06:49 PM Bug #46883 (Triaged): kclient: ghost kernel mount
- 06:47 PM Bug #46648 (In Progress): mds: cannot handle hundreds+ of subtrees
- Zheng is currently working on this.
- 06:47 PM Bug #46507 (Triaged): qa: test_data_scan: "show inode" returns ENOENT
- 06:46 PM Bug #47787 (Triaged): mgr/nfs: exercise host-level HA of NFS-Ganesha by killing the process
- 05:57 PM Bug #47806 (Fix Under Review): mon/MDSMonitor: divide mds identifier and mds real name with dot
- 03:14 AM Bug #47806 (Resolved): mon/MDSMonitor: divide mds identifier and mds real name with dot
- Current health detail outputs mds slow request as below....
- 05:36 PM Backport #42157 (Rejected): nautilus: cephfs-shell: rmdir doesn't complain when directory is not ...
- 04:02 PM Backport #47259: nautilus: client: FAILED assert(dir->readdir_cache[dirp->cache_index] == dn)
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37232
merged - 04:01 PM Backport #47252: nautilus: mds: fix possible crash when the MDS is stopping
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37229
merged - 04:01 PM Backport #47246: nautilus: qa: Replacing daemon mds.a as rank 0 with standby daemon mds.b" in clu...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37228
merged - 04:00 PM Backport #47088: nautilus: mds: recover files after normal session close
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37178
merged - 03:08 PM Backport #47605: nautilus: mds: purge_queue's _calculate_ops is inaccurate
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37481
merged - 03:03 PM Backport #46960 (Resolved): nautilus: cephfs-journal-tool: incorrect read_offset after finding mi...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37479
m... - 03:03 PM Backport #47622: nautilus: various quota failures
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37231
m... - 02:37 AM Backport #47622 (Resolved): nautilus: various quota failures
- 03:03 PM Backport #47254: nautilus: client: Client::open() pass wrong cap mask to path_walk
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37231
m... - 02:36 AM Backport #47254 (Resolved): nautilus: client: Client::open() pass wrong cap mask to path_walk
- 03:02 PM Backport #47090 (Resolved): nautilus: After restarting an mds, its standy-replay mds remained in ...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37179
m... - 03:02 PM Backport #46784 (Resolved): nautilus: mds/CInode: Optimize only pinned by subtrees check
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36965
m... - 11:53 AM Bug #45344 (Fix Under Review): doc: Table Of Contents doesn't work
- 09:53 AM Bug #45344: doc: Table Of Contents doesn't work
- Zac Dover wrote:
> There's strange behavior here.
>
> The top-level menu items link nowhere, but the second-order... - 09:32 AM Bug #45344: doc: Table Of Contents doesn't work
- There's strange behavior here.
The top-level menu items link nowhere, but the second-order menu items link to targ...
10/08/2020
- 08:25 PM Feature #46892 (Pending Backport): pybind/mgr/volumes: Make number of cloner threads configurable
- 08:23 PM Feature #42451 (Resolved): mds: add root_squash
- 08:16 PM Bug #47786: mds: log [ERR] : failed to commit dir 0x100000005f1.1010* object, errno -2
- /ceph/teuthology-archive/pdonnell-2020-10-08_01:40:56-multimds-wip-pdonnell-testing-20201007.214100-distro-basic-smit...
- 03:38 PM Bug #47798 (Duplicate): pybind/mgr/volumes: TypeError: bad operand type for unary -: 'str' for er...
- A stack trace when facing ETIMEDOUT errno during subvolume operations is presented below,...
- 03:23 PM Backport #46960: nautilus: cephfs-journal-tool: incorrect read_offset after finding missing objects
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37479
merged - 03:23 PM Backport #47622: nautilus: various quota failures
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37231
merged - 03:23 PM Backport #47254: nautilus: client: Client::open() pass wrong cap mask to path_walk
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37231
merged - 03:22 PM Backport #47090: nautilus: After restarting an mds, its standy-replay mds remained in the "resolv...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37179
merged - 03:22 PM Backport #46784: nautilus: mds/CInode: Optimize only pinned by subtrees check
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36965
merged - 01:35 PM Bug #46273 (Resolved): mds: deleting a large number of files in a directory causes the file syste...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 01:35 PM Bug #46355 (Resolved): client: directory inode can not call release_callback
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 01:35 PM Bug #46597 (Resolved): qa: Fs cleanup fails with a traceback
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 01:34 PM Bug #47015 (Resolved): mds: decoding of enum types on big-endian systems broken
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 01:34 PM Bug #47201 (Resolved): mds: CDir::_omap_commit(int): Assertion `committed_version == 0' failed.
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 10:53 AM Bug #46434: osdc: FAILED ceph_assert(bh->waitfor_read.empty())
- Saw this issue again in a recent nautilus test run,
https://pulpito.ceph.com/yuriw-2020-10-05_22:19:52-multimds-wip-... - 10:26 AM Backport #47316 (Resolved): octopus: mds: CDir::_omap_commit(int): Assertion `committed_version =...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37034
m... - 10:26 AM Backport #46520 (Resolved): octopus: mds: deleting a large number of files in a directory causes ...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37034
m... - 10:26 AM Backport #46522: octopus: mds: fix hang issue when accessing a file under a lost parent directory
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37020
m... - 10:25 AM Backport #46516 (Resolved): octopus: client: directory inode can not call release_callback
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37017
m... - 10:25 AM Backport #47080 (Resolved): octopus: mds: decoding of enum types on big-endian systems broken
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36813
m... - 10:25 AM Backport #46947 (Resolved): octopus: qa: Fs cleanup fails with a traceback
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36713
m...
10/07/2020
- 08:34 PM Bug #47787 (Triaged): mgr/nfs: exercise host-level HA of NFS-Ganesha by killing the process
- In my own testing, the process is not respawned and the NFS client hangs. I suspect there's some changes necessary to...
- 08:24 PM Bug #47786 (Resolved): mds: log [ERR] : failed to commit dir 0x100000005f1.1010* object, errno -2
- ...
- 07:38 PM Bug #47591 (Resolved): TestNFS: test_exports_on_mgr_restart: command failed with status 32: 'sudo...
- 05:41 PM Documentation #47784 (In Progress): nfs: Remove doc on creating cephfs exports using rook
- 05:37 PM Documentation #47784 (Resolved): nfs: Remove doc on creating cephfs exports using rook
- The doc[1] on creating cephfs exports using dashboard with rook is outdated and using dashboard backend script is bug...
- 05:12 PM Bug #47783 (Fix Under Review): mgr/nfs: Pseudo path prints wrong error message
- 05:09 PM Bug #47783 (Resolved): mgr/nfs: Pseudo path prints wrong error message
- Pseudo path must be an absolute path. But the error message printed is "It should not be absolute path".
10/06/2020
- 05:37 PM Bug #47591 (Fix Under Review): TestNFS: test_exports_on_mgr_restart: command failed with status 3...
- I am still not able to reproduce the issue with latest master branch. Looking at failure logs, I suspect ganesha daem...
- 05:26 PM Bug #46129 (Resolved): mds: fix hang issue when accessing a file under a lost parent directory
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 05:05 PM Bug #47698: mds crashed in try_remove_dentries_for_stray after touching file in strange directory
- Thanks. scan_links fixed these dups (and some Bad nlink). For future reference, it took 2 hours to complete on a clus...
- 06:42 AM Backport #45853: octopus: cephfs-journal-tool: NetHandler create_socket couldn't create socket
- Zheng Yan do you intend to work on this one?
- 05:39 AM Bug #47515 (Fix Under Review): pybind/snap_schedule: deactivating a schedule is ineffective
- 01:49 AM Backport #46522 (Resolved): octopus: mds: fix hang issue when accessing a file under a lost paren...
10/05/2020
- 11:33 PM Backport #47316: octopus: mds: CDir::_omap_commit(int): Assertion `committed_version == 0' failed.
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37034
merged - 11:32 PM Backport #46520: octopus: mds: deleting a large number of files in a directory causes the file sy...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37034
merged - 11:32 PM Backport #46522: octopus: mds: fix hang issue when accessing a file under a lost parent directory
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37020
merged - 11:31 PM Backport #46516: octopus: client: directory inode can not call release_callback
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37017
merged - 11:31 PM Backport #47080: octopus: mds: decoding of enum types on big-endian systems broken
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36813
merged - 11:30 PM Backport #46947: octopus: qa: Fs cleanup fails with a traceback
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36713
merged - 08:20 PM Bug #44638 (Resolved): test_scrub_pause_and_resume (tasks.cephfs.test_scrub_checks.TestScrubContr...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:53 PM Backport #46524: octopus: non-head batch requests may hold authpins and locks
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37022
m... - 02:15 AM Backport #46524 (Resolved): octopus: non-head batch requests may hold authpins and locks
- 07:52 PM Backport #46473: octopus: mds: make threshold for MDS_TRIM warning configurable
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36970
m... - 02:15 AM Backport #46473 (Resolved): octopus: mds: make threshold for MDS_TRIM warning configurable
- 07:50 PM Backport #47017 (Resolved): nautilus: mds: kcephfs parse dirfrag's ndist is always 0
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37177
m... - 03:41 PM Backport #47017: nautilus: mds: kcephfs parse dirfrag's ndist is always 0
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37177
merged - 07:50 PM Backport #47317 (Resolved): nautilus: mds: CDir::_omap_commit(int): Assertion `committed_version ...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/37035
m... - 03:41 PM Backport #47317: nautilus: mds: CDir::_omap_commit(int): Assertion `committed_version == 0' failed.
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37035
merged - 07:50 PM Backport #46941: nautilus: mds: memory leak during cache drop
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36967
m... - 03:48 PM Backport #46941 (Resolved): nautilus: mds: memory leak during cache drop
- 03:40 PM Backport #46941: nautilus: mds: memory leak during cache drop
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36967
merged - 07:50 PM Backport #46787: nautilus: client: in _open() the open ref maybe decreased twice, but only increa...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36966
m... - 03:48 PM Backport #46787 (Resolved): nautilus: client: in _open() the open ref maybe decreased twice, but ...
- 03:39 PM Backport #46787: nautilus: client: in _open() the open ref maybe decreased twice, but only increa...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36966
merged - 07:49 PM Backport #47081 (Resolved): nautilus: mds: decoding of enum types on big-endian systems broken
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36814
m... - 03:39 PM Backport #47081: nautilus: mds: decoding of enum types on big-endian systems broken
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36814
merged - 07:49 PM Backport #46151 (Resolved): nautilus: test_scrub_pause_and_resume (tasks.cephfs.test_scrub_checks...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36168
m... - 03:38 PM Backport #46151: nautilus: test_scrub_pause_and_resume (tasks.cephfs.test_scrub_checks.TestScrubC...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36168
merged - 07:49 PM Backport #46943: nautilus: mds: segv in MDCache::wait_for_uncommitted_fragments
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36968
m... - 03:48 PM Backport #46943 (Resolved): nautilus: mds: segv in MDCache::wait_for_uncommitted_fragments
- 03:36 PM Backport #46943: nautilus: mds: segv in MDCache::wait_for_uncommitted_fragments
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36968
merged - 07:48 PM Backport #46633: nautilus: mds forwarding request 'no_available_op_found'
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36963
m... - 03:36 PM Backport #46633 (Resolved): nautilus: mds forwarding request 'no_available_op_found'
- 03:35 PM Backport #46633: nautilus: mds forwarding request 'no_available_op_found'
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36963
merged - 01:16 PM Bug #44785 (Resolved): non-head batch requests may hold authpins and locks
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 01:15 PM Feature #45906 (Resolved): mds: make threshold for MDS_TRIM warning configurable
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 06:30 AM Bug #47744 (Duplicate): nautilus: CDir::_omap_commit(int): Assertion `committed_version == 0' fai...
- Duplicate of https://tracker.ceph.com/issues/47201
- 06:27 AM Bug #47744 (Duplicate): nautilus: CDir::_omap_commit(int): Assertion `committed_version == 0' fai...
- ...
10/04/2020
10/03/2020
- 04:05 PM Backport #46524: octopus: non-head batch requests may hold authpins and locks
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/37022
merged - 03:58 PM Backport #46473: octopus: mds: make threshold for MDS_TRIM warning configurable
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36970
merged - 12:15 AM Bug #47734 (Fix Under Review): client: hang after statfs
- 12:00 AM Bug #47734 (Resolved): client: hang after statfs
- ...
10/02/2020
- 06:07 PM Bug #47689 (Fix Under Review): rados/upgrade/nautilus-x-singleton fails due to cluster [WRN] evic...
- 06:04 PM Bug #47689 (In Progress): rados/upgrade/nautilus-x-singleton fails due to cluster [WRN] evicting ...
- This appears to be a fairly old failure. Here's a few instances:
https://pulpito.ceph.com/teuthology-2020-07-22_07... - 04:25 PM Bug #47591 (New): TestNFS: test_exports_on_mgr_restart: command failed with status 32: 'sudo moun...
- ...
- 01:07 PM Bug #47698: mds crashed in try_remove_dentries_for_stray after touching file in strange directory
- Dan van der Ster wrote:
> Now I'm trying to repair the metadata on this fs so it fully consistent.
> When I run 'sc... - 11:35 AM Bug #47698: mds crashed in try_remove_dentries_for_stray after touching file in strange directory
- Now I'm trying to repair the metadata on this fs so it fully consistent.
When I run 'scrub start / force recurstive ...
10/01/2020
- 08:08 PM Bug #47642 (Resolved): nautilus: qa/suites/{kcephfs, multimds}: client kernel "testing" builds fo...
- 05:05 PM Bug #47689: rados/upgrade/nautilus-x-singleton fails due to cluster [WRN] evicting unresponsive c...
- /a/teuthology-2020-10-01_07:01:02-rados-master-distro-basic-smithi/5485885
- 03:31 PM Bug #43762: pybind/mgr/volumes: create fails with TypeError
- Jos Collin wrote:
> Victoria Martinez de la Cruz wrote:
> > Adding more context to this
> >
> > This happened af... - 06:36 AM Bug #47565: qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x200000007d5 pending p...
- Patrick Donnelly wrote:
> Sounds good. Please write up a PR for this Xiubo.
Sure, will do. - 02:17 AM Bug #43902: qa: mon_thrash: timeout "ceph quorum_status"
- /ceph/teuthology-archive/pdonnell-2020-09-29_05:23:34-fs-wip-pdonnell-testing-20200929.022151-distro-basic-smithi/547...
09/30/2020
- 09:18 PM Bug #47565: qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x200000007d5 pending p...
- Sounds good. Please write up a PR for this Xiubo.
- 01:42 AM Bug #47565: qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x200000007d5 pending p...
- Patrick Donnelly wrote:
> Xiubo Li wrote:
> > @Patrick,
> >
> > Maybe the MDS shouldn't report the WRN to monito... - 09:10 PM Bug #47307: mds: throttle workloads which acquire caps faster than the client can release
- Dan van der Ster wrote:
> Are you sure that the defaults for recalling aren't overly conservative?
Yes, the proba... - 05:54 PM Bug #47689: rados/upgrade/nautilus-x-singleton fails due to cluster [WRN] evicting unresponsive c...
- /a/teuthology-2020-09-30_07:01:02-rados-master-distro-basic-smithi/5483508/
- 03:41 PM Fix #46645 (Resolved): librados|libcephfs: use latest MonMap when creating from CephContext
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 03:02 PM Bug #47698: mds crashed in try_remove_dentries_for_stray after touching file in strange directory
- b1 was not longer there after we followed the recover_dentries procedure, so it is gone.
- 02:43 PM Bug #47698: mds crashed in try_remove_dentries_for_stray after touching file in strange directory
- try deleting 'd1' using 'rados rmomapkey'. If you have debug_mds=10, it should be easy to get d1's parent dirfrag (co...
- 01:21 PM Bug #47698: mds crashed in try_remove_dentries_for_stray after touching file in strange directory
- Here is the `b1` dir at the start of this issue:...
- 01:19 PM Bug #47698: mds crashed in try_remove_dentries_for_stray after touching file in strange directory
- After finishing the following, the MDS started:...
- 01:03 PM Bug #47698 (New): mds crashed in try_remove_dentries_for_stray after touching file in strange dir...
- We had a directory "b1" which appeared empty but could not be rmdir'd.
The directory also had a very large size, als... - 07:27 AM Bug #47693 (In Progress): qa: snap replicator tests
- 07:24 AM Bug #47693 (Rejected): qa: snap replicator tests
- add tests for snap replicator component
requires PR#36276 - 07:11 AM Backport #46479 (Resolved): octopus: mds: send scrub status to ceph-mgr only when scrub is runnin...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36047
m...
09/29/2020
- 09:46 PM Backport #46479: octopus: mds: send scrub status to ceph-mgr only when scrub is running (or pause...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36047
merged - 08:08 PM Bug #47689 (Resolved): rados/upgrade/nautilus-x-singleton fails due to cluster [WRN] evicting unr...
- ...
- 07:20 PM Backport #47605 (In Progress): nautilus: mds: purge_queue's _calculate_ops is inaccurate
- 05:05 PM Backport #47020 (In Progress): nautilus: client: shutdown race fails with status 141
- 04:56 PM Backport #46960 (In Progress): nautilus: cephfs-journal-tool: incorrect read_offset after finding...
- 02:17 PM Bug #47307: mds: throttle workloads which acquire caps faster than the client can release
- Are you sure that the defaults for recalling aren't overly conservative?
Today debugging a situation with 2 heavy ... - 02:08 PM Bug #47307 (In Progress): mds: throttle workloads which acquire caps faster than the client can r...
- 02:10 PM Bug #47565: qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x200000007d5 pending p...
- Xiubo Li wrote:
> @Patrick,
>
> Maybe the MDS shouldn't report the WRN to monitor when revoking the "Fwbl" caps ?... - 08:58 AM Bug #47565: qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x200000007d5 pending p...
- @Patrick,
Maybe the MDS shouldn't report the WRN to monitor when revoking the "Fwbl" caps ? Since it may need to f... - 08:21 AM Bug #47565: qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x200000007d5 pending p...
- During flush the 0x200000007d5 inode, there also have many other inodes doing the flush on the same osd.6 at the same...
- 03:21 AM Bug #47565: qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x200000007d5 pending p...
- From 5451587/remote/smithi110/log/ceph-client.1.30354.log.gz:
We can see that the client.4606 has received the rev... - 02:45 AM Bug #47565 (In Progress): qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x2000000...
- 02:07 PM Bug #47682: MDS can't release caps faster than clients taking caps
- Dan, see: #47307
- 01:51 PM Bug #47682 (Rejected): MDS can't release caps faster than clients taking caps
- with more effective tuning I think we can manage. cancelling this ticket.
- 10:23 AM Bug #47682: MDS can't release caps faster than clients taking caps
- Our current config is:
mds_recall_global_max_decay_threshold 200000
mds_recall_max_decay_threshold 100000
mds_re... - 10:10 AM Bug #47682: MDS can't release caps faster than clients taking caps
- Update:
* the central cache freelist eventually decreases after an hour or so.
* I suppose the bigger issue is tha... - 08:06 AM Bug #47682 (Rejected): MDS can't release caps faster than clients taking caps
- We have a workload in which a kernel client is stat'ing all files in an FS. This workload triggered a few issues:
...
09/28/2020
- 07:30 PM Backport #47014 (Resolved): octopus: librados|libcephfs: use latest MonMap when creating from Cep...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36705
m... - 07:30 PM Backport #47013 (Resolved): nautilus: librados|libcephfs: use latest MonMap when creating from Ce...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36704
m... - 02:51 PM Backport #47013: nautilus: librados|libcephfs: use latest MonMap when creating from CephContext
- Patrick Donnelly wrote:
> https://github.com/ceph/ceph/pull/36704
merged - 07:29 PM Bug #47563: qa: kernel client closes session improperly causing eviction due to timeout
- I have a patch I'm testing now that seems to also anecdotally fix some of the umount hangs I've seen lately during xf...
- 05:50 PM Bug #47563: qa: kernel client closes session improperly causing eviction due to timeout
- Patrick Donnelly wrote:
>
> I _think_ the concern is that hte client could conceivably dirty the cap the MDS just ... - 05:39 PM Bug #47563: qa: kernel client closes session improperly causing eviction due to timeout
- Jeff Layton wrote:
> Doesn't look like libcephfs does anything saner:
>
> [...]
>
> ...and it looks like the t... - 05:07 PM Bug #47563: qa: kernel client closes session improperly causing eviction due to timeout
- Doesn't look like libcephfs does anything saner:...
- 04:51 PM Bug #47563: qa: kernel client closes session improperly causing eviction due to timeout
- Jeff Layton wrote:
> Hmm, ok. This may be related to another bug I've been chasing where umount hangs waiting for th... - 04:39 PM Bug #47563 (In Progress): qa: kernel client closes session improperly causing eviction due to tim...
- 04:34 PM Bug #47563: qa: kernel client closes session improperly causing eviction due to timeout
- Hmm, ok. This may be related to another bug I've been chasing where umount hangs waiting for the session to close. I ...
- 06:50 PM Bug #47006 (Resolved): mon: required client features adding/removing
- 06:49 PM Feature #47148 (Resolved): mds: get rid of the mds_lock when storing the inode backtrace to meta ...
- 06:47 PM Tasks #47047 (Resolved): client: release the client_lock before copying data in all the reads
- 06:47 PM Bug #47039 (Resolved): client: mutex lock FAILED ceph_assert(nlock > 0)
- 06:42 PM Bug #47679 (New): kceph: kernel does not open session with MDS importing subtree
- ...
- 06:24 PM Bug #47678: mgr: include/interval_set.h: 466: ceph_abort_msg("abort() called")
- https://pulpito.ceph.com/teuthology-2020-09-21_04:15:02-multimds-master-distro-basic-smithi/5454314/
Seems to be a... - 06:17 PM Bug #47678 (New): mgr: include/interval_set.h: 466: ceph_abort_msg("abort() called")
- ...
- 06:11 PM Bug #47294: client: thread hang in Client::_setxattr_maybe_wait_for_osdmap
- Another: /ceph/teuthology-archive/pdonnell-2020-09-26_05:47:56-fs-wip-pdonnell-testing-20200926.000836-distro-basic-s...
- 04:33 PM Feature #47034: mds: readdir for snapshot diff
- Hey Zheng,
CephFS snapshot mirror would make use of rctime approach. That needs PR https://github.com/ceph/ceph/pu... - 03:03 PM Bug #47642 (Fix Under Review): nautilus: qa/suites/{kcephfs, multimds}: client kernel "testing" b...
- 01:40 PM Bug #47662 (Fix Under Review): mds: try to replicate hot dir to restarted MDS
09/27/2020
- 10:59 PM Backport #47014: octopus: librados|libcephfs: use latest MonMap when creating from CephContext
- Patrick Donnelly wrote:
> https://github.com/ceph/ceph/pull/36705
merged - 10:41 AM Bug #47662 (Resolved): mds: try to replicate hot dir to restarted MDS
- Hot dir would be replicated to other active MDSes, but if replica MDS restarted, auth MDS won't replicate this dir ag...
09/25/2020
- 04:40 PM Feature #15070 (Resolved): mon: client: multifs: auth caps on client->mon connections to limit th...
- 02:59 PM Bug #47652: teuthology's misc.sudo_write_file is incompatible with vstart_runner
- > The compatibility was broken by this teuthology PR, since it makes
"this teuthology PR": https://github.com/cep... - 02:58 PM Bug #47652 (Fix Under Review): teuthology's misc.sudo_write_file is incompatible with vstart_runner
- 02:41 PM Bug #47652 (Resolved): teuthology's misc.sudo_write_file is incompatible with vstart_runner
- Here's the traceback -...
- 02:32 PM Feature #46059: vstart_runner.py: optionally rotate logs between tests
- Got some time to work on this finally. Fixed the PR after some scrutiny, ceph API tests pass for this PR now.
- 08:57 AM Backport #47622 (In Progress): nautilus: various quota failures
- 08:43 AM Bug #47643: mds: Segmentation fault in thread 7fcff3078700 thread_name:md_log_replay
- Patrick Donnelly wrote:
> > #x 0x5628d800
>
> I'm not sure this double-deref is indicating anything. Are you sure... - 12:22 AM Cleanup #47325 (Resolved): client: remove unneccessary client_lock for objector->write()
- 12:20 AM Bug #40613 (New): kclient: .handle_message_footer got old message 1 <= 648 0x558ceadeaac0 client_...
- This one is back:...
09/24/2020
- 07:33 PM Bug #46823 (Resolved): nautilus: kceph w/ testing branch: mdsc_handle_session corrupt message mds...
- Fixed upstream.
- 07:17 PM Backport #47622 (Need More Info): nautilus: various quota failures
- 07:16 PM Backport #47623 (In Progress): octopus: various quota failures
- 05:29 PM Bug #47643 (Need More Info): mds: Segmentation fault in thread 7fcff3078700 thread_name:md_log_re...
- > #x 0x5628d800
I'm not sure this double-deref is indicating anything. Are you sure that's a pointer? Would you no... - 04:43 PM Bug #47643 (Need More Info): mds: Segmentation fault in thread 7fcff3078700 thread_name:md_log_re...
- In ceph-14.2.11.394+g9cbbc473c0 (downstream build but mds sources are the same as v14.2.11) we got a report about the...
- 04:33 PM Bug #47642 (Resolved): nautilus: qa/suites/{kcephfs, multimds}: client kernel "testing" builds fo...
- As described in https://tracker.ceph.com/issues/47540, kernel "testing" builds for CentOS 7 are unavailable. This is ...
- 11:31 AM Bug #47591 (Can't reproduce): TestNFS: test_exports_on_mgr_restart: command failed with status 32...
- The mount command does not fail with latest builds: http://pulpito.front.sepia.ceph.com/varsha-2020-09-24_10:49:55-ra...
- 07:29 AM Bug #46769: qa: Refactor cephfs creation/removal code.
- Based on comment https://github.com/ceph/ceph/pull/36368#pullrequestreview-458486627, retaining the behavior of clean...
- 03:44 AM Backport #47608 (In Progress): octopus: mds: OpenFileTable::prefetch_inodes during rejoin can cau...
- https://github.com/ceph/ceph/pull/37383
- 03:43 AM Backport #47609 (In Progress): nautilus: mds: OpenFileTable::prefetch_inodes during rejoin can ca...
- https://github.com/ceph/ceph/pull/37382
09/23/2020
- 07:06 PM Bug #45835: mds: OpenFileTable::prefetch_inodes during rejoin can cause out-of-memory
- Dan van der Ster wrote:
> The fix was merged. Something needed to start the backports process?
@Dan, the "backpor... - 07:05 PM Bug #46583 (Resolved): mds slave request 'no_available_op_found'
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:04 PM Backport #47623 (Resolved): octopus: various quota failures
- https://github.com/ceph/ceph/pull/37369
- 07:04 PM Backport #47622 (Resolved): nautilus: various quota failures
- https://github.com/ceph/ceph/pull/37231
- 03:56 PM Backport #46790 (Rejected): nautilus: mds slave request 'no_available_op_found'
- This isn't really necessary for backport.
- 01:31 PM Backport #46790 (Need More Info): nautilus: mds slave request 'no_available_op_found'
- non-trivial conflicts
@Patrick, could you help find the right assignee for this? - 03:56 PM Backport #46789 (Rejected): octopus: mds slave request 'no_available_op_found'
- This isn't really necessary for backport.
- 01:30 PM Backport #46789 (Need More Info): octopus: mds slave request 'no_available_op_found'
- non-trivial conflicts
@Patrick, could you help find the right assignee for this? - 03:06 PM Bug #47224 (Pending Backport): various quota failures
- 01:22 PM Backport #47608 (Need More Info): octopus: mds: OpenFileTable::prefetch_inodes during rejoin can ...
- extensive changeset with non-trivial conflicts
- 11:17 AM Backport #47608 (Resolved): octopus: mds: OpenFileTable::prefetch_inodes during rejoin can cause ...
- https://github.com/ceph/ceph/pull/37383
- 01:19 PM Backport #47604 (In Progress): octopus: mds: purge_queue's _calculate_ops is inaccurate
- 11:15 AM Backport #47604 (Resolved): octopus: mds: purge_queue's _calculate_ops is inaccurate
- https://github.com/ceph/ceph/pull/37372
- 01:12 PM Backport #47601 (In Progress): octopus: mgr/nfs: Cluster creation throws 'NoneType' object has no...
- 11:14 AM Backport #47601 (Resolved): octopus: mgr/nfs: Cluster creation throws 'NoneType' object has no at...
- https://github.com/ceph/ceph/pull/37371
- 01:09 PM Backport #47260 (In Progress): octopus: client: FAILED assert(dir->readdir_cache[dirp->cache_inde...
- 01:08 PM Backport #47255 (In Progress): octopus: client: Client::open() pass wrong cap mask to path_walk
- 01:02 PM Backport #47253 (In Progress): octopus: mds: fix possible crash when the MDS is stopping
- 01:02 PM Backport #47247 (In Progress): octopus: qa: Replacing daemon mds.a as rank 0 with standby daemon ...
- 12:53 PM Backport #47151 (In Progress): octopus: pybind/mgr/volumes: add debugging for global lock
- 12:52 PM Backport #47147 (In Progress): octopus: pybind/mgr/nfs: Test mounting of exports created with nfs...
- 12:51 PM Backport #47095 (Need More Info): octopus: mds: provide altrenatives to increase the total cephfs...
- non-trivial feature
- 12:50 PM Backport #47089 (In Progress): octopus: After restarting an mds, its standy-replay mds remained i...
- 12:49 PM Backport #47085 (In Progress): octopus: common: validate type CephBool cause 'invalid command json'
- 12:30 PM Backport #47083 (In Progress): octopus: mds: 'forward loop' when forward_all_requests_to_auth is set
- 12:25 PM Feature #47266 (Closed): add a subcommand to change caps in a simpler and clear way
- 12:13 PM Bug #47006 (Fix Under Review): mon: required client features adding/removing
- 12:07 PM Backport #47021 (In Progress): octopus: client: shutdown race fails with status 141
- 12:06 PM Backport #47018 (In Progress): octopus: mds: kcephfs parse dirfrag's ndist is always 0
- 12:06 PM Backport #47016 (In Progress): octopus: mds: fix the decode version
- 12:05 PM Backport #46942 (In Progress): octopus: mds: segv in MDCache::wait_for_uncommitted_fragments
- 12:05 PM Backport #46940 (In Progress): octopus: mds: memory leak during cache drop
- 12:02 PM Backport #46859 (In Progress): octopus: mds: do not raise "client failing to respond to cap relea...
- 12:01 PM Backport #46857 (In Progress): octopus: qa: add debugging for volumes plugin use of libcephfs
- 12:01 PM Backport #46855 (In Progress): octopus: client: static dirent for readdir is not thread-safe
- 11:59 AM Backport #46463 (In Progress): octopus: mgr/volumes: fs subvolume clones stuck in progress when l...
- 11:54 AM Backport #46094 (Need More Info): octopus: cephfs-shell: set proper return value for the tool
- non-trivial conflicts
- 11:18 AM Bug #44408 (Resolved): qa: after the cephfs qa test case quit the mountpoints still exist
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 11:17 AM Backport #47609 (Rejected): nautilus: mds: OpenFileTable::prefetch_inodes during rejoin can cause...
- https://github.com/ceph/ceph/pull/37382
- 11:17 AM Bug #46269 (Resolved): ceph-fuse: ceph-fuse process is terminated by the logratote task and what ...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 11:15 AM Backport #47605 (Resolved): nautilus: mds: purge_queue's _calculate_ops is inaccurate
- https://github.com/ceph/ceph/pull/37481
- 11:10 AM Backport #47087 (In Progress): octopus: mds: recover files after normal session close
- 11:02 AM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
- Hi Jeff,
Have finished code in MDS, and for now I didn't handle the loopup version case. All the version related ... - 01:24 AM Feature #47162 (Fix Under Review): mds: handle encrypted filenames in the MDS for fscrypt
- 08:21 AM Backport #47178 (Resolved): nautilus: qa: after the cephfs qa test case quit the mountpoints stil...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36863
m... - 08:19 AM Backport #47152 (Resolved): nautilus: pybind/mgr/volumes: add debugging for global lock
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36828
m... - 08:19 AM Backport #46948 (Resolved): nautilus: qa: Fs cleanup fails with a traceback
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36714
m... - 08:19 AM Backport #46592 (Resolved): nautilus: ceph-fuse: ceph-fuse process is terminated by the logratote...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/36181
m...
09/22/2020
- 10:27 PM Bug #47591 (Resolved): TestNFS: test_exports_on_mgr_restart: command failed with status 32: 'sudo...
- a/mgfritch-2020-09-21_20:24:35-rados:cephadm-wip-mgfritch-testing-2020-09-21-1034-distro-basic-smithi/5457554/teuthol...
- 08:09 PM Backport #47095: octopus: mds: provide altrenatives to increase the total cephfs subvolume snapsh...
- https://tracker.ceph.com/issues/47158 depends on the backport for this issue.
A simple cherry pick is throwing con... - 08:05 PM Backport #47158: octopus: mgr/volumes: Mark subvolumes with ceph.dir.subvolume vxattr, to improve...
- Also depends on the backport of https://tracker.ceph.com/issues/47095
- 07:49 PM Backport #47178: nautilus: qa: after the cephfs qa test case quit the mountpoints still exist
- Patrick Donnelly wrote:
> https://github.com/ceph/ceph/pull/36863
merged - 06:15 PM Bug #47582: MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- Thank you for the provided information.
I will test the MDS failover in a day. Quick question regarding "mds_log_m... - 04:14 PM Bug #47582: MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- Heilig IOS wrote:
> Still no changes. The "mds_log_max_segments" didn't help. The MDS failover is running for 30 min... - 04:13 PM Bug #47582 (Rejected): MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- (This discussion should move to ceph-users.)
- 02:40 PM Bug #47582: MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- Still no changes. The "mds_log_max_segments" didn't help. The MDS failover is running for 30 minutes already. What el...
- 02:04 PM Bug #47582: MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- I decreased it with these commands:...
- 01:52 PM Bug #47582: MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- Heilig IOS wrote:
> Current value: mds_log_max_segments = 100000
that's the root cause. the value should be small... - 01:47 PM Bug #47582: MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- Current value: mds_log_max_segments = 100000
- 01:34 PM Bug #47582: MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- what is the value of "mds log max segments" config
- 01:24 PM Bug #47582: MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- I have this issue right now. No, there is no "mds behind on trim" warning.
- 01:11 PM Bug #47582: MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- were there "mds behind on trim" warning
- 12:16 PM Bug #47582 (Rejected): MDS failover takes 10-15 hours: Ceph MDS stays in "up:replay" state for hours
- We have 9 nodes Ceph cluster. Ceph version is 15.2.5. The cluster has 175 OSD (HDD) + 3 NVMe for cache tier for "ceph...
- 04:07 PM Backport #47254: nautilus: client: Client::open() pass wrong cap mask to path_walk
- regression: https://tracker.ceph.com/issues/47224
- 04:07 PM Backport #47255: octopus: client: Client::open() pass wrong cap mask to path_walk
- regression: https://tracker.ceph.com/issues/47224
- 03:21 PM Feature #47490: Integration of dashboard with volume/nfs module
- Volume/nfs module doc: https://docs.ceph.com/docs/master/cephfs/fs-nfs-exports
- 03:02 PM Feature #47490 (In Progress): Integration of dashboard with volume/nfs module
- 09:35 AM Feature #47490: Integration of dashboard with volume/nfs module
- Exports and nfs clusters cannot be managed by dashboard and volumes/nfs interface at the same time. Xattrs can be use...
- 02:55 PM Feature #47587 (In Progress): pybind/mgr/nfs: add Rook support
- 02:10 PM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
- Jeff Layton wrote:
> Xiubo Li wrote:
> > Hi Jeff,
> >
> > There is another case for lookup:
> >
> > If the MD... - 12:02 PM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
- Xiubo Li wrote:
> Hi Jeff,
>
> There is another case for lookup:
>
> If the MDS is old version, such as all th... - 11:52 AM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
- I think the MDS should treat these names as opaque. The client should never need to look up a dentry by the binary cr...
- 02:35 AM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
- Hi Jeff,
There is another case for lookup:
If the MDS is old version, such as all the dentries is under `ceph_f... - 12:34 PM Bug #47224 (Fix Under Review): various quota failures
09/21/2020
- 09:53 PM Bug #47294: client: thread hang in Client::_setxattr_maybe_wait_for_osdmap
- All right, I'm going to shove some more debug information in Objecter and Monitor.
- 12:40 AM Bug #47294: client: thread hang in Client::_setxattr_maybe_wait_for_osdmap
- Xiubo Li wrote:
> Patrick Donnelly wrote:
> > Xiubo Li wrote:
> > > Hi Patrick,
> > >
> > > For this let's add ... - 09:04 PM Bug #47526 (Resolved): qa: RuntimeError: FSCID 2 not in map
- 09:02 PM Bug #36389: untar encounters unexpected EPERM on kclient/multimds cluster with thrashing
- ...
- 08:32 PM Bug #47565 (Resolved): qa: "client.4606 isn't responding to mclientcaps(revoke), ino 0x200000007d...
- ...
- 07:47 PM Bug #47563 (Resolved): qa: kernel client closes session improperly causing eviction due to timeout
- ...
- 05:08 PM Bug #45835 (Pending Backport): mds: OpenFileTable::prefetch_inodes during rejoin can cause out-of...
- 03:47 PM Bug #45835: mds: OpenFileTable::prefetch_inodes during rejoin can cause out-of-memory
- The fix was merged. Something needed to start the backports process?
- 03:21 PM Backport #47152: nautilus: pybind/mgr/volumes: add debugging for global lock
- Patrick Donnelly wrote:
> https://github.com/ceph/ceph/pull/36828
merged - 03:20 PM Backport #46948: nautilus: qa: Fs cleanup fails with a traceback
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36714
merged - 03:20 PM Backport #46592: nautilus: ceph-fuse: ceph-fuse process is terminated by the logratote task and w...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/36181
merged - 01:51 PM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
- Jeff Layton wrote:
> Xiubo Li wrote:
> > Ceph has its own base64 encode/decode logic already in src/common/armor.c,... - 01:42 PM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
- Xiubo Li wrote:
> Ceph has its own base64 encode/decode logic already in src/common/armor.c, which is the same with ... - 01:39 PM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
I am planing to append a `fscrypt.alternate_name : ${raw_ciphertext}` pair to the xattr map when doing the create d...- 04:07 AM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
- Ceph has its own base64 encode/decode logic already in src/common/armor.c, which is the same with the kernel does.
09/20/2020
- 11:01 AM Feature #47162: mds: handle encrypted filenames in the MDS for fscrypt
- Jeff Layton wrote:
> Xiubo Li wrote:
> >
> > Yeah, this looks good.
> >
> > BTW, what the alternat_name will s...
Also available in: Atom