Project

General

Profile

Activity

From 04/23/2018 to 05/22/2018

05/22/2018

10:18 PM Bug #21777: src/mds/MDCache.cc: 4332: FAILED assert(mds->is_rejoin())
Zheng, do you think this is also resolved by the fix to #23826? Patrick Donnelly
10:12 PM Bug #24129 (In Progress): qa: test_version_splitting (tasks.cephfs.test_sessionmap.TestSessionMap...
Patrick Donnelly
10:11 PM Documentation #24093 (In Progress): doc: Update *remove a metadata server*
Patrick Donnelly
10:08 PM Bug #24101 (Closed): mds: deadlock during fsstress workunit with 9 actives
Apparently resolved by the revert. Patrick Donnelly
10:03 PM Feature #23689: qa: test major/minor version upgrades
Partially addressed by the QA suite fs:upgrade:snaps. Patrick Donnelly
09:48 PM Feature #24233: Add new command ceph mds status
Why does the command need renamed? Patrick Donnelly
08:49 PM Feature #24233 (Closed): Add new command ceph mds status
Add new command ceph mds status
For more information please check - https://tracker.ceph.com/issues/24217
Changin...
Vikhyat Umrao
09:42 PM Feature #20598: mds: revisit LAZY_IO
See also: https://github.com/ceph/ceph/pull/21067 Patrick Donnelly
08:12 PM Bug #23972: Ceph MDS Crash from client mounting aufs over cephfs
John Spray wrote:
> Any chance you can reproduce this with debuginfo packages installed, so that we can get meaningf...
Sean Sullivan
01:54 PM Backport #24191 (In Progress): mimic: fs: reduce number of helper debug messages at level 5 for c...
Patrick Donnelly
01:53 PM Bug #24177: qa: fsstress workunit does not execute in parallel on same host without clobbering files
I suspect the problem is in unpacking and building ltp. The fsstress commands already use a pid-specific directory. H... Jeff Layton
01:52 PM Backport #24157 (In Progress): mimic: mds: crash when using `config set` on tracked configs
Patrick Donnelly
01:48 PM Backport #24209 (In Progress): mimic: client: deleted inode's Bufferhead which was in STATE::Tx w...
https://github.com/ceph/ceph/pull/22136 Patrick Donnelly
01:48 PM Backport #24187 (In Progress): mimic: mds didn't update file's max_size
https://github.com/ceph/ceph/pull/22137 Patrick Donnelly
01:46 PM Backport #24186 (In Progress): mimic: client: segfault in trim_caps
https://github.com/ceph/ceph/pull/22139 Patrick Donnelly
01:45 PM Backport #24202 (In Progress): mimic: client: fails to respond cap revoke from non-auth mds
Patrick Donnelly
01:44 PM Backport #24206 (In Progress): mimic: mds: broadcast quota to relevant clients when quota is expl...
Patrick Donnelly

05/21/2018

01:28 PM Backport #24049 (In Progress): luminous: ceph-fuse: missing dentries in readdir result
https://github.com/ceph/ceph/pull/22119 Prashant D
01:21 PM Backport #24050 (In Progress): luminous: mds: MClientCaps should carry inode's dirstat
https://github.com/ceph/ceph/pull/22118 Prashant D
08:49 AM Backport #24209 (Resolved): mimic: client: deleted inode's Bufferhead which was in STATE::Tx woul...
https://github.com/ceph/ceph/pull/22136 Nathan Cutler
08:49 AM Backport #24208 (Rejected): jewel: client: deleted inode's Bufferhead which was in STATE::Tx woul...
Nathan Cutler
08:49 AM Backport #24207 (Resolved): luminous: client: deleted inode's Bufferhead which was in STATE::Tx w...
https://github.com/ceph/ceph/pull/22168 Nathan Cutler
08:48 AM Backport #24206 (Resolved): mimic: mds: broadcast quota to relevant clients when quota is explici...
https://github.com/ceph/ceph/pull/22141 Nathan Cutler
08:48 AM Backport #24205 (Resolved): luminous: mds: broadcast quota to relevant clients when quota is expl...
https://github.com/ceph/ceph/pull/22271 Nathan Cutler
08:48 AM Backport #24202 (Resolved): mimic: client: fails to respond cap revoke from non-auth mds
https://github.com/ceph/ceph/pull/22140 Nathan Cutler
08:48 AM Backport #24201 (Resolved): luminous: client: fails to respond cap revoke from non-auth mds
https://github.com/ceph/ceph/pull/22221 Nathan Cutler

05/20/2018

11:56 PM Bug #24133 (Pending Backport): mds: broadcast quota to relevant clients when quota is explicitly set
Patrick Donnelly
11:55 PM Bug #23837 (Pending Backport): client: deleted inode's Bufferhead which was in STATE::Tx would le...
Patrick Donnelly
11:55 PM Bug #24172 (Pending Backport): client: fails to respond cap revoke from non-auth mds
Patrick Donnelly

05/19/2018

10:05 AM Backport #24191 (Resolved): mimic: fs: reduce number of helper debug messages at level 5 for client
https://github.com/ceph/ceph/pull/22154 Nathan Cutler
10:05 AM Backport #24190 (Resolved): luminous: fs: reduce number of helper debug messages at level 5 for c...
https://github.com/ceph/ceph/pull/23014 Nathan Cutler
10:04 AM Backport #24189 (Resolved): luminous: qa: kernel_mount.py umount must handle timeout arg
https://github.com/ceph/ceph/pull/22208 Nathan Cutler
10:04 AM Backport #24188 (Resolved): luminous: kceph: umount on evicted client blocks forever
https://github.com/ceph/ceph/pull/22208 Nathan Cutler
10:04 AM Backport #24187 (Resolved): mimic: mds didn't update file's max_size
https://github.com/ceph/ceph/pull/22137 Nathan Cutler
10:04 AM Backport #24186 (Resolved): mimic: client: segfault in trim_caps
https://github.com/ceph/ceph/pull/22139 Nathan Cutler
09:57 AM Bug #24054 (Pending Backport): kceph: umount on evicted client blocks forever
Zheng Yan
09:56 AM Bug #24053 (Pending Backport): qa: kernel_mount.py umount must handle timeout arg
Zheng Yan
04:45 AM Backport #24185 (Resolved): luminous: client: segfault in trim_caps
https://github.com/ceph/ceph/pull/22201 Patrick Donnelly
04:38 AM Bug #24137 (Pending Backport): client: segfault in trim_caps
Patrick Donnelly

05/18/2018

09:33 PM Bug #24111 (Pending Backport): mds didn't update file's max_size
Patrick Donnelly
09:32 PM Bug #21014 (Pending Backport): fs: reduce number of helper debug messages at level 5 for client
Patrick Donnelly
07:38 PM Bug #24177 (Resolved): qa: fsstress workunit does not execute in parallel on same host without cl...
... Patrick Donnelly
05:10 PM Bug #24172: client: fails to respond cap revoke from non-auth mds
adding mimic backport because the PR targets master Nathan Cutler
12:17 PM Bug #24172: client: fails to respond cap revoke from non-auth mds
this can also explain http://tracker.ceph.com/issues/23350 Zheng Yan
12:10 PM Bug #24172 (Fix Under Review): client: fails to respond cap revoke from non-auth mds
Zheng Yan
12:10 PM Bug #24172: client: fails to respond cap revoke from non-auth mds
https://github.com/ceph/ceph/pull/22080 Zheng Yan
12:06 PM Bug #24172 (Resolved): client: fails to respond cap revoke from non-auth mds
Zheng Yan
02:33 PM Bug #24173: ceph_volume_client: allow atomic update of RADOS objects

Greg Farnum's suggestions to do atomic RADOS object updates,
"If you've already got code that does all these t...
Ramana Raja
02:13 PM Bug #24173 (Resolved): ceph_volume_client: allow atomic update of RADOS objects
The manila driver needs the ceph_volume_client to atomically update contents
of RADOS objects used to store ganesha'...
Ramana Raja
02:09 PM Bug #23393 (Resolved): ceph-ansible: update Ganesha config for nfs_file_gw to use optimal settings
Ramana Raja
12:14 AM Bug #24137 (In Progress): client: segfault in trim_caps
https://github.com/ceph/ceph/pull/22073 Patrick Donnelly

05/17/2018

08:45 AM Backport #23946 (In Progress): luminous: mds: crash when failover
Nathan Cutler
08:44 AM Bug #24137: client: segfault in trim_caps
compile test_trim_caps.cc with the newest libcephfs. set mds_min_caps_per_client to 1, set mds_max_ratio_caps_per_cli... Zheng Yan
04:44 AM Bug #24137: client: segfault in trim_caps
Zheng Yan wrote:
> The problem is that anchor only pins current inode. Client::unlink() still may drop reference of ...
Patrick Donnelly
12:44 AM Bug #24137: client: segfault in trim_caps
The problem is that anchor only pins current inode. Client::unlink() still may drop reference of its parent inode. Zheng Yan
08:41 AM Backport #24157 (Resolved): mimic: mds: crash when using `config set` on tracked configs
https://github.com/ceph/ceph/pull/22153 Nathan Cutler
04:14 AM Documentation #24093 (Fix Under Review): doc: Update *remove a metadata server*
https://github.com/ceph/ceph/pull/22035 Jos Collin
12:55 AM Bug #24052: repeated eviction of idle client until some IO happens
https://github.com/ceph/ceph/pull/22026 Zheng Yan
12:52 AM Bug #24101: mds: deadlock during fsstress workunit with 9 actives
Ivan Guan wrote:
> Zheng Yan wrote:
> > caused by https://github.com/ceph/ceph/pull/21615
>
> Sorry,i don't unde...
Zheng Yan

05/16/2018

09:04 PM Bug #24118 (Pending Backport): mds: crash when using `config set` on tracked configs
Sage Weil
07:59 PM Bug #24138: qa: support picking a random distro using new teuthology $
@Warren - wonder if it easy doable to add `yaml` configuration so if suites ^ run on `rhel` then `-k testing` is used... Yuri Weinstein
05:43 PM Bug #24138: qa: support picking a random distro using new teuthology $
FYI
merged PRs related to this:
https://tracker.ceph.com/issues/24138
https://github.com/ceph/ceph/pull/21932
h...
Yuri Weinstein
05:34 PM Bug #24138: qa: support picking a random distro using new teuthology $

That's it I guess. Should also find a way to make `-k testing` the default unless distro == RHEL.
Patrick Donnelly
05:33 PM Bug #24138: qa: support picking a random distro using new teuthology $
@batrick I assume suites are: `fs`, `kcephfs`, `nutlimds` ? more? Yuri Weinstein
06:16 PM Bug #24137: client: segfault in trim_caps
Zheng Yan wrote:
> [...]
>
> I think above commit isn't quite right. how about patch below
>
> [...]
I'm no...
Patrick Donnelly
11:10 AM Bug #24137: client: segfault in trim_caps
... Zheng Yan
12:47 PM Bug #24101: mds: deadlock during fsstress workunit with 9 actives
Zheng Yan wrote:
> caused by https://github.com/ceph/ceph/pull/21615
Sorry,i don't understand why this pr can cau...
Ivan Guan
03:39 AM Bug #24052 (Fix Under Review): repeated eviction of idle client until some IO happens
Zheng Yan

05/15/2018

11:35 PM Bug #21014 (Fix Under Review): fs: reduce number of helper debug messages at level 5 for client
https://github.com/ceph/ceph/pull/21972 Patrick Donnelly
10:35 PM Backport #23991 (In Progress): luminous: client: hangs on umount if it had an MDS session evicted
Patrick Donnelly
09:37 PM Bug #24028: CephFS flock() on a directory is broken
Марк Коренберг wrote:
> Patrick Donnelly, why you set version to 14 ? Will this change be merged to Luminous ?
Be...
Patrick Donnelly
08:25 PM Bug #24028: CephFS flock() on a directory is broken
Patrick Donnelly, why you set version to 14 ? Will this change be merged to Luminous ? Марк Коренберг
08:23 PM Bug #24028: CephFS flock() on a directory is broken
https://github.com/ceph/ceph/blob/master/src/client/fuse_ll.cc#L1037 ? Марк Коренберг
07:48 PM Bug #24028: CephFS flock() on a directory is broken
Does ceph-fuse not have this problem? Patrick Donnelly
03:38 AM Bug #24028: CephFS flock() on a directory is broken
https://github.com/ceph/ceph-client/commit/ae2a8539ab7bb72f37306a544a555e9fc9ce8221 Zheng Yan
08:04 PM Bug #23837: client: deleted inode's Bufferhead which was in STATE::Tx would lead a assert fail
BZ: https://bugzilla.redhat.com/show_bug.cgi?id=1576908 Patrick Donnelly
08:01 PM Bug #23837: client: deleted inode's Bufferhead which was in STATE::Tx would lead a assert fail
Fixed formatting. Patrick Donnelly
11:37 AM Bug #23837 (Fix Under Review): client: deleted inode's Bufferhead which was in STATE::Tx would le...
https://github.com/ceph/ceph/pull/22001 Zheng Yan
08:04 PM Bug #24087 (Duplicate): client: assert during shutdown after blacklisted
Missed that. Thanks Zheng! Patrick Donnelly
09:58 AM Bug #24087: client: assert during shutdown after blacklisted
dup of http://tracker.ceph.com/issues/23837 Zheng Yan
07:55 PM Bug #24133 (Fix Under Review): mds: broadcast quota to relevant clients when quota is explicitly set
Patrick Donnelly
08:18 AM Bug #24133: mds: broadcast quota to relevant clients when quota is explicitly set
https://github.com/ceph/ceph/pull/21997 Zhi Zhang
08:13 AM Bug #24133 (Resolved): mds: broadcast quota to relevant clients when quota is explicitly set
We found client won't get quota updated for a long time under following case. We found this issue on Luminous, but it... Zhi Zhang
07:41 PM Bug #24138 (Resolved): qa: support picking a random distro using new teuthology $
Similar to https://github.com/ceph/ceph/pull/22008/files Patrick Donnelly
07:38 PM Bug #24137: client: segfault in trim_caps
Reasonable assumption about this crash is either the inode was deleted (in which case the Cap should have been delete... Patrick Donnelly
07:18 PM Bug #24137 (Resolved): client: segfault in trim_caps
... Patrick Donnelly
03:28 PM Backport #24136 (Resolved): luminous: MDSMonitor: uncommitted state exposed to clients/mdss
https://github.com/ceph/ceph/pull/23013 Nathan Cutler
01:45 PM Bug #23768 (Pending Backport): MDSMonitor: uncommitted state exposed to clients/mdss
Mimic PR: https://github.com/ceph/ceph/pull/22005 Patrick Donnelly
03:53 AM Bug #24074: Read ahead in fuse client is broken with large buffer size
try passing '--client_readahead_max_bytes=4194304' option to ceph-fuse Zheng Yan

05/14/2018

10:21 PM Bug #24129 (Fix Under Review): qa: test_version_splitting (tasks.cephfs.test_sessionmap.TestSessi...
https://github.com/ceph/ceph/pull/21992 Patrick Donnelly
08:13 PM Bug #24129 (Resolved): qa: test_version_splitting (tasks.cephfs.test_sessionmap.TestSessionMap) t...
... Patrick Donnelly
10:10 PM Bug #24074 (Need More Info): Read ahead in fuse client is broken with large buffer size
Chuan Qiu wrote:
> If the read is larger than 128K(e.g. 4M as our object size), fuse client will receive read reques...
Patrick Donnelly
08:54 PM Backport #23935 (In Progress): luminous: mds: may send LOCK_SYNC_MIX message to starting MDS
https://github.com/ceph/ceph/pull/21990 Patrick Donnelly
08:36 PM Backport #24130 (In Progress): luminous: mds: race with new session from connection and imported ...
https://github.com/ceph/ceph/pull/21989 Patrick Donnelly
08:33 PM Backport #24130 (Resolved): luminous: mds: race with new session from connection and imported ses...
Patrick Donnelly
08:32 PM Bug #24072 (Pending Backport): mds: race with new session from connection and imported session
Mimic PR: https://github.com/ceph/ceph/pull/21988 Patrick Donnelly
04:29 AM Bug #24072: mds: race with new session from connection and imported session
WIP: https://github.com/ceph/ceph/pull/21966 Patrick Donnelly
06:53 PM Documentation #24093: doc: Update *remove a metadata server*
It should be sufficient to say that the operator can just turn it the MDS off, however that is done for their environ... Patrick Donnelly
05:57 PM Bug #24118 (Fix Under Review): mds: crash when using `config set` on tracked configs
https://github.com/ceph/ceph/pull/21984 Sage Weil
05:46 PM Bug #24118 (Resolved): mds: crash when using `config set` on tracked configs
These configs: https://github.com/ceph/ceph/blob/7dbba9e54282e0a4c3000eb0c1a66e346c7eab98/src/mds/MDSDaemon.cc#L362-L... Patrick Donnelly
04:03 PM Bug #24052: repeated eviction of idle client until some IO happens
The log for that client at 128.142.160.86 is here: ceph-post-file: dd10811e-2790-43e4-b0a9-135725f70209
Thanks for...
Dan van der Ster
01:47 PM Bug #24052: repeated eviction of idle client until some IO happens
It's not expected. could you upload client log with debug_ms=1 Zheng Yan
02:41 PM Bug #23972: Ceph MDS Crash from client mounting aufs over cephfs
Any chance you can reproduce this with debuginfo packages installed, so that we can get meaningful backtraces? John Spray
01:57 PM Bug #24054 (Fix Under Review): kceph: umount on evicted client blocks forever
https://github.com/ceph/ceph/pull/21941 Patrick Donnelly
01:48 PM Bug #23837 (In Progress): client: deleted inode's Bufferhead which was in STATE::Tx would lead a ...
Kicking this back to In Progress. Please see comments in original PR. It has been reverted by https://github.com/ceph... Patrick Donnelly
01:45 PM Bug #24101: mds: deadlock during fsstress workunit with 9 actives
Revert: https://github.com/ceph/ceph/pull/21975 Patrick Donnelly
01:39 PM Bug #24101: mds: deadlock during fsstress workunit with 9 actives
Zheng Yan
11:53 AM Bug #24101: mds: deadlock during fsstress workunit with 9 actives
caused by https://github.com/ceph/ceph/pull/21615 Zheng Yan
01:29 PM Bug #24030 (Fix Under Review): ceph-fuse: double dash meaning
Patrick Donnelly
07:45 AM Bug #24053 (Fix Under Review): qa: kernel_mount.py umount must handle timeout arg
Zheng Yan
07:44 AM Bug #24053: qa: kernel_mount.py umount must handle timeout arg
https://github.com/ceph/ceph/pull/21941 Zheng Yan
04:05 AM Bug #24111 (Fix Under Review): mds didn't update file's max_size
https://github.com/ceph/ceph/pull/21963 Zheng Yan
03:37 AM Bug #24111 (Resolved): mds didn't update file's max_size
http://pulpito.ceph.com/pdonnell-2018-05-04_03:45:51-multimds-master-testing-basic-smithi/2474517/ Zheng Yan
03:59 AM Bug #24039 (Closed): MDSTableServer.cc: 62: FAILED assert(g_conf->mds_kill_mdstable_at != 1)
create new ticket for the fsstress hang http://tracker.ceph.com/issues/24111
close this one
Zheng Yan

05/13/2018

03:01 PM Backport #24108 (Resolved): luminous: MDCache.cc: 5317: FAILED assert(mds->is_rejoin())
https://github.com/ceph/ceph/pull/22171 Nathan Cutler
03:01 PM Backport #24107 (Resolved): luminous: PurgeQueue::_consume() could return true when there were no...
https://github.com/ceph/ceph/pull/22176 Nathan Cutler

05/12/2018

04:19 AM Bug #24072 (In Progress): mds: race with new session from connection and imported session
Patrick Donnelly

05/11/2018

10:14 PM Bug #23837 (Pending Backport): client: deleted inode's Bufferhead which was in STATE::Tx would le...
Mimic PR: https://github.com/ceph/ceph/pull/21954 Patrick Donnelly
10:09 PM Backport #23946: luminous: mds: crash when failover
Prashant, pr21769 is merged. Patrick Donnelly
10:06 PM Bug #24047 (Pending Backport): MDCache.cc: 5317: FAILED assert(mds->is_rejoin())
Mimic PR: https://github.com/ceph/ceph/pull/21952 Patrick Donnelly
10:01 PM Bug #24073 (Pending Backport): PurgeQueue::_consume() could return true when there were no purge ...
Mimic PR: https://github.com/ceph/ceph/pull/21951 Patrick Donnelly
03:50 AM Bug #24073: PurgeQueue::_consume() could return true when there were no purge queue item actually...
dongdong tao wrote:
> Yeah, that‘s what i want to recommend to you, it can work as you expected.
Thank you:-) Tha...
Xuehan Xu
03:46 AM Bug #24073: PurgeQueue::_consume() could return true when there were no purge queue item actually...
Yeah, that‘s what i want to recommend to you, it can work as you expected. dongdong tao
03:04 AM Bug #24073: PurgeQueue::_consume() could return true when there were no purge queue item actually...
dongdong tao wrote:
> Hi Xuehan,
> I'm just curious about that how do you repair your purge queue journal ?
By t...
Xuehan Xu
03:01 AM Bug #24073: PurgeQueue::_consume() could return true when there were no purge queue item actually...
dongdong tao wrote:
> Hi Xuehan,
> I'm just curious about that how do you repair your purge queue journal ?
Actu...
Xuehan Xu
02:34 AM Bug #24073: PurgeQueue::_consume() could return true when there were no purge queue item actually...
Hi Xuehan,
I'm just curious about that how do you repair your purge queue journal ?
dongdong tao
06:06 PM Bug #24101 (Closed): mds: deadlock during fsstress workunit with 9 actives
http://pulpito.ceph.com/pdonnell-2018-05-11_00:47:01-multimds-wip-pdonnell-testing-20180510.225359-testing-basic-smit... Patrick Donnelly
05:52 PM Feature #17230 (Fix Under Review): ceph_volume_client: py3 compatible
https://github.com/ceph/ceph/pull/21948 Rishabh Dave
04:26 AM Documentation #24093 (Resolved): doc: Update *remove a metadata server*
Update: http://docs.ceph.com/docs/master/rados/deployment/ceph-deploy-mds/#remove-a-metadata-server
See:
http://d...
Jos Collin

05/10/2018

10:09 PM Bug #24090 (Resolved): mds: fragmentation in QA is slowing down ops enough for WRNs
http://pulpito.ceph.com/pdonnell-2018-05-08_18:15:09-fs-mimic-testing-basic-smithi/
http://pulpito.ceph.com/pdonnell...
Patrick Donnelly
08:48 PM Bug #24089 (Rejected): mds: print slow requests to debug log when sending health WRN to monitors ...
Nevermind, it is actually printed earlier in the log. Sorry for the noise. Patrick Donnelly
08:46 PM Bug #24089 (Rejected): mds: print slow requests to debug log when sending health WRN to monitors ...
... Patrick Donnelly
08:22 PM Bug #24088 (Duplicate): mon: slow remove_snaps op reported in cluster health log
... Patrick Donnelly
04:57 PM Bug #24087 (Duplicate): client: assert during shutdown after blacklisted
... Patrick Donnelly
11:48 AM Bug #24039: MDSTableServer.cc: 62: FAILED assert(g_conf->mds_kill_mdstable_at != 1)
the pjd: http://pulpito.ceph.com/pdonnell-2018-05-04_03:45:51-multimds-master-testing-basic-smithi/2475062/... Zheng Yan
11:25 AM Feature #22446: mds: ask idle client to trim more caps
Glad to see this :)
- Backport set to mimic,luminous
Thanks.
Webert Lima
09:44 AM Bug #23332: kclient: with fstab entry is not coming up reboot
I still don't think this is kernel issue. please path kernel with below change and try again.... Zheng Yan
06:44 AM Bug #24073: PurgeQueue::_consume() could return true when there were no purge queue item actually...
Xuehan Xu wrote:
> In our online clusters, we encountered the bug #19593. Although we cherry-pick the fixing commits...
Xuehan Xu
04:38 AM Bug #24073: PurgeQueue::_consume() could return true when there were no purge queue item actually...
https://github.com/ceph/ceph/pull/21923 Xuehan Xu
04:38 AM Bug #24073 (Resolved): PurgeQueue::_consume() could return true when there were no purge queue it...
In our online clusters, we encountered the bug #19593. Although we cherry-pick the fixing commits, the purge queue's ... Xuehan Xu
06:39 AM Bug #24074 (Need More Info): Read ahead in fuse client is broken with large buffer size
If the read is larger than 128K(e.g. 4M as our object size), fuse client will receive read requests as multiple ll_re... Chuan Qiu
04:10 AM Backport #23984 (In Progress): luminous: mds: scrub on fresh file system fails
https://github.com/ceph/ceph/pull/21922 Prashant D
04:07 AM Backport #23982 (In Progress): luminous: qa: TestVolumeClient.test_lifecycle needs updated for ne...
https://github.com/ceph/ceph/pull/21921 Prashant D
04:05 AM Bug #23826: mds: assert after daemon restart
Finish context of MDCache::open_undef_inodes_dirfrags() calls rejoin_gather_finish() without check rejoin_gather. I t... Zheng Yan

05/09/2018

09:29 PM Bug #24072 (Resolved): mds: race with new session from connection and imported session
... Patrick Donnelly
09:14 PM Feature #22446 (New): mds: ask idle client to trim more caps
Patrick Donnelly
09:11 PM Documentation #23611: doc: add description of new fs-client auth profile
Blocked by resolution to #23751. Patrick Donnelly
09:08 PM Feature #22370 (In Progress): cephfs: add kernel client quota support
Patrick Donnelly
09:08 PM Feature #22372: kclient: implement quota handling using new QuotaRealm
Zheng, what's the status on those patches? Patrick Donnelly
09:05 PM Bug #23332 (Need More Info): kclient: with fstab entry is not coming up reboot
Zheng Yan wrote:
> kexec in dmesgs looks suspicious. client mounted cephfs, then used kexec to load kernel image aga...
Patrick Donnelly
09:02 PM Bug #23350: mds: deadlock during unlink and export
Well this is aggravating. I think it's time we plan evictions for clients that do not respond to cap release. Patrick Donnelly
08:56 PM Bug #23394 (Rejected): nfs-ganesha: check cache configuration when exporting FSAL_CEPH
Patrick Donnelly
08:52 PM Feature #14456 (Fix Under Review): mon: prevent older/incompatible clients from mounting the file...
https://github.com/ceph/ceph/pull/21885 Patrick Donnelly
07:06 PM Bug #23855: mds: MClientCaps should carry inode's dirstat
Testing; will revert Nathan Cutler
06:57 PM Bug #23291 (Resolved): client: add way to sync setattr operations to MDS
Nathan Cutler
06:57 PM Backport #23474 (Resolved): luminous: client: allow caller to request that setattr request be syn...
Nathan Cutler
02:55 PM Backport #23474: luminous: client: allow caller to request that setattr request be synchronous
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21109
merged
Yuri Weinstein
06:56 PM Bug #23602 (Resolved): mds: handle client requests when mds is stopping
Nathan Cutler
06:56 PM Backport #23632 (Resolved): luminous: mds: handle client requests when mds is stopping
Nathan Cutler
02:54 PM Backport #23632: luminous: mds: handle client requests when mds is stopping
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21346
merged
Yuri Weinstein
06:56 PM Bug #23541 (Resolved): client: fix request send_to_auth was never really used
Nathan Cutler
06:56 PM Backport #23635 (Resolved): luminous: client: fix request send_to_auth was never really used
Nathan Cutler
02:54 PM Backport #23635: luminous: client: fix request send_to_auth was never really used
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21354
merged
Yuri Weinstein
06:13 PM Bug #24040 (Need More Info): mds: assert in CDir::_committed
Patrick Donnelly
02:14 PM Bug #24040: mds: assert in CDir::_committed
Thanks for the report - it looks like you're using a 11.x ("kraken") version, which is no longer receiving bug fixes.... John Spray
01:56 PM Bug #24039: MDSTableServer.cc: 62: FAILED assert(g_conf->mds_kill_mdstable_at != 1)
the fsstress failure: http://pulpito.ceph.com/pdonnell-2018-05-04_03:45:51-multimds-master-testing-basic-smithi/24745... Zheng Yan
01:12 PM Bug #24039: MDSTableServer.cc: 62: FAILED assert(g_conf->mds_kill_mdstable_at != 1)
fsstress failure looks like new bug
pjd failure is similar to http://tracker.ceph.com/issues/23327
two dead tasks...
Zheng Yan
11:55 AM Bug #23327: qa: pjd test sees wrong ctime after unlink
http://pulpito.ceph.com/pdonnell-2018-05-04_03:45:51-multimds-master-testing-basic-smithi/2475062/ Zheng Yan
06:08 AM Backport #23951 (In Progress): luminous: mds: stuck during up:stopping
https://github.com/ceph/ceph/pull/21901 Prashant D
03:36 AM Backport #23946: luminous: mds: crash when failover
Opened backport PR#21900 (https://github.com/ceph/ceph/pull/21900). We need to cherry pick PR#21769 once it gets merg... Prashant D
03:29 AM Backport #23950 (In Progress): luminous: mds: stopping rank 0 cannot shutdown until log is trimmed
https://github.com/ceph/ceph/pull/21899 Prashant D

05/08/2018

10:49 PM Backport #24055 (In Progress): luminous: VolumeClient: allow ceph_volume_client to create 'volume...
Patrick Donnelly
10:45 PM Backport #24055 (Resolved): luminous: VolumeClient: allow ceph_volume_client to create 'volumes' ...
https://github.com/ceph/ceph/pull/21897 Patrick Donnelly
10:43 PM Feature #23695 (Pending Backport): VolumeClient: allow ceph_volume_client to create 'volumes' wit...
Mimic PR: https://github.com/ceph/ceph/pull/21896 Patrick Donnelly
10:37 PM Bug #24054 (Resolved): kceph: umount on evicted client blocks forever
Failed test:
/ceph/teuthology-archive/pdonnell-2018-05-08_01:06:46-kcephfs-mimic-testing-basic-smithi/2494030/teut...
Patrick Donnelly
10:33 PM Bug #24053 (Resolved): qa: kernel_mount.py umount must handle timeout arg
... Patrick Donnelly
10:27 PM Bug #24052 (Resolved): repeated eviction of idle client until some IO happens
We see repeated eviction of idle client sessions. We have client_reconnect_stale on the ceph-fuse clients, and these ... Dan van der Ster
08:56 PM Backport #24050 (Resolved): luminous: mds: MClientCaps should carry inode's dirstat
https://github.com/ceph/ceph/pull/22118 Nathan Cutler
08:56 PM Backport #24049 (Resolved): luminous: ceph-fuse: missing dentries in readdir result
https://github.com/ceph/ceph/pull/22119 Nathan Cutler
08:49 PM Bug #23530 (Resolved): mds: kicked out by monitor during rejoin
Nathan Cutler
08:49 PM Backport #23636 (Resolved): luminous: mds: kicked out by monitor during rejoin
Nathan Cutler
07:47 PM Backport #23636: luminous: mds: kicked out by monitor during rejoin
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21366
merged
Yuri Weinstein
08:49 PM Bug #23452 (Resolved): mds: assertion in MDSRank::validate_sessions
Nathan Cutler
08:48 PM Backport #23637 (Resolved): luminous: mds: assertion in MDSRank::validate_sessions
Nathan Cutler
07:46 PM Backport #23637: luminous: mds: assertion in MDSRank::validate_sessions
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21372
merged
Yuri Weinstein
08:48 PM Bug #23625 (Resolved): mds: sessions opened by journal replay do not get dirtied properly
Nathan Cutler
08:48 PM Backport #23702 (Resolved): luminous: mds: sessions opened by journal replay do not get dirtied p...
Nathan Cutler
07:46 PM Backport #23702: luminous: mds: sessions opened by journal replay do not get dirtied properly
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21441
merged
Yuri Weinstein
08:47 PM Bug #23582 (Resolved): MDSMonitor: mds health warnings printed in bad format
Nathan Cutler
08:47 PM Backport #23703 (Resolved): luminous: MDSMonitor: mds health warnings printed in bad format
Nathan Cutler
07:46 PM Backport #23703: luminous: MDSMonitor: mds health warnings printed in bad format
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21447
merged
Yuri Weinstein
08:47 PM Bug #23380 (Resolved): mds: ceph.dir.rctime follows dir ctime not inode ctime
Nathan Cutler
08:47 PM Backport #23750 (Resolved): luminous: mds: ceph.dir.rctime follows dir ctime not inode ctime
Nathan Cutler
07:45 PM Backport #23750: luminous: mds: ceph.dir.rctime follows dir ctime not inode ctime
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21448
merged
Yuri Weinstein
08:46 PM Bug #23764 (Resolved): MDSMonitor: new file systems are not initialized with the pending_fsmap epoch
Nathan Cutler
08:46 PM Backport #23791 (Resolved): luminous: MDSMonitor: new file systems are not initialized with the p...
Nathan Cutler
07:44 PM Backport #23791: luminous: MDSMonitor: new file systems are not initialized with the pending_fsma...
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21512
merged
Yuri Weinstein
08:46 PM Bug #23714 (Resolved): slow ceph_ll_sync_inode calls after setattr
Nathan Cutler
08:45 PM Backport #23802 (Resolved): luminous: slow ceph_ll_sync_inode calls after setattr
Nathan Cutler
07:44 PM Backport #23802: luminous: slow ceph_ll_sync_inode calls after setattr
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21542
merged
Yuri Weinstein
08:45 PM Bug #23652 (Resolved): client: fix gid_count check in UserPerm->deep_copy_from()
Nathan Cutler
08:44 PM Backport #23771 (Resolved): luminous: client: fix gid_count check in UserPerm->deep_copy_from()
Nathan Cutler
07:43 PM Backport #23771: luminous: client: fix gid_count check in UserPerm->deep_copy_from()
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21596
merged
Yuri Weinstein
08:44 PM Bug #23762 (Resolved): MDSMonitor: MDSMonitor::encode_pending modifies fsmap not pending_fsmap
Nathan Cutler
08:43 PM Backport #23792 (Resolved): luminous: MDSMonitor: MDSMonitor::encode_pending modifies fsmap not p...
Nathan Cutler
07:43 PM Backport #23792: luminous: MDSMonitor: MDSMonitor::encode_pending modifies fsmap not pending_fsmap
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21732
merged
Yuri Weinstein
08:43 PM Bug #23873 (Resolved): cephfs does not count st_nlink for directories correctly?
Nathan Cutler
08:43 PM Bug #19706: Laggy mon daemons causing MDS failover (symptom: failed to set counters on mds daemon...
I don't have reason to believe use of utime_t caused this issue but it's possible this could fix it: https://github.c... Patrick Donnelly
08:43 PM Backport #23987 (Resolved): luminous: cephfs does not count st_nlink for directories correctly?
Nathan Cutler
07:42 PM Backport #23987: luminous: cephfs does not count st_nlink for directories correctly?
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21796
merged
Yuri Weinstein
08:42 PM Bug #23880 (Resolved): mds: scrub code stuck at trimming log segments
Nathan Cutler
08:42 PM Backport #23930 (Resolved): luminous: mds: scrub code stuck at trimming log segments
Nathan Cutler
07:41 PM Backport #23930: luminous: mds: scrub code stuck at trimming log segments
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21840
merged
Yuri Weinstein
08:41 PM Bug #23813 (Resolved): client: "remove_session_caps still has dirty|flushing caps" when thrashing...
Nathan Cutler
08:41 PM Backport #23934 (Resolved): luminous: client: "remove_session_caps still has dirty|flushing caps"...
Nathan Cutler
07:41 PM Backport #23934: luminous: client: "remove_session_caps still has dirty|flushing caps" when thras...
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21844
merged
Yuri Weinstein
06:25 PM Bug #21777 (New): src/mds/MDCache.cc: 4332: FAILED assert(mds->is_rejoin())
Patrick Donnelly
01:36 PM Bug #21777 (Fix Under Review): src/mds/MDCache.cc: 4332: FAILED assert(mds->is_rejoin())
Zheng Yan
12:40 PM Bug #21777: src/mds/MDCache.cc: 4332: FAILED assert(mds->is_rejoin())
-https://github.com/ceph/ceph/pull/21883- Zheng Yan
03:57 AM Bug #21777 (In Progress): src/mds/MDCache.cc: 4332: FAILED assert(mds->is_rejoin())
Zheng Yan
06:23 PM Bug #24047 (Fix Under Review): MDCache.cc: 5317: FAILED assert(mds->is_rejoin())
https://github.com/ceph/ceph/pull/21883 Patrick Donnelly
06:23 PM Bug #24047 (Resolved): MDCache.cc: 5317: FAILED assert(mds->is_rejoin())
... Patrick Donnelly
04:58 PM Bug #24030: ceph-fuse: double dash meaning
https://github.com/ceph/ceph/pull/21889 Jos Collin
04:07 PM Bug #23885 (Resolved): MDSMonitor: overzealous MDS_ALL_DOWN and MDS_UP_LESS_THAN_MAX health warni...
Mimic PR: https://github.com/ceph/ceph/pull/21888 Patrick Donnelly
02:39 PM Bug #24039: MDSTableServer.cc: 62: FAILED assert(g_conf->mds_kill_mdstable_at != 1)
Right, there's something else wrong with the test. Patrick Donnelly
01:35 PM Bug #24039: MDSTableServer.cc: 62: FAILED assert(g_conf->mds_kill_mdstable_at != 1)
These are intentional crashes in table transaction test Zheng Yan
07:10 AM Backport #23936 (In Progress): luminous: cephfs-journal-tool: segfault during journal reset
https://github.com/ceph/ceph/pull/21874 Prashant D

05/07/2018

11:40 PM Bug #24040 (Need More Info): mds: assert in CDir::_committed
... zs 吴
10:56 PM Bug #23894 (Pending Backport): ceph-fuse: missing dentries in readdir result
Mimic: https://github.com/ceph/ceph/pull/21867 Patrick Donnelly
10:47 PM Bug #23855 (Pending Backport): mds: MClientCaps should carry inode's dirstat
Mimic PR: https://github.com/ceph/ceph/pull/21866 Patrick Donnelly
10:40 PM Bug #21777: src/mds/MDCache.cc: 4332: FAILED assert(mds->is_rejoin())
<deleted/> Patrick Donnelly
08:54 PM Bug #21777 (New): src/mds/MDCache.cc: 4332: FAILED assert(mds->is_rejoin())
Deleted: see #24047. Patrick Donnelly
10:00 PM Bug #24039 (Closed): MDSTableServer.cc: 62: FAILED assert(g_conf->mds_kill_mdstable_at != 1)
... Patrick Donnelly
02:41 PM Bug #24002 (Resolved): qa: check snap upgrade on multimds cluster
Patrick Donnelly
01:39 PM Bug #24030: ceph-fuse: double dash meaning
Jos, please take a crack at fixing this. Thanks! Patrick Donnelly
04:38 AM Bug #24030 (Closed): ceph-fuse: double dash meaning
... Jos Collin
01:37 PM Bug #23994 (Need More Info): mds: OSD space is not reclaimed until MDS is restarted
Patrick Donnelly
02:47 AM Bug #23994: mds: OSD space is not reclaimed until MDS is restarted
please try again and dump mds' cache (ceph daemon mds.xxx dump cache /tmp/cachedump.x) Zheng Yan
05:41 AM Backport #23934 (In Progress): luminous: client: "remove_session_caps still has dirty|flushing ca...
https://github.com/ceph/ceph/pull/21844 Prashant D
04:36 AM Bug #23768 (Fix Under Review): MDSMonitor: uncommitted state exposed to clients/mdss
https://github.com/ceph/ceph/pull/21842 Patrick Donnelly
04:02 AM Backport #23931 (In Progress): luminous: qa: test_purge_queue_op_rate: self.assertTrue(phase2_ops...
https://github.com/ceph/ceph/pull/21841 Prashant D
03:59 AM Backport #23930 (In Progress): luminous: mds: scrub code stuck at trimming log segments
https://github.com/ceph/ceph/pull/21840 Prashant D

05/06/2018

08:44 PM Bug #24028: CephFS flock() on a directory is broken
I tested flock() logic on different hosts.
on one host:
flock my_dir sleep 1000
on second:
flock my_dir e...
Марк Коренберг
08:42 PM Bug #24028 (Resolved): CephFS flock() on a directory is broken
Accroding to man page, flock() semantics must work also on directory. Actually it works with, say, Ext4. It does not ... Марк Коренберг

05/04/2018

02:12 PM Bug #23994: mds: OSD space is not reclaimed until MDS is restarted
This was on the kernel client. I tried Ubuntu's 4.13.0-39-generic and 4.15.0-15-generic kernels.
With the fuse cli...
Niklas Hambuechen
01:44 PM Bug #23994: mds: OSD space is not reclaimed until MDS is restarted
What client (kernel or fuse), and what version of the client? John Spray
05:09 AM Bug #23885 (Fix Under Review): MDSMonitor: overzealous MDS_ALL_DOWN and MDS_UP_LESS_THAN_MAX heal...
https://github.com/ceph/ceph/pull/21810 Patrick Donnelly
02:07 AM Bug #24002 (Pending Backport): qa: check snap upgrade on multimds cluster
Patrick Donnelly

05/03/2018

10:33 PM Feature #23695: VolumeClient: allow ceph_volume_client to create 'volumes' without namespace isol...
https://github.com/ceph/ceph/pull/21808 Ramana Raja
09:27 PM Bug #24004 (Resolved): mds: curate priority of perf counters sent to mgr
Make sure we have the most interesting statisitcs available for prometheus for dashboard use. Additionally, see if we... Patrick Donnelly
08:36 PM Bug #24002 (Fix Under Review): qa: check snap upgrade on multimds cluster
https://github.com/ceph/ceph/pull/21805 Patrick Donnelly
08:35 PM Bug #24002 (Resolved): qa: check snap upgrade on multimds cluster
To get an idea how the snap format upgrade works on a previously multimds cluster. (No need to exercise the two MDS s... Patrick Donnelly
07:48 PM Cleanup #24001 (Resolved): MDSMonitor: remove vestiges of `mds deactivate`
For Nautilus. Patrick Donnelly
06:02 PM Backport #23946: luminous: mds: crash when failover
Will also need: https://github.com/ceph/ceph/pull/21769 Patrick Donnelly
05:33 PM Feature #23623 (Resolved): mds: mark allow_snaps true by default
Patrick Donnelly
05:33 PM Documentation #23583 (Resolved): doc: update snapshot doc to account for recent changes
Patrick Donnelly
01:41 PM Backport #23987 (In Progress): luminous: cephfs does not count st_nlink for directories correctly?
Patrick Donnelly
10:28 AM Backport #23987 (Resolved): luminous: cephfs does not count st_nlink for directories correctly?
https://github.com/ceph/ceph/pull/21796 Nathan Cutler
01:27 PM Bug #23393 (Fix Under Review): ceph-ansible: update Ganesha config for nfs_file_gw to use optimal...
Ramana Raja
01:26 PM Bug #23393: ceph-ansible: update Ganesha config for nfs_file_gw to use optimal settings
https://github.com/ceph/ceph-ansible/pull/2556 Ramana Raja
01:02 PM Bug #23994 (Need More Info): mds: OSD space is not reclaimed until MDS is restarted
With my Luminous test cluster on Ubuntu I ran into a situation where I filled up an OSD by putting files on CephFS, a... Niklas Hambuechen
10:29 AM Backport #23991 (Resolved): luminous: client: hangs on umount if it had an MDS session evicted
https://github.com/ceph/ceph/pull/22018 Nathan Cutler
10:29 AM Backport #23990 (Rejected): jewel: client: hangs on umount if it had an MDS session evicted
Nathan Cutler
10:28 AM Backport #23989 (Resolved): luminous: mds: don't report slow request for blocked filelock request
https://github.com/ceph/ceph/pull/22782
follow-on fix: https://github.com/ceph/ceph/pull/26048 went into 12.2.11
Nathan Cutler
10:27 AM Backport #23984 (Resolved): luminous: mds: scrub on fresh file system fails
https://github.com/ceph/ceph/pull/21922 Nathan Cutler
10:27 AM Backport #23982 (Resolved): luminous: qa: TestVolumeClient.test_lifecycle needs updated for new e...
https://github.com/ceph/ceph/pull/21921 Nathan Cutler
12:00 AM Bug #23958: mds: scrub doesn't always return JSON results
Zheng Yan wrote:
> recursive scrub is async, it does not return anything
Good point, thanks. Even so, we shoudl r...
Patrick Donnelly

05/02/2018

11:56 PM Bug #16842 (Can't reproduce): mds: replacement MDS crashes on InoTable release
Patrick Donnelly
10:57 PM Bug #23975 (Pending Backport): qa: TestVolumeClient.test_lifecycle needs updated for new eviction...
Patrick Donnelly
07:53 PM Bug #23975 (Fix Under Review): qa: TestVolumeClient.test_lifecycle needs updated for new eviction...
https://github.com/ceph/ceph/pull/21789 Patrick Donnelly
06:59 PM Bug #23975 (Resolved): qa: TestVolumeClient.test_lifecycle needs updated for new eviction behavior
... Patrick Donnelly
08:50 PM Bug #23768 (New): MDSMonitor: uncommitted state exposed to clients/mdss
Moving this back to fs. This is a different bug Josh. Patrick Donnelly
08:44 PM Bug #23768 (Resolved): MDSMonitor: uncommitted state exposed to clients/mdss
backport is tracked in the fs bug Josh Durgin
06:06 PM Bug #23972 (New): Ceph MDS Crash from client mounting aufs over cephfs

Here is a rough outline of my topology
https://pastebin.com/HQqbMxyj
---
I can reliably crash all (in my case...
Sean Sullivan
05:02 PM Feature #17230 (In Progress): ceph_volume_client: py3 compatible
Patrick Donnelly
04:08 PM Bug #10915 (Pending Backport): client: hangs on umount if it had an MDS session evicted
Patrick Donnelly
02:21 PM Bug #23960 (Pending Backport): mds: scrub on fresh file system fails
Patrick Donnelly
02:20 PM Bug #23873 (Pending Backport): cephfs does not count st_nlink for directories correctly?
Patrick Donnelly
02:20 PM Bug #22428 (Pending Backport): mds: don't report slow request for blocked filelock request
Patrick Donnelly
03:10 AM Bug #23958: mds: scrub doesn't always return JSON results
recursive scrub is async, it does not return anything Zheng Yan

05/01/2018

10:42 PM Bug #23960 (In Progress): mds: scrub on fresh file system fails
https://github.com/ceph/ceph/pull/21762 Patrick Donnelly
10:21 PM Bug #23960 (Resolved): mds: scrub on fresh file system fails
In a fresh vstart cluster:... Patrick Donnelly
04:01 PM Bug #23958 (Resolved): mds: scrub doesn't always return JSON results
On a vstart cluster:... Patrick Donnelly
06:52 AM Backport #23951 (Resolved): luminous: mds: stuck during up:stopping
https://github.com/ceph/ceph/pull/21901 Nathan Cutler
06:52 AM Backport #23950 (Resolved): luminous: mds: stopping rank 0 cannot shutdown until log is trimmed
https://github.com/ceph/ceph/pull/21899 Nathan Cutler
06:29 AM Bug #23826: mds: assert after daemon restart
checking MDSMap::is_rejoining() is not required here. If there are recovering mds which haven't entered rejoin state.... Zheng Yan
12:29 AM Bug #23923 (Pending Backport): mds: stopping rank 0 cannot shutdown until log is trimmed
Patrick Donnelly
12:29 AM Bug #23919 (Pending Backport): mds: stuck during up:stopping
Patrick Donnelly

04/30/2018

09:05 PM Bug #23448 (Resolved): nfs-ganesha: fails to parse rados URLs with '.' in object name
Yes. Jeff Layton
08:51 PM Bug #23448: nfs-ganesha: fails to parse rados URLs with '.' in object name
Is this resolved? Patrick Donnelly
08:00 PM Backport #23946 (Resolved): luminous: mds: crash when failover
https://github.com/ceph/ceph/pull/21900 Nathan Cutler
07:21 PM Bug #23826: mds: assert after daemon restart
Here's one possible way this could happen I think:
1. All MDS are rejoin or later.
2. A up:rejoin MDS does:
3....
Patrick Donnelly
07:00 PM Bug #23826: mds: assert after daemon restart
Adding log from failed MDS.
Looks like it's receiving handle_cache_rejoin_ack message while in replay.
Patrick Donnelly
06:53 PM Bug #23518 (Pending Backport): mds: crash when failover
Patrick Donnelly
01:43 PM Bug #23883: kclient: CephFS kernel client hang
v4.9 is quite old at this point, so it would be helpful to know if this is something that has already been fixed in m... Jeff Layton
06:54 AM Backport #23932 (In Progress): jewel: client: avoid second lock on client_lock
Jos Collin
04:38 AM Backport #23792 (In Progress): luminous: MDSMonitor: MDSMonitor::encode_pending modifies fsmap no...
https://github.com/ceph/ceph/pull/21732 Patrick Donnelly
03:59 AM Backport #23933 (In Progress): luminous: client: avoid second lock on client_lock
Jos Collin

04/29/2018

08:31 PM Backport #23936 (Resolved): luminous: cephfs-journal-tool: segfault during journal reset
https://github.com/ceph/ceph/pull/21874 Nathan Cutler
08:30 PM Backport #23935 (Resolved): luminous: mds: may send LOCK_SYNC_MIX message to starting MDS
https://github.com/ceph/ceph/pull/21990 Nathan Cutler
08:30 PM Backport #23934 (Resolved): luminous: client: "remove_session_caps still has dirty|flushing caps"...
https://github.com/ceph/ceph/pull/21844 Nathan Cutler
08:30 PM Backport #23933 (Resolved): luminous: client: avoid second lock on client_lock
https://github.com/ceph/ceph/pull/21730 Nathan Cutler
08:30 PM Backport #23932 (Resolved): jewel: client: avoid second lock on client_lock
https://github.com/ceph/ceph/pull/21734 Nathan Cutler
08:30 PM Backport #23931 (Resolved): luminous: qa: test_purge_queue_op_rate: self.assertTrue(phase2_ops < ...
https://github.com/ceph/ceph/pull/21841 Nathan Cutler
08:30 PM Backport #23930 (Resolved): luminous: mds: scrub code stuck at trimming log segments
https://github.com/ceph/ceph/pull/21840 Nathan Cutler
08:07 PM Bug #23815 (Pending Backport): client: avoid second lock on client_lock
Patrick Donnelly
08:06 PM Bug #23813 (Pending Backport): client: "remove_session_caps still has dirty|flushing caps" when t...
Patrick Donnelly
08:06 PM Bug #23812 (Pending Backport): mds: may send LOCK_SYNC_MIX message to starting MDS
Patrick Donnelly
08:06 PM Bug #20549 (Pending Backport): cephfs-journal-tool: segfault during journal reset
Patrick Donnelly
08:05 PM Bug #23829 (Pending Backport): qa: test_purge_queue_op_rate: self.assertTrue(phase2_ops < phase1_...
Patrick Donnelly
08:05 PM Bug #23880 (Pending Backport): mds: scrub code stuck at trimming log segments
Patrick Donnelly
01:44 AM Bug #23919 (Fix Under Review): mds: stuck during up:stopping
Zheng Yan wrote:
> I think we should call Locker::_readlock_kick in this case.
https://github.com/ceph/ceph/pull/...
Patrick Donnelly
01:15 AM Bug #23927 (Rejected): qa: test_full failure in test_barrier
https://github.com/ceph/ceph/pull/21668#pullrequestreview-116152567 Patrick Donnelly
12:54 AM Bug #23927: qa: test_full failure in test_barrier
Here too: http://pulpito.ceph.com/pdonnell-2018-04-28_06:20:24-fs-wip-pdonnell-testing-20180428.041811-testing-basic-... Patrick Donnelly
12:49 AM Bug #23927 (Rejected): qa: test_full failure in test_barrier
... Patrick Donnelly
12:36 AM Bug #23923 (Fix Under Review): mds: stopping rank 0 cannot shutdown until log is trimmed
https://github.com/ceph/ceph/pull/21719 Patrick Donnelly

04/28/2018

06:59 PM Bug #23923 (Resolved): mds: stopping rank 0 cannot shutdown until log is trimmed
... Patrick Donnelly
03:53 PM Bug #23883: kclient: CephFS kernel client hang
Hi Wei,
this is a very interesting problem, from your description, i would like to share my thought:
this shoul...
dongdong tao
10:10 AM Bug #23883: kclient: CephFS kernel client hang
client kernel dmesg:... wei jin
10:09 AM Bug #23883: kclient: CephFS kernel client hang
... wei jin
08:02 AM Bug #23883: kclient: CephFS kernel client hang
debug_mds = 10. only for period that mds is recovering Zheng Yan
07:53 AM Bug #23883: kclient: CephFS kernel client hang
Zheng Yan wrote:
> please upload mds log
which level?
after setting debug_mds = 20 and debug_ms = 1, log file is...
wei jin
05:03 AM Bug #23883: kclient: CephFS kernel client hang
please upload mds log Zheng Yan
10:34 AM Bug #22428 (Fix Under Review): mds: don't report slow request for blocked filelock request
https://github.com/ceph/ceph/pull/21715 Zheng Yan
07:50 AM Bug #23919: mds: stuck during up:stopping
I think we should call Locker::_readlock_kick in this case. Zheng Yan
04:02 AM Bug #23919: mds: stuck during up:stopping
/ceph/tmp/pdonnell/bz1566016/0x20000205a64.log.gz
holds the output of
zgrep -C5 0x20000205a64 ceph-mds.magna05...
Patrick Donnelly
03:52 AM Bug #23919: mds: stuck during up:stopping
crux of the issue appears to be here:... Patrick Donnelly
06:34 AM Bug #23920: Multiple ceph-fuse and one ceph-client.admin.log
I am using the method you said to modify,After that I found three questions:
1、when I run ceph-fuse ,There will be...
yuanli zhu
06:04 AM Bug #23920: Multiple ceph-fuse and one ceph-client.admin.log
because I have two ceph-fuse,how can i set config use the command as below for each ceph-fuse:
ceph daemon clien...
yuanli zhu
05:00 AM Bug #23920 (Rejected): Multiple ceph-fuse and one ceph-client.admin.log
config issue. you should set log file config option like
log file = /var/log/ceph/ceph-client.$pid.log
Zheng Yan
02:28 AM Bug #23920 (Rejected): Multiple ceph-fuse and one ceph-client.admin.log
I use the command as below:
/usr/bin/ceph-fuse -c /etc/ceph/ceph.conf /nas/test1 -r /test1
/usr/bin/ceph-fu...
yuanli zhu
04:54 AM Bug #23894 (Fix Under Review): ceph-fuse: missing dentries in readdir result
https://github.com/ceph/ceph/pull/21712 Zheng Yan
01:37 AM Bug #23894: ceph-fuse: missing dentries in readdir result
libcephfs does not handle session stale message properly
steps to reproduce:
1. create two ceph-fuse mounts, mo...
Zheng Yan

04/27/2018

10:27 PM Bug #23919 (Resolved): mds: stuck during up:stopping
... Patrick Donnelly
10:16 PM Backport #23833: luminous: MDSMonitor: crash after assigning standby-replay daemon in multifs setup
I think they are separate issues but I will take a look. Patrick Donnelly
07:42 PM Backport #23833: luminous: MDSMonitor: crash after assigning standby-replay daemon in multifs setup
@Patrick - this one looks like it could benefit from being done in a single PR along with http://tracker.ceph.com/iss... Nathan Cutler
05:34 PM Backport #23833: luminous: MDSMonitor: crash after assigning standby-replay daemon in multifs setup
Travis Nielsen wrote:
> What is the timeline for the backport? Rook would like to see it in 12.2.6. Thanks!
It sh...
Patrick Donnelly
05:16 PM Backport #23833: luminous: MDSMonitor: crash after assigning standby-replay daemon in multifs setup
What is the timeline for the backport? Rook would like to see it in 12.2.6. Thanks! Travis Nielsen
07:40 PM Backport #23792: luminous: MDSMonitor: MDSMonitor::encode_pending modifies fsmap not pending_fsmap
@Patrick could you take this one? Nathan Cutler
05:05 PM Bug #23658: MDSMonitor: crash after assigning standby-replay daemon in multifs setup
When this issue hits, is there a way to recover? For example, to forcefully remove the multiple filesystems that are ... Travis Nielsen
11:58 AM Bug #23873: cephfs does not count st_nlink for directories correctly?
Peter Mauritius wrote:
> The Dovecot mail server does not work properly, if mailbox files are stored on cephfs and a...
Jeff Layton
10:18 AM Bug #23883: kclient: CephFS kernel client hang
Zheng Yan wrote:
> besides, 4.4/4.9 kernel is too old for using multimds.
It is very difficult to upgrade kernel ...
wei jin
03:55 AM Documentation #23897 (In Progress): doc: create snapshot user doc
Include suggested upgrade procedure: https://github.com/ceph/ceph/pull/21374/commits/e05ebd08ea895626f4a2a52805f17e61... Patrick Donnelly
12:50 AM Bug #23894 (Resolved): ceph-fuse: missing dentries in readdir result
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2018-April/026224.html Zheng Yan

04/26/2018

11:54 PM Bug #23883: kclient: CephFS kernel client hang
besides, 4.4/4.9 kernel is too old for using multimds. Zheng Yan
11:47 PM Bug #23883: kclient: CephFS kernel client hang
need mds log to check what happened Zheng Yan
08:09 PM Bug #23883: kclient: CephFS kernel client hang

Patrick Donnelly
10:19 AM Bug #23883 (New): kclient: CephFS kernel client hang
ceph: 12.2.4/12.2.5
os: debian jessie
kernel: 4.9/4.4
After restart all mds(6 in total, 5 active, 1 standby), cl...
wei jin
10:23 PM Backport #23638 (In Progress): luminous: ceph-fuse: getgroups failure causes exception
Patrick Donnelly
08:01 PM Bug #23421: ceph-fuse: stop ceph-fuse if no root permissions?
Jos Collin wrote:
> The hang doesn't exist in the latest code.
>
> The following is my latest finding:
>
> [.....
Patrick Donnelly
10:07 AM Bug #23421: ceph-fuse: stop ceph-fuse if no root permissions?
The hang doesn't exist in the latest code.
The following is my latest finding:...
Jos Collin
05:20 PM Bug #23873: cephfs does not count st_nlink for directories correctly?
The Dovecot mail server does not work properly, if mailbox files are stored on cephfs and a mailbox prefix is configu... Peter Mauritius
04:39 AM Bug #23873: cephfs does not count st_nlink for directories correctly?
Zheng Yan wrote:
> If I remember right, this is not required by POSIX (btrfs does not do this). how NFS behaves depe...
Patrick Donnelly
02:34 AM Bug #23873: cephfs does not count st_nlink for directories correctly?
If I remember right, this is not required by POSIX (btrfs does not do this). how NFS behaves depends on the exported ... Zheng Yan
11:10 AM Bug #23885 (Resolved): MDSMonitor: overzealous MDS_ALL_DOWN and MDS_UP_LESS_THAN_MAX health warni...

This is what we currently get when starting with vstart, which creates MDS daemons before creating the filesystem:
...
John Spray
10:35 AM Bug #23855 (Fix Under Review): mds: MClientCaps should carry inode's dirstat
https://github.com/ceph/ceph/pull/21668 Zheng Yan
09:49 AM Bug #23880 (Fix Under Review): mds: scrub code stuck at trimming log segments
https://github.com/ceph/ceph/pull/21664 Zheng Yan
07:49 AM Bug #23880 (Resolved): mds: scrub code stuck at trimming log segments
/a/pdonnell-2018-04-25_18:15:51-kcephfs-wip-pdonnell-testing-20180425.144904-testing-basic-smithi/2439034 Zheng Yan
01:26 AM Feature #17854: mds: only evict an unresponsive client when another client wants its caps
Rishabh Dave wrote:
> I am planning to start working on this feature. How can I get a client to be unresponsive with...
Zheng Yan
12:49 AM Bug #23332: kclient: with fstab entry is not coming up reboot
kexec in dmesgs looks suspicious. client mounted cephfs, then used kexec to load kernel image again. All issues happe... Zheng Yan

04/25/2018

09:08 PM Feature #17854 (In Progress): mds: only evict an unresponsive client when another client wants it...
Patrick Donnelly
07:39 PM Feature #17854: mds: only evict an unresponsive client when another client wants its caps
I am planning to start working on this feature. How can I get a client to be unresponsive without evicting it? Rishabh Dave
08:24 PM Bug #23873 (Fix Under Review): cephfs does not count st_nlink for directories correctly?
https://github.com/ceph/ceph/pull/21652 Patrick Donnelly
07:42 PM Bug #23873 (Resolved): cephfs does not count st_nlink for directories correctly?
Not sure if this behavior is by intention, but if you create a empty directory on cephfs and call stat on the directo... Danny Al-Gaaf
06:09 PM Bug #23332: kclient: with fstab entry is not coming up reboot
Luis Henriques wrote:
> Actually, the first failure seems to be a bit before:
> [...]
> The client seems to be try...
Shreekara Shastry
04:58 PM Bug #23848 (Rejected): mds: stuck shutdown procedure
Patrick Donnelly
04:06 AM Bug #23848: mds: stuck shutdown procedure
... Patrick Donnelly
04:00 AM Bug #23848 (Rejected): mds: stuck shutdown procedure
The following outputs in an infinite loop:... Patrick Donnelly
01:10 PM Bug #23855 (Resolved): mds: MClientCaps should carry inode's dirstat
inode's dirstat gets updated by request reply, but not by cap message. this is problematic.
For example:
...
MDS...
Zheng Yan
08:33 AM Bug #22428: mds: don't report slow request for blocked filelock request
In case you need more examples, we're seeing this recently on 12.2.4:... Dan van der Ster
02:55 AM Bug #16842: mds: replacement MDS crashes on InoTable release
make we should mark this as "need more info" or "can't reproduce" Zheng Yan
02:08 AM Backport #23698: luminous: mds: load balancer fixes
https://github.com/ceph/ceph/pull/21412 Zheng Yan

04/24/2018

07:50 PM Bug #23829 (Fix Under Review): qa: test_purge_queue_op_rate: self.assertTrue(phase2_ops < phase1_...
Zheng Yan wrote:
> It's test case issue. The test caused so much trouble. I'd like to drop/disable it
Agreed.
...
Patrick Donnelly
12:24 PM Bug #23829: qa: test_purge_queue_op_rate: self.assertTrue(phase2_ops < phase1_ops * 1.25)
It's test case issue. The test caused so much trouble. I'd like to drop/disable it Zheng Yan
07:37 PM Bug #23837 (Fix Under Review): client: deleted inode's Bufferhead which was in STATE::Tx would le...
Patrick Donnelly
10:44 AM Bug #23837: client: deleted inode's Bufferhead which was in STATE::Tx would lead a assert fail
fixed by: https://github.com/ceph/ceph/pull/21615 Ivan Guan
09:45 AM Bug #23837 (Resolved): client: deleted inode's Bufferhead which was in STATE::Tx would lead a ass...
... Ivan Guan
07:07 PM Backport #23671 (In Progress): luminous: mds: MDBalancer using total (all time) request count in ...
https://github.com/ceph/ceph/pull/21412/commits/1a5b7eaac572f1810d0453b053781e6bc8185dd2 Patrick Donnelly
06:55 PM Tasks #23844 (In Progress): client: break client_lock
See past efforts on this. Matt Benjamin did some prototyping on Firefly. Those patches will likely be unusable but co... Patrick Donnelly
11:19 AM Backport #23835 (In Progress): luminous: mds: fix occasional dir rstat inconsistency between mult...
https://github.com/ceph/ceph/pull/21617 Prashant D
05:48 AM Backport #23835 (Resolved): luminous: mds: fix occasional dir rstat inconsistency between multi-M...
https://github.com/ceph/ceph/pull/21617 Nathan Cutler
11:10 AM Backport #23308 (In Progress): luminous: doc: Fix -d option in ceph-fuse doc
Jos Collin
08:24 AM Bug #20549 (Fix Under Review): cephfs-journal-tool: segfault during journal reset
https://github.com/ceph/ceph/pull/21610 Zheng Yan
07:09 AM Feature #23362: mds: add drop_cache command
https://github.com/ceph/ceph/pull/21566 Rishabh Dave
05:47 AM Backport #23834 (Rejected): jewel: MDSMonitor: crash after assigning standby-replay daemon in mul...
Nathan Cutler
05:47 AM Backport #23833 (Resolved): luminous: MDSMonitor: crash after assigning standby-replay daemon in ...
https://github.com/ceph/ceph/pull/22603 Nathan Cutler
04:42 AM Bug #23567 (Resolved): MDSMonitor: successive changes to max_mds can allow hole in ranks
Patrick Donnelly
04:35 AM Bug #23538 (Pending Backport): mds: fix occasional dir rstat inconsistency between multi-MDSes
Patrick Donnelly
04:34 AM Bug #23658 (Pending Backport): MDSMonitor: crash after assigning standby-replay daemon in multifs...
Patrick Donnelly
04:33 AM Bug #23799 (Resolved): MDSMonitor: creates invalid transition from up:creating to up:shutdown
Patrick Donnelly
04:32 AM Bug #23800 (Resolved): MDSMonitor: setting fs down twice will wipe old_max_mds
Patrick Donnelly

04/23/2018

08:26 PM Bug #23829 (Resolved): qa: test_purge_queue_op_rate: self.assertTrue(phase2_ops < phase1_ops * 1.25)
... Patrick Donnelly
07:51 PM Bug #20549: cephfs-journal-tool: segfault during journal reset
Another: http://pulpito.ceph.com/pdonnell-2018-04-23_17:22:02-kcephfs-wip-pdonnell-testing-20180423.033341-testing-ba... Patrick Donnelly
05:51 PM Bug #23814 (Rejected): mds: newly active mds aborts may abort in handle_file_lock
Patrick Donnelly
08:40 AM Bug #23814: mds: newly active mds aborts may abort in handle_file_lock
I think this is related to #23812. The patch for #23812 makes mds skip sending lock message to 'starting' mds. The sk... Zheng Yan
05:50 PM Bug #23812: mds: may send LOCK_SYNC_MIX message to starting MDS
https://github.com/ceph/ceph/pull/21601 Patrick Donnelly
05:10 PM Backport #22860 (Resolved): luminous: osdc: "FAILED assert(bh->last_write_tid > tid)" in powercyc...
Looks like a different assertion so perhaps a new bug. I'll create a separate issue for this. Patrick Donnelly
03:49 PM Backport #22860 (In Progress): luminous: osdc: "FAILED assert(bh->last_write_tid > tid)" in power...
... Sage Weil
03:36 PM Backport #23151 (In Progress): luminous: doc: update ceph-fuse with FUSE options
Jos Collin
01:54 PM Bug #23826 (Duplicate): mds: assert after daemon restart
... Patrick Donnelly
01:26 PM Backport #23475 (In Progress): luminous: ceph-fuse: trim ceph-fuse -V output
Jos Collin
11:54 AM Backport #23771 (In Progress): luminous: client: fix gid_count check in UserPerm->deep_copy_from()
Jos Collin
11:50 AM Backport #23771: luminous: client: fix gid_count check in UserPerm->deep_copy_from()
https://github.com/ceph/ceph/pull/21596 Jos Collin
10:45 AM Bug #23813 (Fix Under Review): client: "remove_session_caps still has dirty|flushing caps" when t...
https://github.com/ceph/ceph/pull/21593 Zheng Yan
08:53 AM Bug #23518 (Fix Under Review): mds: crash when failover
https://github.com/ceph/ceph/pull/21592 Zheng Yan
07:49 AM Bug #23815: client: avoid second lock on client_lock
supriti singh wrote:
> supriti singh wrote:
> > In function ll_get_stripe_osd client_lock is taken. But its acquire...
supriti singh
03:52 AM Backport #23818 (In Progress): luminous: client: add option descriptions and review levels (e.g. ...
https://github.com/ceph/ceph/pull/21589 Prashant D
 

Also available in: Atom