Project

General

Profile

Activity

From 05/23/2018 to 06/21/2018

06/21/2018

12:53 PM Feature #24604 (Resolved): Implement "cephfs-journal-tool event splice" equivalent for purge queue
cephfs-journal-tool recently got the ability to scan the purge queue via the --journal=purge_queue argument.
Howev...
John Spray

06/20/2018

11:06 PM Bug #24522 (New): blogbench.sh failed in upgrade:luminous-x-mimic-distro-basic-smithi
Whoops! Patrick Donnelly
09:58 PM Bug #24522: blogbench.sh failed in upgrade:luminous-x-mimic-distro-basic-smithi
Patrick, this is a different script(blogbench.sh) not pjd.sh, so might not be a duplicate. Neha Ojha
09:00 PM Bug #24522 (Duplicate): blogbench.sh failed in upgrade:luminous-x-mimic-distro-basic-smithi
Thanks Neha! Patrick Donnelly
08:39 PM Bug #24522: blogbench.sh failed in upgrade:luminous-x-mimic-distro-basic-smithi
Following is the problem:... Neha Ojha
10:01 PM Bug #24137 (Resolved): client: segfault in trim_caps
Nathan Cutler
10:00 PM Backport #24185 (Resolved): luminous: client: segfault in trim_caps
Nathan Cutler
07:59 PM Backport #24185: luminous: client: segfault in trim_caps
Patrick Donnelly wrote:
> https://github.com/ceph/ceph/pull/22201
merged
Yuri Weinstein
10:00 PM Backport #24331 (Resolved): luminous: mon: mds health metrics sent to cluster log indpeendently
Nathan Cutler
07:58 PM Backport #24331: luminous: mon: mds health metrics sent to cluster log indpeendently
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/22558
merged
Yuri Weinstein
07:57 PM Backport #23792: luminous: MDSMonitor: MDSMonitor::encode_pending modifies fsmap not pending_fsmap
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21732
merged
Yuri Weinstein
05:10 PM Feature #17854 (Fix Under Review): mds: only evict an unresponsive client when another client wan...
Patrick Donnelly
08:22 AM Bug #24579: client: returning garbage (?) for readdir
ffsb issue. below patch can fix it. I don't know how to update http://download.ceph.com/qa/ffsb.tar.bz2... Zheng Yan
07:57 AM Bug #24512: Raw used space leak
some additional info:
- mounted with 'mount -t ceph'
- default config but:
--- 2 mds servers active
--- ram per O...
Thomas De Maet

06/19/2018

07:22 PM Documentation #24580 (Resolved): doc: complete documentation for `ceph fs` administration commands
Current skeleton: http://docs.ceph.com/docs/luminous/cephfs/administration/ Patrick Donnelly
07:22 PM Bug #24240: qa: 1 mutations had unexpected outcomes
not indicated for backport to luminous because, as per Zheng, luminous does not have open file table. Nathan Cutler
06:56 PM Bug #24512: Raw used space leak
I checked 'stored' vs. 'allocated' counters under bluestore section. 'stored' is the actual amount written to bluesto... Igor Fedotov
12:28 PM Bug #24512: Raw used space leak
Here they are from 3 hosts (the link with hosts is in the df tree) !
Thanks!
Thomas De Maet
11:13 AM Bug #24512: Raw used space leak
Would you share performance counters dump for several (3-5) OSDs, preferably from different nodes? And corresponding ... Igor Fedotov
08:08 AM Bug #24512: Raw used space leak
sorry, wrong ceph version: 12.2.5-407 (luminous stable)
I'm still very interested by any answer. If I try filestor...
Thomas De Maet
06:09 PM Bug #24579: client: returning garbage (?) for readdir
This seems to only happen on Ubuntu 18.04:... Patrick Donnelly
06:05 PM Bug #24579: client: returning garbage (?) for readdir
Here too:
/ceph/teuthology-archive/teuthology-2018-06-18_20:06:42-powercycle-master-distro-basic-smithi/2678660
...
Patrick Donnelly
06:03 PM Bug #24579 (Resolved): client: returning garbage (?) for readdir
... Patrick Donnelly
04:18 PM Bug #24441 (Closed): Ceph fs new cephfs command failed when meta pool already contains some objects
This is not a bug -- the check was added to avoid people accidentally getting corrupt filesystems by trying to use a ... John Spray
03:37 AM Feature #24464: cephfs: file-level snapshots
I think using rados snapshot to support is too expensive. Zheng Yan
02:55 AM Bug #24557 (Fix Under Review): client: segmentation fault in handle_client_reply
https://github.com/ceph/ceph/pull/22611 Zheng Yan

06/18/2018

09:19 PM Bug #24557 (Resolved): client: segmentation fault in handle_client_reply
... Patrick Donnelly
07:21 PM Backport #23833 (In Progress): luminous: MDSMonitor: crash after assigning standby-replay daemon ...
Patrick Donnelly
05:45 PM Feature #17230: ceph_volume_client: py3 compatible
There are high-level pushes to py3 in future (minor-)releases of Ceph/Openstack. RHCS 3.X (Luminous) will need to be ... Patrick Donnelly
05:35 PM Feature #17230: ceph_volume_client: py3 compatible
Question for all, and particularly for Patrick: why is a luminous backport of this needed, and is it worth the risk? Nathan Cutler
05:42 PM Bug #24518 (Duplicate): "pjd.sh: line 7: cd: too many arguments" in fs suite
Yuri, you're using an old qa-suite branch. Thanks to Neha for noticing the cause. Patrick Donnelly
05:29 PM Bug #24518: "pjd.sh: line 7: cd: too many arguments" in fs suite
Line in question: https://github.com/ceph/ceph/blob/2d2293948066cae8d656dfe91bdb6695958a52e9/qa/workunits/suites/pjd.... Patrick Donnelly

06/15/2018

04:01 PM Backport #24541 (Resolved): mimic: qa: 1 mutations had unexpected outcomes
https://github.com/ceph/ceph/pull/22841 Nathan Cutler
04:01 PM Backport #24540 (Resolved): luminous: multimds pjd open test fails
https://github.com/ceph/ceph/pull/22783 Nathan Cutler
04:01 PM Backport #24539 (Resolved): mimic: multimds pjd open test fails
https://github.com/ceph/ceph/pull/22819 Nathan Cutler
04:01 PM Backport #24538 (Resolved): luminous: common/DecayCounter: set last_decay to current time when de...
https://github.com/ceph/ceph/pull/22779 Nathan Cutler
04:01 PM Backport #24537 (Resolved): mimic: common/DecayCounter: set last_decay to current time when decod...
https://github.com/ceph/ceph/pull/22816 Nathan Cutler
04:01 PM Backport #24536 (Rejected): jewel: client: _ll_drop_pins travel inode_map may access invalid ‘nex...
Nathan Cutler
04:01 PM Backport #24535 (Resolved): luminous: client: _ll_drop_pins travel inode_map may access invalid ‘...
https://github.com/ceph/ceph/pull/22786 Nathan Cutler
04:01 PM Backport #24534 (Resolved): mimic: client: _ll_drop_pins travel inode_map may access invalid ‘nex...
https://github.com/ceph/ceph/pull/22791 Nathan Cutler
03:45 PM Bug #24533 (Fix Under Review): PurgeQueue sometimes ignores Journaler errors
https://github.com/ceph/ceph/pull/22580 John Spray
03:33 PM Bug #24533 (Resolved): PurgeQueue sometimes ignores Journaler errors
We check journaler.get_error() in PurgeQueue::_recover, but never later in _consume -- if something like a decode err... John Spray
02:06 PM Bug #24491 (Pending Backport): client: _ll_drop_pins travel inode_map may access invalid ‘next’ i...
Patrick Donnelly
02:04 PM Bug #24440 (Pending Backport): common/DecayCounter: set last_decay to current time when decoding ...
Patrick Donnelly
02:03 PM Bug #24269 (Pending Backport): multimds pjd open test fails
Patrick Donnelly
02:02 PM Bug #24240 (Pending Backport): qa: 1 mutations had unexpected outcomes
Patrick Donnelly

06/14/2018

10:02 AM Bug #24284 (Fix Under Review): cephfs: allow prohibiting user snapshots in CephFS
https://github.com/ceph/ceph/pull/22560 Zheng Yan
05:52 AM Backport #24331 (In Progress): luminous: mon: mds health metrics sent to cluster log indpeendently
https://github.com/ceph/ceph/pull/22558 Prashant D
01:44 AM Backport #24330: mimic: mon: mds health metrics sent to cluster log indpeendently
mimic backport PR https://github.com/ceph/ceph/pull/22265 is already opened. So closing PR#22540. Prashant D
12:24 AM Bug #24306: mds: use intrusive_ptr to manage Message life-time
https://github.com/ceph/ceph/pull/22555 Patrick Donnelly

06/13/2018

09:53 PM Bug #24522 (Resolved): blogbench.sh failed in upgrade:luminous-x-mimic-distro-basic-smithi
Seems bionic specific
Run: http://pulpito.ceph.com/yuriw-2018-06-12_20:54:23-upgrade:luminous-x-mimic-distro-basic-s...
Yuri Weinstein
09:05 PM Bug #24518: "pjd.sh: line 7: cd: too many arguments" in fs suite
Also seems like in run:
http://pulpito.ceph.com/yuriw-2018-06-12_21:34:02-powercycle-mimic-distro-basic-smithi/
Job...
Yuri Weinstein
08:27 PM Bug #24518 (Duplicate): "pjd.sh: line 7: cd: too many arguments" in fs suite
This seems to be bionic specific
Run: http://pulpito.ceph.com/yuriw-2018-06-12_21:09:43-fs-master-distro-basic-smith...
Yuri Weinstein
08:54 PM Bug #24520 (Duplicate): "[WRN] MDS health message (mds.0): 2 slow requests are blocked > 30 sec""...
Run: http://pulpito.ceph.com/yuriw-2018-06-12_21:34:02-powercycle-mimic-distro-basic-smithi/
Jobs: '2660103', '26600...
Yuri Weinstein
08:22 PM Bug #24517 (Duplicate): "Loading libcephfs-jni: Failure!" in fs suite
This seems to be rhel specific
Run: http://pulpito.ceph.com/yuriw-2018-06-12_21:09:43-fs-master-distro-basic-smithi/...
Yuri Weinstein
05:16 PM Bug #23697 (Resolved): mds: load balancer fixes
Nathan Cutler
05:15 PM Backport #23698 (Resolved): luminous: mds: load balancer fixes
Nathan Cutler
05:15 PM Bug #21745 (Resolved): mds: MDBalancer using total (all time) request count in load statistics
Nathan Cutler
05:13 PM Backport #23671 (Resolved): luminous: mds: MDBalancer using total (all time) request count in loa...
Nathan Cutler
05:11 PM Feature #23695 (Resolved): VolumeClient: allow ceph_volume_client to create 'volumes' without nam...
Nathan Cutler
05:11 PM Backport #24055 (Resolved): luminous: VolumeClient: allow ceph_volume_client to create 'volumes' ...
Nathan Cutler
02:57 PM Feature #21571: mds: limit number of snapshots (global and subtree)
There should be a global limit (if necessary for performance) and subtree limits (from #24429) so that operators can ... Patrick Donnelly
02:56 PM Feature #24429 (Duplicate): fs: implement snapshot count limit by subtree
Patrick Donnelly
02:55 PM Backport #24296: mimic: repeated eviction of idle client until some IO happens
Zheng Yan wrote:
> just replace 'cbegin()' with begin()
Thanks, Zheng. Did just that.
Nathan Cutler
02:54 PM Backport #24296 (In Progress): mimic: repeated eviction of idle client until some IO happens
Nathan Cutler
02:06 PM Bug #19438 (Won't Fix): ceph mds error "No space left on device"
dirfrags are not stable on jewel. Closing this. Patrick Donnelly
01:44 PM Bug #24512 (New): Raw used space leak
Hello
I'm testing an setup of cephfs over a EC pool with 21 data + 3 coding chunks ([EC_]stripe_unit of 16k).
All...
Thomas De Maet
12:52 PM Feature #24465 (Fix Under Review): client: allow client to leave state intact on MDS when tearing...
https://github.com/ceph/ceph/pull/22543 Zheng Yan
04:14 AM Backport #24330 (In Progress): mimic: mon: mds health metrics sent to cluster log indpeendently
-https://github.com/ceph/ceph/pull/22540- Prashant D

06/12/2018

07:50 AM Bug #23665 (Resolved): ceph-fuse: return proper exit code
Nathan Cutler
07:49 AM Bug #22933 (Resolved): client: add option descriptions and review levels (e.g. LEVEL_DEV)
Nathan Cutler
02:23 AM Bug #24491 (Fix Under Review): client: _ll_drop_pins travel inode_map may access invalid ‘next’ i...
https://github.com/ceph/ceph/pull/22512 Zheng Yan
02:09 AM Bug #24491: client: _ll_drop_pins travel inode_map may access invalid ‘next’ iterator
Thank for reporting this. Could you fix this issue in a way similar to https://github.com/ceph/ceph/pull/22073? Zheng Yan

06/11/2018

03:50 PM Bug #24491 (Resolved): client: _ll_drop_pins travel inode_map may access invalid ‘next’ iterator
We have encounter a process crash when using libcephfs.
the call stack is below:
#0 0x00007fdef24941f7 in raise ...
Yangkuan Liu
01:42 PM Bug #24400 (Can't reproduce): CephFS - All MDS went offline and required repair of filesystem
reopen this ticket if you encounter this issue again Zheng Yan
01:39 PM Bug #24369 (Resolved): luminous: checking quota while holding cap ref may deadlock
Zheng Yan

06/10/2018

09:15 AM Bug #23538 (Resolved): mds: fix occasional dir rstat inconsistency between multi-MDSes
Nathan Cutler

06/09/2018

02:24 PM Backport #23698: luminous: mds: load balancer fixes
Zheng Yan wrote:
> https://github.com/ceph/ceph/pull/21412
merged
Yuri Weinstein
02:24 PM Backport #24055: luminous: VolumeClient: allow ceph_volume_client to create 'volumes' without nam...
Patrick Donnelly wrote:
> https://github.com/ceph/ceph/pull/21897
merged
Yuri Weinstein
02:23 PM Bug #24369: luminous: checking quota while holding cap ref may deadlock
Zheng Yan wrote:
> https://github.com/ceph/ceph/pull/22354
merged
Yuri Weinstein
11:36 AM Bug #23815 (Resolved): client: avoid second lock on client_lock
Nathan Cutler
11:35 AM Bug #23829 (Resolved): qa: test_purge_queue_op_rate: self.assertTrue(phase2_ops < phase1_ops * 1.25)
Nathan Cutler
11:21 AM Bug #20549 (Resolved): cephfs-journal-tool: segfault during journal reset
Nathan Cutler
11:20 AM Bug #23923 (Resolved): mds: stopping rank 0 cannot shutdown until log is trimmed
Nathan Cutler
11:17 AM Bug #23919 (Resolved): mds: stuck during up:stopping
Nathan Cutler
11:16 AM Bug #23960 (Resolved): mds: scrub on fresh file system fails
Nathan Cutler
11:15 AM Bug #23812 (Resolved): mds: may send LOCK_SYNC_MIX message to starting MDS
Nathan Cutler
11:14 AM Bug #23855 (Resolved): mds: MClientCaps should carry inode's dirstat
Nathan Cutler
11:13 AM Bug #23894 (Resolved): ceph-fuse: missing dentries in readdir result
Nathan Cutler
11:12 AM Bug #23518 (Resolved): mds: crash when failover
Nathan Cutler
11:11 AM Bug #24073 (Resolved): PurgeQueue::_consume() could return true when there were no purge queue it...
Nathan Cutler
11:11 AM Bug #24047 (Resolved): MDCache.cc: 5317: FAILED assert(mds->is_rejoin())
Nathan Cutler
07:51 AM Bug #24467 (Fix Under Review): mds: low wrlock efficiency due to dirfrags traversal
https://github.com/ceph/ceph/pull/22486 Kefu Chai
07:03 AM Bug #24467 (Resolved): mds: low wrlock efficiency due to dirfrags traversal
Recently, when trying to create/remove massive files/dirs(7x10^6) within a common directory, we found that as the cre... Xuehan Xu

06/08/2018

09:08 PM Feature #24461: cephfs: improve file create performance buffering file unlink/create operations
We've talked about this quite a lot in the past. I thought we had a tracker ticket for it, but on searching the most ... Greg Farnum
06:21 PM Feature #24461: cephfs: improve file create performance buffering file unlink/create operations
Neat. NFS and SMB have directory delegations/leases, but I haven't studied the topic in detail.
So the idea is to ...
Jeff Layton
05:10 PM Feature #24461 (Resolved): cephfs: improve file create performance buffering file unlink/create o...
**Serialized single-client** file creation (e.g. untar/rsync) is an area CephFS (and most distributed file systems) c... Patrick Donnelly
07:08 PM Feature #24465 (Resolved): client: allow client to leave state intact on MDS when tearing down ob...
When ganesha shuts down cleanly, it'll tear down all of its filehandle objects and release the files that it has open... Jeff Layton
05:50 PM Feature #24464 (New): cephfs: file-level snapshots
Use-case is to support dropbox-style versioning of files. Patrick Donnelly
05:46 PM Feature #24463 (Resolved): kclient: add btime support
Patrick Donnelly
05:43 PM Feature #24462 (New): MDSMonitor: check for mixed version MDS
And create a health error if it detects this. Patrick Donnelly
09:00 AM Bug #24173 (In Progress): ceph_volume_client: allow atomic update of RADOS objects
https://github.com/ceph/ceph/pull/22455 Rishabh Dave

06/07/2018

01:30 PM Backport #24296: mimic: repeated eviction of idle client until some IO happens
just replace 'cbegin()' with begin() Zheng Yan
01:07 PM Backport #24296 (Need More Info): mimic: repeated eviction of idle client until some IO happens
While backporting changes related to tracker 24052, getting cbegin not found compilation error :
/home/pdvian/backpo...
Prashant D
01:11 PM Bug #24435 (Resolved): doc: incorrect snaprealm format upgrade process in mimic release note
Patrick Donnelly
01:07 PM Bug #24435 (Pending Backport): doc: incorrect snaprealm format upgrade process in mimic release note
Patrick Donnelly
01:11 PM Backport #24451 (Rejected): mimic: doc: incorrect snaprealm format upgrade process in mimic relea...
Nevermind, this doc doesn't exist in mimic. Patrick Donnelly
01:08 PM Backport #24451 (Rejected): mimic: doc: incorrect snaprealm format upgrade process in mimic relea...
Patrick Donnelly
08:23 AM Feature #24444 (Resolved): cephfs: make InodeStat, DirStat, LeaseStat versioned
Make InodeStat/DirStat/LeaseStat versioned, so client can decode InodeStat in request reply without checking mds feat... Zheng Yan
07:34 AM Feature #20598 (Fix Under Review): mds: revisit LAZY_IO
https://github.com/ceph/ceph/pull/22450 Zheng Yan
06:31 AM Bug #24441: Ceph fs new cephfs command failed when meta pool already contains some objects
ceph version 10.2.10:
when meta pool has objects.Run ceph fs new cephfs meta data,it can create fs successed.
...
鹏 张
06:23 AM Bug #24441 (Closed): Ceph fs new cephfs command failed when meta pool already contains some objects
ceph fs new cephfs meta4 data
Error EINVAL: pool 'meta4' already contains some objects. Use an empty pool instead.
鹏 张
03:04 AM Bug #24440: common/DecayCounter: set last_decay to current time when decoding decay counter
https://github.com/ceph/ceph/pull/22357 Zhi Zhang
03:03 AM Bug #24440 (Resolved): common/DecayCounter: set last_decay to current time when decoding decay co...
Recently we found mds load might become zero on another MDS under multi-MDSes scenario. The ceph version is Luminous.... Zhi Zhang

06/06/2018

10:28 PM Documentation #24093 (Resolved): doc: Update *remove a metadata server*
Patrick Donnelly
09:23 PM Bug #24435 (Fix Under Review): doc: incorrect snaprealm format upgrade process in mimic release note
https://github.com/ceph/ceph/pull/22445 Patrick Donnelly
09:17 PM Bug #24435 (In Progress): doc: incorrect snaprealm format upgrade process in mimic release note
Patrick Donnelly
01:55 PM Bug #24435 (Resolved): doc: incorrect snaprealm format upgrade process in mimic release note
The commands to upgrade snaprealm format in release note are
ceph daemon <mds of rank 0> scrub_path /
ceph daemon...
Zheng Yan
08:49 AM Bug #24028: CephFS flock() on a directory is broken
In fuse filesystem, flock on directory is handled by VFS, there is nothing ceph-fuse can do. Zheng Yan
08:12 AM Bug #24028: CephFS flock() on a directory is broken
In that case flock() syscall over FUSEd directory should return an ENOTSUPP?. In any case we must not allow unsafe lo... Марк Коренберг
07:46 AM Bug #24028: CephFS flock() on a directory is broken
ceph-fuse does not support file lock on directory. It's limitation of fuse kernel module. Zheng Yan
07:12 AM Bug #24400: CephFS - All MDS went offline and required repair of filesystem
http://tracker.ceph.com/issues/17177 can explain this issue. full filesystem scrub should repair incorrect dirstat/rs... Zheng Yan
06:24 AM Bug #24400: CephFS - All MDS went offline and required repair of filesystem
Zheng Yan wrote:
> there are lots of inodes have incorrect dirstat/rstat. have you ever run 'journal reset' before t...
James Wilkins
02:16 AM Bug #24400: CephFS - All MDS went offline and required repair of filesystem
there are lots of inodes have incorrect dirstat/rstat. have you ever run 'journal reset' before the crash Zheng Yan
02:07 AM Feature #24430 (Resolved): libcephfs: provide API to change umask
The current use-case will be the CephFS shell. Patrick Donnelly

06/05/2018

09:05 PM Feature #24429 (Duplicate): fs: implement snapshot count limit by subtree
e.g. don't let a subtree have more than 7 snapshots. This should be configurable via an xattr.
Idea is from Dan va...
Patrick Donnelly
06:06 PM Feature #24426 (New): mds: add second level cache backed by local SSD or NVRAM
Idea is to have a second level to the MDS cache to improve access time and reduce reads on the metadata pool. This wo... Patrick Donnelly
02:47 PM Bug #24284: cephfs: allow prohibiting user snapshots in CephFS
> change default of mds_snap_max_uid to 0
Use-cases such as Manila let the users mount with root so this will be i...
Dan van der Ster
02:19 AM Bug #24284: cephfs: allow prohibiting user snapshots in CephFS
maybe we can use 'auth string' Zheng Yan
10:42 AM Bug #24403: mon failed to return metadata for mds
I have updated first telegeo02 with no different result (as mds on telegeo02 was standby as last one rebooted)
The...
Thomas De Maet
09:14 AM Feature #22446: mds: ask idle client to trim more caps
Can I get few implementation specific details to get started working on this issue?
And for clarity on my side, we...
Rishabh Dave
08:27 AM Bug #24400: CephFS - All MDS went offline and required repair of filesystem
Zheng Yan wrote:
> do have have full log (the time mds started replay to mds crash). thanks
Full MDS log starting...
James Wilkins
12:50 AM Bug #23032 (Resolved): mds: underwater dentry check in CDir::_omap_fetched is racy
Zheng Yan
12:49 AM Backport #23157 (Resolved): luminous: mds: underwater dentry check in CDir::_omap_fetched is racy
Zheng Yan
12:49 AM Backport #22696 (Resolved): luminous: client: dirty caps may never get the chance to flush
Zheng Yan

06/04/2018

10:57 PM Bug #24400: CephFS - All MDS went offline and required repair of filesystem
do have have full log (the time mds started replay to mds crash). thanks Zheng Yan
02:06 PM Bug #24400: CephFS - All MDS went offline and required repair of filesystem
Zheng Yan wrote:
> do you have mds log just before the crash
Excellent timing - we've just finished trawling thro...
James Wilkins
01:55 PM Bug #24400: CephFS - All MDS went offline and required repair of filesystem
do you have mds log just before the crash Zheng Yan
08:02 AM Bug #24400: CephFS - All MDS went offline and required repair of filesystem
Forgot to say - one of the logs was taken with debug enabled (thus the size). Can provide whole log if needed
James Wilkins
07:45 AM Bug #24400 (Can't reproduce): CephFS - All MDS went offline and required repair of filesystem
Hi,
Raising this incase we can get some more insight and/or it helps others.
We have a 12.2.5 cluster provising...
James Wilkins
09:14 PM Bug #24241 (New): NFS-Ganesha libcephfs: Assert failure in object_locator_to_pg
Patrick Donnelly
06:15 PM Feature #24263: client/mds: create a merkle tree of objects to allow efficient generation of diff...
Sage Weil wrote:
> A few questions:
>
> - What is the sha1 of? The object's content? That isn't necessarily kno...
Patrick Donnelly
05:59 PM Feature #24263: client/mds: create a merkle tree of objects to allow efficient generation of diff...
John Spray wrote:
> Patrick Donnelly wrote:
> > John Spray wrote:
> > > I'm a fan. Questions that spring to mind:...
Patrick Donnelly
02:20 PM Bug #24403: mon failed to return metadata for mds
The "sen2agriprod" server actually runs on centOS7 (kernel 3.10.0) which is in the recommended platforms.
If you t...
Thomas De Maet
01:30 PM Bug #24403: mon failed to return metadata for mds
please try newer kernel Zheng Yan
10:04 AM Bug #24403 (Resolved): mon failed to return metadata for mds
Hello,
Redigging an error found into the ceph-users mailing list: http://lists.ceph.com/pipermail/ceph-users-ceph....
Thomas De Maet
01:41 PM Bug #24306 (In Progress): mds: use intrusive_ptr to manage Message life-time
Patrick Donnelly
09:34 AM Bug #24172 (Resolved): client: fails to respond cap revoke from non-auth mds
Nathan Cutler
05:39 AM Bug #23214 (Resolved): doc: Fix -d option in ceph-fuse doc
Jos Collin
05:36 AM Bug #23248 (Resolved): ceph-fuse: trim ceph-fuse -V output
Jos Collin
01:24 AM Backport #23704 (Resolved): luminous: ceph-fuse: broken directory permission checking
Zheng Yan
01:24 AM Backport #23770 (Resolved): luminous: ceph-fuse: return proper exit code
Zheng Yan
01:22 AM Backport #23818 (Resolved): luminous: client: add option descriptions and review levels (e.g. LEV...
Zheng Yan
01:22 AM Backport #23475 (Resolved): luminous: ceph-fuse: trim ceph-fuse -V output
Zheng Yan
01:21 AM Backport #23835 (Resolved): luminous: mds: fix occasional dir rstat inconsistency between multi-M...
Zheng Yan
01:21 AM Backport #23638 (Resolved): luminous: ceph-fuse: getgroups failure causes exception
Zheng Yan
01:20 AM Backport #23933 (Resolved): luminous: client: avoid second lock on client_lock
Zheng Yan
01:17 AM Backport #23931 (Resolved): luminous: qa: test_purge_queue_op_rate: self.assertTrue(phase2_ops < ...
Zheng Yan
01:16 AM Backport #23936 (Resolved): luminous: cephfs-journal-tool: segfault during journal reset
Zheng Yan
01:16 AM Backport #23950 (Resolved): luminous: mds: stopping rank 0 cannot shutdown until log is trimmed
Zheng Yan
01:15 AM Backport #23951 (Resolved): luminous: mds: stuck during up:stopping
Zheng Yan
01:15 AM Backport #23984 (Resolved): luminous: mds: scrub on fresh file system fails
Zheng Yan
01:14 AM Backport #23935 (Resolved): luminous: mds: may send LOCK_SYNC_MIX message to starting MDS
Zheng Yan
01:13 AM Backport #23991 (Resolved): luminous: client: hangs on umount if it had an MDS session evicted
Zheng Yan
01:13 AM Backport #24050 (Resolved): luminous: mds: MClientCaps should carry inode's dirstat
Zheng Yan
01:12 AM Backport #24049 (Resolved): luminous: ceph-fuse: missing dentries in readdir result
Zheng Yan
01:12 AM Backport #23946 (Resolved): luminous: mds: crash when failover
Zheng Yan
01:10 AM Backport #24107 (Resolved): luminous: PurgeQueue::_consume() could return true when there were no...
Zheng Yan
01:09 AM Backport #24108 (Resolved): luminous: MDCache.cc: 5317: FAILED assert(mds->is_rejoin())
Zheng Yan
01:03 AM Backport #24130 (Resolved): luminous: mds: race with new session from connection and imported ses...
Zheng Yan
01:02 AM Backport #24188 (Resolved): luminous: kceph: umount on evicted client blocks forever
Zheng Yan
01:01 AM Backport #24201 (Resolved): luminous: client: fails to respond cap revoke from non-auth mds
Zheng Yan
01:00 AM Backport #24207 (Resolved): luminous: client: deleted inode's Bufferhead which was in STATE::Tx w...
Zheng Yan
12:59 AM Bug #24289 (Resolved): mds memory leak
Zheng Yan
12:57 AM Backport #23982 (Resolved): luminous: qa: TestVolumeClient.test_lifecycle needs updated for new e...
Zheng Yan
12:55 AM Backport #24205 (Resolved): luminous: mds: broadcast quota to relevant clients when quota is expl...
Zheng Yan
12:53 AM Backport #24189 (Resolved): luminous: qa: kernel_mount.py umount must handle timeout arg
Zheng Yan
12:52 AM Backport #24341 (Resolved): luminous: mds memory leak
Zheng Yan

06/02/2018

06:18 AM Backport #23308 (Resolved): luminous: doc: Fix -d option in ceph-fuse doc
Nathan Cutler

06/01/2018

11:44 AM Documentation #24093 (Fix Under Review): doc: Update *remove a metadata server*
https://github.com/ceph/ceph/pull/22338 Jos Collin
02:43 AM Bug #24369 (Fix Under Review): luminous: checking quota while holding cap ref may deadlock
https://github.com/ceph/ceph/pull/22354 Zheng Yan
12:58 AM Bug #24369: luminous: checking quota while holding cap ref may deadlock
For example:
mds revokes an inode's Fw
mds freezes the subtree that contains the inode
client::_write() calls ...
Zheng Yan
12:52 AM Bug #24369 (Resolved): luminous: checking quota while holding cap ref may deadlock
Zheng Yan
02:10 AM Backport #24372 (Rejected): luminous: mds: root inode's snaprealm doesn't get journalled correctly
Zheng Yan
02:10 AM Backport #24372: luminous: mds: root inode's snaprealm doesn't get journalled correctly
luminous does not support snapshot Zheng Yan
02:07 AM Backport #24372 (Rejected): luminous: mds: root inode's snaprealm doesn't get journalled correctly
Zheng Yan
02:08 AM Bug #24370 (Duplicate): luminous: root dir's new snapshot lost when restart mds
dup of https://tracker.ceph.com/issues/24372 Zheng Yan
01:59 AM Bug #24370 (Duplicate): luminous: root dir's new snapshot lost when restart mds
affect version: luminous & mimic
reproduce steps:
1. ceph-fuse mount /cephfuse
2. write a file /cephfuse/file1
...
Min Chen
02:06 AM Backport #24340 (Resolved): mimic: mds memory leak
Zheng Yan

05/31/2018

01:47 PM Bug #24241: NFS-Ganesha libcephfs: Assert failure in object_locator_to_pg
If you have time, it's probably worthwhile to roll a new testcase for ceph_ll_get_stripe_osd for this sort of thing. ... Jeff Layton
11:57 AM Backport #24345 (Resolved): mimic: mds: root inode's snaprealm doesn't get journalled correctly
Nathan Cutler

05/30/2018

12:04 PM Backport #24345 (In Progress): mimic: mds: root inode's snaprealm doesn't get journalled correctly
https://github.com/ceph/ceph/pull/22322 Zheng Yan
11:52 AM Backport #24345 (Resolved): mimic: mds: root inode's snaprealm doesn't get journalled correctly
Zheng Yan
12:04 PM Bug #24343 (Resolved): mds: root inode's snaprealm doesn't get journalled correctly
https://github.com/ceph/ceph/pull/22320 Zheng Yan
11:25 AM Bug #24343 (Resolved): mds: root inode's snaprealm doesn't get journalled correctly
Zheng Yan
03:47 AM Backport #24341 (In Progress): luminous: mds memory leak
https://github.com/ceph/ceph/pull/22310 Zheng Yan
03:43 AM Backport #24341 (Resolved): luminous: mds memory leak
https://github.com/ceph/ceph/pull/22310 Zheng Yan
03:42 AM Backport #24340 (In Progress): mimic: mds memory leak
https://github.com/ceph/ceph/pull/22309 Zheng Yan
03:38 AM Backport #24340 (Resolved): mimic: mds memory leak
https://github.com/ceph/ceph/pull/22309 Zheng Yan
03:36 AM Bug #24289 (Pending Backport): mds memory leak
Zheng Yan

05/29/2018

09:08 PM Bug #24284: cephfs: allow prohibiting user snapshots in CephFS
We should actually discuss what kind of interface admins want. Dan van der Ster certainly has thoughts; others might ... Greg Farnum
05:33 PM Feature #24263: client/mds: create a merkle tree of objects to allow efficient generation of diff...
A few questions:
- What is the sha1 of? The object's content? That isn't necessarily known (e.g. 4 MB object whe...
Sage Weil
09:58 AM Bug #22269 (Resolved): ceph-fuse: failure to remount in startup test does not handle client_die_o...
Nathan Cutler
09:58 AM Backport #22378 (Resolved): jewel: ceph-fuse: failure to remount in startup test does not handle ...
Nathan Cutler
09:57 AM Backport #23932 (Resolved): jewel: client: avoid second lock on client_lock
Nathan Cutler
09:45 AM Backport #24189: luminous: qa: kernel_mount.py umount must handle timeout arg
Prashant D wrote:
> This tracker should be closed as duplicate tracker for #24188.
Here's what I see happening he...
Nathan Cutler
09:40 AM Backport #24331 (Resolved): luminous: mon: mds health metrics sent to cluster log indpeendently
https://github.com/ceph/ceph/pull/22558 Nathan Cutler
09:40 AM Backport #24330 (Resolved): mimic: mon: mds health metrics sent to cluster log indpeendently
https://github.com/ceph/ceph/pull/22265 Nathan Cutler

05/28/2018

04:39 PM Feature #24233 (Closed): Add new command ceph mds status
Vikhyat Umrao
04:38 PM Feature #24233: Add new command ceph mds status
Patrick Donnelly wrote:
>
> Why can't this information be from `ceph fs status --format=json`? I'm not really se...
Vikhyat Umrao
03:47 AM Backport #24205 (In Progress): luminous: mds: broadcast quota to relevant clients when quota is e...
https://github.com/ceph/ceph/pull/22271 Prashant D
12:49 AM Bug #24269 (Fix Under Review): multimds pjd open test fails
https://github.com/ceph/ceph/pull/22266 Zheng Yan

05/27/2018

10:20 PM Bug #24308 (Pending Backport): mon: mds health metrics sent to cluster log indpeendently
mimic backport: https://github.com/ceph/ceph/pull/22265 Sage Weil

05/25/2018

08:39 PM Backport #24311 (Resolved): luminous: pjd: cd: too many arguments
https://github.com/ceph/ceph/pull/22883 Nathan Cutler
08:39 PM Backport #24310 (Resolved): mimic: pjd: cd: too many arguments
https://github.com/ceph/ceph/pull/22882 Nathan Cutler
07:03 PM Bug #24307 (Pending Backport): pjd: cd: too many arguments
Josh Durgin
04:35 PM Bug #24307: pjd: cd: too many arguments
https://github.com/ceph/ceph/pull/22233 Neha Ojha
04:21 PM Bug #24307 (Fix Under Review): pjd: cd: too many arguments
-https://github.com/ceph/ceph/pull/22251- Sage Weil
04:20 PM Bug #24307 (Resolved): pjd: cd: too many arguments
... Sage Weil
04:44 PM Bug #24308 (Fix Under Review): mon: mds health metrics sent to cluster log indpeendently
Sage Weil
04:44 PM Bug #24308: mon: mds health metrics sent to cluster log indpeendently
https://github.com/ceph/ceph/pull/22252 Sage Weil
04:42 PM Bug #24308 (Resolved): mon: mds health metrics sent to cluster log indpeendently
We generate a health warning, which has its own logging infrastructure. But MDSMonitor is *also* sending them to wrn... Sage Weil
03:23 PM Bug #24306 (Resolved): mds: use intrusive_ptr to manage Message life-time
We're regularly getting bugs relating to messages not getting released. Latest one is #24289.
Use a boost::intrusi...
Patrick Donnelly
03:10 PM Feature #24233: Add new command ceph mds status
Vikhyat Umrao wrote:
> Thanks John and Patrick for the feedback. I think rename is not needed let us get a new comma...
Patrick Donnelly
02:51 PM Feature #24305 (Resolved): client/mds: allow renaming across quota boundaries
Issue here: https://github.com/ceph/ceph/blob/77b35faa36f83d837a5fe2685efcd4b9be59406a/src/client/Client.cc#L12214-L1... Patrick Donnelly
11:03 AM Backport #24296 (Resolved): mimic: repeated eviction of idle client until some IO happens
https://github.com/ceph/ceph/pull/22550 Nathan Cutler
11:03 AM Backport #24295 (Resolved): luminous: repeated eviction of idle client until some IO happens
https://github.com/ceph/ceph/pull/22780 Nathan Cutler
10:10 AM Feature #24263: client/mds: create a merkle tree of objects to allow efficient generation of diff...
Patrick Donnelly wrote:
> John Spray wrote:
> > I'm a fan. Questions that spring to mind:
> >
> > - Do we apply...
John Spray
08:16 AM Bug #24289 (Fix Under Review): mds memory leak
https://github.com/ceph/ceph/pull/22240 Zheng Yan
08:09 AM Bug #24289 (Resolved): mds memory leak
forget to call message->put() in some cases Zheng Yan
04:05 AM Bug #24052 (Pending Backport): repeated eviction of idle client until some IO happens
Patrick Donnelly
03:07 AM Feature #24286 (Resolved): tools: create CephFS shell
> The Ceph file system (CephFS) provides for kernel driver and FUSE client access. In testing and trivial system admi... Patrick Donnelly
02:54 AM Bug #24240 (Fix Under Review): qa: 1 mutations had unexpected outcomes
https://github.com/ceph/ceph/pull/22234 Zheng Yan
02:31 AM Bug #24284: cephfs: allow prohibiting user snapshots in CephFS
Zheng Yan wrote:
> change default of mds_snap_max_uid to 0
Okay, but we should enforce that as a file system opti...
Patrick Donnelly
02:25 AM Bug #24284: cephfs: allow prohibiting user snapshots in CephFS
change default of mds_snap_max_uid to 0 Zheng Yan
02:28 AM Feature #24285 (Resolved): mgr: add module which displays current usage of file system (`fs top`)
It would ideally provide a list of sessions doing I/O, what kind of I/O, bandwidth of reads/writes, etc. Also the sam... Patrick Donnelly
02:24 AM Documentation #23775 (Resolved): PendingReleaseNotes: add notes for major Mimic features
Patrick Donnelly
02:15 AM Feature #9659 (Duplicate): MDS: support cache eviction
Patrick Donnelly
01:46 AM Bug #23715 (Closed): "Scrubbing terminated -- not all pgs were active and clean" in fs-jewel-dist...
Problem seems to have gone away. Closing. Patrick Donnelly

05/24/2018

10:43 PM Feature #14456: mon: prevent older/incompatible clients from mounting the file system
We're moving this to target 13.2.1. Patrick Donnelly
10:40 PM Documentation #23775: PendingReleaseNotes: add notes for major Mimic features
https://github.com/ceph/ceph/pull/22232 Patrick Donnelly
10:27 PM Bug #24284 (Resolved): cephfs: allow prohibiting user snapshots in CephFS
Since snapshots can be used to circumvent (accidentally or not) the quotas as snapshot file data that has since been ... Patrick Donnelly
09:03 PM Feature #22370 (Resolved): cephfs: add kernel client quota support
Patrick Donnelly
08:28 PM Backport #22378: jewel: ceph-fuse: failure to remount in startup test does not handle client_die_...
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21162
merged
Yuri Weinstein
08:28 PM Backport #23932: jewel: client: avoid second lock on client_lock
Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21734
merged
Yuri Weinstein
07:21 PM Backport #24209 (Resolved): mimic: client: deleted inode's Bufferhead which was in STATE::Tx woul...
Patrick Donnelly
07:21 PM Bug #24111 (Resolved): mds didn't update file's max_size
Patrick Donnelly
07:21 PM Backport #24187 (Resolved): mimic: mds didn't update file's max_size
Patrick Donnelly
07:20 PM Backport #24254 (Resolved): mimic: kceph: umount on evicted client blocks forever
Patrick Donnelly
07:20 PM Backport #24255 (Resolved): mimic: qa: kernel_mount.py umount must handle timeout arg
Patrick Donnelly
07:17 PM Backport #24186 (Resolved): mimic: client: segfault in trim_caps
Patrick Donnelly
07:15 PM Backport #24202 (Resolved): mimic: client: fails to respond cap revoke from non-auth mds
Patrick Donnelly
07:14 PM Backport #24206 (Resolved): mimic: mds: broadcast quota to relevant clients when quota is explici...
Patrick Donnelly
07:14 PM Bug #24118 (Resolved): mds: crash when using `config set` on tracked configs
Patrick Donnelly
07:13 PM Backport #24157 (Resolved): mimic: mds: crash when using `config set` on tracked configs
Patrick Donnelly
07:13 PM Backport #24191 (Resolved): mimic: fs: reduce number of helper debug messages at level 5 for client
Patrick Donnelly
05:05 PM Feature #24263: client/mds: create a merkle tree of objects to allow efficient generation of diff...
John Spray wrote:
> I'm a fan. Questions that spring to mind:
>
> - Do we apply this to all files, or only large...
Patrick Donnelly
09:39 AM Feature #24263: client/mds: create a merkle tree of objects to allow efficient generation of diff...
I'm a fan. Questions that spring to mind:
- Do we apply this to all files, or only large ones based on some heuri...
John Spray
02:11 PM Backport #24201 (In Progress): luminous: client: fails to respond cap revoke from non-auth mds
https://github.com/ceph/ceph/pull/22221 Prashant D
01:47 PM Bug #24240: qa: 1 mutations had unexpected outcomes
The test case corrupted open file table’s omap header. One field in omap header is ‘num_objects’. The corrupted heade... Zheng Yan
01:36 PM Bug #24241: NFS-Ganesha libcephfs: Assert failure in object_locator_to_pg
Patrick Donnelly wrote:
> What version of Ceph are you using?
I run vstart cluster from master (last commit in on...
supriti singh
05:01 AM Bug #24241 (Need More Info): NFS-Ganesha libcephfs: Assert failure in object_locator_to_pg
What version of Ceph are you using? Patrick Donnelly
01:36 PM Feature #24233: Add new command ceph mds status
Thanks John and Patrick for the feedback. I think rename is not needed let us get a new command which can give status... Vikhyat Umrao
09:57 AM Bug #23084 (Resolved): doc: update ceph-fuse with FUSE options
Kefu Chai
09:57 AM Backport #23151 (Resolved): luminous: doc: update ceph-fuse with FUSE options
Kefu Chai
09:49 AM Backport #24189 (In Progress): luminous: qa: kernel_mount.py umount must handle timeout arg
This tracker should be closed as duplicate tracker for #24188. Prashant D
09:42 AM Backport #24188 (In Progress): luminous: kceph: umount on evicted client blocks forever
https://github.com/ceph/ceph/pull/22208 Prashant D
07:48 AM Bug #24269 (Resolved): multimds pjd open test fails
http://qa-proxy.ceph.com/teuthology/pdonnell-2018-05-23_14:53:33-multimds-wip-pdonnell-testing-20180522.181319-mimic-... Zheng Yan
04:53 AM Backport #24185 (In Progress): luminous: client: segfault in trim_caps
Patrick Donnelly
02:21 AM Bug #23972: Ceph MDS Crash from client mounting aufs over cephfs
The crash was at "mdr->tracedn = mdr->dn[ 0].back()", because mdr->dn[ 0] is empty. request that triggered the crash ... Zheng Yan

05/23/2018

08:55 PM Feature #24263 (New): client/mds: create a merkle tree of objects to allow efficient generation o...
Idea is that the collection of objects representing a file would be arranged as a merkle tree. Any write to an object... Patrick Donnelly
07:09 PM Backport #24255 (In Progress): mimic: qa: kernel_mount.py umount must handle timeout arg
Patrick Donnelly
06:31 PM Backport #24255 (Resolved): mimic: qa: kernel_mount.py umount must handle timeout arg
https://github.com/ceph/ceph/pull/22138 Nathan Cutler
07:08 PM Backport #24254 (In Progress): mimic: kceph: umount on evicted client blocks forever
Patrick Donnelly
06:31 PM Backport #24254 (Resolved): mimic: kceph: umount on evicted client blocks forever
https://github.com/ceph/ceph/pull/22138 Nathan Cutler
12:44 PM Backport #24107 (In Progress): luminous: PurgeQueue::_consume() could return true when there were...
https://github.com/ceph/ceph/pull/22176 Prashant D
10:50 AM Feature #24233: Add new command ceph mds status
So I guess Vikhyat is suggesting an "MDS" command to match those for other daemons, but that wouldn't just be a renam... John Spray
08:11 AM Bug #23826 (Duplicate): mds: assert after daemon restart
Checked again, it's likely fixed by https://github.com/ceph/ceph/pull/21883/commits/0a38a499b86c0ee13aa0e783a8359bcce... Zheng Yan
08:08 AM Backport #24108 (In Progress): luminous: MDCache.cc: 5317: FAILED assert(mds->is_rejoin())
Zheng Yan
08:07 AM Backport #24108: luminous: MDCache.cc: 5317: FAILED assert(mds->is_rejoin())
https://github.com/ceph/ceph/pull/22171 Zheng Yan
07:19 AM Backport #23946: luminous: mds: crash when failover
@Nathan @Patrick I have cherry-picked pr21769 as well. Please review pr21900. Prashant D
07:17 AM Bug #24241 (New): NFS-Ganesha libcephfs: Assert failure in object_locator_to_pg
When calling ceph_ll_get_stripe_osd from nfs-ganesha fsal ceph in file mds.c, assertion failure causes segmentation f... supriti singh
03:46 AM Backport #24207 (In Progress): luminous: client: deleted inode's Bufferhead which was in STATE::T...
https://github.com/ceph/ceph/pull/22168 Zheng Yan
03:02 AM Bug #24239 (Fix Under Review): cephfs-journal-tool: Importing a zero-length purge_queue journal b...
https://github.com/ceph/ceph/pull/22144 Patrick Donnelly
02:33 AM Bug #24239 (Resolved): cephfs-journal-tool: Importing a zero-length purge_queue journal breaks it...
When we were importing a zero-length purge_queue journal exported previously, the last object and
the following one ...
yupeng chen
02:57 AM Bug #24236 (Fix Under Review): cephfs-journal-tool: journal inspect reports DAMAGED for purge que...
https://github.com/ceph/ceph/pull/22146 Patrick Donnelly
02:19 AM Bug #24236: cephfs-journal-tool: journal inspect reports DAMAGED for purge queue when it's empty
https://github.com/ceph/ceph/pull/22146
before fix, in a new created cluster fs, run:
$cephfs-journal-tool --jour...
cory gu
02:18 AM Bug #24236 (Fix Under Review): cephfs-journal-tool: journal inspect reports DAMAGED for purge que...
When purge queue is empty, joural inspect still report DAMAGED
journal integrity.
cory gu
02:46 AM Bug #24240 (Resolved): qa: 1 mutations had unexpected outcomes
... Patrick Donnelly
02:46 AM Bug #24238 (Fix Under Review): test gets ENOSPC from bluestore block device
https://github.com/ceph/ceph/pull/22165 Sage Weil
02:43 AM Bug #24238: test gets ENOSPC from bluestore block device
The underlying block device was thinly provisioned and ran out of space. Asserting on ENOSPC from a block IO is expe... Sage Weil
02:42 AM Bug #24238: test gets ENOSPC from bluestore block device
Sage Weil
02:32 AM Bug #24238 (Resolved): test gets ENOSPC from bluestore block device
... Patrick Donnelly
01:36 AM Feature #22372 (Resolved): kclient: implement quota handling using new QuotaRealm
Zheng Yan
 

Also available in: Atom