Activity
From 10/09/2016 to 11/07/2016
11/07/2016
- 02:29 PM Bug #17799: cephfs-data-scan: doesn't know how to handle files with pool_namespace layouts
- the backtrace objects are always in default namespace. I think data scan tool can't calculate correct size for files ...
- 02:25 PM Bug #17801 (Fix Under Review): Cleanly reject "session evict" command when in replay
- https://github.com/ceph/ceph/pull/11813
- 11:43 AM Bug #17193 (Resolved): truncate can cause unflushed snapshot data lose
- This is no longer failing when running against the testing kernel.
11/04/2016
- 02:32 PM Bug #17531 (Pending Backport): mds fails to respawn if executable has changed
- 02:13 PM Bug #17801 (Resolved): Cleanly reject "session evict" command when in replay
Currently we crash like this (from ceph-users):...- 02:07 PM Fix #15134 (Fix Under Review): multifs: test case exercising mds_thrash for multiple filesystems
- PR adding support to mds_thrash.py: https://github.com/ceph/ceph-qa-suite/pull/1175
Need to check if we have a tes... - 02:06 PM Feature #10792 (In Progress): qa: enable thrasher for MDS cluster size (vary max_mds)
- Pre-requisite PR: https://github.com/ceph/ceph-qa-suite/pull/1175
multimds testing with the thrasher will be added... - 01:41 PM Bug #17800 (Resolved): ceph_volume_client.py : Error: Can't handle arrays of non-strings
- When using Ceph (python-cephfs-10.2.3+git.1475228057.755cf99 , SLE12SP2) together with OpenStack Manila and trying to...
- 01:18 PM Bug #17799 (New): cephfs-data-scan: doesn't know how to handle files with pool_namespace layouts
Not actually sure how we currently behave.
Do we see the data objects and inject files with incorrect layouts? ...- 01:06 PM Bug #17798 (Resolved): Clients without pool-changing caps shouldn't be allowed to change pool_nam...
The purpose of the 'p' flag in MDS client auth caps is to enable creating clients that cannot set the pool part of ...- 01:01 PM Bug #17797 (Fix Under Review): rmxattr on ceph.[dir|file].layout.pool_namespace doesn't work
- https://github.com/ceph/ceph/pull/11783
- 11:05 AM Bug #17797 (Resolved): rmxattr on ceph.[dir|file].layout.pool_namespace doesn't work
Currently it's obvious how to set the namespace but much less so how to clear it (i.e. revert to default namespace)...
11/02/2016
- 08:22 AM Bug #17747 (Fix Under Review): ceph-mds: remove "--journal-check" help text
- *master PR*: https://github.com/ceph/ceph/pull/11739
11/01/2016
- 05:16 PM Bug #17563: extremely slow ceph_fsync calls
- PR to fix the userland side of things is here:
https://github.com/ceph/ceph/pull/11710 - 01:10 PM Bug #17563: extremely slow ceph_fsync calls
- It looks like ceph-fuse has the same problem with fsync. Here's a POSIX API reproducer that shows similar improvement...
- 03:53 PM Bug #17115 (Resolved): kernel panic when running IO with cephfs and resource pool becomes full
- http://tracker.ceph.com/issues/17770
- 03:53 PM Feature #17770 (New): qa: test kernel client against "full" pools/filesystems
- We test the uclient against full pools to validate behavior. We discovered in #17115 that we don't for the kernel cli...
- 03:49 PM Bug #17240 (Closed): inode_permission error with kclient when running client IO with recovery ope...
- When the RADOS cluster has blocked IO, the kernel client is going to have blocked IO. That's just life. :(
- 03:37 PM Bug #17656 (Need More Info): cephfs: high concurrent causing slow request
- Just from the description it sounds like we're backing up while the MDS purges deleted files from RADOS. You can adju...
- 03:08 PM Bug #7750: Attempting to mount a kNFS export of a sub-directory of a CephFS filesystem fails with...
- NFS-Utils-1.2.8
NFS server on Ubuntu 16.04 - 03:06 PM Bug #7750: Attempting to mount a kNFS export of a sub-directory of a CephFS filesystem fails with...
- I still can't reproduce it. This does not seem like kernel issue. which version of nfs-utils do you use?
- 12:13 PM Bug #7750: Attempting to mount a kNFS export of a sub-directory of a CephFS filesystem fails with...
- I've reproduced this bug in CephFS jewel. Can't mount via NFS NON root CephFS dir.
Get error on client 'Stale NFS fi... - 09:24 AM Bug #17747: ceph-mds: remove "--journal-check" help text
- You should use the cephfs-journal-tool for dealing with this stuff now. The journal-check oneshot-replay mode got rem...
- 08:39 AM Bug #17620: Data Integrity Issue with kernel client vs fuse client
- I'm not familiar with docker,how did the jobs failed.(what's the symptom)
10/31/2016
- 06:29 PM Bug #17620: Data Integrity Issue with kernel client vs fuse client
- Thanks for that. I'm working on getting to the point where I can test that.
In the meantime further testing has i... - 05:05 PM Bug #4212: mds: open_snap_parents isn't called all the times it needs to be
- Having all past_parents open is hard because of dir renames. Say you do
/a/b/c
and snapshot /a/b, then rename... - 02:06 PM Bug #17747: ceph-mds: remove "--journal-check" help text
- running ceph-mds -d -i ceph --journal-check 0 gives me following output:
--conf/-c FILE read configuration fr... - 09:51 AM Bug #17747 (Resolved): ceph-mds: remove "--journal-check" help text
- Hi,
running ceph-mds -d -i ceph --journal-check 0 gives me following input:
--conf/-c FILE read configurat...
10/28/2016
- 05:31 PM Bug #17563: extremely slow ceph_fsync calls
- Ok, thanks. That makes sense. I've got a patchset that works as a PoC, but it's pretty ugly and could use some cleanu...
- 11:23 AM Bug #17548 (Resolved): should userland ceph_llseek do permission checking?
- Fixed in commit db2e7e0811679b4c284e105536ebf3327cc02ffc.
- 10:35 AM Bug #17731 (Can't reproduce): MDS stuck in stopping with other rank's strays
Kraken v11.0.2
Seen on a max_mds=2 MDS cluster with a fuse client doing an rsync -av --delete on a dir that incl...
10/27/2016
- 07:50 AM Backport #17720 (Resolved): jewel: MDS: false "failing to respond to cache pressure" warning
- https://github.com/ceph/ceph/pull/11856
10/26/2016
- 01:37 PM Bug #17562 (Resolved): backtrace check fails when scrubbing directory created by fsstress
- 12:44 PM Bug #17716 (Resolved): MDS: false "failing to respond to cache pressure" warning
- Creating this ticket for a PR that went in without a ticket on it so that we can backport.
https://github.com/ceph... - 07:56 AM Backport #17705: jewel: ceph_volume_client: recovery of partial auth update is broken
- h3. previous description
I run into the following traceback when the volume_client tries
to recover from partia... - 07:18 AM Backport #17705: jewel: ceph_volume_client: recovery of partial auth update is broken
- https://github.com/ceph/ceph/pull/11656
https://github.com/ceph/ceph-qa-suite/pull/1221 - 05:21 AM Backport #17705 (In Progress): jewel: ceph_volume_client: recovery of partial auth update is broken
- 05:20 AM Backport #17705 (Resolved): jewel: ceph_volume_client: recovery of partial auth update is broken
- https://github.com/ceph/ceph/pull/11656
- 07:07 AM Backport #17706 (Resolved): jewel: multimds: mds entering up:replay and processing down mds aborts
- https://github.com/ceph/ceph/pull/11857
- 05:16 AM Bug #17216 (Pending Backport): ceph_volume_client: recovery of partial auth update is broken
10/25/2016
- 07:53 PM Bug #17670 (Pending Backport): multimds: mds entering up:replay and processing down mds aborts
- 01:46 PM Backport #17697 (Resolved): jewel: MDS long-time blocked ops. ceph-fuse locks up with getattr of ...
- https://github.com/ceph/ceph/pull/11858
- 01:42 PM Bug #17620: Data Integrity Issue with kernel client vs fuse client
- I have fixed a bug that may cause this issue. could you have a try https://github.com/ceph/ceph-client/commits/testing
- 11:16 AM Bug #17275 (Pending Backport): MDS long-time blocked ops. ceph-fuse locks up with getattr of file
- 11:12 AM Bug #17691 (Resolved): bad backtrace on inode
- Merged https://github.com/ceph/ceph-qa-suite/pull/1218
- 10:46 AM Bug #17691 (In Progress): bad backtrace on inode
- Sorry, that's happening because I merged my backtrace repair PR before the ceph-qa-suite piece, so the log message is...
- 03:08 AM Bug #17691 (Resolved): bad backtrace on inode
- Seen this in testing:
http://pulpito.ceph.com/pdonnell-2016-10-25_02:25:11-fs:recovery-master---basic-mira/493889/...
10/24/2016
- 05:49 PM Bug #17563: extremely slow ceph_fsync calls
- The client is waiting for an ack to a cap *flush*, not to get caps granted. Usually flushes happen asynchronously (ju...
- 05:36 PM Bug #17563: extremely slow ceph_fsync calls
- OTOH...do we even need a flag at all here? Under what circumstances is it beneficial to delay granting and recalling ...
- 01:45 PM Feature #17639 (Resolved): Repair file backtraces during forward scrub
10/22/2016
- 10:52 PM Bug #17670 (Fix Under Review): multimds: mds entering up:replay and processing down mds aborts
- https://github.com/ceph/ceph/pull/11611
- 10:46 PM Bug #17670 (Resolved): multimds: mds entering up:replay and processing down mds aborts
- ...
10/21/2016
- 09:48 PM Feature #17249 (Resolved): cephfs tool for finding files that use named PGs
- 01:58 PM Bug #17620: Data Integrity Issue with kernel client vs fuse client
- I suspect the zeros are from stale page cache data. If you encounter the issue again, please drop the kernel page cac...
- 01:39 PM Bug #17275 (Fix Under Review): MDS long-time blocked ops. ceph-fuse locks up with getattr of file
- https://github.com/ceph/ceph/pull/11593
- 01:21 PM Bug #17275: MDS long-time blocked ops. ceph-fuse locks up with getattr of file
- created http://tracker.ceph.com/issues/17660
- 01:03 PM Bug #17275: MDS long-time blocked ops. ceph-fuse locks up with getattr of file
- got a getattr long lock, but this time client has long-running objecter requests. I will be filling a ticket for that...
- 05:33 AM Bug #17656: cephfs: high concurrent causing slow request
- William, could you describe the issue from the Ceph's perspective in detail?
- 01:46 AM Bug #17656 (Need More Info): cephfs: high concurrent causing slow request
- background:
we use cephfs as CDN backend, when CDN vendor prefetch video files in cephfs, it will cause high concu...
10/20/2016
- 08:44 PM Bug #17620: Data Integrity Issue with kernel client vs fuse client
- On further testing, it seems I can only make this happen when doing multi-threaded downloads from multiple hosts. I h...
- 04:39 PM Bug #17591 (Resolved): Samba failures: smbtorture, dbench.sh, fsstress.sh
- 02:53 PM Bug #17636 (Resolved): MDS crash on creating: interval_set<inodeno_t> segfaults with new encoding
- 10:31 AM Bug #17636 (Fix Under Review): MDS crash on creating: interval_set<inodeno_t> segfaults with new ...
- https://github.com/ceph/ceph/pull/11577
- 10:23 AM Bug #17636: MDS crash on creating: interval_set<inodeno_t> segfaults with new encoding
- Should also note that the backtraces are like:...
- 09:35 AM Bug #17636 (Resolved): MDS crash on creating: interval_set<inodeno_t> segfaults with new encoding
The map bound_encode method does this:...- 02:43 PM Feature #17643 (New): Verify and repair layout xattr during forward scrub
- We should do exactly the same thing we do for backtrace, but for the layout that we write into 'layout' xattr. Repai...
- 02:36 PM Feature #17639 (Fix Under Review): Repair file backtraces during forward scrub
- https://github.com/ceph/ceph/pull/11578
https://github.com/ceph/ceph-qa-suite/pull/1218 - 02:31 PM Feature #17639 (Resolved): Repair file backtraces during forward scrub
Instead of just recording that a backtrace was missing or invalid, repair it in the same way we do in verify_diri_b...- 02:22 PM Feature #15619 (Resolved): Repair InoTable during forward scrub
- ...
- 10:38 AM Bug #17216 (Resolved): ceph_volume_client: recovery of partial auth update is broken
- 08:32 AM Bug #17216 (Pending Backport): ceph_volume_client: recovery of partial auth update is broken
- 09:35 AM Bug #17606 (Need More Info): multimds: assertion failure during directory migration
- debug level of the log is too low. This is dup of a long-standing bug http://tracker.ceph.com/issues/8405. I can't fi...
- 08:43 AM Bug #17629 (Duplicate): ceph_volume_client: TypeError with wrong argument count
- The fix for this just merged in master yesterday.
Python says "2 args, 3 given" because the 3 is counting the "sel... - 12:33 AM Bug #17629 (Duplicate): ceph_volume_client: TypeError with wrong argument count
- From: http://pulpito.ceph.com/pdonnell-2016-10-19_23:26:58-fs:recovery-master---basic-mira/486359/...
- 12:43 AM Bug #14195: test_full_fclose fails (tasks.cephfs.test_full.TestClusterFull)
- I just ran across this in http://pulpito.ceph.com/pdonnell-2016-10-19_23:26:58-fs:recovery-master---basic-mira/486355...
10/19/2016
- 05:39 PM Bug #17621 (Rejected): Hadoop does a bad job closing files and we end up holding too many caps
- (Just scribbling this in the tracker for a record.)
We've had several reports on the mailing list of Hadoop client... - 05:32 PM Feature #12141: cephfs-data-scan: File size correction from backward scan
- Unassigning as I believe this was dropped?
- 04:48 PM Bug #17620: Data Integrity Issue with kernel client vs fuse client
- No I only see:
Oct 19 15:34:52 phx-r2-r3-comp5 kernel: [683656.055247] libceph: client259809 fsid <redacted>
Oc... - 04:28 PM Bug #17620: Data Integrity Issue with kernel client vs fuse client
- Were there any ceph related kernel message on the client hosts? (such as "ceph: mds0 caps stale")
- 03:49 PM Bug #17620 (Resolved): Data Integrity Issue with kernel client vs fuse client
- ceph version 10.2.3 (ecc23778eb545d8dd55e2e4735b53cc93f92e65b)
kernel: 4.4.0-42-generic #62~14.04.1-Ubuntu SMP
I ... - 03:07 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- The cephx user, not the *nix user. I'd be fine with the other stuff — we already run (all our?) nightlies with our ow...
- 03:04 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- I don't quite understand. The reproducer that does the ceph_ll_mkdir does it as an unprivileged user, so why is it al...
- 02:46 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- There we go, that case would concern me.
Contrary to our discussion in standup today, I think MDSAuthCaps::is_capa... - 12:25 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- No, the order has already been established, because each operation _really_ starts with the permission check. But ok,...
- 04:08 AM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- No, I mean: I don't think this counts as a bug. To get in this situation we have two outstanding requests:
1) chmod
... - 02:54 PM Backport #17617 (Resolved): jewel: [cephfs] fuse client crash when adding a new osd
- https://github.com/ceph/ceph/pull/11860
- 02:53 PM Backport #17615 (Resolved): jewel: mds: false "failing to respond to cache pressure" warning
- https://github.com/ceph/ceph/pull/11861
- 01:14 PM Bug #17606: multimds: assertion failure during directory migration
- Fixed!
- 03:49 AM Bug #17606: multimds: assertion failure during directory migration
- I can't access /ceph/cephfs-perf/tmp/2016-10-18/ceph-mds.ceph-mds0.log. please change permission
- 11:56 AM Feature #16016 (Resolved): Populate DamageTable from forward scrub
- 10:24 AM Bug #17270 (Pending Backport): [cephfs] fuse client crash when adding a new osd
- 10:21 AM Bug #17611 (Resolved): mds: false "failing to respond to cache pressure" warning
Creating this ticket retroactively for a ticketless PR that needs backporting: https://github.com/ceph/ceph/pull/11373- 07:04 AM Bug #17172 (Resolved): Failure in snaptest-git-ceph.sh
10/18/2016
- 11:35 PM Feature #17604: MDSMonitor: raise health warning when there are no standbys but there should be
- I'll take this one.
- 11:44 AM Feature #17604 (Resolved): MDSMonitor: raise health warning when there are no standbys but there ...
We could have a per-filesystem setting for how many standbys it wants to be available. Then if the available set o...- 09:22 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- FWIW, "little holes" like this are worse in that they'll often strike when things are really redlined. If the race wi...
- 09:02 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- The client has to send out both a cap drop message and the user's request.
Those will be ordered, and if the cap dro... - 08:34 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- Will they be ordered after the reply comes in, or just ordered wrt how they go out onto the wire?
IOW, is there so... - 08:13 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- Luckily the messages are ordered so once a client sends off a request to the MDS, any caps it drops will be ordered a...
- 07:22 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- I don't think that's sufficient.
My understanding is that once you drop the client_lock mutex (as the client does ... - 06:54 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- The MDS does enforce cephx-based permissions, but unless you've deliberately constructed a limited cephx key it's goi...
- 04:27 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- Ok, and the reason setting fuse_default_permissions didn't work for me yesterday was due to a typo. When I set that o...
- 03:26 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- Ok, here's a "backported" test program. Both mkdirs still succeed, but when I set fuse_default_permissions to false o...
- 01:02 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- No, I don't think so. Here is ceph_userperm_new in my tree:...
- 03:22 PM Bug #17606 (Resolved): multimds: assertion failure during directory migration
- This is from an experiment on Linode with 9 active MDS, 32 OSD, and 128 clients building the kernel....
- 09:47 AM Bug #17591 (Fix Under Review): Samba failures: smbtorture, dbench.sh, fsstress.sh
- https://github.com/ceph/ceph/pull/11526
- 09:23 AM Bug #17591 (In Progress): Samba failures: smbtorture, dbench.sh, fsstress.sh
- samba daemon changes its own euid/egid before calling libcephfs functions. But libcephfs setup a default_perm on init...
- 02:44 AM Bug #17563: extremely slow ceph_fsync calls
- No method so far, need to extend the MClientCaps message
10/17/2016
- 09:45 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- And at a guess, you might be using the uid,gid I presume you have there (0,0) to create folders with the ownership sp...
- 09:39 PM Bug #17594: cephfs: permission checking not working (MDS should enforce POSIX permissions)
- Quick sanity check: is fuse_default_permissions set to false? Otherwise the assumption is that fuse (or nfs, samba) ...
- 09:02 PM Bug #17594 (In Progress): cephfs: permission checking not working (MDS should enforce POSIX permi...
- Frank Filz noticed that cephfs doesn't seem to be enforcing permissions properly, particularly on mkdir.
This test... - 05:06 PM Bug #17591: Samba failures: smbtorture, dbench.sh, fsstress.sh
- Yep, still failing on master.
- 02:23 PM Bug #17591: Samba failures: smbtorture, dbench.sh, fsstress.sh
- Rerun of failures on master (a1fd258)
http://pulpito.ceph.com/jspray-2016-10-17_14:22:43-samba-master---basic-mira/ - 02:19 PM Bug #17591: Samba failures: smbtorture, dbench.sh, fsstress.sh
- Failures are on a4ce1f56c1e22d4d3c190403339d5082d66fc89c, last passes were on cfc0a16e048a5868b137e2b7b89c7ae105218bc...
- 02:16 PM Bug #17591 (Resolved): Samba failures: smbtorture, dbench.sh, fsstress.sh
Samba tests were some of the most stable, they are suddenly failing on master.
http://pulpito.ceph.com/teutholog...- 03:14 PM Bug #17286 (Resolved): Failure in dirfrag.sh
- 03:14 PM Bug #17271 (Resolved): Failure in snaptest-git-ceph.sh
- 03:14 PM Bug #17253 (Resolved): Crash in Client::_invalidate_kernel_dcache when reconnecting during unmount
- 03:14 PM Bug #17173 (Resolved): Duplicate damage table entries
- 03:14 PM Feature #16973 (Resolved): Log path as well as ino when detecting metadata damage
- 03:14 PM Bug #16668 (Resolved): client: nlink count is not maintained correctly
- 02:38 PM Bug #17547 (Resolved): ceph-fuse 10.2.3 segfault
- 02:32 PM Bug #17547: ceph-fuse 10.2.3 segfault
- looks good so far. I guess we can close this. Thanks!
- 01:42 PM Bug #17547: ceph-fuse 10.2.3 segfault
- Henrik: did that build fix the issue for you?
- 12:31 PM Backport #16946 (Resolved): jewel: client: nlink count is not maintained correctly
- 12:30 PM Backport #17244 (Resolved): jewel: Failure in snaptest-git-ceph.sh
- 12:30 PM Backport #17246 (Resolved): jewel: Log path as well as ino when detecting metadata damage
- 12:30 PM Backport #17474 (Resolved): jewel: Failure in dirfrag.sh
- 12:30 PM Backport #17476 (Resolved): jewel: Failure in snaptest-git-ceph.sh
- 12:30 PM Backport #17477 (Resolved): jewel: Crash in Client::_invalidate_kernel_dcache when reconnecting d...
- 12:30 PM Backport #17479 (Resolved): jewel: Duplicate damage table entries
- 09:06 AM Bug #17562 (Fix Under Review): backtrace check fails when scrubbing directory created by fsstress
- https://github.com/ceph/ceph/pull/11517/
10/16/2016
- 10:45 AM Bug #17563: extremely slow ceph_fsync calls
- Yep, exactly. I noticed this while running the cthon04 testsuite against it. It copies some source files into place a...
10/15/2016
- 10:06 PM Bug #17564: close race window when handling writes on a file descriptor opened with O_APPEND
- Good point. Yeah, a retry loop may be the best we can do in that case. I'll have to read up on rados asserts to make ...
- 01:16 AM Bug #17564: close race window when handling writes on a file descriptor opened with O_APPEND
- Are we *allowed* to fail writes like that? :/
It doesn't actually close the race in all cases, but for sane use it... - 09:39 AM Backport #17582 (Resolved): jewel: monitor assertion failure when deactivating mds in (invalid) f...
- https://github.com/ceph/ceph/pull/11862
- 09:34 AM Bug #16610 (Resolved): Jewel: segfault in ObjectCacher::FlusherThread
- 08:28 AM Bug #17468: CephFs: IO Pauses for more than a 40 seconds, while running write intensive IOs
- Further analysis of I/O drop points to correlation of MDS config parameter mds_log_max_segments and I/O drop. Using d...
10/14/2016
- 11:39 PM Bug #17563: extremely slow ceph_fsync calls
- Oh right, this doesn't even have a way to ask the MDS for immediate service — notice how it just puts already-sent re...
- 11:27 PM Bug #17562: backtrace check fails when scrubbing directory created by fsstress
- It looks to me like compare() is doing the right thing, but that the scrubbing code is declaring an error if the on-d...
- 11:19 PM Bug #17562: backtrace check fails when scrubbing directory created by fsstress
- Well, in inode_backtrace_t::compare() we're trying to determine and return three different things, as described in th...
- 08:06 PM Backport #17131 (Resolved): jewel: Jewel: segfault in ObjectCacher::FlusherThread
- 05:06 PM Bug #16255 (Resolved): ceph-create-keys: sometimes blocks forever if mds "allow" is set
- 02:26 PM Backport #17347 (Resolved): jewel: ceph-create-keys: sometimes blocks forever if mds "allow" is set
- 10:04 AM Bug #17518 (Pending Backport): monitor assertion failure when deactivating mds in (invalid) fscid 0
- 07:10 AM Bug #17525 (Duplicate): "[ FAILED ] LibCephFS.ThreesomeInterProcessRecordLocking" in smoke
- 07:03 AM Bug #17525: "[ FAILED ] LibCephFS.ThreesomeInterProcessRecordLocking" in smoke
- In jewel http://pulpito.ceph.com/loic-2016-10-13_17:04:37-fs-jewel-backports-distro-basic-smithi/471638/...
10/13/2016
- 07:00 PM Bug #17563: extremely slow ceph_fsync calls
- This program is a reproducer. You can build it with something like ...
- 01:10 PM Bug #17563: extremely slow ceph_fsync calls
- Is that something we can change? Slow fsync() performance is particularly awful for applications.
In any case, the... - 11:46 AM Bug #17563: extremely slow ceph_fsync calls
- The reason is that client fsync does not force MDS flush its journal. fsync may wait up to a MDS tick if there is no ...
- 10:22 AM Bug #17563 (Resolved): extremely slow ceph_fsync calls
- I've been seeing problems with very slow fsyncs vs. a vstart cluster when I run ganesha on top of it. This is the gan...
- 03:51 PM Bug #17466 (Resolved): MDSMonitor: non-existent standby_for_fscid not caught
- 03:51 PM Bug #17197 (Resolved): ceph-fuse crash in Client::get_root_ino
- 03:51 PM Bug #17105 (Resolved): multimds: allow_multimds not required when max_mds is set in ceph.conf at ...
- 03:51 PM Bug #16764 (Resolved): ceph-fuse crash on force unmount with file open
- 03:51 PM Bug #16066 (Resolved): client: FAILED assert(root_ancestor->qtree == __null)
- 01:27 PM Bug #17564: close race window when handling writes on a file descriptor opened with O_APPEND
- For the record -- I really don't care for O_APPEND semantics, but...
_write() does this currently:... - 11:41 AM Bug #17564: close race window when handling writes on a file descriptor opened with O_APPEND
- Not that easy to do. MDS does not always issue caps what client wants. In this case, client wants Fsw, but MDS may on...
- 10:59 AM Bug #17564 (New): close race window when handling writes on a file descriptor opened with O_APPEND
- This comment is in _write() in the userland client code:...
- 01:02 PM Backport #16313 (Resolved): jewel: client: FAILED assert(root_ancestor->qtree == __null)
- 12:45 PM Bug #17468: CephFs: IO Pauses for more than a 40 seconds, while running write intensive IOs
- John, there are two test cases.
1. Load files into file system. We created 900k files of 1k size
2. Run I/O tests s... - 09:15 AM Bug #17562 (Resolved): backtrace check fails when scrubbing directory created by fsstress
- when scrubbing created by fsstress, there are lots of false backtrace errors. The bug is in inode_backtrace_t::compar...
- 06:54 AM Backport #17207 (Resolved): jewel: ceph-fuse crash on force unmount with file open
- 06:44 AM Backport #17206 (Resolved): jewel: ceph-fuse crash in Client::get_root_ino
- 06:44 AM Backport #17264 (Resolved): jewel: multimds: allow_multimds not required when max_mds is set in c...
- 06:43 AM Backport #17557 (Resolved): jewel: MDSMonitor: non-existent standby_for_fscid not caught
10/12/2016
- 04:25 PM Feature #17276 (Resolved): stick client PID in client_metadata
- 04:22 PM Bug #17531 (Resolved): mds fails to respawn if executable has changed
10/11/2016
- 08:57 AM Backport #17244 (In Progress): jewel: Failure in snaptest-git-ceph.sh
- 08:56 AM Backport #17246 (In Progress): jewel: Log path as well as ino when detecting metadata damage
- 08:52 AM Backport #17347 (In Progress): jewel: ceph-create-keys: sometimes blocks forever if mds "allow" i...
- 08:51 AM Backport #17474 (In Progress): jewel: Failure in dirfrag.sh
- 08:50 AM Backport #17476 (In Progress): jewel: Failure in snaptest-git-ceph.sh
- 08:45 AM Backport #17477 (In Progress): jewel: Crash in Client::_invalidate_kernel_dcache when reconnectin...
- 08:40 AM Backport #17478 (In Progress): jewel: MDS goes damaged on blacklist (failed to read JournalPointe...
- 08:34 AM Backport #17479 (In Progress): jewel: Duplicate damage table entries
- 08:32 AM Backport #17557 (In Progress): jewel: MDSMonitor: non-existent standby_for_fscid not caught
- 08:08 AM Backport #17557 (Resolved): jewel: MDSMonitor: non-existent standby_for_fscid not caught
- https://github.com/ceph/ceph/pull/11389
10/10/2016
- 09:15 PM Bug #17548: should userland ceph_llseek do permission checking?
- ...
- 08:47 PM Bug #17548: should userland ceph_llseek do permission checking?
- Actually, now I'm confused. What's the failing test, if this isn't a case users should run into anyway?
- 08:42 PM Bug #17548: should userland ceph_llseek do permission checking?
- There are several ways, but yeah...it comes down to being careful to close out old file descriptors (or use O_CLOEXEC...
- 08:40 PM Bug #17548: should userland ceph_llseek do permission checking?
- Yeah, sorry. I know it's possible for a process to drop permissions somehow or other.
- 08:36 PM Bug #17548: should userland ceph_llseek do permission checking?
- That's just the way UNIX (and hence POSIX) works. UNIX pipelines require file descriptor inheritance.
I'm not sure... - 08:12 PM Bug #17548: should userland ceph_llseek do permission checking?
- But...isn't that insecure by design? Why should we assume an open FD is still valid on a setuid'ed fork(), for instance?
- 08:07 PM Bug #17548: should userland ceph_llseek do permission checking?
- So does fstat, and we don't do any special permission checking there either.
I guess my view is that if you're iss... - 07:56 PM Bug #17548: should userland ceph_llseek do permission checking?
- We probably want to follow the kernel. But when I read the man pages and think about security, it seems like it shoul...
- 03:53 PM Bug #17548: should userland ceph_llseek do permission checking?
- Seems reasonable to remove.
- 11:48 AM Bug #17548 (Resolved): should userland ceph_llseek do permission checking?
- One of the test failures here:
http://qa-proxy.ceph.com/teuthology/jlayton-2016-10-08_00:13:32-fs-wip-jlayton-... - 01:51 PM Bug #17468: CephFs: IO Pauses for more than a 40 seconds, while running write intensive IOs
- Vishal: are you still investigating this or do you need some input from others?
- 01:46 PM Bug #17522 (In Progress): ceph_readdirplus_r does not acquire caps before sending back attributes
- 12:35 PM Bug #17547: ceph-fuse 10.2.3 segfault
- I am preparing new build with a patch from https://github.com/ceph/ceph/pull/10921
- 12:12 PM Bug #17547: ceph-fuse 10.2.3 segfault
- It is happening on 10.2.3 cluster too (just checked another cluster). It's just that I took this segfault from client...
- 12:11 PM Bug #17547: ceph-fuse 10.2.3 segfault
- Seems likely to be same issue fixed by https://github.com/ceph/ceph/pull/10921 (i.e. duplicate of http://tracker.ceph...
- 12:10 PM Bug #17547: ceph-fuse 10.2.3 segfault
- > ceph-fuse 10.2.3 segfaults on 10.2.2 Ceph cluster.
Are you asserting that it *doesn't* segfault on a 10.2.3 ceph... - 11:38 AM Bug #17547 (Resolved): ceph-fuse 10.2.3 segfault
- ceph-fuse 10.2.3 segfaults on 10.2.2 Ceph cluster.
ceph-fuse has #17275 and https://github.com/ceph/ceph/pull/1086... - 11:24 AM Bug #17466 (Pending Backport): MDSMonitor: non-existent standby_for_fscid not caught
- Jewel backport PR here: https://github.com/ceph/ceph/pull/11389
Also available in: Atom