Activity
From 12/19/2016 to 01/17/2017
01/17/2017
- 05:38 PM Bug #18574 (Pending Backport): cephfs test failures (ceph.com/qa is broken, should be download.ce...
- 04:16 PM Bug #18574 (Fix Under Review): cephfs test failures (ceph.com/qa is broken, should be download.ce...
- https://github.com/ceph/ceph/pull/12964
- 04:14 PM Bug #18574 (Resolved): cephfs test failures (ceph.com/qa is broken, should be download.ceph.com/qa)
- Like http://tracker.ceph.com/issues/18542 but for the remaining references to ceph.com.
- 03:13 PM Bug #18532: mds: forward scrub failing to repair dir stats (was: subdir with corrupted dirstat is...
- maybe there is a bad remote link in the directory
- 11:13 AM Bug #16842: mds: replacement MDS crashes on InoTable release
- Hmm, this one got lost somehow, targeting to Luminous so that it gets at least looked at.
- 08:35 AM Backport #18566 (Resolved): kraken: MDS crashes on missing metadata object
- https://github.com/ceph/ceph/pull/14565
- 08:35 AM Backport #18565 (Resolved): jewel: MDS crashes on missing metadata object
- https://github.com/ceph/ceph/pull/13119
- 08:35 AM Backport #18562 (Resolved): kraken: Test Failure: kcephfs test_client_recovery.TestClientRecovery
- https://github.com/ceph/ceph/pull/14564
- 08:34 AM Backport #18552 (Resolved): kraken: ceph-fuse crash during snapshot tests
- https://github.com/ceph/ceph/pull/14563
- 08:34 AM Backport #18551 (Resolved): jewel: ceph-fuse crash during snapshot tests
- https://github.com/ceph/ceph/pull/13120
01/16/2017
- 09:45 PM Backport #18540 (Fix Under Review): kraken: Test failure: test_session_reject (tasks.cephfs.test_...
- 09:45 PM Backport #18540 (Resolved): kraken: Test failure: test_session_reject (tasks.cephfs.test_sessionm...
- https://github.com/ceph/ceph/pull/12951
- 03:09 PM Bug #18363 (Can't reproduce): Test failure: test_ops_throttle (tasks.cephfs.test_strays.TestStrays)
- This appears to have been something intermittent, and the deletion code is going to change anyway so I'm going to ski...
- 03:08 PM Bug #18362 (Duplicate): Test failure: test_evict_client (tasks.cephfs.test_volume_client.TestVolu...
- 02:18 PM Feature #18537: libcephfs cache invalidation upcalls
- My intent in this experiment was, yes, that upcalls be synchronous. To be brief I'd say that my understanding was, if...
- 12:42 PM Feature #18537: libcephfs cache invalidation upcalls
- This really seems like the wrong approach to me. Is this callback going to be synchronous or async? Imagine you get a...
- 12:24 PM Feature #18537 (Rejected): libcephfs cache invalidation upcalls
- Matt Benjamin did some work in this area:
https://github.com/linuxbox2/nfs-ganesha/tree/ceph-invalidates
https://... - 12:24 PM Feature #18490: client: implement delegation support in userland cephfs
- Ah, I think I had (incorrectly) assumed that the work Matt did on invalidations before had been merged, but if that's...
- 12:19 PM Bug #18530: ceph tell mds prints warning about ms_handle_reset
- I think the OSD does do roughly the same sort of thing (hidden inside Objecter), so it might be instructive to look a...
- 12:15 PM Bug #18532: mds: forward scrub failing to repair dir stats (was: subdir with corrupted dirstat is...
- Without having looked into this in detail yet, my presumption would be that the bug is that the repair code isn't fix...
01/15/2017
- 03:39 PM Feature #17855 (Fix Under Review): Don't evict a slow client if it's the only client
- https://github.com/ceph/ceph/pull/12935
01/14/2017
- 01:33 AM Bug #18532 (New): mds: forward scrub failing to repair dir stats (was: subdir with corrupted dirs...
- Somehow a path in the long-running cluster got a corrupted number of files/subdirs, and responds to "rm -rf" with "ca...
- 12:42 AM Backport #18531 (Resolved): kraken: speed up readdir by skipping unwanted dn
- https://github.com/ceph/ceph/pull/13028
- 12:28 AM Feature #18514: qa: don't use a node for each kclient
- Probably, though I'm not really sure. I just found https://github.com/ceph/ceph-qa-suite/pull/1156/ in the depths of ...
- 12:26 AM Bug #17193: truncate can cause unflushed snapshot data lose
- So do we think this is fixed or not? Need to undo https://github.com/ceph/ceph-qa-suite/pull/1156/commits/5f1abf9c310...
- 12:00 AM Bug #18530 (New): ceph tell mds prints warning about ms_handle_reset
- Client::ms_handle_reset logs at level 0, and ceph tell mds seems to always print two of those log messages. I *think...
01/13/2017
- 04:00 PM Feature #18490: client: implement delegation support in userland cephfs
- Matt B. also had some upcall/invalidate work that may be relevant here that he has in these branches:
https://gith... - 01:39 PM Backport #18520 (In Progress): jewel: speed up readdir by skipping unwanted dn
- 01:30 PM Backport #18520 (Resolved): jewel: speed up readdir by skipping unwanted dn
- https://github.com/ceph/ceph/pull/12921
- 01:29 PM Bug #18519: speed up readdir by skipping unwanted dn
- https://github.com/ceph/ceph/pull/12870
- 01:29 PM Bug #18519 (Resolved): speed up readdir by skipping unwanted dn
- we hit MDS CPU bottleneck (100% on one core as it is single thread) in our cephFS production enviroment.
Troublesh... - 11:46 AM Feature #18514: qa: don't use a node for each kclient
- What's the proposed solution here? To isolate the tests that require killing mounts in a directory with a different ...
- 07:07 AM Feature #18514 (Resolved): qa: don't use a node for each kclient
- https://github.com/ceph/ceph-qa-suite/pull/1156/commits/c5f6dfc14f47cca251dcac5c53f6369fd36ace1a
Right now each ke... - 11:24 AM Bug #18396 (Pending Backport): Test Failure: kcephfs test_client_recovery.TestClientRecovery
- 11:23 AM Bug #18306 (Pending Backport): segfault in handle_client_caps
- 11:23 AM Bug #18361 (Pending Backport): Test failure: test_session_reject (tasks.cephfs.test_sessionmap.Te...
- 11:22 AM Bug #18179 (Pending Backport): MDS crashes on missing metadata object
- 11:22 AM Bug #18460 (Pending Backport): ceph-fuse crash during snapshot tests
- 02:59 AM Feature #18513 (Resolved): MDS: scrub: forward scrub reports missing backtraces on new files as d...
- It appears that running a recursive repair scrub_path results in reports of MDS damage on files that are new enough t...
- 01:09 AM Feature #18509: MDS: damage reporting by ino number is useless
- Path string is certainly the one I was thinking of.
- 12:58 AM Feature #18509: MDS: damage reporting by ino number is useless
The log message reporting the path is still there:...
01/12/2017
- 11:30 PM Feature #18509 (Resolved): MDS: damage reporting by ino number is useless
- We had two damaged directories on the long-running cluster, but examining the directories in question (other than thr...
- 10:54 PM Feature #18490: client: implement delegation support in userland cephfs
- This is basically what we've discussed previously in this area. My main concern is just designing an interface that c...
- 10:51 PM Bug #18461: failed to reconnect caps during snapshot tests
01/11/2017
- 08:07 PM Feature #18490: client: implement delegation support in userland cephfs
- I've created an nfs-ganesha category to match our Samba category.
- 05:52 PM Feature #18490 (Resolved): client: implement delegation support in userland cephfs
- To properly implement NFSv4 delegations in ganesha, we need something that operates a little like Linux's fcntl(..., ...
- 05:49 PM Feature #11950 (Fix Under Review): Strays enqueued for purge cause MDCache to exceed size limit
- https://github.com/ceph/ceph/pull/12786
- 05:00 PM Feature #18489 (New): mds: Multi-MDS-aware dirfrag split/join test
- Similar to the existing dirfrag tests, but do some import/exporting of the resulting fragments so that they span mult...
- 01:04 PM Bug #18487 (Resolved): Crash in MDCache::split_dir -- FAILED assert(dir->is_auth())
- ...
- 12:40 PM Feature #18477: O_TMPFILE support in libcephfs
- Yeah, with Linux' O_TMPFILE you can definitely do I/O to the inode before it's linked, and I think it'd be good to mi...
- 11:49 AM Feature #18477: O_TMPFILE support in libcephfs
- I was assuming that when doing it ephemerally we would not be allowing any data IO operations on the inode until it w...
- 11:50 AM Feature #18483: Forward scrub ops are not in Op Tracker
- See also http://tracker.ceph.com/issues/17852
- 01:22 AM Feature #18483 (New): Forward scrub ops are not in Op Tracker
- We started a forward scrub on the LRC today to look for any busted rstats since we appear to have leaked data somewhe...
01/10/2017
- 10:38 PM Feature #18477: O_TMPFILE support in libcephfs
- I'm pretty skeptical that doing it ephemerally (without initially setting it up as a journaled stray) is a feasible s...
- 08:50 PM Feature #18477: O_TMPFILE support in libcephfs
- The stray would end up getting journaled, probably never written to backing store as long as the link operation came ...
- 07:19 PM Feature #18477: O_TMPFILE support in libcephfs
- I think it makes sense to optimize for the success case here. In most cases, the link will be successful and it'll en...
- 07:12 PM Feature #18477: O_TMPFILE support in libcephfs
- Main decision here is probably whether it should be a stray or some new mechanism.
Strays feel like overkill here ... - 05:39 PM Feature #18477 (New): O_TMPFILE support in libcephfs
- nfs-ganesha could make use of the ability to create a disconnected inode (pinned only by an open file descriptor) tha...
- 02:40 PM Feature #18475 (Resolved): qa: run xfstests in the nightlies
- We have manually run xfstests against ceph-fuse and kceph before, but apparently don't do so in the nightlies. Jeff r...
- 02:16 PM Support #16526 (Resolved): cephfs client side quotas - nfs-ganesha
- Yep, I think so.
- 01:53 PM Support #16526: cephfs client side quotas - nfs-ganesha
- In a Ganesha (V2.5-dev-6) and Ceph (latest Jewel) setup, I set `client quota = true` in the client section of the cep...
- 09:45 AM Bug #18460 (Fix Under Review): ceph-fuse crash during snapshot tests
- https://github.com/ceph/ceph/pull/12859
- 03:40 AM Bug #18461 (Fix Under Review): failed to reconnect caps during snapshot tests
- 03:40 AM Bug #18461: failed to reconnect caps during snapshot tests
- https://github.com/ceph/ceph/pull/12852
01/09/2017
- 01:58 PM Backport #18462 (In Progress): jewel: Decode errors on backtrace will crash MDS
- 11:21 AM Backport #18462 (Resolved): jewel: Decode errors on backtrace will crash MDS
- https://github.com/ceph/ceph/pull/12836
- 01:57 PM Backport #18463 (In Progress): kraken: Decode errors on backtrace will crash MDS
- 11:21 AM Backport #18463 (Resolved): kraken: Decode errors on backtrace will crash MDS
- https://github.com/ceph/ceph/pull/12835
- 01:02 PM Bug #18396 (Fix Under Review): Test Failure: kcephfs test_client_recovery.TestClientRecovery
- kernel_mount.py does implement force umount
https://github.com/ceph/ceph/pull/12833 - 08:36 AM Bug #18396: Test Failure: kcephfs test_client_recovery.TestClientRecovery
- http://qa-proxy.ceph.com/teuthology/teuthology-2017-01-05_11:20:01-kcephfs-kraken-testing-basic-smithi/691532/
http:... - 11:04 AM Bug #18311 (Pending Backport): Decode errors on backtrace will crash MDS
- 08:31 AM Bug #18461 (Resolved): failed to reconnect caps during snapshot tests
- http://qa-proxy.ceph.com/teuthology/teuthology-2017-01-07_17:15:02-fs-master---basic-smithi/698957/
- 07:47 AM Bug #18460 (Resolved): ceph-fuse crash during snapshot tests
- http://qa-proxy.ceph.com/teuthology/teuthology-2017-01-05_11:10:02-fs-kraken---basic-smithi/691432/teuthology.log
...
01/08/2017
- 08:04 PM Bug #11124 (Fix Under Review): MDSMonitor: refuse to do "fs new" on metadata pools containing obj...
- https://github.com/ceph/ceph/pull/12825
01/06/2017
- 03:42 PM Backport #18439 (In Progress): kraken: TestVolumeClient.test_evict_client failure creating pidfile
- 03:40 PM Backport #18439 (Resolved): kraken: TestVolumeClient.test_evict_client failure creating pidfile
- https://github.com/ceph/ceph/pull/12813
- 03:29 PM Bug #18309 (Pending Backport): TestVolumeClient.test_evict_client failure creating pidfile
- 07:59 AM Bug #18306 (Fix Under Review): segfault in handle_client_caps
- https://github.com/ceph/ceph/pull/12808
01/05/2017
- 01:50 PM Bug #18306: segfault in handle_client_caps
- 01:49 PM Bug #18306: segfault in handle_client_caps
- http://qa-proxy.ceph.com/teuthology/zyan-2017-01-05_11:43:04-fs-wip-zyan-testing-testing-basic-mira/692034/teuthology...
- 05:21 AM Feature #18425: mds: add the option to use tcmalloc directly
- https://github.com/ceph/ceph/pull/12792
By default, MDS is still using boost::pool. - 05:19 AM Feature #18425 (Resolved): mds: add the option to use tcmalloc directly
- As describable in http://tracker.ceph.com/issues/18225, MDS uses boost:pool by default and unused memory is usually m...
01/04/2017
- 01:57 PM Bug #16914: multimds: pathologically slow deletions in some tests
- Right, but users are going to do this -- it needs to work.
- 10:31 AM Backport #18413 (In Progress): jewel: lookup of /.. in jewel returns -ENOENT
- 10:26 AM Backport #18413 (Resolved): jewel: lookup of /.. in jewel returns -ENOENT
- https://github.com/ceph/ceph/pull/12783
01/03/2017
- 08:13 PM Bug #16397: nfsd selinux denials causing knfs tests to fail
- Ok, sorry for the delay on this. Finally got around to opening a RHBZ:
https://bugzilla.redhat.com/show_bug.cg... - 05:54 PM Bug #18408: lookup of /.. in jewel returns -ENOENT
- The gory details of this problem are available here:
https://bugzilla.redhat.com/show_bug.cgi?id=1408989 - 05:52 PM Bug #18408 (Pending Backport): lookup of /.. in jewel returns -ENOENT
- 05:52 PM Bug #18408 (Resolved): lookup of /.. in jewel returns -ENOENT
- This is a problem for nfs-ganesha, which needs to be able to perform a lookup of ".." in the root directory in order ...
- 02:38 PM Bug #17563 (Resolved): extremely slow ceph_fsync calls
- Kernel patches merged for v4.10.
- 02:27 PM Bug #18131: ceph-fuse not clearing setuid/setgid bits on chown
- Sorry for the late response. At the very least, we need these commits:...
- 09:36 AM Bug #18396 (Resolved): Test Failure: kcephfs test_client_recovery.TestClientRecovery
- http://pulpito.ceph.com/teuthology-2016-12-31_11:20:02-kcephfs-kraken-testing-basic-smithi/678546/
Looks like the ... - 09:09 AM Bug #18361: Test failure: test_session_reject (tasks.cephfs.test_sessionmap.TestSessionMap)
- https://github.com/ceph/ceph/pull/12757
- 03:55 AM Bug #18157 (Resolved): ceph-fuse segfaults on daemonize
- 03:47 AM Bug #18179 (Fix Under Review): MDS crashes on missing metadata object
- https://github.com/ceph/ceph/pull/12749
- 03:02 AM Bug #18047 (Resolved): assertion in MDSMap::get_up_features()
- 02:19 AM Bug #18362: Test failure: test_evict_client (tasks.cephfs.test_volume_client.TestVolumeClient)
- http://qa-proxy.ceph.com/teuthology/teuthology-2016-12-31_17:15:02-fs-master---basic-smithi/679504/
http://qa-proxy....
12/30/2016
- 06:10 AM Bug #18211 (Resolved): test_snapshot_remove (tasks.cephfs.test_strays.TestStrays) failed at data ...
- 06:08 AM Bug #18363: Test failure: test_ops_throttle (tasks.cephfs.test_strays.TestStrays)
- Oops, The log files no long exist. looks like someone is actively deleting the logs
12/29/2016
- 09:58 AM Bug #18363 (Can't reproduce): Test failure: test_ops_throttle (tasks.cephfs.test_strays.TestStrays)
- http://qa-proxy.ceph.com/teuthology/teuthology-2016-12-24_17:15:02-fs-master---basic-smithi/663459/teuthology.log
- 09:55 AM Bug #18362 (Duplicate): Test failure: test_evict_client (tasks.cephfs.test_volume_client.TestVolu...
- http://qa-proxy.ceph.com/teuthology/teuthology-2016-12-27_11:10:01-fs-kraken---basic-smithi/669343/
probably it'... - 09:53 AM Bug #18361: Test failure: test_session_reject (tasks.cephfs.test_sessionmap.TestSessionMap)
- http://qa-proxy.ceph.com/teuthology/teuthology-2016-12-27_11:10:01-fs-kraken---basic-smithi/669334/
- 09:49 AM Bug #18361 (Fix Under Review): Test failure: test_session_reject (tasks.cephfs.test_sessionmap.Te...
- https://github.com/ceph/ceph/pull/12708
- 09:16 AM Bug #18361 (Resolved): Test failure: test_session_reject (tasks.cephfs.test_sessionmap.TestSessio...
12/28/2016
- 10:19 AM Bug #16914: multimds: pathologically slow deletions in some tests
- The reason is that you use "rm -rf delete_me/*" to delete files. ceph-fuse needs to do a lookup "delete_me" for each ...
12/27/2016
- 11:33 AM Bug #11482 (Resolved): kclient: intermittent log warnings "client.XXXX isn't responding to mclien...
- 11:32 AM Backport #13932 (Rejected): hammer: kclient: intermittent log warnings "client.XXXX isn't respond...
12/24/2016
- 02:45 PM Bug #16914: multimds: pathologically slow deletions in some tests
- I have a nice simple reproducer for this now (even with fuse default permissions = false it has the slowdown).
It'...
12/22/2016
- 05:26 PM Bug #18309: TestVolumeClient.test_evict_client failure creating pidfile
- Alternative approach: https://github.com/ceph/ceph/pull/12628
- 02:17 PM Bug #18314 (Resolved): commit 41d46e492 "osd/ReplicatedPG: limit omap request by bytes" breaks la...
- 01:58 PM Bug #18314: commit 41d46e492 "osd/ReplicatedPG: limit omap request by bytes" breaks large directory
- http://tracker.ceph.com/issues/18334 to track a proper fix for OMAP_GETKEYS
- 10:14 AM Backport #17478 (Resolved): jewel: MDS goes damaged on blacklist (failed to read JournalPointer: ...
- 10:14 AM Backport #17582 (Resolved): jewel: monitor assertion failure when deactivating mds in (invalid) f...
- 10:14 AM Backport #17615 (Resolved): jewel: mds: false "failing to respond to cache pressure" warning
- 10:14 AM Backport #17617 (Resolved): jewel: [cephfs] fuse client crash when adding a new osd
- 10:14 AM Backport #17697 (Resolved): jewel: MDS long-time blocked ops. ceph-fuse locks up with getattr of ...
- 10:14 AM Backport #17706 (Resolved): jewel: multimds: mds entering up:replay and processing down mds aborts
- 10:14 AM Backport #17720 (Resolved): jewel: MDS: false "failing to respond to cache pressure" warning
- 10:13 AM Backport #17841 (Resolved): jewel: mds fails to respawn if executable has changed
- 10:13 AM Backport #17885 (Resolved): jewel: "[ FAILED ] LibCephFS.InterProcessLocking" in jewel v10.2.4
12/21/2016
- 08:02 PM Bug #18309 (Fix Under Review): TestVolumeClient.test_evict_client failure creating pidfile
- -https://github.com/ceph/ceph/pull/12606-
- 06:20 PM Bug #18309: TestVolumeClient.test_evict_client failure creating pidfile
- The problem is that global_init_prefork is calling pidfile_write, and we started using that from the client in 83aaa5...
- 06:17 PM Backport #18308 (Resolved): ceph-fuse not clearing setuid/setgid bits on chown
- 05:20 PM Backport #18308 (New): ceph-fuse not clearing setuid/setgid bits on chown
- 04:36 PM Backport #18308: ceph-fuse not clearing setuid/setgid bits on chown
- h3. original description
I had some test failures that showed up in my most recent fs suite run here:
http... - 06:17 PM Bug #18131: ceph-fuse not clearing setuid/setgid bits on chown
- *master PR*: https://github.com/ceph/ceph/pull/12331
- 04:38 PM Bug #18131: ceph-fuse not clearing setuid/setgid bits on chown
- @Jeff: Which commit fixes the issue/should be backported to jewel?
- 05:18 PM Bug #18254 (Resolved): path restricted cephx caps not working correctly
- 05:16 PM Bug #18254: path restricted cephx caps not working correctly
- *master PR*: https://github.com/ceph/ceph/pull/12505
- 04:52 PM Bug #18254: path restricted cephx caps not working correctly
- @Jeff: We have a system/service in place for backporting bugfixes to our stable releases. Patches backported via this...
- 05:18 PM Backport #18307: path restricted cephx caps not working correctly
- (removed attachments that are available at #18254)
- 05:17 PM Backport #18307 (Resolved): path restricted cephx caps not working correctly
- 04:39 PM Backport #18307 (New): path restricted cephx caps not working correctly
- h3. original description
Ramana noticed this first while testing my ganesha patches to allow restricting exports. ... - 02:24 PM Bug #18314 (Fix Under Review): commit 41d46e492 "osd/ReplicatedPG: limit omap request by bytes" b...
- https://github.com/ceph/ceph/pull/12599
- 08:37 AM Bug #18314: commit 41d46e492 "osd/ReplicatedPG: limit omap request by bytes" breaks large directory
- ...
- 08:32 AM Bug #18314 (Resolved): commit 41d46e492 "osd/ReplicatedPG: limit omap request by bytes" breaks la...
- OPTION(osd_max_omap_bytes_per_request, OPT_U64, 4<<20)
4M can only carry about 5k dentries. It's too small - 11:14 AM Bug #15921 (Can't reproduce): segfault in cephfs-journal-tool (TestJournalRepair failure)
- Haven't seen this failure in a long time.
- 11:11 AM Bug #2375 (Closed): rrdtoll data malfuntion..
- Ancient, closing.
- 11:09 AM Bug #1206 (Closed): NFS reexport file creation lags 1-3 seconds
- Closing this because it's ancient (and if NFS creates were super-slow we'd notice on the knfs suite)
12/20/2016
- 06:54 PM Backport #18307: path restricted cephx caps not working correctly
- PR is up here:
https://github.com/ceph/ceph/pull/12592 - 01:00 PM Backport #18307 (Resolved): path restricted cephx caps not working correctly
- https://github.com/ceph/ceph/pull/12592
- 06:06 PM Bug #18311 (Fix Under Review): Decode errors on backtrace will crash MDS
- https://github.com/ceph/ceph/pull/12588
- 03:16 PM Bug #18311 (Resolved): Decode errors on backtrace will crash MDS
- Noticed by inspection:...
- 05:46 PM Bug #18225 (Resolved): MDS doesn't release memory after exceeding its cache size limit
- 05:34 PM Bug #9935 (Fix Under Review): client: segfault on ceph_rmdir path "/"
- https://github.com/ceph/ceph/pull/12550
- 01:19 PM Bug #18309 (Resolved): TestVolumeClient.test_evict_client failure creating pidfile
- Consistent on master
http://pulpito.ceph.com/jspray-2016-12-19_21:05:25-fs-master-distro-basic-smithi/648157
I ... - 01:13 PM Backport #18308 (Resolved): ceph-fuse not clearing setuid/setgid bits on chown
- https://github.com/ceph/ceph/pull/12591
- 01:01 PM Bug #18131 (Pending Backport): ceph-fuse not clearing setuid/setgid bits on chown
- 12:59 PM Bug #18254 (Pending Backport): path restricted cephx caps not working correctly
- 12:21 PM Bug #18254: path restricted cephx caps not working correctly
- Patch merged. We'll also want this backported to jewel.
- 11:16 AM Bug #18306 (Resolved): segfault in handle_client_caps
- http://pulpito.ceph.com/jspray-2016-12-19_21:05:25-fs-master-distro-basic-smithi/648247...
Also available in: Atom