Activity
From 06/09/2016 to 07/08/2016
07/08/2016
- 10:36 PM Feature #16419: add statx-like interface to libcephfs
- Possibly. The thing is that the btime should only ever change due to an deliberate setattr call. It's unlike the othe...
- 10:20 PM Feature #16419: add statx-like interface to libcephfs
- We need to be able to serve an accurate btime. I suppose we could break our rules and assume it won't get changed in ...
- 09:54 PM Feature #16419: add statx-like interface to libcephfs
- Aside from the stuff Greg noticed in his latest review pass, I noticed a number of flaws in the original patchset and...
- 08:02 PM Feature #16419: add statx-like interface to libcephfs
- Changing the description since this has ballooned a bit in scope. We want to add btime support and a change_attribute...
- 10:21 PM Bug #16640 (Won't Fix): libcephfs: Java bindings failing to load on CentOS
- http://qa-proxy.ceph.com/teuthology/jspray-2016-07-08_05:19:56-fs-master-distro-basic-mira/302088/teuthology.log
<... - 12:49 PM Feature #16631 (New): ObjectCacher cache size stats for ceph-fuse
- Currently the perf stats from ObjectCacher don't include the actual size of the cache (get_stat_clean, get_stat_dirty...
- 08:26 AM Bug #16588 (Fix Under Review): ceph mds dump show incorrect number of metadata pools.
- https://github.com/ceph/ceph/pull/10202
- 07:28 AM Backport #16625 (In Progress): jewel: Failing file operations on kernel based cephfs mount point ...
- 07:18 AM Backport #16625 (Resolved): jewel: Failing file operations on kernel based cephfs mount point lea...
- https://github.com/ceph/ceph/pull/10199
- 07:27 AM Backport #16626 (In Progress): hammer: Failing file operations on kernel based cephfs mount point...
- 07:18 AM Backport #16626 (Resolved): hammer: Failing file operations on kernel based cephfs mount point le...
- https://github.com/ceph/ceph/pull/10198
- 07:06 AM Bug #16013: Failing file operations on kernel based cephfs mount point leaves unaccessible file b...
- *master PR*: https://github.com/ceph/ceph/pull/8778
- 07:05 AM Bug #16013 (Pending Backport): Failing file operations on kernel based cephfs mount point leaves ...
07/07/2016
- 09:53 PM Backport #16621 (Resolved): jewel: mds: `session evict` tell command blocks forever with async me...
- https://github.com/ceph/ceph/pull/10501
- 09:53 PM Backport #16620 (Resolved): jewel: Fix shutting down mds timed-out due to deadlock
- https://github.com/ceph/ceph/pull/10500
- 08:58 PM Bug #16592: Jewel: monitor asserts on "mon/MDSMonitor.cc: 2796: FAILED assert(info.state == MDSMa...
- Should note that this is maybe related to: http://tracker.ceph.com/issues/15591
- 05:44 PM Bug #16610: Jewel: segfault in ObjectCacher::FlusherThread
- Log is now here: /ceph/post/i16610/client.log
- 02:04 PM Bug #16610 (Resolved): Jewel: segfault in ObjectCacher::FlusherThread
- ...
- 03:10 PM Feature #15942: MDS: use FULL_TRY Objecter flag instead of relying on an exemption from full chec...
- Related: https://github.com/ceph/ceph/pull/9087
- 03:09 PM Cleanup #16144 (Resolved): Remove cephfs-data-scan tmap_upgrade
- 03:08 PM Cleanup #16195 (In Progress): mds: Don't spam log with standby_replay_restart messages
- 03:05 PM Bug #16288 (Pending Backport): mds: `session evict` tell command blocks forever with async messen...
- 03:04 PM Bug #16396 (Pending Backport): Fix shutting down mds timed-out due to deadlock
- 01:05 PM Feature #16570 (In Progress): MDS health warning for failure to enforce cache size limit
- 01:04 PM Bug #15485 (Duplicate): drop /usr/bin/cephfs
- 11:28 AM Bug #16588: ceph mds dump show incorrect number of metadata pools.
- h3. original description
Ceph mds dump shows metadata pool count as 2, even though only one metadata pool is prese... - 08:54 AM Bug #16588: ceph mds dump show incorrect number of metadata pools.
- Hi Xiaoxi,
You are right about the bug. The metadata_pool field should be left blank. I have changed the descripti... - 08:48 AM Bug #16588: ceph mds dump show incorrect number of metadata pools.
- Rohith Radhakrishnan wrote:
> Ceph mds dump shows metadata_pool id as 0. When no FS is present, then metadata_pool ... - 08:34 AM Bug #16588: ceph mds dump show incorrect number of metadata pools.
- Hmm, yes, this is because metadata_pool is initialized to 0 , this seems worth to fix.
The bug is , when no FS pr... - 06:50 AM Bug #16588: ceph mds dump show incorrect number of metadata pools.
- ceph osd pool stats
*there are no pools!*
ems@rack2-client-3:~$ ceph mds dump
dumped fsmap epoch 3
fs_name ceph... - 06:42 AM Bug #16588: ceph mds dump show incorrect number of metadata pools.
- on what basis is the pool id generated? There are no existing pools. So shouldn't the count start with 0 or 1?
Als...
07/06/2016
- 03:33 PM Feature #15406 (In Progress): Add versioning to CephFSVolumeClient interface
- 06:29 AM Bug #16588 (Rejected): ceph mds dump show incorrect number of metadata pools.
- This is not a bug.
The numbers following "data_pools" and "metadata_pool" are not count, but the pool ids.
root... - 03:44 AM Bug #16588: ceph mds dump show incorrect number of metadata pools.
07/05/2016
- 09:00 PM Bug #16042 (Fix Under Review): MDS Deadlock on shutdown active rank while busy with metadata IO
- PR: https://github.com/ceph/ceph/pull/10142
- 05:44 PM Bug #16592 (Need More Info): Jewel: monitor asserts on "mon/MDSMonitor.cc: 2796: FAILED assert(in...
- We've seen a few reports on the ceph-user mailing lists of the latest jewel....
- 11:42 AM Bug #16588 (Resolved): ceph mds dump show incorrect number of metadata pools.
- Ceph mds dump shows metadata_pool id as 0. When no FS is present, then metadata_pool id should be left blank.
ceph...
07/02/2016
- 07:48 AM Backport #16320 (In Progress): jewel: fs: fuse mounted file systems fails SAMBA CTDB ping_pong rw...
- 07:35 AM Backport #16313 (In Progress): jewel: client: FAILED assert(root_ancestor->qtree == __null)
- 07:31 AM Backport #16215 (In Progress): jewel: client: crash in unmount when fuse_use_invalidate_cb is ena...
- 07:29 AM Backport #16515 (In Progress): jewel: Session::check_access() is buggy
- 07:26 AM Backport #16560 (In Progress): jewel: mds: enforce a dirfrag limit on entries
- 07:22 AM Backport #16037: jewel: MDSMonitor::check_subs() is very buggy
- QA suite backported in https://github.com/ceph/ceph-qa-suite/pull/1075
- 07:11 AM Backport #16037 (In Progress): jewel: MDSMonitor::check_subs() is very buggy
07/01/2016
- 08:28 PM Feature #15069 (In Progress): MDS: multifs: enable two filesystems to point to same pools if one ...
- 08:19 PM Cleanup #16144 (Fix Under Review): Remove cephfs-data-scan tmap_upgrade
- https://github.com/ceph/ceph/pull/10100
- 07:39 PM Cleanup #16144 (In Progress): Remove cephfs-data-scan tmap_upgrade
- 12:07 PM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
- Thanks Kefu, I guess the lockdep one is either a cephfs or msgr issue so we'll keep this ticket open to look into it.
- 03:39 AM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
- LibCephFS.Fchown is fixed by https://github.com/ceph/ceph/pull/10081,
but we still have... - 11:11 AM Feature #15066: multifs: Allow filesystems to be assigned RADOS namespace as well as pool for met...
- Just in case I lose it, the draft code for splitting messengers was here: https://github.com/jcsp/ceph/tree/wip-15399...
- 10:48 AM Feature #16570 (Resolved): MDS health warning for failure to enforce cache size limit
This can have many causes, but it is a sign that something is not wrong, and a possible precursor to the MDS dying ...
06/30/2016
- 08:47 PM Bug #16042: MDS Deadlock on shutdown active rank while busy with metadata IO
- I'm able to reproduce this with vstart.sh and `cp -a /usr ...`. I'm seeing this every 10 seconds:...
- 04:35 PM Backport #16560 (Resolved): jewel: mds: enforce a dirfrag limit on entries
- https://github.com/ceph/ceph/pull/10104
- 02:21 PM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
- Jeff points out that we can also get it to blow up with just a passing test like bin/ceph_test_libcephfs --gtest_fil...
- 01:53 PM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
- ...
- 01:51 PM Bug #16556 (New): LibCephFS.InterProcessLocking failing on master and jewel
- Maybe related to https://github.com/ceph/ceph/pull/9995 ?
Failures on master here: http://pulpito.ceph.com/jspray-... - 02:19 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
- Ahh, the reason I could reproduce this yesterday is because the client box was running a v4.5 kernel. With a v4.7-rc5...
- 11:47 AM Bug #16164 (Pending Backport): mds: enforce a dirfrag limit on entries
06/29/2016
- 06:26 PM Support #16528: Stuck with CephFS with 1M files in one dir
- Thank you!
Raised "mds cache size" to 3M and it took couple of minutes to list this dir. - 05:40 PM Support #16528 (Closed): Stuck with CephFS with 1M files in one dir
- Assuming your MDS server has enough memory (it probably does), turn up the "mds cache size" to a number larger than 1...
- 04:48 PM Support #16528 (Closed): Stuck with CephFS with 1M files in one dir
- I'm pretty much stukc with cephfs (jewel 10.2.2) with 1 million 0 byte files in one dir left behind from unsuccessful...
- 06:24 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
- The fio threads at this point are all sitting in ceph_get_caps:...
- 05:38 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
- Ok, the mds session evict command definitely did the trick. Once I issued that (while running a fio test in another s...
- 05:38 PM Support #16526: cephfs client side quotas - nfs-ganesha
- How are you evaluating that the quotas are ignored? There isn't any integration, certainly, but the Ceph client libra...
- 02:56 PM Support #16526 (Resolved): cephfs client side quotas - nfs-ganesha
- I am not sure If this is best logged on the nfs-ganesha project or here.
Ceph quotas are configured using virtual ... - 10:45 AM Feature #16523 (Resolved): Assert directory fragmentation is occuring during stress tests
- Currently we enable fragmentation and set a low (100) frag size limit, but nothing actually validates that there is a...
- 06:40 AM Backport #16515 (Resolved): jewel: Session::check_access() is buggy
- https://github.com/ceph/ceph/pull/10105
- 02:10 AM Bug #16358: Session::check_access() is buggy
- Yes, it could happen for normal case (newly created file). We should backport it
- 12:11 AM Bug #16358: Session::check_access() is buggy
- Whoops, yes. Luckily only for users of hard links, but that's good enough reason!
- 12:02 AM Bug #16358 (Pending Backport): Session::check_access() is buggy
- Seems like this could be serious enough to backport (Zheng: this could happen in normal use, right?)
- 01:11 AM Bug #16367 (In Progress): libcephfs: UID parsing breaks root squash (Ganesha FSAL)
- My basic approach here is to just stop automatically setting UID/GID within the Client class code base at all. It cur...
06/28/2016
- 09:53 PM Bug #16358 (Resolved): Session::check_access() is buggy
- 08:55 PM Bug #16407 (Rejected): LibCephFS.UseUnmounted failed
- 01:10 AM Bug #16407: LibCephFS.UseUnmounted failed
- @John Spray, This is my fault, please closed it.
Thanks. - 07:57 PM Feature #11171: Path filtering on "dump cache" asok
- For test, see https://github.com/ceph/ceph-qa-suite/pull/1066
- 02:27 PM Bug #16397 (Can't reproduce): nfsd selinux denials causing knfs tests to fail
- Ok, talked with Bruce (knfsd maintainer) and the SELinux folks and the consensus is that we have no clue as to why th...
- 01:11 PM Bug #16397: nfsd selinux denials causing knfs tests to fail
- Anyway, the first AVC denial is here:
avc: denied { add_name } for pid=22038 comm="rpc.mountd" name="channel" ... - 12:38 PM Bug #16397: nfsd selinux denials causing knfs tests to fail
- Ok, looking at the log, I do see the SELinux denials. I am new teuthology though...
So you have ubuntu boxes that ... - 01:21 PM Cleanup #15923 (In Progress): MDS: remove TMAP2OMAP check and move Objecter into MDSRank
- 01:21 PM Cleanup #16035 (In Progress): Remove "cephfs" CLI
06/27/2016
- 08:01 PM Bug #16288 (In Progress): mds: `session evict` tell command blocks forever with async messenger (...
- Still no reproducer, but
https://github.com/ceph/ceph/pull/9971
may help. - 01:44 PM Bug #16407: LibCephFS.UseUnmounted failed
- Can you update us? Where are you you seeing the issue and is there a new fix PR?
- 09:13 AM Bug #16042: MDS Deadlock on shutdown active rank while busy with metadata IO
- Could it be via following paths to call MDSDaemon::ms_handle_reset() like async msgr?
One mds thread: ... -> Simpl... - 03:44 AM Bug #16186: kclient: drops requests without poking system calls on reconnect
- there is a 'ceph daemon mds.xxx session evict' command, which makes mds close client session. (use 'ceph daemon mds.x...
06/25/2016
- 05:32 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
- Ok, I tried reproducing this by issuing a stat() while outbound traffic from the client was blocked (on a v4.7-rc4 ke...
06/24/2016
- 08:21 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
- I don't suppose we have a way to reproduce this, do we? Maybe drive a lot of MDS ops and continually stop and restart...
- 05:08 PM Feature #11171 (Fix Under Review): Path filtering on "dump cache" asok
- https://github.com/ceph/ceph/pull/9925
- 10:15 AM Bug #16042: MDS Deadlock on shutdown active rank while busy with metadata IO
- Interesting, #16396 is with async messenger (and is probably the issue we're seeing in current master testing), but w...
- 03:12 AM Bug #16042: MDS Deadlock on shutdown active rank while busy with metadata IO
- Hi guys,
Looks like this issue is very similar to this one here: http://tracker.ceph.com/issues/16396 - 10:07 AM Feature #16468 (Resolved): kclient: Exclude ceph.* xattr namespace in listxattr
- See this thread: http://www.spinics.net/lists/ceph-devel/msg30948.html
Some userspaces tools (notably rsync) try t... - 10:06 AM Feature #16467 (New): ceph-fuse: Exclude ceph.* xattr namespace in listxattr
- See this thread: http://www.spinics.net/lists/ceph-devel/msg30948.html
Some userspaces tools (notably rsync) try t...
06/23/2016
- 07:44 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
- Well, if we have unsafe requests the MDS will in fact have committed them (assuming the MDS didn't crash or something...
- 01:53 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
- If the mds has torn down the client's session, then I don't see what can reasonably be done other than to return an e...
- 06:33 PM Bug #16288: mds: `session evict` tell command blocks forever with async messenger (TestVolumeClie...
- Not to take away Doug's thunder, but I gather he's been unable to reproduce it. The AsyncMessenger may have already b...
- 05:44 PM Bug #15921: segfault in cephfs-journal-tool (TestJournalRepair failure)
- As far as I can tell, we don't even have the backtrace of the segfault in either of those logs, and the sha1 isn't av...
- 01:20 PM Bug #16013 (Resolved): Failing file operations on kernel based cephfs mount point leaves unaccess...
- 11:59 AM Bug #16367: libcephfs: UID parsing breaks root squash (Ganesha FSAL)
- I don't know if I should open a new issue for this, but it looks like even with another ID something is still wrong:
... - 04:51 AM Bug #16396: Fix shutting down mds timed-out due to deadlock
- https://github.com/ceph/ceph/pull/9884
06/22/2016
- 09:09 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
- But if we restart requests from scratch, we're dramatically re-ordering them. We can seemingly send files back in tim...
- 09:01 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
- I think it is working the way it is supposed to work.
We skip unsafe requests because the mds already got them and... - 08:59 PM Bug #16407: LibCephFS.UseUnmounted failed
- You appear to have closed your own PR. And generally speaking we pass around negative error numbers, so readdir() is ...
- 08:44 AM Bug #16407: LibCephFS.UseUnmounted failed
- https://github.com/ceph/ceph/pull/9860
- 07:36 AM Bug #16407 (Rejected): LibCephFS.UseUnmounted failed
- 2016-06-22T15:03:06.176 INFO:tasks.workunit.client.0.plana146.stdout:[ RUN ] LibCephFS.StripeUnitGran
2016-06-2... - 08:55 PM Support #16043 (Closed): MDS is crashed
- 07:40 PM Feature #16228: Create teuthology task for Samba ping_pong test
- (Copied from #16417) See Greg's draft https://github.com/gregsfortytwo/ceph-qa-suite/tree/wip-pingpong
- 07:40 PM Feature #16417 (Duplicate): test pingpong on ceph-fuse
- 05:10 PM Feature #16417 (Duplicate): test pingpong on ceph-fuse
- See #12653. We should integrate pingpong into our nightly test suite, to verify consistency on the kernel client and ...
- 06:10 PM Feature #16419: add statx-like interface to libcephfs
- Yeah, that's what I mean. We have ceph_ll_getattr now (afaict), so we need something like a ceph_ll_getattrx (that na...
- 06:01 PM Feature #16419: add statx-like interface to libcephfs
- Jeff Layton wrote:
> What I'm thinking is that we should add something along the lines of what David Howells has pro... - 05:39 PM Feature #16419: add statx-like interface to libcephfs
- What I'm thinking is that we should add something along the lines of what David Howells has proposed for the new stat...
- 05:35 PM Feature #16419 (Resolved): add statx-like interface to libcephfs
- samba, in particular, can make use of the birthtime for an inode. Have ceph track the btime in the inode and provide ...
- 01:01 PM Feature #15615: CephFSVolumeClient: List authorized IDs by share
- https://github.com/ceph/ceph/pull/9864
06/21/2016
- 02:03 PM Bug #16397: nfsd selinux denials causing knfs tests to fail
- Ahh, hmm -- just noticed the "add name" deinal too. Does the path "/proc/net/rpc/auth.unix.ip/channel" even exist? Ma...
- 01:46 PM Bug #16397: nfsd selinux denials causing knfs tests to fail
- Looks unrelated to anything ceph-specific. My guess is that this is an selinux policy bug, since rpc.mountd should be...
- 11:56 AM Bug #16397 (Resolved): nfsd selinux denials causing knfs tests to fail
- http://pulpito.ceph.com/teuthology-2016-06-20_17:35:01-knfs-master-testing-basic-mira/267607/
- 11:26 AM Support #16043: MDS is crashed
- I execute...
- 06:05 AM Support #16043: MDS is crashed
- Yes, i try reset journal and sessions.
I run:... - 01:34 AM Support #16043: MDS is crashed
- Yep. So looking through the log, I now see
>mds.2.journal ESession.replay sessionmap 0 < 18884 close client.166758... - 09:36 AM Bug #16396: Fix shutting down mds timed-out due to deadlock
- -https://github.com/ceph/ceph/pull/9841-
- 09:31 AM Bug #16396 (Resolved): Fix shutting down mds timed-out due to deadlock
- This issue was found in jewel when restarting/stopping mds. It took long time for mds to completely stop until mds th...
- 09:02 AM Bug #16288 (New): mds: `session evict` tell command blocks forever with async messenger (TestVolu...
- Oops, I meant to paste to begin with. I think it was this one:
/a/jspray-2016-06-13_14:56:46-fs-wip-jcsp-testing-qu...
06/20/2016
- 08:12 PM Bug #16367: libcephfs: UID parsing breaks root squash (Ganesha FSAL)
- 07:57 PM Bug #16367: libcephfs: UID parsing breaks root squash (Ganesha FSAL)
- Yeah, I expect that Frank's report is the root cause, but wanted to see to make sure. :)
- 08:56 AM Bug #16367: libcephfs: UID parsing breaks root squash (Ganesha FSAL)
- Now easier to read:...
- 08:55 AM Bug #16367: libcephfs: UID parsing breaks root squash (Ganesha FSAL)
I have ceph mounted under /mnt/nfs/ceph:
[root@test2202 test]# pwd
/mnt/nfs/ceph/test
[root@test2202 test]# ls ...- 08:08 PM Bug #16288 (Need More Info): mds: `session evict` tell command blocks forever with async messenge...
- 08:08 PM Bug #16288: mds: `session evict` tell command blocks forever with async messenger (TestVolumeClie...
- John, do you have any logs? The only failure of this test I can find is http://qa-proxy.ceph.com/teuthology/teutholog...
- 07:12 PM Bug #16288 (In Progress): mds: `session evict` tell command blocks forever with async messenger (...
- 01:31 PM Bug #16042 (In Progress): MDS Deadlock on shutdown active rank while busy with metadata IO
- 09:35 AM Support #16043: MDS is crashed
- Greg, I sent message with link to my debug log on your email. Service for ceph-post-file working has becomes unstable...
06/17/2016
- 08:46 PM Bug #16164: mds: enforce a dirfrag limit on entries
- PR here: https://github.com/ceph/ceph/pull/9789
- 05:50 PM Bug #16367 (Need More Info): libcephfs: UID parsing breaks root squash (Ganesha FSAL)
- 05:49 PM Bug #16367: libcephfs: UID parsing breaks root squash (Ganesha FSAL)
- Can you please:
1) run ls -lha on the director you're testing in
2) do your tests
3) run ls -lha on all the releva... - 03:15 PM Bug #16367 (Resolved): libcephfs: UID parsing breaks root squash (Ganesha FSAL)
- Testing with ganesha 2.4-o-dev20 and libcephfs 10.2.1:
I did set root squash on in the ganesha.conf, but as root I c... - 05:14 PM Support #16043: MDS is crashed
- Please set "debug mds = 20" and "debug mds log = 20" in your ceph.conf, turn it on, and then upload the mds log file ...
- 04:04 AM Bug #16358 (Fix Under Review): Session::check_access() is buggy
- https://github.com/ceph/ceph/pull/9769
- 03:53 AM Bug #16358 (Resolved): Session::check_access() is buggy
- It calls CInode::make_path_string(path, false, in->get_projected_parent_dn()). The second argument 'false' makes the ...
06/16/2016
- 03:14 PM Bug #16255: ceph-create-keys: sometimes blocks forever if mds "allow" is set
- > The loop you're seeing presumably is only occurring when /etc/ceph/ceph.client-admin.keyring has been removed.
e... - 03:05 PM Bug #16255: ceph-create-keys: sometimes blocks forever if mds "allow" is set
- The difference between @"allow"@ and @"allow *"@ is that the @"*"@ is necessary in more recent versions to issue 'tel...
- 02:39 PM Fix #16276: Update TestSessionMap.test_mount_conn_close for async messenger
- NB back out part of https://github.com/ceph/ceph-qa-suite/pull/1054 when fixing this, it's switched back to simple me...
- 02:29 PM Fix #16276: Update TestSessionMap.test_mount_conn_close for async messenger
- http://pulpito.ceph.com/gregf-2016-06-10_19:20:53-fs-greg-fs-testing-610---basic-mira/250875/
- 02:39 PM Bug #16288: mds: `session evict` tell command blocks forever with async messenger (TestVolumeClie...
- NB back out part of https://github.com/ceph/ceph-qa-suite/pull/1054 when fixing this, it's switched back to simple me...
- 02:38 PM Bug #16288: mds: `session evict` tell command blocks forever with async messenger (TestVolumeClie...
- This deadlocks and lockdep makes it crash in our nightlies; we should fix it quickly! :)
- 02:37 PM Feature #14271 (Resolved): directory listing: do not reset when fragmenting
- 02:33 PM Support #16043: MDS is crashed
- ...
- 02:31 PM Support #16043: MDS is crashed
- I upgraded my cluster to 10.2.2, situation not changed.
- 01:57 PM Support #16043 (Need More Info): MDS is crashed
- This probably isn't an issue any more, but if it is upgrade to 10.2.2 and report back if it's still an issue.
- 02:26 PM Feature #11171 (In Progress): Path filtering on "dump cache" asok
- 02:21 PM Backport #16284 (Resolved): jewel: directory listing: do not reset when fragmenting
- This was done as part of #16251.
- 11:54 AM Bug #16298 (Resolved): mds: failure in tasks/migration.yaml
- 11:15 AM Bug #16322: ceph mds getting killed for no reason
- $gdb /usr/local/bin/ceph-mds
If gdb does not say "no debugging symbols found", the debug package is properly insta... - 09:45 AM Bug #16322: ceph mds getting killed for no reason
- Zheng Yan wrote:
> Your ceph-mds does not contain debuginfo, please install debuginfo package first. then start ceph... - 02:20 AM Bug #16322: ceph mds getting killed for no reason
- Your ceph-mds does not contain debuginfo, please install debuginfo package first. then start ceph-mds manually with c...
- 07:39 AM Backport #16136: jewel: MDSMonitor fixes
- Original description:
These two commits:
https://github.com/ceph/ceph/pull/9418/commits/24b82bafffced97384135e5...
06/15/2016
- 06:31 PM Bug #16042: MDS Deadlock on shutdown active rank while busy with metadata IO
- This is rearing its head in general testing now:
http://pulpito.ceph.com/jspray-2016-06-15_05:28:02-fs-wip-jcsp-test... - 02:01 PM Bug #16322: ceph mds getting killed for no reason
- log: http://95.211.209.196/imgs/ceph-mds.mds01.log
- 01:48 PM Bug #16322: ceph mds getting killed for no reason
- kernel: 4.2.0-36-generic
- 01:46 PM Bug #16322: ceph mds getting killed for no reason
(...)
Loaded symbols for /lib/x86_64-linux-gnu/libnss_files.so.2
Reading symbols from /usr/lib/x86_64-linux-gnu/n...- 01:41 PM Bug #16322: ceph mds getting killed for no reason
- I am not very experience with gdb, sorry. Should I use it in ceph-mds ?
I will paste the whole log (it has a lot of ... - 12:16 PM Bug #16322: ceph mds getting killed for no reason
- could you enable coredump and use gdb to check which line causes the crash
- 11:50 AM Bug #16322: ceph mds getting killed for no reason
- add:
2016-06-15 03:15:51.017714 7f582103f700 -1 *** Caught signal (Aborted) **
in thread 7f582103f700 thread_nam... - 11:50 AM Bug #16322 (Can't reproduce): ceph mds getting killed for no reason
- Hello,
my ceph mds get killed for no reason (normally they do the active failover).
Log:
ceph version 10.2.1 (... - 10:11 AM Bug #15920: mds/StrayManager.cc: 520: FAILED assert(dnl->is_primary())
- John, thank you very much! Yeah, I saw that it was going to miss 10.2.2. Thank you for making this exception! I'll st...
- 08:06 AM Backport #16320 (Resolved): jewel: fs: fuse mounted file systems fails SAMBA CTDB ping_pong rw te...
- https://github.com/ceph/ceph/pull/10108
- 08:04 AM Backport #16313 (Resolved): jewel: client: FAILED assert(root_ancestor->qtree == __null)
- https://github.com/ceph/ceph/pull/10107
- 02:21 AM Bug #16160 (Resolved): PJD failures on Jewel
- http://qa-proxy.ceph.com/teuthology/teuthology-2016-06-13_17:25:02-kcephfs-master-testing-basic-mira/257158/teutholog...
06/14/2016
- 04:30 PM Bug #12653 (Pending Backport): fuse mounted file systems fails SAMBA CTDB ping_pong rw test with ...
- 04:29 PM Documentation #16300: doc: fuse_disable_pagecache
- NB while doing this would be useful to ask performance team to measure how much impact this really has
- 04:28 PM Documentation #16300 (Resolved): doc: fuse_disable_pagecache
- http://tracker.ceph.com/issues/12653
https://github.com/ceph/ceph/pull/5521/commits/0f11ec237d4692d313a038ed61aa07a3... - 04:24 PM Backport #16299 (Resolved): jewel: mds: fix SnapRealm::have_past_parents_open()
- https://github.com/ceph/ceph/pull/10499
- 04:22 PM Bug #16298 (Fix Under Review): mds: failure in tasks/migration.yaml
- https://github.com/ceph/ceph/pull/9697
- 04:20 PM Bug #16298 (Resolved): mds: failure in tasks/migration.yaml
- http://pulpito.ceph.com/jspray-2016-06-14_01:19:46-fs-wip-jcsp-testing-20160610-distro-basic-mira/257906
- 04:19 PM Bug #15920: mds/StrayManager.cc: 520: FAILED assert(dnl->is_primary())
I've pushed a jewel-15920 branch for you with the fix cherry-picked onto it. (don't usually do this, but it's fair...- 01:56 PM Bug #15920: mds/StrayManager.cc: 520: FAILED assert(dnl->is_primary())
- Good morning everyone!
Considering that a backport is done, though not merged yet, is there away for me to get a g... - 02:50 PM Backport #13927 (New): hammer: cephfs-java ftruncate unit test failure
- 02:50 PM Backport #13927: hammer: cephfs-java ftruncate unit test failure
- One attempted backport https://github.com/ceph/ceph/pull/6754 was closed.
- 12:30 PM Bug #16067 (Resolved): client: InvalidWrite in put_qtree
- (resolved via http://tracker.ceph.com/issues/16066, track backport there)
- 12:30 PM Bug #16066 (Pending Backport): client: FAILED assert(root_ancestor->qtree == __null)
- 10:13 AM Bug #16288 (Resolved): mds: `session evict` tell command blocks forever with async messenger (Tes...
I'm assuming for the moment that this is an MDS bug rather than something getting dropped in the new messenger code...- 07:19 AM Backport #16284 (Resolved): jewel: directory listing: do not reset when fragmenting
- https://github.com/ceph/ceph/pull/9655
- 12:24 AM Bug #16042: MDS Deadlock on shutdown active rank while busy with metadata IO
- I just saw this (or similar shutdown bug) for the first time in an automated test: http://qa-proxy.ceph.com/teutholog...
- 12:20 AM Fix #16276 (New): Update TestSessionMap.test_mount_conn_close for async messenger
When the default messenger changed from simple to async, this test started failing[1]. It's because it is using th...
06/13/2016
- 10:42 AM Feature #14271 (Pending Backport): directory listing: do not reset when fragmenting
- 08:27 AM Backport #16252 (Resolved): jewel: Client: reports that readahead is not working
- 05:05 AM Backport #16252 (In Progress): jewel: Client: reports that readahead is not working
- 05:00 AM Backport #16252 (Resolved): jewel: Client: reports that readahead is not working
- https://github.com/ceph/ceph/pull/9656
- 08:27 AM Bug #16024 (Resolved): Client: reports that readahead is not working
- 08:23 AM Backport #16251 (Resolved): jewel: client: simultaneous readdirs are very racy
- 04:55 AM Backport #16251 (In Progress): jewel: client: simultaneous readdirs are very racy
- 04:54 AM Backport #16251 (Resolved): jewel: client: simultaneous readdirs are very racy
- https://github.com/ceph/ceph/pull/9655
- 08:23 AM Bug #15508 (Resolved): client: simultaneous readdirs are very racy
- 05:41 AM Bug #16255 (Resolved): ceph-create-keys: sometimes blocks forever if mds "allow" is set
- The documentations at:
http://docs.ceph.com/docs/master/dev/mon-bootstrap/
tells to create the client.admin key...
06/12/2016
- 09:45 PM Bug #16024 (Pending Backport): Client: reports that readahead is not working
- Backport PR: https://github.com/ceph/ceph/pull/9656
- 09:35 PM Bug #15508 (Pending Backport): client: simultaneous readdirs are very racy
- Backport PR: https://github.com/ceph/ceph/pull/9655
06/10/2016
- 09:55 AM Feature #16228 (New): Create teuthology task for Samba ping_pong test
The Samba ping_pong test validates the interaction between multiple clients accessing the same data.
Related:
h...
06/09/2016
- 09:03 PM Bug #16067: client: InvalidWrite in put_qtree
- Greg: yes, I expect the big quotatree patch will fix both.
- 04:32 PM Bug #16067: client: InvalidWrite in put_qtree
- Any chance this is because of #16066, or at least resolved by the associated PR?
- 07:40 PM Feature #16219 (New): test: smallfile benchmark tool
- Run this metadata tester in our nightlies.
https://github.com/bengland2/smallfile
>smallfile is a python-based ... - 05:38 PM Backport #16215 (Resolved): jewel: client: crash in unmount when fuse_use_invalidate_cb is enabled
- https://github.com/ceph/ceph/pull/10106
- 09:49 AM Cleanup #15922 (Resolved): MDS: remove TMAP support from CDir
- 09:36 AM Bug #16137 (Pending Backport): client: crash in unmount when fuse_use_invalidate_cb is enabled
Also available in: Atom