Project

General

Profile

Activity

From 07/27/2016 to 08/25/2016

08/25/2016

10:47 PM Backport #17126 (Resolved): mds: fix double-unlock on shutdown
Loïc Dachary
07:28 PM Feature #11172 (In Progress): mds: inode filtering on 'dump cache' asok
Douglas Fuller
05:33 PM Feature #12274 (Fix Under Review): mds: start forward scrubs from all subtree roots, skip non-aut...
https://github.com/ceph/ceph/pull/10876 Douglas Fuller
05:31 PM Backport #16946: jewel: client: nlink count is not maintained correctly
FYI: Github is annoying and does some kind of timestamp sort when displaying commits. I'm not sure if it's the origin... Greg Farnum
05:17 PM Backport #16946: jewel: client: nlink count is not maintained correctly
@Jeff this is a very unusual situation and I apologize for the noise. It turns out that github does not display the c... Loïc Dachary
03:13 PM Backport #16946 (In Progress): jewel: client: nlink count is not maintained correctly
Jeff Layton
03:13 PM Backport #16946: jewel: client: nlink count is not maintained correctly
You want the latter approach, and you want to pick them in the order they were originally committed, in case we need ... Jeff Layton
02:54 PM Backport #16946 (Need More Info): jewel: client: nlink count is not maintained correctly
Actually, you were right to ask, my question was about something else :-) It's good to know that the four commits are... Loïc Dachary
02:40 PM Backport #16946 (New): jewel: client: nlink count is not maintained correctly
This is perfect, thank you ! Loïc Dachary
02:38 PM Backport #16946 (In Progress): jewel: client: nlink count is not maintained correctly
Jeff Layton
12:11 PM Backport #16946: jewel: client: nlink count is not maintained correctly
Yes. I think you'll want the entire patch pile from that PR. These 4 patches at least:
https://github.com/ceph/cep...
Jeff Layton
11:59 AM Backport #16946 (Need More Info): jewel: client: nlink count is not maintained correctly
git cherry-pick -x https://github.com/ceph/ceph/pull/10386/commits/f3605d39e53b3ff777eb64538abfa62a5f98a4f2 which is ... Loïc Dachary
04:59 PM Bug #17074 (Closed): "SELinux denials" in knfs-master-testing-basic-smithi
per IRC
(09:54:34 AM) yuriw: loicd dgalloway can we say that old tests for hammer ran in ovh never had SELinux enabl...
Yuri Weinstein
04:53 PM Bug #17074: "SELinux denials" in knfs-master-testing-basic-smithi
the suite defensively passed in previous point releases
http://pulpito.ovh.sepia.ceph.com:8081/teuthology-2016-04-24...
Yuri Weinstein
04:47 PM Bug #17074 (Need More Info): "SELinux denials" in knfs-master-testing-basic-smithi
I don't think CephFS/knfs tests and SELinux ever worked on Hammer. Yuri, can you find evidence they did or else close... Greg Farnum
04:55 PM Feature #4142 (Duplicate): MDS: forward scrub: Implement cross-MDS scrubbing
Douglas Fuller
04:25 PM Bug #16592 (Need More Info): Jewel: monitor asserts on "mon/MDSMonitor.cc: 2796: FAILED assert(in...
Moving this down and setting Need More Info based on Patrick's investigation and the new asserts; let me know if that... Greg Farnum
04:23 PM Bug #15903: smbtorture failing on pipe_number test
We aren't seeing this in regular nightlies; marking it down. Greg Farnum
03:28 PM Bug #17113: MDS EImport crashing with mds/journal.cc: 2929: FAILED assert(mds->sessionmap.get_ver...
It's not super-likely the rebooting client actually caused this problem. If it did, it was only incidentally, and it'... Greg Farnum
06:55 AM Bug #17113: MDS EImport crashing with mds/journal.cc: 2929: FAILED assert(mds->sessionmap.get_ver...
Full log was uploaded ceph-post-file: 610fd186-9150-4e6b-8050-37dc314af39b
Before I recover, I'd really like to se...
Tomasz Torcz
12:35 PM Bug #16655 (Resolved): ceph-fuse is not linked to libtcmalloc
Loïc Dachary
12:35 PM Bug #15705 (Resolved): ceph status mds output ignores active MDS when there is a standby replay
Loïc Dachary
11:56 AM Backport #15968 (Resolved): jewel: ceph status mds output ignores active MDS when there is a stan...
Loïc Dachary
11:54 AM Backport #15968 (In Progress): jewel: ceph status mds output ignores active MDS when there is a s...
Loïc Dachary
11:56 AM Backport #16697 (Resolved): jewel: ceph-fuse is not linked to libtcmalloc
Loïc Dachary
11:54 AM Backport #16697 (In Progress): jewel: ceph-fuse is not linked to libtcmalloc
Loïc Dachary
11:56 AM Backport #17131 (In Progress): jewel: Jewel: segfault in ObjectCacher::FlusherThread
Loïc Dachary
06:27 AM Backport #17131 (Resolved): jewel: Jewel: segfault in ObjectCacher::FlusherThread
https://github.com/ceph/ceph/pull/10864 Loïc Dachary
07:23 AM Bug #15702 (Resolved): mds: wrongly treat symlink inode as normal file/dir when symlink inode is ...
Loïc Dachary
07:20 AM Backport #16083 (Resolved): jewel: mds: wrongly treat symlink inode as normal file/dir when symli...
Loïc Dachary
01:11 AM Bug #16610 (Pending Backport): Jewel: segfault in ObjectCacher::FlusherThread
This got merged to master forever ago. Guess it should get backported too. Greg Farnum

08/24/2016

11:41 PM Bug #17105 (Fix Under Review): multimds: allow_multimds not required when max_mds is set in ceph....
PR: https://github.com/ceph/ceph/pull/10848 Patrick Donnelly
09:55 PM Bug #17096 (Won't Fix): Pool name is not displayed after changing CephFS File layout using extend...
I think this is just a result of not having the current OSDMap yet. If you're doing IO on the client, you're unlikely... Greg Farnum
08:59 PM Backport #17126 (Resolved): mds: fix double-unlock on shutdown
https://github.com/ceph/ceph/pull/10847 Loïc Dachary
06:00 PM Bug #17113 (Need More Info): MDS EImport crashing with mds/journal.cc: 2929: FAILED assert(mds->s...
It looks like you're running with multiple active MDSes, which is not currently recommended. We saw this in #16043 as... Greg Farnum
09:44 AM Bug #17113 (Can't reproduce): MDS EImport crashing with mds/journal.cc: 2929: FAILED assert(mds->...
I have tiny CEPH cluster (3xmon, 8xosd, 2xmds) with ceph-mds-10.2.2-2.fc24.x86_64.
Recently, one of the clients usin...
Tomasz Torcz
04:24 PM Bug #17115: kernel panic when running IO with cephfs and resource pool becomes full
We increased the pool size to a higher size. but system is in same state
Steps done:-
=========================...
Rohith Radhakrishnan
01:36 PM Bug #17115: kernel panic when running IO with cephfs and resource pool becomes full
These are warning (write blocked for too long) instead of panic. When pool is full, write osd requests get paused. If... Zheng Yan
01:12 PM Bug #17115 (Resolved): kernel panic when running IO with cephfs and resource pool becomes full
Steps:-
Create a data pool with limited quota size and start running IO from client. After the pool becomes full, ...
Rohith Radhakrishnan
03:53 PM Bug #16288 (Resolved): mds: `session evict` tell command blocks forever with async messenger (Tes...
Loïc Dachary
08:41 AM Support #17079: Io runs only on one pool even though 2 pools are attached to cephfs FS.
You are right. I could do that. Rohith Radhakrishnan
07:17 AM Support #17079: Io runs only on one pool even though 2 pools are attached to cephfs FS.
There is no option to do that. Your requirement is strange, why not enlarge quota of the first pool. Zheng Yan
05:58 AM Support #17079: Io runs only on one pool even though 2 pools are attached to cephfs FS.
@Zheng: What I would like to achieve is after adding 2 pools to a ceph FS, I should be able to redirect the objects f... Rohith Radhakrishnan

08/23/2016

06:23 PM Bug #17105: multimds: allow_multimds not required when max_mds is set in ceph.conf at startup
I think we want to force users to set multi-mds flags explicitly, not implicitly via the initial config. I'm fine wit... Greg Farnum
06:02 PM Bug #17105 (Resolved): multimds: allow_multimds not required when max_mds is set in ceph.conf at ...
Problem:... Patrick Donnelly
04:08 PM Bug #17099 (Closed): MDS command for listing mds_cache_size
The config option can be shown through the standard config interface. The counter values are exported via the perf co... Greg Farnum
07:52 AM Bug #17099 (Closed): MDS command for listing mds_cache_size
Not able to find mds_cache_size listed anywhere. For e.g in ceph mds dump or elsewhere. If currently there is no way ... Rohith Radhakrishnan
01:44 PM Backport #16621 (Resolved): jewel: mds: `session evict` tell command blocks forever with async me...
Loïc Dachary
01:27 PM Bug #17096: Pool name is not displayed after changing CephFS File layout using extended attributes
Just saw the note: *Note When reading layouts, the pool will usually be indicated by name. However, in rare cases whe... Rohith Radhakrishnan
07:39 AM Bug #16396 (Resolved): Fix shutting down mds timed-out due to deadlock
Loïc Dachary
07:39 AM Bug #16358 (Resolved): Session::check_access() is buggy
Loïc Dachary
07:39 AM Bug #16164 (Resolved): mds: enforce a dirfrag limit on entries
Loïc Dachary
07:39 AM Bug #16137 (Resolved): client: crash in unmount when fuse_use_invalidate_cb is enabled
Loïc Dachary
07:39 AM Bug #16042 (Resolved): MDS Deadlock on shutdown active rank while busy with metadata IO
Loïc Dachary
07:39 AM Bug #16022 (Resolved): MDSMonitor::check_subs() is very buggy
Loïc Dachary
07:39 AM Bug #16013 (Resolved): Failing file operations on kernel based cephfs mount point leaves unaccess...
Loïc Dachary
07:39 AM Bug #12653 (Resolved): fuse mounted file systems fails SAMBA CTDB ping_pong rw test with v9.0.2
Loïc Dachary
06:51 AM Backport #16037 (Resolved): jewel: MDSMonitor::check_subs() is very buggy
Loïc Dachary
06:51 AM Backport #16215 (Resolved): jewel: client: crash in unmount when fuse_use_invalidate_cb is enabled
Loïc Dachary
06:51 AM Backport #16299 (Resolved): jewel: mds: fix SnapRealm::have_past_parents_open()
Loïc Dachary
06:51 AM Backport #16320 (Resolved): jewel: fs: fuse mounted file systems fails SAMBA CTDB ping_pong rw te...
Loïc Dachary
06:51 AM Backport #16515 (Resolved): jewel: Session::check_access() is buggy
Loïc Dachary
06:50 AM Backport #16560 (Resolved): jewel: mds: enforce a dirfrag limit on entries
Loïc Dachary
06:50 AM Backport #16620 (Resolved): jewel: Fix shutting down mds timed-out due to deadlock
Loïc Dachary
06:50 AM Backport #16625 (Resolved): jewel: Failing file operations on kernel based cephfs mount point lea...
Loïc Dachary
06:50 AM Backport #16797 (Resolved): jewel: MDS Deadlock on shutdown active rank while busy with metadata IO
Loïc Dachary

08/22/2016

06:45 PM Bug #17096 (Won't Fix): Pool name is not displayed after changing CephFS File layout using extend...
Steps-
1)Create a pool and a metadata pool and create a new cephfs using the pools and mount the file system from ...
Rohith Radhakrishnan
11:47 AM Support #17079: Io runs only on one pool even though 2 pools are attached to cephfs FS.
Tried setting a non-default pool using "SETFATT", but I am not able to set more than one pool to a directory at a tim... Rohith Radhakrishnan

08/19/2016

04:18 PM Bug #14716: "Thread.cc: 143: FAILED assert(status == 0)" in fs-hammer---basic-smithi
Same in hammer 0.94.8
http://qa-proxy.ceph.com/teuthology/yuriw-2016-08-18_20:11:00-fs-master---basic-smithi/373246/...
Yuri Weinstein
01:18 PM Support #17079: Io runs only on one pool even though 2 pools are attached to cephfs FS.
the first pool is default pool. see http://docs.ceph.com/docs/master/cephfs/file-layouts/ for how to store file in no... Zheng Yan
11:23 AM Support #17079 (New): Io runs only on one pool even though 2 pools are attached to cephfs FS.
Steps:-
1) Create a pool and a metadata pool and create a new cephfs using the pools.
2) Now create another data ...
Rohith Radhakrishnan

08/18/2016

08:44 PM Bug #17074: "SELinux denials" in knfs-master-testing-basic-smithi
Not a result of environmental issue or system misconfiguration. David Galloway
08:21 PM Bug #17074 (Closed): "SELinux denials" in knfs-master-testing-basic-smithi
This is point release tests hammer 0.94.8
Run: http://pulpito.front.sepia.ceph.com/yuriw-2016-08-17_20:57:47-knfs-...
Yuri Weinstein
06:47 AM Bug #17069: multimds: slave rmdir assertion failure
strange. have you ever use snapshot on the testing cluster? Zheng Yan

08/17/2016

08:06 PM Bug #17069 (Closed): multimds: slave rmdir assertion failure
... Patrick Donnelly
04:09 PM Backport #16946: jewel: client: nlink count is not maintained correctly
https://github.com/ceph/ceph/pull/10386/commits/f3605d39e53b3ff777eb64538abfa62a5f98a4f2 conflicts Loïc Dachary

08/16/2016

02:20 PM Feature #16419: add statx-like interface to libcephfs
Ok, smaller set of changes is now merged. Now we have the larger set to contend with. I've gone ahead and rolled some... Jeff Layton

08/15/2016

06:48 PM Feature #12274 (In Progress): mds: start forward scrubs from all subtree roots, skip non-auth met...
Douglas Fuller

08/13/2016

11:09 AM Feature #16419: add statx-like interface to libcephfs
I have a PR up with a smaller set of changes here:
https://github.com/ceph/ceph/pull/10691
This is just cha...
Jeff Layton

08/12/2016

11:18 AM Bug #16640: libcephfs: Java bindings failing to load on CentOS

So, the PR had a passing test run:
https://github.com/ceph/ceph-qa-suite/pull/1084
http://pulpito.ceph.com/jspray...
John Spray
01:43 AM Bug #16983: mds: handle_client_open failing on open
It's already fixed by https://github.com/ceph/ceph/pull/8778 Zheng Yan

08/11/2016

05:04 PM Bug #16013: Failing file operations on kernel based cephfs mount point leaves unaccessible file b...
User seeing an assertion failure in the MDS in v10.2.1:
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-A...
Patrick Donnelly
05:03 PM Bug #16983: mds: handle_client_open failing on open
Zheng, I think you fixed this in 4d15eb12298e007744486e28924a6f0ae071bd06 from PR #8778.
Here's the issue from cep...
Patrick Donnelly

08/10/2016

06:18 PM Bug #16983 (Resolved): mds: handle_client_open failing on open
Randy Orr reported an assertion failure on the ceph-users list:... Patrick Donnelly
11:33 AM Feature #16419: add statx-like interface to libcephfs
Test run mostly passed last night, with only failures for unrelated problems -- the known problem with valgrind on ce... Jeff Layton

08/09/2016

08:04 PM Feature #15069 (Fix Under Review): MDS: multifs: enable two filesystems to point to same pools if...
https://github.com/ceph/ceph/pull/10636 Douglas Fuller
08:04 PM Feature #15068 (Fix Under Review): fsck: multifs: enable repair tools to read from one filesystem...
Douglas Fuller
08:04 PM Feature #15068: fsck: multifs: enable repair tools to read from one filesystem and write to another
https://github.com/ceph/ceph/pull/10636 Douglas Fuller
07:47 PM Feature #16419: add statx-like interface to libcephfs
Found it. I had transposed the size and change_attr args in one call to update_inode_file_bits. fsx now seems to be O... Jeff Layton
03:22 PM Feature #16419: add statx-like interface to libcephfs
Mostly working now, but I'm seeing occasional problems with truncating files. I bisected the problem down to a one li... Jeff Layton
06:14 PM Feature #16973 (Resolved): Log path as well as ino when detecting metadata damage
Currently our cluster log messages look like this:... John Spray
04:21 PM Bug #16909 (Fix Under Review): Stopping an MDS rank does not stop standby-replays for that rank
https://github.com/ceph/ceph/pull/10628 John Spray
04:20 PM Bug #16919 (Fix Under Review): MDS: Standby replay daemons don't drop purged strays
https://github.com/ceph/ceph/pull/10606 John Spray
01:29 PM Bug #16925: multimds: cfuse (?) hang on fsx.sh workunit
this can either be caused by hang MDS request or be caused by hang read/write (MDS does not properly issue Frw caps t... Zheng Yan
10:42 AM Bug #16954: Metadata damage reported with snapshots+smallcache+dirfrags ("object missing on disk")
Hmm, no failures in that re-run, so it's not quite completely reproducible. John Spray

08/08/2016

02:45 PM Bug #16926: multimds: kclient fails to mount
(pass "-k testing" when scheduling runs that will use a kclient, to ensure you're getting a nice recent cephfs kernel) John Spray
01:49 PM Bug #16914: multimds: pathologically slow deletions in some tests
retest with fuse default permissions set differently because it's doing too many getattr at the moment John Spray
12:51 PM Support #16884: rename() doesn't work between directories
Donatas: currently, renaming files in and out of trees with different quotas is going to give you EXDEV. You can wor... John Spray
06:41 AM Support #16884: rename() doesn't work between directories
guys, so what's the summary about this 'feature'? Donatas Abraitis
10:48 AM Bug #16954: Metadata damage reported with snapshots+smallcache+dirfrags ("object missing on disk")
... John Spray
10:36 AM Bug #16954: Metadata damage reported with snapshots+smallcache+dirfrags ("object missing on disk")
Given that it happened twice in one job, seems a decent change it's reproducible, let's see:
http://pulpito.ceph.com...
John Spray
10:31 AM Bug #16954 (New): Metadata damage reported with snapshots+smallcache+dirfrags ("object missing on...

http://pulpito.ceph.com/jspray-2016-08-07_16:42:13-fs-wip-prompt-frag-distro-basic-mira/353833...
John Spray
08:44 AM Bug #14681 (Resolved): Wrong ceph get mdsmap assertion
Loïc Dachary
08:44 AM Bug #14319 (Resolved): Double decreased the count to trim caps which will cause failing to respon...
Loïc Dachary
08:42 AM Bug #16154 (Resolved): mds: lock waiters are not finished in the same order that they were added
Loïc Dachary
08:42 AM Bug #15920 (Resolved): mds/StrayManager.cc: 520: FAILED assert(dnl->is_primary())
Loïc Dachary
08:42 AM Bug #15723 (Resolved): client: fstat cap release
Loïc Dachary
08:42 AM Bug #15689 (Resolved): Confusing MDS log message when shut down with stalled journaler reads
Loïc Dachary
08:42 AM Feature #15615 (Resolved): CephFSVolumeClient: List authorized IDs by share
Loïc Dachary
08:41 AM Feature #15406 (Resolved): Add versioning to CephFSVolumeClient interface
Loïc Dachary
08:34 AM Bug #11482: kclient: intermittent log warnings "client.XXXX isn't responding to mclientcaps(revoke)"
infernalis is EOL Loïc Dachary
08:33 AM Bug #15050 (Resolved): deleting striped file in cephfs doesn't free up file's space
Loïc Dachary
08:32 AM Bug #14144 (Resolved): standy-replay MDS does not cleanup finished replay threads
Loïc Dachary
08:28 AM Backport #15281 (Rejected): infernalis: standy-replay MDS does not cleanup finished replay threads
Loïc Dachary
08:28 AM Backport #15057 (Rejected): infernalis: deleting striped file in cephfs doesn't free up file's space
Loïc Dachary
08:28 AM Backport #14843 (Rejected): infernalis: test_object_deletion fails (tasks.cephfs.test_damage.Test...
Loïc Dachary
08:28 AM Backport #14761 (Rejected): infernalis: ceph-fuse does not mount at boot on Debian Jessie
Loïc Dachary
08:28 AM Backport #14690 (Rejected): infernalis: Client::_fsync() on a given file does not wait unsafe req...
Loïc Dachary
08:28 AM Backport #13890 (Rejected): infernalis: Race in TestSessionMap.test_version_splitting
Loïc Dachary
08:25 AM Backport #16299: jewel: mds: fix SnapRealm::have_past_parents_open()
https://github.com/ceph/ceph/pull/9447 Loïc Dachary
08:22 AM Backport #14668 (Resolved): hammer: Wrong ceph get mdsmap assertion
Loïc Dachary
08:22 AM Backport #15056 (Resolved): hammer: deleting striped file in cephfs doesn't free up file's space
Loïc Dachary
08:21 AM Backport #15512 (Resolved): hammer: Double decreased the count to trim caps which will cause fail...
Loïc Dachary
08:21 AM Backport #15898 (Resolved): jewel: Confusing MDS log message when shut down with stalled journale...
Loïc Dachary
08:21 AM Backport #16041 (Resolved): jewel: mds/StrayManager.cc: 520: FAILED assert(dnl->is_primary())
Loïc Dachary
08:21 AM Backport #16082 (Resolved): hammer: mds: wrongly treat symlink inode as normal file/dir when syml...
Loïc Dachary
08:21 AM Backport #16135 (Resolved): jewel: MDS: fix getattr starve setattr
Loïc Dachary
08:21 AM Backport #16136 (Resolved): jewel: MDSMonitor fixes
Loïc Dachary
08:21 AM Backport #16152 (Resolved): jewel: fs: client: fstat cap release
Loïc Dachary
08:20 AM Backport #16626 (Resolved): hammer: Failing file operations on kernel based cephfs mount point le...
Loïc Dachary
08:19 AM Backport #16830 (Resolved): jewel: CephFSVolumeClient: List authorized IDs by share
Loïc Dachary
08:19 AM Backport #16831 (Resolved): jewel: Add versioning to CephFSVolumeClient interface
Loïc Dachary

08/05/2016

09:04 PM Backport #16946 (Resolved): jewel: client: nlink count is not maintained correctly
https://github.com/ceph/ceph/pull/10877 Nathan Cutler
02:57 PM Bug #16919 (In Progress): MDS: Standby replay daemons don't drop purged strays
John Spray
02:57 PM Bug #16909 (In Progress): Stopping an MDS rank does not stop standby-replays for that rank
John Spray

08/04/2016

08:04 PM Bug #16771: mon crash in MDSMonitor::prepare_beacon on ARM
so, I tried to run ceph outside of docker to run gdb on ceph-mon, but I don't know what I suppose to see.
$ gdb /usr...
stephane beuret
07:26 PM Bug #16926 (Rejected): multimds: kclient fails to mount
In many test cases, the kernel client fails to mount with EIO:
http://pulpito.ceph.com/pdonnell-2016-08-03_12:43:1...
Patrick Donnelly
06:58 PM Bug #16925 (Can't reproduce): multimds: cfuse (?) hang on fsx.sh workunit
http://pulpito.ceph.com/pdonnell-2016-07-18_20:02:54-multimds-master---basic-mira/321794/... Patrick Donnelly
05:23 PM Bug #16924: Crash replaying EExport
git blame points at b7e698a52bf7838f8e37842074c510a6561f165b from Zheng.
> mds: no bloom filter for replica dir
>...
Greg Farnum
04:49 PM Bug #16924 (Resolved): Crash replaying EExport
... John Spray
01:55 PM Bug #16919: MDS: Standby replay daemons don't drop purged strays
The standby does have all the information about which files are open (since they get journaled), right? Or do we only... Greg Farnum
09:59 AM Bug #16919 (Resolved): MDS: Standby replay daemons don't drop purged strays

This is not fatal, because the inodes will ultimately end up at the top of the LRU list and get trimmed, but it's a...
John Spray
12:24 PM Documentation #16906: doc: clarify path restriction instructions
@John Spray
fixup:
https://github.com/ceph/ceph/pull/10573/commits/d1277f116cd297bae8da7b3e1a7000d3f99c6a51
huanwen ren
10:41 AM Bug #16920 (New): mds.inodes* perf counters sound like the number of inodes but they aren't

These counters actually reflect the LRU, which is a collection of dentries, not inodes.
mds_mem.ino on the other...
John Spray

08/03/2016

09:08 PM Bug #16914 (Resolved): multimds: pathologically slow deletions in some tests
http://qa-proxy.ceph.com/teuthology/pdonnell-2016-07-18_20:02:54-multimds-master---basic-mira/321823/teuthology.log
...
Patrick Donnelly
08:52 PM Bug #16886: multimds: kclient hang (?) in tests
Another blogbench:
http://pulpito.ceph.com/pdonnell-2016-07-29_08:28:00-multimds-master---basic-mira/339886/
<p...
Patrick Donnelly
08:01 PM Feature #16419: add statx-like interface to libcephfs
I have a prototype set for this, but I now think that the handling of the change_attr is wrong for directories. I'm g... Jeff Layton
01:56 PM Bug #16909 (Resolved): Stopping an MDS rank does not stop standby-replays for that rank

Run vstart with MDS=2 and -s flag
Set max_mds to 2
See that you get two active daemons and two standby-replays
S...
John Spray
12:57 PM Bug #16640: libcephfs: Java bindings failing to load on CentOS
Yes, you really don't want to load the unversioned library at runtime. It's possible that you'll end up picking up a ... Jeff Layton
11:21 AM Bug #16881: RuntimeError: Files in flight high water is unexpectedly low (0 / 6)
They are racy, but this particular case is a bit odd:... John Spray
11:08 AM Bug #16668 (Pending Backport): client: nlink count is not maintained correctly
John Spray
11:02 AM Documentation #16906: doc: clarify path restriction instructions
So there's no bug here as such, it's just that the instructions don't explicitly tell you to write out your client ke... John Spray
08:08 AM Documentation #16906 (Resolved): doc: clarify path restriction instructions
I do path restriction follow:http://docs.ceph.com/docs/master/cephfs/client-auth/... huanwen ren
10:45 AM Bug #16880 (Duplicate): saw valgrind issue <kind>Leak_StillReachable</kind> in /var/log/ceph/va...
See:
http://tracker.ceph.com/issues/14794
http://tracker.ceph.com/issues/15356
(aka The Mystery Of the Valgrind ...
John Spray
10:41 AM Bug #16876 (Duplicate): java.lang.UnsatisfiedLinkError: Can't load library: /usr/lib/jni/libcephf...
This should be fixed in master now (there was a backed-out change for 16640 in teuthology, then finally the fix was h... John Spray
10:39 AM Bug #16879 (Resolved): scrub: inode wrongly marked free: 0x10000000002
John Spray

08/02/2016

07:08 PM Bug #16186 (Duplicate): kclient: drops requests without poking system calls on reconnect
Jeff Layton
07:08 PM Bug #16186: kclient: drops requests without poking system calls on reconnect
I'm going to go ahead and close this out, and pursue the follow-up work in tracker #15255.
Jeff Layton
03:33 PM Bug #16668 (Resolved): client: nlink count is not maintained correctly
Ok, PR is now merged! Jeff Layton
11:13 AM Bug #16879 (Fix Under Review): scrub: inode wrongly marked free: 0x10000000002
https://github.com/ceph/ceph-qa-suite/pull/1107 John Spray

08/01/2016

11:15 PM Feature #10627: teuthology: qa: enable Samba runs on RHEL
Passing this to John to watch. Greg Farnum
08:50 PM Support #16884: rename() doesn't work between directories
Zheng, what are the limits and requirements of that quota root EXDEV?
I think it's probably required and can't cha...
Greg Farnum
08:15 PM Support #16884: rename() doesn't work between directories
Debug output is:
todir->snapid:-2 todir->quota.is_enable:0 fromdir->snapid:-2 fromdir->quota->max_files:20000 return...
Donatas Abraitis
08:12 PM Support #16884: rename() doesn't work between directories
What about removing this block at all? Or is it required too much? Donatas Abraitis
08:04 PM Support #16884: rename() doesn't work between directories
Looks like this part is failing: https://github.com/ceph/ceph/blob/0080b6bc92cefdd2115c904fd0c83ae83c9c2f01/src/clien... Donatas Abraitis
08:00 PM Support #16884: rename() doesn't work between directories
More details, please. Cross-directory rename definitely works in general. What's the output of "mount"? What versions... Greg Farnum
07:03 PM Support #16884 (Closed): rename() doesn't work between directories
Hi folks!
looks like rename() just doesn't work between directories. Here is the snippet FTP daemon does:
#incl...
Donatas Abraitis
08:39 PM Bug #16886: multimds: kclient hang (?) in tests
Updated title/description. Patrick Donnelly
07:54 PM Bug #16886: multimds: kclient hang (?) in tests
Well I feel silly. This is actually more general but wasn't obvious by how I had organized the failures. I'm going to... Patrick Donnelly
07:32 PM Bug #16886 (Can't reproduce): multimds: kclient hang (?) in tests
There are strange pauses which are showing up in several tests for the kclient:
* http://qa-proxy.ceph.com/teuthol...
Patrick Donnelly
06:57 PM Bug #16771: mon crash in MDSMonitor::prepare_beacon on ARM
root@ceph1:/# ps -ef
UID PID PPID C STIME TTY TIME CMD
ceph 1 0 0 18:51 ? 00:...
stephane beuret
05:42 AM Bug #16771: mon crash in MDSMonitor::prepare_beacon on ARM
I must admit that I have trouble putting it in place. I do not know enough how to use gdb, and as my ceph-mon is in a... stephane beuret
04:42 PM Bug #16879: scrub: inode wrongly marked free: 0x10000000002
This test also fails with the master branch (as of earlier this morning):
http://pulpito.ceph.com/jlayton-2016...
Jeff Layton
03:08 PM Bug #16879: scrub: inode wrongly marked free: 0x10000000002
Rebased onto current master branch, and still seeing the error. Rerunning the test now on a branch without any of my ... Jeff Layton
01:00 PM Bug #16879: scrub: inode wrongly marked free: 0x10000000002
Reran the test and it failed again: (btw: thanks Nathan for the pointer to how to filter out failures and rerun only ... Jeff Layton
12:34 PM Bug #16879: scrub: inode wrongly marked free: 0x10000000002
Ahh thanks, Nathan. Ok, this is a recently-added test and my local ceph-qa-suite was missing it. A git pull fixed tha... Jeff Layton
11:36 AM Bug #16879: scrub: inode wrongly marked free: 0x10000000002
I found it by looking at the "task" function in "tasks/cephfs_test_runner.py" - it says: ... Nathan Cutler
11:35 AM Bug #16879: scrub: inode wrongly marked free: 0x10000000002
Hi Jeff, this:
https://github.com/ceph/ceph-qa-suite/blob/master/tasks/cephfs/test_forward_scrub.py
Nathan Cutler
10:51 AM Bug #16879: scrub: inode wrongly marked free: 0x10000000002
Message comes from CInode::validate_disk_state, but I haven't yet been able to figure out where the test itself comes... Jeff Layton
10:23 AM Bug #16879 (Resolved): scrub: inode wrongly marked free: 0x10000000002
I ran the "fs" testsuite on a branch that has a pile of small, userland client-side patches. One of the tests (tasks/... Jeff Layton
01:34 PM Bug #16807: Crash in handle_slave_rename_prep
Zheng Yan
12:56 PM Bug #16876: java.lang.UnsatisfiedLinkError: Can't load library: /usr/lib/jni/libcephfs_jni.so
Of course, it may be that I reached the box too late and the filesystem had been changed. I'm not sure how to tell. E... Jeff Layton
10:57 AM Bug #16832 (Resolved): libcephfs failure at shutdown (Attempt to free invalid pointer)
I haven't seen this in the latest test runs, so I'm going to go ahead and close this out under the assumption that is... Jeff Layton
10:40 AM Bug #16881: RuntimeError: Files in flight high water is unexpectedly low (0 / 6)
Comments in test_strays.py seem to indicate that this test is racy anyway:... Jeff Layton
10:39 AM Bug #16881 (Resolved): RuntimeError: Files in flight high water is unexpectedly low (0 / 6)
During a test_files_throttle test run, I hit the following error:... Jeff Layton
10:34 AM Bug #16880: saw valgrind issue <kind>Leak_StillReachable</kind> in /var/log/ceph/valgrind/clien...
client.0.log is here:
http://qa-proxy.ceph.com/teuthology/jlayton-2016-07-29_18:51:42-fs-wip-jlayton-nlink---b...
Jeff Layton
10:27 AM Bug #16880 (Duplicate): saw valgrind issue <kind>Leak_StillReachable</kind> in /var/log/ceph/va...
One of my "fs" test runs over the weekend failed with this:... Jeff Layton
02:36 AM Bug #16768: multimds: check_rstat assertion failure
Here's another instance of the assertion failure on a more recent master branch:
http://qa-proxy.ceph.com/teutholo...
Patrick Donnelly

07/31/2016

09:12 PM Bug #16876 (Duplicate): java.lang.UnsatisfiedLinkError: Can't load library: /usr/lib/jni/libcephf...
I had a failed fs testsuite run, and a couple of the jobs failed with what looks like the error below:... Jeff Layton

07/29/2016

10:18 PM Documentation #16743 (Resolved): client: config settings missing in documentation
Patrick Donnelly
12:18 PM Backport #16797 (In Progress): jewel: MDS Deadlock on shutdown active rank while busy with metada...
Abhishek Varshney
12:13 PM Bug #16842: mds: replacement MDS crashes on InoTable release
Min Chen: can you describe the client part of how to reproduce this? What does the client have to be doing to reprod... John Spray
12:08 PM Backport #16621 (In Progress): jewel: mds: `session evict` tell command blocks forever with async...
Abhishek Varshney
11:48 AM Backport #16620 (In Progress): jewel: Fix shutting down mds timed-out due to deadlock
Abhishek Varshney
11:44 AM Backport #16299 (In Progress): jewel: mds: fix SnapRealm::have_past_parents_open()
Abhishek Varshney
11:00 AM Cleanup #15923 (Resolved): MDS: remove TMAP2OMAP check and move Objecter into MDSRank
John Spray
10:58 AM Cleanup #16195 (Resolved): mds: Don't spam log with standby_replay_restart messages
John Spray
10:44 AM Bug #16857 (Duplicate): Crash in Client::_invalidate_kernel_dcache
... John Spray
03:18 AM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
It's not only Ceph's locks, mutexes, etc. that we need to be aware of or concerned with. I have seen multiple occurre... Brad Hubbard
03:08 AM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
Kefu Chai
03:05 AM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
John,
i marked PR#10472 as a fix of this issue. it does. but i would like to keep this issue open, because:
by ...
Kefu Chai

07/28/2016

06:33 PM Bug #16842: mds: replacement MDS crashes on InoTable release
This looks more complicated than that to reproduce. The code that's crashing is timing out a client connection that d... Greg Farnum
06:25 AM Bug #16842 (Can't reproduce): mds: replacement MDS crashes on InoTable release
ceph version 10.2.0-2638-gf7fc985
reproduce step:
1. new fs and start mds.a
2. start mds.b
3. kill mds.a
fai...
Min Chen
01:58 PM Bug #16844 (Duplicate): hammer: libcephfs-java/test.sh fails
Nathan Cutler
11:09 AM Bug #16844 (Duplicate): hammer: libcephfs-java/test.sh fails
Failing consistently on hammer-backports branch:
http://pulpito.ceph.com/smithfarm-2016-07-25_05:09:12-fs-hammer-ba...
Nathan Cutler
11:20 AM Bug #16556 (Fix Under Review): LibCephFS.InterProcessLocking failing on master and jewel
https://github.com/ceph/ceph/pull/10472 Kefu Chai
08:40 AM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
... Kefu Chai

07/27/2016

09:07 PM Bug #16829: ceph-mds crashing constantly
1) Possibly, but we're more likely to just lock out pools which have data in them.
2) You mean you intermingled RB...
Greg Farnum
08:50 PM Bug #16829: ceph-mds crashing constantly
So, two things:
1) can ceph-mds be made more resiliant when finding data from non-existing filesystems?
2) I ca...
Tomasz Torcz
06:31 PM Bug #16829 (Closed): ceph-mds crashing constantly
It looks like you did "fs rm" and "fs new" but kept the same metadata pool in RADOS. That doesn't work; you can resol... Greg Farnum
12:50 PM Bug #16829 (Closed): ceph-mds crashing constantly
I'm using CEPH packages from Fedora 24: ceph-mds-10.2.2-2.fc24.x86_64
I've created simple cephfs once, stored some...
Tomasz Torcz
04:41 PM Bug #16832: libcephfs failure at shutdown (Attempt to free invalid pointer)
@Jeff: Just in case you don't know it yet, here is a trick for rescheduling failed and dead jobs from a previous run ... Nathan Cutler
04:17 PM Bug #16832: libcephfs failure at shutdown (Attempt to free invalid pointer)
The other two test failures -- one was a segfault in ceph_test_libcephfs:
(gdb) bt
#0 0x00007f771c5aaa63 in lock...
Jeff Layton
02:58 PM Bug #16832: libcephfs failure at shutdown (Attempt to free invalid pointer)
> 1) some binary segfaulted, but I don't seem to be able to track down the core to see what actually failed:
<dgallo...
David Galloway
02:47 PM Bug #16832 (Resolved): libcephfs failure at shutdown (Attempt to free invalid pointer)
My fs test run had 3 failures:
1) some binary segfaulted, but I don't seem to be able to track down the core to se...
Jeff Layton
02:33 PM Feature #15406 (Pending Backport): Add versioning to CephFSVolumeClient interface
Ramana Raja
02:32 PM Feature #15615 (Pending Backport): CephFSVolumeClient: List authorized IDs by share
Ramana Raja
01:54 PM Backport #16831 (In Progress): jewel: Add versioning to CephFSVolumeClient interface
Ramana Raja
01:53 PM Backport #16831 (Resolved): jewel: Add versioning to CephFSVolumeClient interface
https://github.com/ceph/ceph/pull/10453
https://github.com/ceph/ceph-qa-suite/pull/1100
Ramana Raja
01:23 PM Backport #16830 (In Progress): jewel: CephFSVolumeClient: List authorized IDs by share
Ramana Raja
01:03 PM Backport #16830 (Resolved): jewel: CephFSVolumeClient: List authorized IDs by share
https://github.com/ceph/ceph/pull/10453
https://github.com/ceph/ceph-qa-suite/pull/1100
Ramana Raja
11:58 AM Cleanup #16035 (Resolved): Remove "cephfs" CLI
Nathan Cutler
11:01 AM Cleanup #16035: Remove "cephfs" CLI
Mop-up *master PR*: https://github.com/ceph/ceph/pull/10444 Nathan Cutler
11:00 AM Cleanup #16035 (Fix Under Review): Remove "cephfs" CLI
Nathan Cutler
11:22 AM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
Looking into what's happening in the case of running LibCephFS.InterProcessLocking on its own, I see that the forked ... John Spray
05:52 AM Bug #16556: LibCephFS.InterProcessLocking failing on master and jewel
tested with latest master, still fails.
@greg, do we have a fix for this issue now?
Kefu Chai
10:59 AM Cleanup #16808 (Resolved): Merge "ceph-fs-common" into "ceph-common"
Nathan Cutler
06:27 AM Cleanup #16808: Merge "ceph-fs-common" into "ceph-common"
https://github.com/ceph/ceph-qa-suite/pull/1098 is still open Nathan Cutler
05:32 AM Cleanup #16808 (Resolved): Merge "ceph-fs-common" into "ceph-common"
Kefu Chai
 

Also available in: Atom