Activity
From 05/28/2015 to 06/26/2015
06/26/2015
- 04:21 PM Feature #12107: mds: use versioned wire protocol; obviate CEPH_MDS_PROTOCOL
- Mmmm. For over-the-wire encodings that don't go to disk, it's only about whether cross-version daemons can communicat...
- 01:24 PM Bug #12175 (Resolved): Fix ceph-fuse --help
- Problems with this:
* it starts running after printing the help output, instead of quitting
* it omits useful/i... - 01:03 PM Feature #12138: cephfs-data-scan: write inode backtraces when injecting to lost+found
- Hmm. When we do move something into lost+found, are we sure we can't make use of the backtrace any more, or might it ...
- 10:55 AM Bug #12123: testlibcephfs: segfault in preadv/pwritev tests
- Please use the new pull request:
https://github.com/ceph/ceph/pull/5084 - 03:14 AM Bug #12123: testlibcephfs: segfault in preadv/pwritev tests
- A pull request for fixing this issue has been submitted, pending on review.
- 03:13 AM Bug #12123 (Fix Under Review): testlibcephfs: segfault in preadv/pwritev tests
- https://github.com/ceph/ceph/pull/5083
- 01:25 AM Bug #12172 (Resolved): tasks.cephfs.test_auto_repair.TestMDSAutoRepair fails
- http://qa-proxy.ceph.com/teuthology/teuthology-2015-06-23_23:04:02-fs-next---basic-multi/947532/teuthology.log
06/25/2015
- 01:22 PM Feature #12107: mds: use versioned wire protocol; obviate CEPH_MDS_PROTOCOL
- yes, it's better to not rely on CEPH_MDS_PROTOCOL. (CEPH_MDS_PROTOCOL will make the rework easier)
- 11:03 AM Feature #12107: mds: use versioned wire protocol; obviate CEPH_MDS_PROTOCOL
- Oh right, so as Zheng reminds me in his patch for #12105, we do have the CEPH_MDS_PROTOCOL value for these changes.
... - 11:01 AM Bug #12105 (Resolved): CInode misses oldest_snap field during migration
- ...
- 08:36 AM Bug #12105 (Fix Under Review): CInode misses oldest_snap field during migration
- 08:39 AM Bug #12019 (Resolved): multiple_rsync failure
- 02:30 AM Bug #11989 (Resolved): Cephfs Kernel Client data corruption
06/24/2015
- 11:03 PM Feature #4161 (Fix Under Review): MDS: add file layout to head object
- https://github.com/ceph/ceph/pull/5070
- 01:43 PM Feature #4161 (In Progress): MDS: add file layout to head object
- 12:46 PM Feature #12145 (Resolved): cephfs-data-scan: pgls filter for 0th file objects
- Currently, when iterating over 0th objects, we are actually listing all the objects and selecting the ones we care ab...
- 12:44 PM Feature #12144 (Resolved): cephfs-data-scan: integrated with sharded pgls
- This bit was taken out pending the new-style sharded pgls that should be in infernalis.
cephfs-data-scan should ta... - 12:39 PM Feature #12143 (New): cephfs-data-scan: Tool for orchestrating multiple workers
To run at any kind of scale, this tool requires multiple workers executing across multiple clients.
It would be ...- 12:35 PM Feature #12142 (New): cephfs-data-scan: Structured output of errors and operations done
Need a machine-consumable list of:
* Any I/O or encoding errors encountered (i.e. objects that might need manual ...- 12:33 PM Feature #12141 (New): cephfs-data-scan: File size correction from backward scan
- Currently, if a dentry already exists at the backtrace location which points to the inode, we do nothing. We should ...
- 12:30 PM Feature #12140 (New): cephfs-data-scan: Use ancestor file layouts when injecting inodes
- Currently we synthesize layouts on a best effort basis when injecting. We should also look for ancestor layouts whic...
- 12:29 PM Feature #12139 (New): cephfs-data-scan: cache fragtrees during injection
In order to inject dentries, we have to learn the fragtree of the directory we're injecting into. The process exis...- 12:28 PM Feature #12138 (New): cephfs-data-scan: write inode backtraces when injecting to lost+found
- Currently, for cases where we inject linkage for an inode into /lost+found, the inode's backtrace potentially still p...
- 12:26 PM Feature #12137 (New): cephfs-data-scan: backward scan of dirfrag objects, inject orphans
Similar procedure to what we already have for data objects: inject linkage for orphaned (detected a la #12133) dirf...- 12:24 PM Feature #12136 (New): fsck: snapshots: Enumerate snapshots during scan_extents
Optionally, issue a RADOS op to list all the snapshots for every object seen, and accumulate these into a set on th...- 12:20 PM Feature #12135 (New): cephfs-data-scan: Layout override by path
We can't always guess file layouts correctly. Provide a mechanism for users to manually specify the file layout to...- 12:18 PM Feature #12134 (New): cephfs-data-scan: Filter on ino/path/dname expression
Sometimes the user might want to go and recover only certain files from a damaged filesystem, or they might know th...- 12:17 PM Feature #12133 (Resolved): cephfs-data-scan: Filter on inodes not touched by forward scrub
Where forward scrub has marked those inodes that it has touched, add an option (possibly the default) to cephfs-dat...- 12:15 PM Feature #12132 (Resolved): cephfs-data-scan: Cleanup phase
A phase to remove the xattrs created during the scan_extents phase. They are small and harmless, but we should at ...- 12:13 PM Feature #12131 (New): cephfs-data-scan: Update InoTable after injection
Currently, inodes are injected without any consideration to whether their number is still regarded as free in inota...- 12:11 PM Feature #12130 (New): cephfs-data-scan: Accumulate dirfrag sizes on injection
Use a RADOS class to increment an xattr for the number of dentries injected into a fragment object during repair. ...
06/23/2015
- 01:38 PM Bug #12123: testlibcephfs: segfault in preadv/pwritev tests
- Emailed JevonQ to ask him to take a look at this
- 10:39 AM Bug #12123 (Resolved): testlibcephfs: segfault in preadv/pwritev tests
- http://qa-proxy.ceph.com/teuthology/teuthology-2015-06-19_23:04:01-fs-master---basic-multi/941831/...
- 10:26 AM Bug #9994: ceph-qa-suite: nfs mount timeouts
- Still happening:
http://qa-proxy.ceph.com/teuthology/teuthology-2015-06-19_23:10:01-knfs-next-testing-basic-multi/94...
06/22/2015
- 12:00 PM Bug #12105: CInode misses oldest_snap field during migration
- Probably can't do much about this safely without addressing the larger encoding issues here (#12107)
- 11:02 AM Bug #12105 (Resolved): CInode misses oldest_snap field during migration
See CInode::_encode_base -- because encoding is duplicated here wrt InodeStore.
Need to either update that fn or...- 11:46 AM Feature #12107: mds: use versioned wire protocol; obviate CEPH_MDS_PROTOCOL
- This code needs a rework to support versioning, because the outer message encoding in e.g. handle_discover_reply oper...
- 11:37 AM Feature #12107 (Resolved): mds: use versioned wire protocol; obviate CEPH_MDS_PROTOCOL
- 11:05 AM Feature #12106 (New): CInodes encoded unversioned in dirfrags
- Where we encode CInodes in the omap values of a dirfrag, we do it without any ENCODE_START decorators (InodeStoreBase...
06/21/2015
- 02:19 PM Bug #12088: cephfs client crash after enable readahead mechanism through setting conf option 'cli...
- While diving into the source code, we found the code path which will cause the crash. Described as the following step...
06/19/2015
- 07:13 PM Backport #12098 (Resolved): kernel_untar_build fails on EL7
- https://github.com/ceph/ceph/pull/5119
- 07:12 PM Backport #12097 (Resolved): kernel_untar_build fails on EL7
- https://github.com/ceph/ceph/pull/6000
- 03:39 PM Bug #12094 (Duplicate): "Segmentation fault" in smoke-master-distro-basic-multi run
- Run: http://pulpito.ceph.com/teuthology-2015-06-19_05:00:05-smoke-master-distro-basic-multi/
Job: 940481
Logs: http... - 11:03 AM Bug #12088 (Resolved): cephfs client crash after enable readahead mechanism through setting conf ...
- I run fio tool to test the randread performance of cephfs. Ceph client will crash, when I enable readahead on cep...
06/18/2015
- 10:16 AM Bug #11989: Cephfs Kernel Client data corruption
- Zheng Yan wrote:
> please try the attached patch
i have tried your patch with my test case with 300GB of data and... - 03:52 AM Bug #11989: Cephfs Kernel Client data corruption
- please try the attached patch
06/17/2015
- 01:32 PM Bug #11989: Cephfs Kernel Client data corruption
- I reproduced this locally
- 09:24 AM Bug #11784: ceph-fuse hang on unmount (stuck dentry refs)
- Hmm, there shouldn't have been any activity on the mount by this point. Maybe we've got some other kind of bug, though.
- 12:58 AM Bug #11784: ceph-fuse hang on unmount (stuck dentry refs)
- ...
06/16/2015
- 03:33 PM Bug #11758 (Pending Backport): kernel_untar_build fails on EL7
- 03:33 PM Bug #11758 (Resolved): kernel_untar_build fails on EL7
- 03:16 PM Bug #11758 (Fix Under Review): kernel_untar_build fails on EL7
- https://github.com/ceph/ceph/pull/4967
- 01:52 PM Bug #11985 (In Progress): MDS asserts in objecter when transitioning from replay to DNE
- 01:49 PM Bug #11541 (Resolved): MDS is crashed (mds/CDir.cc: 1391: FAILED assert(!is_complete()))
- 11:22 AM Bug #11913 (Resolved): Failure in TestClusterFull.test_barrier
- commit:bf9a9a2d9ff2be129b303d535899f60ad49f7c23
- 10:23 AM Bug #12019: multiple_rsync failure
- Yeah, I think this was me being silly when reading the log, I read straight from the rsync invocation to the error, w...
- 10:09 AM Bug #12019: multiple_rsync failure
- 10:09 AM Bug #12019: multiple_rsync failure
- commit:0804655725d84d866a32826203638fcfd71d4b51
Since we're using sudo to copy we presumably need it to delete. I ... - 07:13 AM Bug #12019 (Fix Under Review): multiple_rsync failure
- https://github.com/ceph/ceph/pull/4964
- 07:10 AM Bug #11989: Cephfs Kernel Client data corruption
- could you please provide me a list of corrupt blocks (offset and size of corrupt block). Besides, could you please tr...
06/15/2015
- 09:56 AM Bug #12019 (Resolved): multiple_rsync failure
- Related to #11781?
This is running with the recent change to source files from a dir in /tmp instead of directly f... - 05:50 AM Bug #11989: Cephfs Kernel Client data corruption
- Zheng Yan wrote:
> are there any suspected message when this happens?
dmesg is silent, ceph logs in /var/log/ceph... - 01:57 AM Bug #11989: Cephfs Kernel Client data corruption
- are there any suspected message when this happens?
06/12/2015
- 07:31 PM Backport #11999 (Resolved): cephfs Dumper tries to load whole journal into memory at once
- https://github.com/ceph/ceph/pull/5120
- 07:06 PM Bug #11989: Cephfs Kernel Client data corruption
- I imagine this is a result of some kind of memory exhaustion, but I'm not sure how best to diagnose it or if there ar...
- 01:22 PM Bug #11989 (Resolved): Cephfs Kernel Client data corruption
- Hi. i get random data corruption with the cephfs kernel client. i do streaming from a non-ceph machine using "cat <fi...
- 06:19 PM Bug #11986: logs changing during tarball generation at end of job
- Actually this job doesn't have log rotation enabled at all.
- 12:34 PM Bug #11986 (Closed): logs changing during tarball generation at end of job
http://pulpito.ceph.com/teuthology-2015-06-08_23:04:01-fs-master---basic-multi/926320/...- 12:31 PM Bug #11985 (Resolved): MDS asserts in objecter when transitioning from replay to DNE
Seen once:
http://pulpito.ceph.com/teuthology-2015-06-08_23:04:01-fs-master---basic-multi/926330/...
06/11/2015
- 01:56 PM Feature #3826 (Resolved): uclient: Be more aggressive about checking for pools we can't write to
- Zheng did this a few months ago; we now write to a test object in every newly-seen pool.
- 01:55 PM Feature #4885 (Resolved): dumper: do an incremental log dump
- https://github.com/ceph/ceph/pull/4835
- 01:52 PM Feature #11588 (Resolved): teuthology: set up log rotate for MDS logs
- https://github.com/ceph/ceph-qa-suite/pull/452
- 01:28 PM Bug #11959 (Resolved): qa-suite: /usr copy needs more perms
- 10:13 AM Bug #11959 (Fix Under Review): qa-suite: /usr copy needs more perms
- That's kind of yucky of virtual box to put those in /usr/lib but whatever!
https://github.com/ceph/ceph/pull/4930 - 05:23 AM Bug #11959 (Resolved): qa-suite: /usr copy needs more perms
- /a/ubuntu-2015-06-10_10:18:19-fs-greg-fs-testing---basic-multi/928341/teuthology.log...
- 01:15 PM Bug #11913 (Fix Under Review): Failure in TestClusterFull.test_barrier
- https://github.com/ceph/ceph-qa-suite/pull/457
06/10/2015
- 08:34 PM Support #11923: MDS init script starts multiple instances when MDS is referenced in ceph.conf
- Can you attach the ceph.conf and an ls -al of the instance directory? This will happen if the section is in ceph.co...
- 02:19 PM Bug #10950 (Resolved): Unable to remove MDS host: error handling
- Merged to master in commit:5441f89c022aa1f4df084a4280e45c5c5b278f00
- 02:18 PM Bug #11746 (Pending Backport): cephfs Dumper tries to load whole journal into memory at once
- Merged to master in commit:04a11f0f2f6d46091d6868ba1cc2fec7a4e7a99c
- 01:19 PM Feature #11950 (Resolved): Strays enqueued for purge cause MDCache to exceed size limit
If your purge operations are going slowly (either because of throttle or because of slow data pool), and you do lot...
06/08/2015
- 09:55 PM Support #11923 (Resolved): MDS init script starts multiple instances when MDS is referenced in ce...
- MDS component init script does not seem to be able to properly differentiate between auto-detected instances and inst...
- 08:35 PM Bug #6613: samba is crashing in teuthology
- Seeing in release firefly v0.80.9 validation
Run: http://pulpito.ceph.com/ubuntu-2015-06-08_11:05:13-samba-firefly... - 01:53 PM Bug #11913: Failure in TestClusterFull.test_barrier
- (Referring to ceph-qa-suite/tasks/cephfs/test_full.py::test_barrier().)
So mount.a is doing open_no_data("alpha") ... - 11:21 AM Bug #11913 (Resolved): Failure in TestClusterFull.test_barrier
http://pulpito.ceph.com/teuthology-2015-06-05_23:04:02-fs-master---basic-multi/922578/
mount_a is getting the ne...
06/07/2015
- 07:45 AM Backport #11737 (In Progress): MDS is crashed (mds/CDir.cc: 1391: FAILED assert(!is_complete()))
06/06/2015
- 09:52 AM Feature #11903 (Resolved): libcephfs: implement read/write interface support "struct iovec"
- Now if qemu virtfs directly call libcephfs api, it need to retranslate iov structure to continuous buffer. We should ...
06/05/2015
- 06:23 PM Bug #11807 (Resolved): workunits: stop rsyncing with /usr as the source material
- 3f4dfb80e32fba86e9076a2915197487bf899906
We can reopen if we decide this needs a backport. - 06:14 PM Bug #11504 (Resolved): CephFS restriction on removing cache tiers is overly strict
- commit:5a3da7b891bf18e0f164352dd0010c7969233d72
06/03/2015
- 10:51 AM Feature #11859 (Resolved): MDS "damage table" for recording scrub/fetch errors
Creating a ticket for this ongoing work in the interests of clarity.
The idea is that when we get an error in fe...- 09:42 AM Bug #10950 (Fix Under Review): Unable to remove MDS host: error handling
- https://github.com/ceph/ceph/pull/4838
- 09:19 AM Bug #11835 (Fix Under Review): FuseMount.umount_wait can hang
- https://github.com/ceph/ceph-qa-suite/pull/453
- 09:07 AM Bug #11746 (Fix Under Review): cephfs Dumper tries to load whole journal into memory at once
- https://github.com/ceph/ceph/pull/4835
06/02/2015
- 09:08 PM Bug #11758: kernel_untar_build fails on EL7
- let's update the tarball to use a newer kernel.
- 05:55 PM Bug #11807 (Fix Under Review): workunits: stop rsyncing with /usr as the source material
- https://github.com/ceph/ceph/pull/4831
- 01:49 PM Bug #11807: workunits: stop rsyncing with /usr as the source material
- John suggests we can grab /usr/doc instead, maybe.
- 01:54 PM Bug #11300 (Resolved): client-limits and mds-full failures on clog warning/errors
- We think we already did that...sorry, no commit refs handy.
- 01:44 PM Bug #10950: Unable to remove MDS host: error handling
- This is just unfriendly error handling. John says one of the other commands has a spurious argument that needs to get...
- 01:42 PM Bug #11784: ceph-fuse hang on unmount (stuck dentry refs)
- I copied the ceph-client log into that folder, although we're missing the server logs.
- 01:36 PM Bug #11294: samba: DISCONNECTED inode warning
- seems like dup of #11835
06/01/2015
- 09:55 PM Bug #11835 (Resolved): FuseMount.umount_wait can hang
Currently code in FuseMount.umount assumes that the write to /sys/fuse/connections/X/abort causes the process to te...- 09:49 PM Bug #11784: ceph-fuse hang on unmount (stuck dentry refs)
- We saw this again today, and it's definitely inode refs this time:
http://pulpito-rdu.front.sepia.ceph.com/gregf-201...
05/31/2015
- 10:14 PM Bug #10248: messenger: failed Pipe;:connect::assert(m) in Hadoop client
- Yes — the related issue chain has been seen a few times, more recently.
- 06:05 PM Bug #10248: messenger: failed Pipe;:connect::assert(m) in Hadoop client
- is it still valid ?
05/29/2015
- 06:15 PM Bug #11792: mds: recursive statistics are either inaccurate or too "chunky"
- Well, they're delayed, yes. But the few times I've bothered to measure the delay is very short (a few seconds at most...
- 07:13 AM Bug #11792: mds: recursive statistics are either inaccurate or too "chunky"
- thanks to delay propagation, recursive statistics are never accurate
05/28/2015
- 06:16 PM Bug #11781: multiple_rsync failure with fuse client
- #11807. I'm not actually sure what good options there are for replacement, but it won't get lost.
- 12:24 PM Bug #11781: multiple_rsync failure with fuse client
- Hmm, maybe using /usr/ as a source of workload files is something we just need to stop doing, there's no particular r...
- 12:12 PM Bug #11781 (Rejected): multiple_rsync failure with fuse client
- ...
- 06:15 PM Bug #11807 (Resolved): workunits: stop rsyncing with /usr as the source material
- We've had various issues with /usr. The permissions on it are fiddly, its size can vary, and apparently some times th...
- 11:47 AM Bug #11789: knfs mount fails with "getfh failed: Function not implemented"
- I don't think this is likely to be a regression from previous hammer release, so not a blocker imho
- 06:17 AM Bug #11783 (In Progress): protocol: flushing caps on MDS restart can go bad
- this is a message ordering issue when MDS failover.
chown marks Ax dirty
client flushes and releases Ax cap
cho... - 01:02 AM Bug #11790 (Resolved): "cannot utime" errors in from tar in knfs workloads
- it's nfs bug in 4.1-rc2 kernel. now the testing branch has been rebased to 4.1-rc5, this issue should disappear
Also available in: Atom