Ceph : Issues
https://tracker.ceph.com/
https://tracker.ceph.com/favicon.ico
2020-04-29T08:39:38Z
Ceph
Redmine
CephFS - Bug #45320 (Fix Under Review): client: Other UID don't write permission when the file is...
https://tracker.ceph.com/issues/45320
2020-04-29T08:39:38Z
huanwen ren
<p>The problem appears in the ceph-fuse mount,there is no problem with the kernel mount</p>
<p>Example:<br /><pre>
[root@cephfs102 sky]# chmod 04777 star.txt
[root@cephfs102 sky]# ls -lh
total 512
-rwsrwxrwx 1 root root 4 Apr 29 17:51 star.txt
[root@cephfs102 sky]# su sky
[sky@cephfs102 sky]$ ls
star.txt
[sky@cephfs102 sky]$ echo "aaa">star.txt
bash: echo: write error: Operation not permitted
</pre></p>
<p>"04777" similar setuid</p>
CephFS - Bug #43410 (Won't Fix): mds:When the directory level is above 3000, the following assert...
https://tracker.ceph.com/issues/43410
2019-12-23T11:10:14Z
huanwen ren
<p>When I use the script to continuously create directories, <br />but the directory level is above 3000, the following assertions appear:</p>
<pre>
Program received signal SIGSEGV, Segmentation fault.
[Switching to Thread 0x7fffeb5a5700 (LWP 118867)]
0x0000555555a6ffb6 in CInode::adjust_nested_auth_pins (this=0x55555ef66300, a=a@entry=1, by=by@entry=0x55555fdc2300)
at /usr/src/debug/ceph-12.2.9-2-59-g1608bbd/src/mds/CInode.cc:2462
2462 dout(35) << "adjust_nested_auth_pins by " << by
</pre>
<pre>
Program received signal SIGSEGV, Segmentation fault.
[Switching to Thread 0x7fffeb139700 (LWP 1622410)]
ratelimit_check (s=0x5555563e9b50 <CInode::adjust_nested_auth_pins(int, void*)::__rs>) at /usr/src/debug/ceph-12.2.9-2-59-g1608bbd/src/common/ratelimit.h:50
50 auto now = std::chrono::steady_clock::now();
</pre>
CephFS - Bug #43409 (Closed): mds:When the directory level is above 3000, the following assertion...
https://tracker.ceph.com/issues/43409
2019-12-23T11:09:06Z
huanwen ren
<p>When I use the script to continuously create directories, <br />but the directory level is above 3000, the following assertions appear:</p>
<pre>
Program received signal SIGSEGV, Segmentation fault.
[Switching to Thread 0x7fffeb5a5700 (LWP 118867)]
0x0000555555a6ffb6 in CInode::adjust_nested_auth_pins (this=0x55555ef66300, a=a@entry=1, by=by@entry=0x55555fdc2300)
at /usr/src/debug/ceph-12.2.9-2-59-g1608bbd/src/mds/CInode.cc:2462
2462 dout(35) << "adjust_nested_auth_pins by " << by
</pre>
<pre>
ratelimit_check (s=0x5555563e9b50 <CInode::adjust_nested_auth_pins(int, void*)::__rs>) at /usr/src/debug/ceph-12.2.9-2-59-g1608bbd/src/common/ratelimit.h:50
50 auto now = std::chrono::steady_clock::now();
</pre>
CephFS - Bug #41868 (Resolved): mds: mds returns -5 error when the deleted file does not exist
https://tracker.ceph.com/issues/41868
2019-09-16T12:04:46Z
huanwen ren
<p>There are 2 nfs-ganehsa ends:<br />1.The A side uses readdir to get all the file information in a directory,<br /> and uses ceph_ll_lookup_inode to find the mds.</p>
<p>2.In addition, the B side just deletes the "aaa" file in this directory.</p>
<p>3.At this time, the A end will go to the MDS through ceph_ll_lookup_inode <br /> to find the ino of the "aaa" file that has been deleted, resulting in a bug of -5.</p>
<p>The call chain looks like this:</p>
<p>(1) NFS:<br /> ceph_ll_lookup_inode ---></p>
<p>(2) Client:<br /> Client::ll_lookup_inode ---> Client::_lookup_ino---><br /> message: CEPH_MDS_OP_LOOKUPINO</p>
<p>(3) MDS:<br /> Server::handle_client_lookup_ino ---><br /> MDCache::open_ino---><br /> MDCache::do_open_ino---><br /> C_IO_MDC_OpenInoBacktraceFetched---><br /> MDCache::_open_ino_backtrace_fetched---><br /> "if (backtrace.ancestors.empty()) return -EIO"</p>
<p>Fixup:</p>
<p>For files that do not exist, we should set it to -ENOENT instead of -EIO, because this error code will be returned to the NFS client and transparently passed to the application layer. The application layer may handle -EIO and -ENOENT differently. And lead to misjudgment</p>
CephFS - Bug #41148 (Resolved): client: _readdir_cache_cb() may use the readdir_cache already clear
https://tracker.ceph.com/issues/41148
2019-08-07T06:56:51Z
huanwen ren
<p>Calling function A means to get dir information from the cache, but in the while loop, <br />the contents of readdir_cache are not static.</p>
<p>For example, _getattr() will trigger insert_trace to update readdir_cache, such as: <br />1. insert_trace()->clear_dir_complete_and_ordered() or <br />2. insert_trace()->add_update_inode()->add_update_cap()-> check_cap_issue()->clear_dir_complete_and_ordered()</p>
<p>So, there is a possibility that dn is empty after the _getattr() call.</p>
CephFS - Bug #40615 (Resolved): ceph-fuse: mount does not support the fallocate()
https://tracker.ceph.com/issues/40615
2019-07-02T06:36:48Z
huanwen ren
<p>ceph version: 14.2.1<br />fuse version: 2.9.2-6</p>
<p>err info:<br /><pre>
tmpfs 6577748 0 6577748 0% /run/user/0
ceph-fuse 2781802496 20508672 2761293824 1% /mycephfs
[root@cephfs102 ~]# fallocate -l 10G /mycephfs/filedata
fallocate: /mycephfs/filedata: fallocate failed: Operation not supported
[root@cephfs102 ~]#
</pre></p>
CephFS - Bug #38822 (Resolved): mds: there is an assertion when calling Beacon::shutdown()
https://tracker.ceph.com/issues/38822
2019-03-20T06:01:50Z
huanwen ren
<p>The construction condition is: <br />when you use systemctl start a non-existing ceph-mds@xx object</p>
<p>Error info:<br /><pre>
2019-03-19 20:26:57.000329 7f5b6671b1c0 0 set uid:gid to 167:167 (ceph:ceph)
2019-03-19 20:26:57.000356 7f5b6671b1c0 0 ceph version 12.2.9(9e300932ef8a8916fb3fda78c58691a6ab0f4217) luminous (stable), process ceph-mds, pid 458996
2019-03-19 20:26:57.004965 7f5b6671b1c0 0 pidfile_write: ignore empty --pid-file
2019-03-19 20:26:57.052413 7f5b6671b1c0 -1 auth: unable to find a keyring on /var/lib/ceph/mds/ceph-star/keyring: (2) No such file or directory
2019-03-19 20:26:57.052438 7f5b6671b1c0 -1 monclient: ERROR: missing keyring, cannot use cephx for authentication
2019-03-19 20:26:57.052445 7f5b6671b1c0 -1 mds.star ERROR: failed to get monmap: (2) No such file or directory
2019-03-19 20:26:57.052453 7f5b6671b1c0 1 mds.star suicide! Wanted state up:boot
2019-03-19 20:26:57.055098 7f5b6671b1c0 -1 *** Caught signal (Aborted) **
in thread 7f5b6671b1c0 thread_name:ceph-mds
ceph version 12.2.9(9e300932ef8a8916fb3fda78c58691a6ab0f4217) luminous (stable)
1: (()+0x611581) [0x561a08abd581]
2: (()+0xf6d0) [0x7f5b648116d0]
3: (gsignal()+0x37) [0x7f5b63a36277]
4: (abort()+0x148) [0x7f5b63a37968]
5: (__gnu_cxx::__verbose_terminate_handler()+0x165) [0x7f5b643457d5]
6: (()+0x5e746) [0x7f5b64343746]
7: (()+0x5e773) [0x7f5b64343773]
8: (()+0x5e993) [0x7f5b64343993]
9: (std::__throw_system_error(int)+0x90) [0x7f5b64398c70]
10: (std::thread::join()+0x18) [0x7f5b64399e28]
11: (Beacon::shutdown()+0xa8) [0x561a087b1dd8]
12: (MDSDaemon::suicide()+0x241) [0x561a08769f61]
13: (MDSDaemon::init()+0x17be) [0x561a0876e86e]
14: (main()+0xb90) [0x561a087575f0]
15: (__libc_start_main()+0xf5) [0x7f5b63a22445]
16: (()+0x2b8163) [0x561a08764163]
NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.
</pre></p>
CephFS - Bug #38376 (Rejected): client:Repeated release of put_request() in make_request()
https://tracker.ceph.com/issues/38376
2019-02-19T06:28:08Z
huanwen ren
<p>code repetition:</p>
<p>Client::make_request()
{<br /> ...<br /> if (!request->reply) {<br /> ceph_assert(request->aborted());<br /> ceph_assert(!request->got_unsafe);<br /> r = request->get_abort_code();<br /> request->item.remove_myself();<br /> <strong>unregister_request(request);</strong><br /> <strong>put_request(request);</strong><br /> return r;<br /> }<br /> ...<br />}</p>
<p>void Client::unregister_request(MetaRequest *req)
{<br /> mds_requests.erase(req->tid);<br /> if (req->tid == oldest_tid) {<br /> map<ceph_tid_t, MetaRequest*>::iterator p = mds_requests.upper_bound(oldest_tid);<br /> while (true) {<br /> if (p == mds_requests.end()) {<br /> oldest_tid = 0;<br /> break;<br /> }<br /> if (p->second->get_op() != CEPH_MDS_OP_SETFILELOCK) {<br /> oldest_tid = p->first;<br /> break;<br /> }<br /> ++p;<br /> }<br /> }<br /> <strong>put_request(req);</strong><br />}</p>
CephFS - Bug #36669 (Rejected): client: displayed as the capacity of all OSDs when there are mult...
https://tracker.ceph.com/issues/36669
2018-11-01T07:25:36Z
huanwen ren
<p>When using ceph-fuse to mount the cephfs file directory, if there are multiple data pools built in FS, the capacity of all OSDs is displayed. If there is only one data pool in FS, it displays the correct capacity information.</p>
<p>Can be constructed as follows:<br /><pre>
[root@hw110 tecs]# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
9.00TiB 7.85TiB 1.15TiB 12.82
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
nfs_metadata 1 171MiB 0.02 1.03TiB 105
nfs_data 2 3.59GiB 0.23 1.55TiB 2341
pool1 9 779GiB 51.86 723GiB 202249
cinder 10 0B 0 723GiB 0
glance 11 0B 0 723GiB 0
[root@hw110 tecs]# ceph fs ls
name: ceph-fs, metadata pool: nfs_metadata, data pools: [nfs_data pool1 ]
[root@hw110 tecs]# df -kh
Filesystem Size Used Avail Use% Mounted on
/dev/mapper/vg_sys-lv_root 99G 8.9G 85G 10% /
devtmpfs 32G 0 32G 0% /dev
tmpfs 32G 57M 32G 1% /dev/shm
tmpfs 32G 215M 31G 1% /run
tmpfs 32G 0 32G 0% /sys/fs/cgroup
/dev/sda2 976M 190M 720M 21% /boot
/dev/xxx1p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-12
/dev/xxx12p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-9
/dev/xxx10p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-14
/dev/xxx14p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-11
/dev/xxx11p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-10
/dev/xxx13p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-13
/dev/xxx0p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-8
/dev/xxx15p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-15
tmpfs 6.3G 0 6.3G 0% /run/user/1001
ceph-fuse 9.1T 1.2T 7.9T 13% /share-fs/export
tmpfs 6.3G 0 6.3G 0% /run/user/0
[root@hw110 tecs]#
[root@hw110 tecs]#
[root@hw110 tecs]# ceph fs rm_data_pool ceph-fs pool1
removed data pool 9 from fsmap
[root@hw110 tecs]#
[root@hw110 tecs]#
[root@hw110 tecs]# df -k
Filesystem 1K-blocks Used Available Use% Mounted on
/dev/mapper/vg_sys-lv_root 103080888 9324112 88497512 10% /
devtmpfs 32702636 0 32702636 0% /dev
tmpfs 32714180 57432 32656748 1% /dev/shm
tmpfs 32714180 220012 32494168 1% /run
tmpfs 32714180 0 32714180 0% /sys/fs/cgroup
/dev/sda2 999320 193816 736692 21% /boot
/dev/xxx1p1 95968 5500 90468 6% /var/lib/ceph/osd/ceph-12
/dev/xxx12p1 95968 5500 90468 6% /var/lib/ceph/osd/ceph-9
/dev/xxx10p1 95968 5500 90468 6% /var/lib/ceph/osd/ceph-14
/dev/xxx14p1 95968 5500 90468 6% /var/lib/ceph/osd/ceph-11
/dev/xxx11p1 95968 5500 90468 6% /var/lib/ceph/osd/ceph-10
/dev/xxx13p1 95968 5500 90468 6% /var/lib/ceph/osd/ceph-13
/dev/xxx0p1 95968 5500 90468 6% /var/lib/ceph/osd/ceph-8
/dev/xxx15p1 95968 5500 90468 6% /var/lib/ceph/osd/ceph-15
tmpfs 6542836 0 6542836 0% /run/user/1001
ceph-fuse 1666256896 3764224 1662492672 1% /share-fs/export
tmpfs 6542836 0 6542836 0% /run/user/0
[root@hw110 tecs]# df -kh
Filesystem Size Used Avail Use% Mounted on
/dev/mapper/vg_sys-lv_root 99G 8.9G 85G 10% /
devtmpfs 32G 0 32G 0% /dev
tmpfs 32G 57M 32G 1% /dev/shm
tmpfs 32G 215M 31G 1% /run
tmpfs 32G 0 32G 0% /sys/fs/cgroup
/dev/sda2 976M 190M 720M 21% /boot
/dev/xxx1p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-12
/dev/xxx12p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-9
/dev/xxx10p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-14
/dev/xxx14p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-11
/dev/xxx11p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-10
/dev/xxx13p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-13
/dev/xxx0p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-8
/dev/xxx15p1 94M 5.4M 89M 6% /var/lib/ceph/osd/ceph-15
tmpfs 6.3G 0 6.3G 0% /run/user/1001
ceph-fuse 1.6T 3.6G 1.6T 1% /share-fs/export
tmpfs 6.3G 0 6.3G 0% /run/user/0
[root@hw110 tecs]# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
9.00TiB 7.85TiB 1.15TiB 12.82
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
nfs_metadata 1 171MiB 0.02 1.03TiB 105
nfs_data 2 3.59GiB 0.23 1.55TiB 2341
pool1 9 779GiB 51.86 723GiB 202249
cinder 10 0B 0 723GiB 0
glance 11 0B 0 723GiB 0
</pre></p>
CephFS - Bug #36368 (Resolved): cephfs/tool: cephfs-shell have "no attribute 'decode'" err
https://tracker.ceph.com/issues/36368
2018-10-10T01:37:41Z
huanwen ren
<p>cephfs-shell have "no attribute 'decode'" err, as follow:</p>
<pre>
[root@star /]# python3 cephfs-shell
CephFS:~/>>> ls
'str' object has no attribute 'decode'
Traceback (most recent call last):
File "cephfs-shell", line 367, in onecmd
res = Cmd.onecmd(self, line)
File "/usr/lib/python3.4/site-packages/cmd2/cmd2.py", line 2050, in onecmd
stop = func(statement)
File "/usr/lib/python3.4/site-packages/cmd2/cmd2.py", line 241, in cmd_wrapper
return func(instance, args)
File "cephfs-shell", line 571, in do_ls
items = [i for i in items if not i.d_name.decode(
File "cephfs-shell", line 571, in <listcomp>
items = [i for i in items if not i.d_name.decode(
AttributeError: 'str' object has no attribute 'decode'
</pre>
RADOS - Support #36351 (New): mon: OSDMonitor.cc: 380: FAILED assert(err == 0)12.2.2
https://tracker.ceph.com/issues/36351
2018-10-09T02:48:30Z
huanwen ren
<p>I have a CEPH cluster which contains 3 mons, due to abnormal power failure, one mon service starts abnormally. The exception information is as follows:</p>
<pre>
2018-10-09 18:01:23.124782 7fc1ab4cfe40 1 mon.Ceph03@-1(probing) e1 preinit fsid ad403a3c-78e3-11e8-982b-52540056dc48
2018-10-09 18:01:23.124985 7fc1ab4cfe40 10 mon.Ceph03@-1(probing) e1 check_fsid cluster_uuid contains 'ad403a3c-78e3-11e8-982b-52540056dc48'
2018-10-09 18:01:23.125013 7fc1ab4cfe40 10 mon.Ceph03@-1(probing) e1 features compat={},rocompat={},incompat={1=initial feature set (~v.18),3=single paxos with k/v store (v0.?),4=support erasure code pools,5=new-style osdmap encoding,6=support isa/lrc erasure code,7=support shec erasure code,8=support monmap features,9=luminous ondisk layout}
2018-10-09 18:01:23.125029 7fc1ab4cfe40 10 mon.Ceph03@-1(probing) e1 calc_quorum_requirements required_features 153140804152475648
2018-10-09 18:01:23.125035 7fc1ab4cfe40 10 mon.Ceph03@-1(probing) e1 required_features 153140804152475648
2018-10-09 18:01:23.125049 7fc1ab4cfe40 10 mon.Ceph03@-1(probing) e1 has_ever_joined = 1
2018-10-09 18:01:23.125090 7fc1ab4cfe40 10 mon.Ceph03@-1(probing) e1 sync_last_committed_floor 0
2018-10-09 18:01:23.125096 7fc1ab4cfe40 10 mon.Ceph03@-1(probing) e1 init_paxos
2018-10-09 18:01:23.125152 7fc1ab4cfe40 1 mon.Ceph03@-1(probing).mds e0 Unable to load 'last_metadata'
2018-10-09 18:01:23.125210 7fc1ab4cfe40 10 mon.Ceph03@-1(probing).health init
2018-10-09 18:01:23.125218 7fc1ab4cfe40 10 mon.Ceph03@-1(probing) e1 refresh_from_paxos
2018-10-09 18:01:23.125252 7fc1ab4cfe40 1 mon.Ceph03@-1(probing).pg v0 on_upgrade discarding in-core PGMap
2018-10-09 18:01:23.125289 7fc1ab4cfe40 10 mon.Ceph03@-1(probing).pg v0 update_from_paxos deleted, clearing in-memory PGMap
2018-10-09 18:01:23.125336 7fc1ab4cfe40 10 mon.Ceph03@-1(probing).mds e0 update_from_paxos version 1, my e 0
2018-10-09 18:01:23.125379 7fc1ab4cfe40 10 mon.Ceph03@-1(probing).mds e0 update_from_paxos got 1
2018-10-09 18:01:23.125405 7fc1ab4cfe40 4 mon.Ceph03@-1(probing).mds e1 new map
2018-10-09 18:01:23.125412 7fc1ab4cfe40 0 mon.Ceph03@-1(probing).mds e1 print_map
e1
enable_multiple, ever_enabled_multiple: 0,0
compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=file layout v2}
legacy client fscid: -1
No filesystems configured
2018-10-09 18:01:23.125438 7fc1ab4cfe40 10 mon.Ceph03@-1(probing).mds e1 update_logger
2018-10-09 18:01:23.125565 7fc1ab4cfe40 15 mon.Ceph03@-1(probing).osd e0 update_from_paxos paxos e 1794, my e 0
2018-10-09 18:01:23.125792 7fc1ab4cfe40 7 mon.Ceph03@-1(probing).osd e0 update_from_paxos loading latest full map e1767
2018-10-09 18:01:23.126303 7fc1ab4cfe40 7 mon.Ceph03@-1(probing).osd e1767 update_from_paxos loading creating_pgs last_scan_epoch 1793 with 0 pgs
2018-10-09 18:01:23.126312 7fc1ab4cfe40 10 mon.Ceph03@-1(probing).osd e1767 update_from_paxos pgservice is mgrstat
2018-10-09 18:01:23.129140 7fc1ab4cfe40 -1 /clove/vm/zstor/ceph/rpmbuild/BUILD/ceph-12.2.2/src/mon/OSDMonitor.cc: In function 'virtual void OSDMonitor::update_from_paxos(bool*)' thread 7fc1ab4cfe40 time 2018-10-09 18:01:23.126328
/clove/vm/zstor/ceph/rpmbuild/BUILD/ceph-12.2.2/src/mon/OSDMonitor.cc: 380: FAILED assert(err == 0)
ceph version 12.2.2-16-6-g9e6bce0 (9e6bce0774b1d5d61c9327cc7c032b9cfea145bc) luminous (stable)
1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x110) [0x55aa5ecc2690]
2: (OSDMonitor::update_from_paxos(bool*)+0x2e5f) [0x55aa5ebdf05f]
3: (PaxosService::refresh(bool*)+0x1ae) [0x55aa5ebacc9e]
4: (Monitor::refresh_from_paxos(bool*)+0x193) [0x55aa5ea7e263]
5: (Monitor::init_paxos()+0x115) [0x55aa5ea7e695]
6: (Monitor::preinit()+0x9c6) [0x55aa5ea7f0b6]
7: (main()+0x4738) [0x55aa5e9ae668]
8: (__libc_start_main()+0xf5) [0x7fc1a81a4c05]
9: (()+0x37292e) [0x55aa5ea5392e]
NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.
</pre>