Project

General

Profile

Actions

Bug #44100

closed

cephfs rsync kworker high load.

Added by none none about 4 years ago. Updated about 1 year ago.

Status:
Resolved
Priority:
Normal
Assignee:
Category:
-
Target version:
-
% Done:

0%

Source:
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(FS):
Labels (FS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

I have an rsync backup running which became 10 hours. When I test with rsync one instance, it looks like it processes quite few files fast, then 20 minutes nothing happens and then processes 4 files and finishes correctly (without rsync errors)

I already posted some related questions on the mailing list
https://www.mail-archive.com/ceph-users@ceph.io/msg01718.html

echo l > /proc/sysrq-trigger

[6310351.557432] CPU: 5 PID: 3453020 Comm: kworker/5:1 Not tainted 3.10.0-1062.4.3.el7.x86_64 #1
[6310351.557433] Hardware name: Supermicro X9DRi-LN4+/X9DR3-LN4+/X9DRi-LN4+/X9DR3-LN4+, BIOS 3.0b 05/27/2014
[6310351.557435] Workqueue: ceph-msgr ceph_con_workfn [libceph]
[6310351.557436] task: ffffa0d29fbe20e0 ti: ffffa0c697b48000 task.ti: ffffa0c697b48000
[6310351.557437] RIP: 0010:[<ffffffffc08d7f70>] [<ffffffffc08d7f70>] build_snap_context+0x20/0x290 [ceph]
[6310351.557438] RSP: 0018:ffffa0c697b4baa8 EFLAGS: 00000292
[6310351.557439] RAX: ffffa0c699949188 RBX: ffffa0c699949100 RCX: 00000000fffffff8
[6310351.557440] RDX: ffffa0c697b4bb68 RSI: ffffa0c697b4bb68 RDI: ffffa0c699949100
[6310351.557441] RBP: ffffa0c697b4bad8 R08: ffffa0c699948e00 R09: 0000000000027247
[6310351.557442] R10: 0000000000000036 R11: 0000000000000000 R12: ffffa0c697b4bb68
[6310351.557443] R13: ffffa0c697b4bb68 R14: ffffa0cda3a37b00 R15: ffffa0cda3a37b00
[6310351.557444] FS: 0000000000000000(0000) GS:ffffa0d2cfb40000(0000) knlGS:0000000000000000
[6310351.557445] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[6310351.557446] CR2: 00007fae085ac000 CR3: 0000001813010000 CR4: 00000000000627e0
[6310351.557447] Call Trace:
[6310351.557448] [<ffffffffc08d820c>] rebuild_snap_realms+0x2c/0x90 [ceph]
[6310351.557449] [<ffffffffc08d822b>] rebuild_snap_realms+0x4b/0x90 [ceph]
[6310351.557450] [<ffffffffc08d91fc>] ceph_update_snap_trace+0x3ec/0x530 [ceph]
[6310351.557452] [<ffffffffc08e2239>] handle_reply+0x359/0xc60 [ceph]
[6310351.557453] [<ffffffffc08e48ba>] dispatch+0x11a/0xb00 [ceph]
[6310351.557454] [<ffffffffb042e56a>] ? kernel_recvmsg+0x3a/0x50
[6310351.557455] [<ffffffffc05fcff4>] try_read+0x544/0x1300 [libceph]
[6310351.557456] [<ffffffffafee13ce>] ? account_entity_dequeue+0xae/0xd0
[6310351.557457] [<ffffffffafee4d5c>] ? dequeue_entity+0x11c/0x5e0
[6310351.557458] [<ffffffffb042e417>] ? kernel_sendmsg+0x37/0x50
[6310351.557459] [<ffffffffc05fdfb4>] ceph_con_workfn+0xe4/0x1530 [libceph]
[6310351.557460] [<ffffffffb057f568>] ? __schedule+0x448/0x9c0
[6310351.557461] [<ffffffffafebe21f>] process_one_work+0x17f/0x440
[6310351.557462] [<ffffffffafebf336>] worker_thread+0x126/0x3c0
[6310351.557463] [<ffffffffafebf210>] ? manage_workers.isra.26+0x2a0/0x2a0
[6310351.557464] [<ffffffffafec61f1>] kthread+0xd1/0xe0
[6310351.557465] [<ffffffffafec6120>] ? insert_kthread_work+0x40/0x40
[6310351.557466] [<ffffffffb058cd37>] ret_from_fork_nospec_begin+0x21/0x21
[6310351.557467] [<ffffffffafec6120>] ? insert_kthread_work+0x40/0x40
[6310351.557469] Code: 41 5e 5d c3 0f 1f 80 00 00 00 00 66 66 66 66 90 55 48 89 e5 41 57 41 56 41 55 41 54 49 89 f4 53 48 89 fb 48 83 ec 08 4c 8b 7f 70 <44> 8b 6f 68 44 03 6f 58 4d 85 ff 74 14 49 8b 87 b8 00 00 00 48


Related issues 1 (0 open1 closed)

Related to CephFS - Bug #54106: kclient: hang during workunit cleanupDuplicateXiubo Li

Actions
Actions

Also available in: Atom PDF