Project

General

Profile

Actions

Bug #64502

open

pacific/quincy/v18.2.0: client: ceph-fuse fails to unmount after upgrade to main

Added by Patrick Donnelly 3 months ago. Updated 8 days ago.

Status:
New
Priority:
Urgent
Assignee:
Category:
Correctness/Safety
Target version:
% Done:

0%

Source:
Development
Tags:
Backport:
squid,reef,quincy
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(FS):
Client, ceph-fuse
Labels (FS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

Every ceph-fuse mount for quincy fails to unmount for reef->main:

https://pulpito.ceph.com/pdonnell-2024-02-19_18:28:45-fs:upgrade:mds_upgrade_sequence-wip-batrick-testing-20240215.160715-distro-default-smithi/

2024-02-19T19:17:36.535 INFO:tasks.cephfs.fuse_mount:Running fusermount -u on ubuntu@smithi060.front.sepia.ceph.com...
2024-02-19T19:17:36.535 INFO:teuthology.orchestra.run:Running command with timeout 300
2024-02-19T19:17:36.535 DEBUG:teuthology.orchestra.run.smithi060:> sudo fusermount -u /home/ubuntu/cephtest/mnt.0
2024-02-19T19:17:36.562 INFO:teuthology.orchestra.run:waiting for 300

From: /teuthology/pdonnell-2024-02-19_18:28:45-fs:upgrade:mds_upgrade_sequence-wip-batrick-testing-20240215.160715-distro-default-smithi/7566635/teuthology.log

2024-02-19T19:17:36.799+0000 7f9fa7fff640 20 client.14548 tick
2024-02-19T19:17:36.799+0000 7f9fa7fff640 20 client.14548 collect_and_send_metrics
2024-02-19T19:17:36.799+0000 7f9fa7fff640 20 client.14548 collect_and_send_global_metrics
2024-02-19T19:17:36.799+0000 7f9fa7fff640  1 -- 192.168.0.1:0/854663557 --> [v2:172.21.15.60:6826/3594652577,v1:172.21.15.60:6827/3594652577] -- client_metrics [client_metric_type: READ_LATENCY latency: 5.996942, avg_latency: 0.000330, sq_sum: 86627012816408144, count=17901][client_metric_type: WRITE_LATENCY latency: 23.710221, avg_latency: 0.000407, sq_sum: 1890169673992666112, count=56281][client_metric_type: METADATA_LATENCY latency: 238.430933, avg_latency: 0.005247, sq_sum: 13600282437617256448, count=45341][client_metric_type: CAP_INFO cap_hits: 831286 cap_misses: 14792 num_caps: 0][client_metric_type: DENTRY_LEASE dlease_hits: 67 dlease_misses: 154700 num_dentries: 0][client_metric_type: OPENED_FILES opened_files: 0 total_inodes: 1][client_metric_type: PINNED_ICAPS pinned_icaps: 1 total_inodes: 1][client_metric_type: OPENED_INODES opened_inodes: 0 total_inodes: 1][client_metric_type: READ_IO_SIZES total_ops: 22272 total_size: 3731108728][client_metric_type: WRITE_IO_SIZES total_ops: 56281 total_size: 4270138133] v1 -- 0x7f9fa000b9e0 con 0x5637e76f0e80
2024-02-19T19:17:36.799+0000 7f9fa7fff640 20 client.14548 trim_cache size 1 max 16384
2024-02-19T19:17:36.799+0000 7f9fa7fff640 20 client.14548 upkeep thread waiting interval 1.000000000s

...

2024-02-19T20:23:30.865+0000 7f9fc8e36480  2 client.14548 unmounting

From: /teuthology/pdonnell-2024-02-19_18:28:45-fs:upgrade:mds_upgrade_sequence-wip-batrick-testing-20240215.160715-distro-default-smithi/7566635/remote/smithi060/log/ceph-client.0.log.gz

During teardown of the cluster the unmount eventual proceeds but it's not clear what was blocking it. I think something was holding the RWRef preventing unmount from proceeding.


Related issues 4 (3 open1 closed)

Related to CephFS - Bug #64503: client: log message when unmount call is receivedPending BackportPatrick Donnelly

Actions
Related to CephFS - Bug #64440: mds: reversed encoding of MDSMap max_xattr_size/bal_rank_mask v18.2.1 <-> mainPending BackportPatrick Donnelly

Actions
Related to CephFS - Bug #61265: qa: tasks.cephfs.fuse_mount:process failed to terminate after unmountDuplicateMilind Changire

Actions
Related to CephFS - Bug #50260: pacific: qa: "rmdir: failed to remove '/home/ubuntu/cephtest': Directory not empty"New

Actions
Actions

Also available in: Atom PDF