Project

General

Profile

Actions

Bug #20731

closed

"[ERR] : Health check failed: 1 mds daemon down (MDS_FAILED)" in upgrade:jewel-x-luminous-distro-basic-smithi

Added by Yuri Weinstein almost 7 years ago. Updated over 6 years ago.

Status:
Resolved
Priority:
Normal
Category:
Testing
Target version:
-
% Done:

0%

Source:
Q/A
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
upgrade/jewel-x
Component(FS):
Labels (FS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

Run: http://pulpito.ceph.com/teuthology-2017-07-19_04:23:05-upgrade:jewel-x-luminous-distro-basic-smithi/
Jobs: 38
Logs: http://qa-proxy.ceph.com/teuthology/teuthology-2017-07-19_04:23:05-upgrade:jewel-x-luminous-distro-basic-smithi/1418248/teuthology.log

2017-07-20T10:21:45.074 INFO:teuthology.orchestra.run.smithi053:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 0 ceph --cluster ceph --admin-daemon /var/run/ceph/ceph-mds.a.asok status'
2017-07-20T10:21:45.153 INFO:teuthology.orchestra.run.smithi053.stdout:{
2017-07-20T10:21:45.154 INFO:teuthology.orchestra.run.smithi053.stdout:    "cluster_fsid": "da23cb82-ed76-4a15-bc25-62655a802f68",
2017-07-20T10:21:45.154 INFO:teuthology.orchestra.run.smithi053.stdout:    "whoami": -1,
2017-07-20T10:21:45.154 INFO:teuthology.orchestra.run.smithi053.stdout:    "id": 34214,
2017-07-20T10:21:45.154 INFO:teuthology.orchestra.run.smithi053.stdout:    "want_state": "up:boot",
2017-07-20T10:21:45.154 INFO:teuthology.orchestra.run.smithi053.stdout:    "state": "???",
2017-07-20T10:21:45.154 INFO:teuthology.orchestra.run.smithi053.stdout:    "mdsmap_epoch": 7,
2017-07-20T10:21:45.154 INFO:teuthology.orchestra.run.smithi053.stdout:    "osdmap_epoch": 0,
2017-07-20T10:21:45.154 INFO:teuthology.orchestra.run.smithi053.stdout:    "osdmap_epoch_barrier": 0
2017-07-20T10:21:45.154 INFO:teuthology.orchestra.run.smithi053.stdout:}
2017-07-20T10:21:45.155 INFO:tasks.cephfs.filesystem:_json_asok output: {
    "cluster_fsid": "da23cb82-ed76-4a15-bc25-62655a802f68",
    "whoami": -1,
    "id": 34214,
    "want_state": "up:boot",
    "state": "???",
    "mdsmap_epoch": 7,
    "osdmap_epoch": 0,
    "osdmap_epoch_barrier": 0
}

2017-07-20T10:21:45.628 INFO:tasks.workunit.client.0.smithi166.stdout:  462: throughput=4.99MB/sec pending data=0
2017-07-20T10:21:45.628 INFO:tasks.workunit.client.0.smithi166.stdout:WRITE : oid=obj-gwXbIyU8kOZleGD off=4003488 len=812660
2017-07-20T10:21:45.634 INFO:tasks.workunit.client.0.smithi166.stdout:op 2583 completed, throughput=4.99MB/sec
2017-07-20T10:21:45.634 INFO:tasks.workunit.client.0.smithi166.stdout:READ : oid=obj-HyoohrFQ3v7y4c1 off=2436879 len=2091857
2017-07-20T10:21:45.643 INFO:tasks.workunit.client.0.smithi166.stdout:op 2584 completed, throughput=5MB/sec
2017-07-20T10:21:45.643 INFO:tasks.workunit.client.0.smithi166.stdout:READ : oid=obj-QUMumdfpGt-Dz41 off=8482169 len=291285
2017-07-20T10:21:45.645 INFO:tasks.workunit.client.0.smithi166.stdout:op 2585 completed, throughput=5MB/sec
2017-07-20T10:21:45.645 INFO:tasks.workunit.client.0.smithi166.stdout:READ : oid=obj-XfuTvcwAWPXVwcQ off=2398965 len=474934
2017-07-20T10:21:46.156 INFO:teuthology.orchestra.run.smithi053:Running: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph fs dump --format=json'
2017-07-20T10:21:46.334 INFO:tasks.ceph.mon.a.smithi053.stderr:2017-07-20 10:21:46.337950 7f6b18faf700 -1 log_channel(cluster) log [ERR] : Health check failed: 1 mds daemon down (MDS_FAILED)
Actions #1

Updated by Patrick Donnelly over 6 years ago

  • Project changed from Ceph to CephFS
  • Category set to Testing
  • Status changed from New to In Progress
  • Assignee set to Patrick Donnelly

Obviously this error is expected when restarting the MDS. We should whitelist the warning.

Actions #3

Updated by Patrick Donnelly over 6 years ago

  • Status changed from In Progress to Resolved
Actions

Also available in: Atom PDF