Project

General

Profile

Actions

Bug #45900

open

"ERROR: (22) Invalid argument" in powercycle

Added by Yuri Weinstein almost 4 years ago. Updated over 3 years ago.

Status:
New
Priority:
Normal
Assignee:
Category:
-
Target version:
-
% Done:

0%

Source:
Q/A
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

Run: http://pulpito.ceph.com/teuthology-2020-06-02_11:15:03-powercycle-master-testing-basic-smithi/
Job: 511856
Logs: /a/teuthology-2020-06-02_11:15:03-powercycle-master-testing-basic-smithi/5111856/teuthology.log

2020-06-04T02:30:30.240 INFO:teuthology.orchestra.run.smithi116.stdout:osd.2: {}
2020-06-04T02:30:30.240 INFO:teuthology.orchestra.run.smithi116.stderr:osd.0: osd_enable_op_tracker = 'false'
2020-06-04T02:30:30.241 INFO:teuthology.orchestra.run.smithi116.stderr:Error ENXIO: problem getting command descriptions from osd.1
2020-06-04T02:30:30.241 INFO:teuthology.orchestra.run.smithi116.stderr:osd.1: problem getting command descriptions from osd.1
2020-06-04T02:30:30.241 INFO:teuthology.orchestra.run.smithi116.stderr:osd.2: osd_enable_op_tracker = 'false'
2020-06-04T02:30:30.243 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:30.241+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:30.251 DEBUG:teuthology.orchestra.run:got remote process result: 6
2020-06-04T02:30:30.251 INFO:tasks.thrashosds.thrasher:Failed to tell all osds, ignoring
2020-06-04T02:30:30.344 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:30.341+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:30.445 INFO:tasks.ceph.osd.0.smithi005.stderr:2020-06-04T02:30:30.443+0000 7f5832534700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 0  (PID: 21975) UID: 0
2020-06-04T02:30:30.471 INFO:teuthology.orchestra.run.smithi005.stdout:ERROR: (22) Invalid argument
2020-06-04T02:30:30.471 INFO:teuthology.orchestra.run.smithi005.stdout:op_tracker tracking is not enabled now, so no ops are tracked currently, even those get stuck. Please enable "osd_enable_op_tracker", and the tracker will start to track new ops received afterwards.
2020-06-04T02:30:30.485 INFO:teuthology.orchestra.run.smithi003:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 30 ceph --cluster ceph --admin-daemon /var/run/ceph/ceph-osd.2.asok dump_ops_in_flight
2020-06-04T02:30:30.547 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:30.544+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:30.648 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:30.645+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:30.708 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.osd.1 is failed for ~303s
2020-06-04T02:30:30.709 INFO:tasks.daemonwatchdog.daemon_watchdog:BARK! unmounting mounts and killing all daemons
2020-06-04T02:30:30.709 INFO:tasks.cephfs.fuse_mount:Running fusermount -u on ubuntu@smithi116.front.sepia.ceph.com...
2020-06-04T02:30:30.710 INFO:teuthology.orchestra.run:Running command with timeout 1800
2020-06-04T02:30:30.710 INFO:teuthology.orchestra.run.smithi116:> (cd /home/ubuntu/cephtest && exec sudo fusermount -u /home/ubuntu/cephtest/mnt.0)
2020-06-04T02:30:30.735 INFO:teuthology.orchestra.run.smithi003.stdout:ERROR: (22) Invalid argument
2020-06-04T02:30:30.736 INFO:teuthology.orchestra.run.smithi003.stdout:op_tracker tracking is not enabled now, so no ops are tracked currently, even those get stuck. Please enable "osd_enable_op_tracker", and the tracker will start to track new ops received afterwards.
2020-06-04T02:30:30.746 INFO:tasks.workunit.client.0.smithi116.stdout:Wrote -1 instead of 4096 bytes.
2020-06-04T02:30:30.747 INFO:tasks.workunit.client.0.smithi116.stdout:Probably out of disk space
2020-06-04T02:30:30.747 INFO:tasks.workunit.client.0.smithi116.stderr:write: Transport endpoint is not connected
2020-06-04T02:30:30.748 DEBUG:teuthology.orchestra.run:got remote process result: 1
2020-06-04T02:30:30.749 INFO:tasks.workunit:Stopping ['suites/ffsb.sh'] on client.0...
2020-06-04T02:30:30.749 INFO:teuthology.orchestra.run.smithi116:> sudo rm -rf -- /home/ubuntu/cephtest/workunits.list.client.0 /home/ubuntu/cephtest/clone.client.0
2020-06-04T02:30:30.750 INFO:tasks.ceph.osd.0.smithi005.stderr:2020-06-04T02:30:30.746+0000 7f5832534700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 0  (PID: 21975) UID: 0
2020-06-04T02:30:30.751 INFO:teuthology.orchestra.run.smithi003:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 30 ceph --cluster ceph --admin-daemon /var/run/ceph/ceph-osd.2.asok dump_blocked_ops
2020-06-04T02:30:30.850 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:30.847+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:30.950 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:30.948+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:30.988 INFO:teuthology.orchestra.run.smithi003.stdout:ERROR: (22) Invalid argument
2020-06-04T02:30:30.989 INFO:teuthology.orchestra.run.smithi003.stdout:op_tracker tracking is not enabled now, so no ops are tracked currently, even those get stuck. Please enable "osd_enable_op_tracker", and the tracker will start to track new ops received afterwards.
2020-06-04T02:30:31.003 INFO:teuthology.orchestra.run.smithi003:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 30 ceph --cluster ceph --admin-daemon /var/run/ceph/ceph-osd.2.asok dump_historic_ops
2020-06-04T02:30:31.045 INFO:teuthology.orchestra.run.smithi116:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph --log-early osd unset nodeep-scrub
2020-06-04T02:30:31.051 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:31.048+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:31.086 INFO:teuthology.orchestra.run:waiting for 900
2020-06-04T02:30:31.087 INFO:teuthology.orchestra.run:Running command with timeout 300
2020-06-04T02:30:31.087 INFO:teuthology.orchestra.run.smithi116:> (cd /home/ubuntu/cephtest && exec rmdir -- /home/ubuntu/cephtest/mnt.0)
2020-06-04T02:30:31.156 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:31.151+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:31.227 INFO:tasks.cephfs.mount:Cleaning the 'ceph-ns--home-ubuntu-cephtest-mnt.0' netns for '/home/ubuntu/cephtest/mnt.0'
2020-06-04T02:30:31.228 INFO:tasks.cephfs.mount:Removing the netns 'ceph-ns--home-ubuntu-cephtest-mnt.0'
2020-06-04T02:30:31.228 INFO:teuthology.orchestra.run:Running command with timeout 300
2020-06-04T02:30:31.228 INFO:teuthology.orchestra.run.smithi116:> sudo bash -c 'ip link set brx.0 down'
2020-06-04T02:30:31.237 INFO:teuthology.orchestra.run.smithi003.stdout:ERROR: (22) Invalid argument
2020-06-04T02:30:31.238 INFO:teuthology.orchestra.run.smithi003.stdout:op_tracker tracking is not enabled now, so no ops are tracked currently, even those get stuck. Please enable "osd_enable_op_tracker", and the tracker will start to track new ops received afterwards.
2020-06-04T02:30:31.253 INFO:teuthology.orchestra.run.smithi005:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 30 ceph --cluster ceph --admin-daemon /var/run/ceph/ceph-osd.0.asok dump_ops_in_flight
2020-06-04T02:30:31.256 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:31.253+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:31.339 INFO:teuthology.orchestra.run:Running command with timeout 300
2020-06-04T02:30:31.340 INFO:teuthology.orchestra.run.smithi116:> sudo bash -c 'ip link delete brx.0'
2020-06-04T02:30:31.355 INFO:tasks.ceph.osd.2.smithi003.stderr:2020-06-04T02:30:31.353+0000 7fa068b4a700 -1 received  signal: Hangup from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2  (PID: 22070) UID: 0
2020-06-04T02:30:31.428 ERROR:teuthology.run_tasks:Saw exception from tasks.
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/run_tasks.py", line 91, in run_tasks
    manager = run_one_task(taskname, ctx=ctx, config=config)
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/run_tasks.py", line 70, in run_one_task
    return task(**kwargs)
  File "/home/teuthworker/src/git.ceph.com_ceph_master/qa/tasks/workunit.py", line 140, in task
    cleanup=cleanup)
  File "/home/teuthworker/src/git.ceph.com_ceph_master/qa/tasks/workunit.py", line 290, in _spawn_on_all_clients
    timeout=timeout)
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/parallel.py", line 87, in __exit__
    for result in self:
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/parallel.py", line 101, in __next__
    resurrect_traceback(result)
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/parallel.py", line 37, in resurrect_traceback
    reraise(*exc_info)
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/six.py", line 703, in reraise
    raise value
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/parallel.py", line 24, in capture_traceback
    return func(*args, **kwargs)
  File "/home/teuthworker/src/git.ceph.com_ceph_master/qa/tasks/workunit.py", line 415, in _run_tests
    label="workunit test {workunit}".format(workunit=workunit)
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/orchestra/remote.py", line 206, in run
    r = self._runner(client=self.ssh, name=self.shortname, **kwargs)
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/orchestra/run.py", line 475, in run
    r.wait()
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/orchestra/run.py", line 162, in wait
    self._raise_for_status()
  File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/teuthology/orchestra/run.py", line 184, in _raise_for_status
    node=self.hostname, label=self.label
teuthology.exceptions.CommandFailedError: Command failed (workunit test suites/ffsb.sh) on smithi116 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=cca6533da2dbb756769bf3640b19705a1d0ea1fa TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/ffsb.sh'
2020-06-04T02:30:31.490 ERROR:teuthology.run_tasks: Sentry event: http://sentry.ceph.com/sepia/teuthology/?q=4fdfbaaaade6448d8cf112d6741b265f


Related issues 1 (1 open0 closed)

Related to Ceph - Bug #45902: thrashosds hits watchdog_daemon_timeout during powercycleNewBrad Hubbard

Actions
Actions #1

Updated by Brad Hubbard almost 4 years ago

  • Related to Bug #45902: thrashosds hits watchdog_daemon_timeout during powercycle added
Actions #2

Updated by Brad Hubbard almost 4 years ago

  • Assignee set to Brad Hubbard
Actions

Also available in: Atom PDF