Actions
Bug #8214
closedCrash in Thread.cc "common/Thread.cc: 110: FAILED assert(ret == 0)" in upgrade:dumpling-x:stress-split-firefly-distro-basic-vps suite
Status:
Can't reproduce
Priority:
Normal
Assignee:
-
Category:
-
Target version:
-
% Done:
0%
Source:
other
Tags:
Backport:
Regression:
Severity:
2 - major
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
Crash signature (v2):
Description
This look similar to #8156, but has no out of memory problem.
2014-04-25T06:33:52.251 DEBUG:teuthology.orchestra.run:Running [10.214.138.144]: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --admin-daemon /var/run/ceph/ceph-osd.2.asok config set filestore_inject_stall 3' 2014-04-25T06:34:11.263 INFO:teuthology.task.thrashosds.thrasher:in_osds: [5, 2, 3, 4, 0] out_osds: [1] dead_osds: [] live_osds: [5, 1, 0, 3, 4, 2] 2014-04-25T06:34:11.263 INFO:teuthology.task.thrashosds.thrasher:choose_action: min_in 3 min_out 0 min_live 2 min_dead 0 2014-04-25T06:34:11.264 INFO:teuthology.task.thrashosds.thrasher:Growing pool unique_pool_2 2014-04-25T06:34:11.264 DEBUG:teuthology.orchestra.run:Running [10.214.138.144]: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph pg dump --format=json' 2014-04-25T06:34:17.736 INFO:teuthology.orchestra.run.err:[10.214.138.144]: dumped all in format json 2014-04-25T06:34:19.659 INFO:teuthology.task.thrashosds.ceph_manager:increase pool size by 10 2014-04-25T06:34:19.660 DEBUG:teuthology.orchestra.run:Running [10.214.138.144]: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd pool set unique_pool_2 pg_num 96' 2014-04-25T06:34:27.951 INFO:teuthology.orchestra.run.err:[10.214.138.144]: set pool 139 pg_num to 96 2014-04-25T06:34:34.207 INFO:teuthology.task.thrashosds.thrasher:in_osds: [5, 2, 3, 4, 0] out_osds: [1] dead_osds: [] live_osds: [5, 1, 0, 3, 4, 2] 2014-04-25T06:34:34.207 INFO:teuthology.task.thrashosds.thrasher:choose_action: min_in 3 min_out 0 min_live 2 min_dead 0 2014-04-25T06:34:34.207 INFO:teuthology.task.thrashosds.thrasher:Adding osd 1 2014-04-25T06:34:34.207 DEBUG:teuthology.orchestra.run:Running [10.214.138.144]: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd in 1' 2014-04-25T06:35:24.476 INFO:teuthology.orchestra.run.err:[10.214.138.144]: Thread::try_create(): pthread_create failed with error 11common/Thread.cc: In function 'void Thread::create(size_t)' thread 7f9537a71700 time 2014-04-25 13:35:24.323605 2014-04-25T06:35:24.476 INFO:teuthology.orchestra.run.err:[10.214.138.144]: common/Thread.cc: 110: FAILED assert(ret == 0) 2014-04-25T06:35:24.493 INFO:teuthology.orchestra.run.err:[10.214.138.144]: ceph version 0.80-rc1-16-g2708c3c (2708c3c559d99e6f3b557ee1d223efa3745f655c) 2014-04-25T06:35:24.494 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 1: (Thread::create(unsigned long)+0x8a) [0x7f9538f2f53a] 2014-04-25T06:35:24.494 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 2: (SafeTimer::init()+0x7f) [0x7f9538f0723f] 2014-04-25T06:35:24.494 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 3: (librados::RadosClient::connect()+0xcc2) [0x7f9538e69832] 2014-04-25T06:35:24.494 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 4: (ffi_call_unix64()+0x4c) [0x7f953aaa6adc] 2014-04-25T06:35:24.494 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 5: (ffi_call()+0x1fc) [0x7f953aaa640c] 2014-04-25T06:35:24.495 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 6: (_ctypes_callproc()+0x48e) [0x7f953acbd5fe] 2014-04-25T06:35:24.495 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 7: (()+0x15f9e) [0x7f953acbef9e] 2014-04-25T06:35:24.495 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 8: (PyEval_EvalFrameEx()+0x1f36) [0x52e1e6] 2014-04-25T06:35:24.495 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 9: (PyEval_EvalFrameEx()+0xc82) [0x52cf32] 2014-04-25T06:35:24.496 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 10: (PyEval_EvalFrameEx()+0xc82) [0x52cf32] 2014-04-25T06:35:24.496 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 11: /usr/bin/python() [0x56d0aa] 2014-04-25T06:35:24.496 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 12: /usr/bin/python() [0x4d9854] 2014-04-25T06:35:24.496 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 13: (PyEval_CallObjectWithKeywords()+0x6b) [0x4da20b] 2014-04-25T06:35:24.497 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 14: /usr/bin/python() [0x5872b2] 2014-04-25T06:35:24.497 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 15: (()+0x8182) [0x7f953c115182] 2014-04-25T06:35:24.497 INFO:teuthology.orchestra.run.err:[10.214.138.144]: 16: (clone()+0x6d) [0x7f953be4230d] 2014-04-25T06:35:24.497 INFO:teuthology.orchestra.run.err:[10.214.138.144]: NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. 2014-04-25T06:35:24.498 INFO:teuthology.orchestra.run.err:[10.214.138.144]: terminate called after throwing an instance of 'ceph::FailedAssertion' 2014-04-25T06:36:36.953 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: finishing write tid 1 to vpm08810175-36 2014-04-25T06:36:36.960 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: finishing write tid 2 to vpm08810175-36 2014-04-25T06:36:41.252 INFO:teuthology.task.ceph.osd.0.err:[10.214.138.144]: daemon-helper: command crashed with signal 6 2014-04-25T06:37:05.450 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: finishing write tid 3 to vpm08810175-36 2014-04-25T06:37:07.245 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: 3266: oids not in use 500 2014-04-25T06:37:07.245 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: Snapping 2014-04-25T06:37:08.809 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: 3267: oids not in use 500 2014-04-25T06:37:08.809 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: Snapping 2014-04-25T06:37:10.823 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: 3268: oids not in use 500 2014-04-25T06:37:10.823 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: RollingBack 194 to 352 2014-04-25T06:37:10.928 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: 3269: oids not in use 500 2014-04-25T06:37:10.928 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: Snapping 2014-04-25T06:37:13.206 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: 3270: oids not in use 500 2014-04-25T06:37:13.206 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: Deleting 62 current snap is 359 2014-04-25T06:37:13.207 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: 3271: oids not in use 500 2014-04-25T06:37:13.207 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: Snapping 2014-04-25T06:37:15.519 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: 3272: oids not in use 500 2014-04-25T06:37:15.519 INFO:teuthology.task.rados.rados.0.out:[10.214.138.133]: Writing 217 current snap is 360
archive_path: /var/lib/teuthworker/archive/teuthology-2014-04-24_20:35:03-upgrade:dumpling-x:stress-split-firefly-distro-basic-vps/213371 description: upgrade/dumpling-x/stress-split/{0-cluster/start.yaml 1-dumpling-install/dumpling.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/snaps-few-objects.yaml 6-next-mon/monb.yaml 7-workload/radosbench.yaml 8-next-mon/monc.yaml 9-workload/{rados_api_tests.yaml rbd-python.yaml rgw-s3tests.yaml snaps-many-objects.yaml} distros/ubuntu_14.04.yaml} email: null job_id: '213371' kernel: &id001 kdb: true sha1: distro last_in_suite: false machine_type: vps name: teuthology-2014-04-24_20:35:03-upgrade:dumpling-x:stress-split-firefly-distro-basic-vps nuke-on-error: true os_type: ubuntu os_version: '14.04' overrides: admin_socket: branch: firefly ceph: conf: mon: debug mon: 20 debug ms: 1 debug paxos: 20 mon warn on legacy crush tunables: false osd: debug filestore: 20 debug journal: 20 debug ms: 1 debug osd: 20 log-whitelist: - slow request - wrongly marked me down - objects unfound and apparently lost - log bound mismatch sha1: 2708c3c559d99e6f3b557ee1d223efa3745f655c ceph-deploy: branch: dev: firefly conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: debug mon: 1 debug ms: 20 debug paxos: 20 osd default pool size: 2 install: ceph: sha1: 2708c3c559d99e6f3b557ee1d223efa3745f655c s3tests: branch: master workunit: sha1: 2708c3c559d99e6f3b557ee1d223efa3745f655c owner: scheduled_teuthology@teuthology roles: - - mon.a - mon.b - mds.a - osd.0 - osd.1 - osd.2 - - osd.3 - osd.4 - osd.5 - mon.c - - client.0 targets: ubuntu@vpm086.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCtEDwzUe+qPsDo5zamTc3vPTySPm6Zdp0tF3N13+6P3uNCXChdAQQHm9UwcymTaZjHGr0ou1JKgxQ4bG/GxuOtB5XB0h84h/LZkDboYGzppzYXpxqOPA4fsa7wE/CkB1L1ECvaUbMbd/bu7etC9vX5BInL9fbmQZx/ZwUvJ5uPUAkVXpZHHFxwh0pxHNsWkv3Rjygg6A8PoD0z7ty2ZboKt7Dee5xGcohSI+cJCBXHiO176tI/JasB+ohmMTYqhlgBnvvcBlRnoyCN0b72v4CXPP8DT06QEPKk3ojnP2LNT/7LfxCMp7Yw4xpQ4dyttmhUAN3lxmTqxv3wZW5HxBV7 ubuntu@vpm087.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCpNfpBMj6XnPPsR4+mIsfICXIjy26Kefpc+8hi9aoY3xpUQUo1OIaepwBFHA+0QnIv0W1OdSbiJm8f1y7JFD1HtVEx6RbxNibN9cr/GVTjEAr8eMwn4y0u6xdZ9JnwINxCkJt0PhbQzMn6x/mWf0M2GcJjF50nfUDBdDiLZ7EI0A6+iYPe5I6lrVM3qjXVrDPO4ox0Wn6P+f87/KOCHlSM5GEFuCOnD1QnEFPwwjV4uidYEbbDWtH+5mzPOWr09Z/WMXK6M/eLSb8cVbNlV+VkA4sJhWJ9URiMftBA+XiFv8ymw/p36bj9Pl3LS0jpG4uFL3EppbsBCLLzwiETjKqp ubuntu@vpm088.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDPYGsveQQeLY6gsvgfOqJQLDW7oLHMGBKnFSRTPFRYay8Dza5GaAw8sqayeXKXYIjTS8JCj6stZncK7Enc9du70lJ14jtRQffLatpz4045MqE1S6sTNWKCFZjn8KNUVRsggum1Eb76+4oEt1q7y+a4ZETyTwKv6n8IRXvdk51XL77kZV5tw+Dpz+WwxK6ufERrxfyE2AbbC+USkLAThLonw2P8D2xf3ctlsjrFSYN5aPRD7kjl6H8AM9+ipMjaXR/0yY7oJrdaBuMLZhLeshu3L5IGChV/3uDCnFSDtchAs7a2VcSwtHPLlgsCzMXDvYbGu2Ttd2rg+vYCYwnGHnY5 tasks: - internal.lock_machines: - 3 - vps - internal.save_config: null - internal.check_lock: null - internal.connect: null - internal.check_conflict: null - internal.check_ceph_data: null - internal.vm_setup: null - kernel: *id001 - internal.base: null - internal.archive: null - internal.coredump: null - internal.sudo: null - internal.syslog: null - internal.timer: null - chef: null - clock.check: null - install: branch: dumpling - ceph: fs: xfs - install.upgrade: osd.0: null - ceph.restart: daemons: - osd.0 - osd.1 - osd.2 - thrashosds: chance_pgnum_grow: 1 chance_pgpnum_fix: 1 thrash_primary_affinity: false timeout: 1200 - ceph.restart: daemons: - mon.a wait-for-healthy: false wait-for-osds-up: true - rados: clients: - client.0 objects: 50 op_weights: delete: 50 read: 100 rollback: 50 snap_create: 50 snap_remove: 50 write: 100 ops: 4000 - ceph.restart: daemons: - mon.b wait-for-healthy: false wait-for-osds-up: true - radosbench: clients: - client.0 time: 1800 - install.upgrade: mon.c: null - ceph.restart: daemons: - mon.c wait-for-healthy: false wait-for-osds-up: true - ceph.wait_for_mon_quorum: - a - b - c - workunit: branch: dumpling clients: client.0: - rados/test-upgrade-firefly.sh - workunit: branch: dumpling clients: client.0: - rbd/test_librbd_python.sh - rgw: client.0: idle_timeout: 300 - swift: client.0: rgw_server: client.0 - rados: clients: - client.0 objects: 500 op_weights: delete: 50 read: 100 rollback: 50 snap_create: 50 snap_remove: 50 write: 100 ops: 4000 teuthology_branch: firefly verbose: true worker_log: /var/lib/teuthworker/archive/worker_logs/worker.vps.30578
description: upgrade/dumpling-x/stress-split/{0-cluster/start.yaml 1-dumpling-install/dumpling.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/snaps-few-objects.yaml 6-next-mon/monb.yaml 7-workload/radosbench.yaml 8-next-mon/monc.yaml 9-workload/{rados_api_tests.yaml rbd-python.yaml rgw-s3tests.yaml snaps-many-objects.yaml} distros/ubuntu_14.04.yaml} duration: 16598.548552036285 failure_reason: 'Command crashed: ''adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph osd in 1''' flavor: basic owner: scheduled_teuthology@teuthology success: false
Actions