Project

General

Profile

Actions

Bug #57757

open

ECUtil: terminate called after throwing an instance of 'ceph::buffer::v15_2_0::end_of_buffer'

Added by Laura Flores over 1 year ago. Updated over 1 year ago.

Status:
Fix Under Review
Priority:
Normal
Category:
-
Target version:
-
% Done:

0%

Source:
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(RADOS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

/a/yuriw-2022-09-29_16:44:24-rados-wip-lflores-testing-distro-default-smithi/7048173/remote/smithi133/crash/posted/2022-09-29T23:32:11.008544Z_f3f12160-378c-4631-a08e-53563a23d684/meta

{
    "crash_id": "2022-09-29T23:32:11.008544Z_f3f12160-378c-4631-a08e-53563a23d684",
    "timestamp": "2022-09-29T23:32:11.008544Z",
    "process_name": "ceph-osd",
    "entity_name": "osd.9",
    "ceph_version": "18.0.0-131-gb2711ca9",
    "utsname_hostname": "smithi133",
    "utsname_sysname": "Linux",
    "utsname_release": "5.4.0-126-generic",
    "utsname_version": "#142-Ubuntu SMP Fri Aug 26 12:12:57 UTC 2022",
    "utsname_machine": "x86_64",
    "os_name": "Ubuntu",
    "os_id": "ubuntu",
    "os_version_id": "20.04",
    "os_version": "20.04.5 LTS (Focal Fossa)",
    "backtrace": [
        "/lib/x86_64-linux-gnu/libpthread.so.0(+0x14420) [0x7f06c745d420]",
        "gsignal()",
        "abort()",
        "ceph-osd(+0xd0d130) [0x55b3c6c14130]",
        "(__cxxabiv1::__terminate(void (*)())+0xc) [0x55b3c7e4770c]",
        "ceph-osd(+0x1f40777) [0x55b3c7e47777]",
        "ceph-osd(+0x1f408d9) [0x55b3c7e478d9]",
        "ceph-osd(+0xce37c1) [0x55b3c6bea7c1]",
        "(ECUtil::decode(ECUtil::stripe_info_t const&, std::shared_ptr<ceph::ErasureCodeInterface>&, std::map<int, ceph::buffer::v15_2_0::list, std::less<int>, std::allocator<std::pair<int const, ceph::buffer::v15_2_0::list> > >&, std::map<int, ceph::buffer::v15_2_0::list*, std::less<int>, std::allocator<std::pair<int const, ceph::buffer::v15_2_0::list*> > >&)+0x49b) [0x55b3c6f831cb]",
        "(ECBackend::handle_recovery_read_complete(hobject_t const&, boost::tuples::tuple<unsigned long, unsigned long, std::map<pg_shard_t, ceph::buffer::v15_2_0::list, std::less<pg_shard_t>, std::allocator<std::pair<pg_shard_t const, ceph::buffer::v15_2_0::list> > >, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type>&, std::optional<std::map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, ceph::buffer::v15_2_0::list, std::less<void>, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, ceph::buffer::v15_2_0::list> > > >, RecoveryMessages*)+0x431) [0x55b3c70db3e1]",
        "(OnRecoveryReadComplete::finish(std::pair<RecoveryMessages*, ECBackend::read_result_t&>&)+0x71) [0x55b3c7100981]",
        "(ECBackend::complete_read_op(ECBackend::ReadOp&, RecoveryMessages*)+0x8d) [0x55b3c70cc9ad]",
        "(ECBackend::handle_sub_read_reply(pg_shard_t, ECSubReadReply&, RecoveryMessages*, ZTracer::Trace const&)+0xdc2) [0x55b3c70e4032]",
        "(ECBackend::_handle_message(boost::intrusive_ptr<OpRequest>)+0x2d4) [0x55b3c70e4894]",
        "(PGBackend::handle_message(boost::intrusive_ptr<OpRequest>)+0x56) [0x55b3c6e9c726]",
        "(PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, ThreadPool::TPHandle&)+0x4d2) [0x55b3c6e38682]",
        "(OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, ThreadPool::TPHandle&)+0x1c0) [0x55b3c6c93860]",
        "(ceph::osd::scheduler::PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, ThreadPool::TPHandle&)+0x69) [0x55b3c6f87629]",
        "(OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0xb28) [0x55b3c6cacf38]",
        "(ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x434) [0x55b3c73d7f14]",
        "(ShardedThreadPool::WorkThreadSharded::entry()+0x14) [0x55b3c73db004]",
        "/lib/x86_64-linux-gnu/libpthread.so.0(+0x8609) [0x7f06c7451609]",
        "clone()" 
    ]
}

Teuthology log:

2022-09-29T23:32:10.984 INFO:tasks.ceph.osd.8.smithi133.stderr:2022-09-29T23:32:10.978+0000 7fe938179700 -1 received  signal: Hangup from /usr/bin/python3 /usr/bin/daemon-helper kill ceph-osd -f --cluster ceph -i 8  (PID: 26823) UID: 0
2022-09-29T23:32:10.992 INFO:tasks.ceph.osd.9.smithi133.stderr:terminate called after throwing an instance of 'ceph::buffer::v15_2_0::end_of_buffer'
2022-09-29T23:32:10.993 INFO:tasks.ceph.osd.9.smithi133.stderr:  what():  End of buffer [buffer:2]
2022-09-29T23:32:10.993 INFO:tasks.ceph.osd.9.smithi133.stderr:*** Caught signal (Aborted) **
2022-09-29T23:32:10.993 INFO:tasks.ceph.osd.9.smithi133.stderr: in thread 7f06a5796700 thread_name:tp_osd_tp
2022-09-29T23:32:11.006 INFO:teuthology.orchestra.run.smithi012.stdout:{
2022-09-29T23:32:11.006 INFO:teuthology.orchestra.run.smithi012.stdout:    "ops": [],
2022-09-29T23:32:11.006 INFO:teuthology.orchestra.run.smithi012.stdout:    "num_ops": 0
2022-09-29T23:32:11.006 INFO:teuthology.orchestra.run.smithi012.stdout:}
2022-09-29T23:32:11.010 INFO:tasks.ceph.osd.9.smithi133.stderr: ceph version 18.0.0-131-gb2711ca9 (b2711ca9ca68627a0666c08c8c65c3a27a2da3f2) quincy (dev)
2022-09-29T23:32:11.010 INFO:tasks.ceph.osd.9.smithi133.stderr: 1: /lib/x86_64-linux-gnu/libpthread.so.0(+0x14420) [0x7f06c745d420]
2022-09-29T23:32:11.011 INFO:tasks.ceph.osd.9.smithi133.stderr: 2: gsignal()
2022-09-29T23:32:11.011 INFO:tasks.ceph.osd.9.smithi133.stderr: 3: abort()
2022-09-29T23:32:11.011 INFO:tasks.ceph.osd.9.smithi133.stderr: 4: ceph-osd(+0xd0d130) [0x55b3c6c14130]
2022-09-29T23:32:11.011 INFO:tasks.ceph.osd.9.smithi133.stderr: 5: (__cxxabiv1::__terminate(void (*)())+0xc) [0x55b3c7e4770c]
2022-09-29T23:32:11.012 INFO:tasks.ceph.osd.9.smithi133.stderr: 6: ceph-osd(+0x1f40777) [0x55b3c7e47777]
2022-09-29T23:32:11.012 INFO:tasks.ceph.osd.9.smithi133.stderr: 7: ceph-osd(+0x1f408d9) [0x55b3c7e478d9]
2022-09-29T23:32:11.012 INFO:tasks.ceph.osd.9.smithi133.stderr: 8: ceph-osd(+0xce37c1) [0x55b3c6bea7c1]
2022-09-29T23:32:11.012 INFO:tasks.ceph.osd.9.smithi133.stderr: 9: (ECUtil::decode(ECUtil::stripe_info_t const&, std::shared_ptr<ceph::ErasureCodeInterface>&, std::map<int, ceph::buffer::v15_2_0::list, std::less<int>, std::allocator<std::pair<int const, ceph::buffer::v15_2_0::list> > >&, std::map<int, ceph::buffer::v15_2_0::list*, std::less<int>, std::allocator<std::pair<int const, ceph::buffer::v15_2_0::list*> > >&)+0x49b) [0x55b3c6f831cb]
2022-09-29T23:32:11.013 INFO:tasks.ceph.osd.9.smithi133.stderr: 10: (ECBackend::handle_recovery_read_complete(hobject_t const&, boost::tuples::tuple<unsigned long, unsigned long, std::map<pg_shard_t, ceph::buffer::v15_2_0::list, std::less<pg_shard_t>, std::allocator<std::pair<pg_shard_t const, ceph::buffer::v15_2_0::list> > >, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type, boost::tuples::null_type>&, std::optional<std::map<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, ceph::buffer::v15_2_0::list, std::less<void>, std::allocator<std::pair<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const, ceph::buffer::v15_2_0::list> > > >, RecoveryMessages*)+0x431) [0x55b3c70db3e1]
2022-09-29T23:32:11.013 INFO:tasks.ceph.osd.9.smithi133.stderr: 11: (OnRecoveryReadComplete::finish(std::pair<RecoveryMessages*, ECBackend::read_result_t&>&)+0x71) [0x55b3c7100981]
2022-09-29T23:32:11.013 INFO:tasks.ceph.osd.9.smithi133.stderr: 12: (ECBackend::complete_read_op(ECBackend::ReadOp&, RecoveryMessages*)+0x8d) [0x55b3c70cc9ad]
2022-09-29T23:32:11.013 INFO:tasks.ceph.osd.9.smithi133.stderr: 13: (ECBackend::handle_sub_read_reply(pg_shard_t, ECSubReadReply&, RecoveryMessages*, ZTracer::Trace const&)+0xdc2) [0x55b3c70e4032]
2022-09-29T23:32:11.014 INFO:tasks.ceph.osd.9.smithi133.stderr: 14: (ECBackend::_handle_message(boost::intrusive_ptr<OpRequest>)+0x2d4) [0x55b3c70e4894]
2022-09-29T23:32:11.014 INFO:tasks.ceph.osd.9.smithi133.stderr: 15: (PGBackend::handle_message(boost::intrusive_ptr<OpRequest>)+0x56) [0x55b3c6e9c726]
2022-09-29T23:32:11.014 INFO:tasks.ceph.osd.9.smithi133.stderr: 16: (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, ThreadPool::TPHandle&)+0x4d2) [0x55b3c6e38682]
2022-09-29T23:32:11.014 INFO:tasks.ceph.osd.9.smithi133.stderr: 17: (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, ThreadPool::TPHandle&)+0x1c0) [0x55b3c6c93860]
2022-09-29T23:32:11.015 INFO:tasks.ceph.osd.9.smithi133.stderr: 18: (ceph::osd::scheduler::PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, ThreadPool::TPHandle&)+0x69) [0x55b3c6f87629]
2022-09-29T23:32:11.015 INFO:tasks.ceph.osd.9.smithi133.stderr: 19: (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0xb28) [0x55b3c6cacf38]
2022-09-29T23:32:11.015 INFO:tasks.ceph.osd.9.smithi133.stderr: 20: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x434) [0x55b3c73d7f14]
2022-09-29T23:32:11.015 INFO:tasks.ceph.osd.9.smithi133.stderr: 21: (ShardedThreadPool::WorkThreadSharded::entry()+0x14) [0x55b3c73db004]
2022-09-29T23:32:11.015 INFO:tasks.ceph.osd.9.smithi133.stderr: 22: /lib/x86_64-linux-gnu/libpthread.so.0(+0x8609) [0x7f06c7451609]
2022-09-29T23:32:11.016 INFO:tasks.ceph.osd.9.smithi133.stderr: 23: clone()

Found a very similar instance of this crash tracked in Telemetry. The affected version there is v14.2.19.
http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?var-sig_v2=ab549128220593a3ca9e0e5dacabaa9ffae5833e27ebd2cb413eeb3c1ac669ee&orgId=1

There were a few other crashes that looked similar (link pasted below), but the one above is the closest to the crash logged in Teuthology. All versions are 14:
http://telemetry.front.sepia.ceph.com:4000/d/Nvj6XTaMk/spec-search?orgId=1&var-substr_1=ECUtil::decode&var-substr_2=ECBackend::handle_recovery_read_complete&var-substr_3=OSD::dequeue_op&var-majors_affected=&var-minors_affected=&var-assert_function=&var-assert_condition=&var-total_results=4&var-sig_v1=&var-sig_v2=&var-daemons=&var-daemons=ceph-osd&var-only_new_fingerprints=false&var-status_description=All&var-only_open=false

Actions

Also available in: Atom PDF