Actions
Bug #60304
opencrash: __pthread_rwlock_wrlock()
Status:
New
Priority:
Normal
Assignee:
-
Target version:
-
% Done:
0%
Source:
Telemetry
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
c926009580f88ce0a4412923032310d706de3a751e9b60f60903bfd19752650f
Crash signature (v2):
Description
Sanitized backtrace:
__pthread_rwlock_wrlock() BlueStore::_txc_add_transaction(BlueStore::TransContext*, ceph::os::Transaction*) BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, std::vector<ceph::os::Transaction, std::allocator<ceph::os::Transaction> >&, boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*) ObjectStore::queue_transaction(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, ceph::os::Transaction&&, boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*) non-virtual thunk to PrimaryLogPG::queue_transaction(ceph::os::Transaction&&, boost::intrusive_ptr<OpRequest>) ECBackend::dispatch_recovery_messages(RecoveryMessages&, int) ECBackend::_handle_message(boost::intrusive_ptr<OpRequest>) PGBackend::handle_message(boost::intrusive_ptr<OpRequest>) PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, ThreadPool::TPHandle&) OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, ThreadPool::TPHandle&) ceph::osd::scheduler::PGRecoveryMsg::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, ThreadPool::TPHandle&) OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*) ShardedThreadPool::shardedthreadpool_worker(unsigned int) ShardedThreadPool::WorkThreadSharded::entry()
Crash dump sample:
{ "backtrace": [ "/lib64/libpthread.so.0(+0x12cf0) [0x7fb93134ccf0]", "__pthread_rwlock_wrlock()", "(BlueStore::_txc_add_transaction(BlueStore::TransContext*, ceph::os::Transaction*)+0x2e6) [0x5569919afe76]", "(BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, std::vector<ceph::os::Transaction, std::allocator<ceph::os::Transaction> >&, boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x2df) [0x5569919c0a5f]", "(ObjectStore::queue_transaction(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, ceph::os::Transaction&&, boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0xae) [0x5569914218fe]", "(non-virtual thunk to PrimaryLogPG::queue_transaction(ceph::os::Transaction&&, boost::intrusive_ptr<OpRequest>)+0x4e) [0x55699158533e]", "(ECBackend::dispatch_recovery_messages(RecoveryMessages&, int)+0x11ad) [0x556991814bdd]", "(ECBackend::_handle_message(boost::intrusive_ptr<OpRequest>)+0x59b) [0x556991825ceb]", "(PGBackend::handle_message(boost::intrusive_ptr<OpRequest>)+0x56) [0x5569915d82c6]", "(PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, ThreadPool::TPHandle&)+0x512) [0x55699156a9f2]", "(OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, ThreadPool::TPHandle&)+0x1c0) [0x5569913c0900]", "(ceph::osd::scheduler::PGRecoveryMsg::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, ThreadPool::TPHandle&)+0x69) [0x5569916bf389]", "(OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x115f) [0x5569913d3dbf]", "(ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x435) [0x556991b318c5]", "(ShardedThreadPool::WorkThreadSharded::entry()+0x14) [0x556991b33fe4]", "/lib64/libpthread.so.0(+0x81ca) [0x7fb9313421ca]", "clone()" ], "ceph_version": "17.2.5", "crash_id": "2023-01-26T06:00:10.775498Z_133f09dd-66ef-4ada-800b-5e6e5db78b51", "entity_name": "osd.ffd2bc8f23828fbba58dbe542083b2bed089f9ac", "os_id": "centos", "os_name": "CentOS Stream", "os_version": "8", "os_version_id": "8", "process_name": "ceph-osd", "stack_sig": "c926009580f88ce0a4412923032310d706de3a751e9b60f60903bfd19752650f", "timestamp": "2023-01-26T06:00:10.775498Z", "utsname_machine": "x86_64", "utsname_release": "5.3.18-150300.59.87-default", "utsname_sysname": "Linux", "utsname_version": "#1 SMP Thu Jul 21 14:31:28 UTC 2022 (cc90276)" }
Actions