Actions
Bug #23395
closedqa/standalone/special/ceph_objectstore_tool.py causes ceph-mon core dump
Status:
Can't reproduce
Priority:
Normal
Assignee:
David Zafman
Category:
EC Pools
Target version:
-
% Done:
0%
Source:
Development
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(RADOS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):
Description
I assume erasure code profile handling must have changed. It shouldn't crash but we may need a test change too.
-24> 2018-03-16 18:50:36.548 7f332a738700 10 mon.a@0(leader).osd e2 check_osdmap_sub 0x5575810cf420 next 0 (onetime) -23> 2018-03-16 18:50:36.548 7f332a738700 1 -- 127.0.0.1:7400/0 --> 127.0.0.1:0/417199523 -- osd_map(2..2 src has 1..2) v4 -- 0x5575809143c0 con 0 -22> 2018-03-16 18:50:36.552 7f332a738700 1 -- 127.0.0.1:7400/0 <== client.4122 127.0.0.1:0/417199523 7 ==== mon_command({"prefix": "get_command_descriptions"} v 0) v1 ==== 80+0+0 (2064402080 0 0) 0x5575812361c0 con 0x557580a2e300 -21> 2018-03-16 18:50:36.552 7f332a738700 20 mon.a@0(leader) e1 _ms_dispatch existing session 0x557581235d40 for client.? 127.0.0.1:0/417199523 -20> 2018-03-16 18:50:36.552 7f332a738700 20 mon.a@0(leader) e1 caps allow * -19> 2018-03-16 18:50:36.556 7f332a738700 2 mon.a@0(leader) e1 send_reply 0x55758117c080 0x5575812358c0 mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 -18> 2018-03-16 18:50:36.556 7f332a738700 1 -- 127.0.0.1:7400/0 --> 127.0.0.1:0/417199523 -- mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 -- 0x5575812358c0 con 0 -17> 2018-03-16 18:50:36.636 7f332a738700 1 -- 127.0.0.1:7400/0 <== client.4122 127.0.0.1:0/417199523 8 ==== mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1 ==== 150+0+0 (2277797133 0 0) 0x5575812358c0 con 0x557580a2e300 -16> 2018-03-16 18:50:36.636 7f332a738700 20 mon.a@0(leader) e1 _ms_dispatch existing session 0x557581235d40 for client.? 127.0.0.1:0/417199523 -15> 2018-03-16 18:50:36.636 7f332a738700 20 mon.a@0(leader) e1 caps allow * -14> 2018-03-16 18:50:36.636 7f332a738700 0 mon.a@0(leader) e1 handle_command mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1 -13> 2018-03-16 18:50:36.636 7f332a738700 20 is_capable service=osd command=osd erasure-code-profile set read write on cap allow * -12> 2018-03-16 18:50:36.636 7f332a738700 20 allow so far , doing grant allow * -11> 2018-03-16 18:50:36.636 7f332a738700 20 allow all -10> 2018-03-16 18:50:36.636 7f332a738700 10 mon.a@0(leader) e1 _allowed_command capable -9> 2018-03-16 18:50:36.636 7f332a738700 0 log_channel(audit) log [INF] : from='client.? 127.0.0.1:0/417199523' entity='client.admin' cmd=[{"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"}]: dispatch -8> 2018-03-16 18:50:36.636 7f332a738700 10 log_client _send_to_monlog to self -7> 2018-03-16 18:50:36.636 7f332a738700 10 log_client log_queue is 1 last_log 35 sent 34 num 1 unsent 1 sending 1 -6> 2018-03-16 18:50:36.636 7f332a738700 10 log_client will send 2018-03-16 18:50:36.640420 mon.a mon.0 127.0.0.1:7400/0 35 : audit [INF] from='client.? 127.0.0.1:0/417199523' entity='client.admin' cmd=[{"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"}]: dispatch -5> 2018-03-16 18:50:36.636 7f332a738700 1 -- 127.0.0.1:7400/0 --> 127.0.0.1:7400/0 -- log(1 entries from seq 35 at 2018-03-16 18:50:36.640420) v1 -- 0x55758124db80 con 0 -4> 2018-03-16 18:50:36.636 7f332a738700 10 mon.a@0(leader).paxosservice(osdmap 1..2) dispatch 0x5575812358c0 mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1 from client.4122 127.0.0.1:0/417199523 con 0x557580a2e300 -3> 2018-03-16 18:50:36.636 7f332a738700 5 mon.a@0(leader).paxos(paxos active c 1..19) is_readable = 1 - now=2018-03-16 18:50:36.640445 lease_expire=0.000000 has v0 lc 19 -2> 2018-03-16 18:50:36.636 7f332a738700 10 mon.a@0(leader).osd e2 preprocess_query mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1 from client.4122 127.0.0.1:0/417199523 -1> 2018-03-16 18:50:36.636 7f332a738700 7 mon.a@0(leader).osd e2 prepare_update mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1 from client.4122 127.0.0.1:0/417199523 0> 2018-03-16 18:50:36.640 7f332a738700 -1 /home/dzafman/ceph/src/mon/OSDMonitor.cc: In function 'int OSDMonitor::parse_erasure_code_profile(const std::vector<std::__cxx11::basic_string<char> >&, std::map<std::__cxx11::basic_string<char>, std::__cxx11::basic_string<char> >*, std::ostream*)' thread 7f332a738700 time 2018-03-16 18:50:36.640570 /home/dzafman/ceph/src/mon/OSDMonitor.cc: 5501: FAILED assert((*erasure_code_profile_map).count("plugin"))
Updated by David Zafman about 6 years ago
../qa/run-standalone.sh ceph_objectstore_tool.py
--- ../qa/standalone/special/ceph_objectstore_tool.py ---
vstarting....DONE
Wait for health_ok...DONE
Created Replicated pool #1
ps shows this is being executed:
/src/ceph/build/bin/ceph osd erasure-code-profile set testecprofile crush-failure-domain=osd
Updated by David Zafman almost 6 years ago
- Status changed from New to Can't reproduce
Actions