Project

General

Profile

Actions

Bug #23395

closed

qa/standalone/special/ceph_objectstore_tool.py causes ceph-mon core dump

Added by David Zafman about 6 years ago. Updated almost 6 years ago.

Status:
Can't reproduce
Priority:
Normal
Assignee:
David Zafman
Category:
EC Pools
Target version:
-
% Done:

0%

Source:
Development
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(RADOS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

I assume erasure code profile handling must have changed. It shouldn't crash but we may need a test change too.

   -24> 2018-03-16 18:50:36.548 7f332a738700 10 mon.a@0(leader).osd e2 check_osdmap_sub 0x5575810cf420 next 0 (onetime)
   -23> 2018-03-16 18:50:36.548 7f332a738700  1 -- 127.0.0.1:7400/0 --> 127.0.0.1:0/417199523 -- osd_map(2..2 src has 1..2) v4 -- 0x5575809143c0 con 0
   -22> 2018-03-16 18:50:36.552 7f332a738700  1 -- 127.0.0.1:7400/0 <== client.4122 127.0.0.1:0/417199523 7 ==== mon_command({"prefix": "get_command_descriptions"} v 0) v1 ==== 80+0+0 (2064402080 0 0) 0x5575812361c0 con 0x557580a2e300
   -21> 2018-03-16 18:50:36.552 7f332a738700 20 mon.a@0(leader) e1 _ms_dispatch existing session 0x557581235d40 for client.? 127.0.0.1:0/417199523
   -20> 2018-03-16 18:50:36.552 7f332a738700 20 mon.a@0(leader) e1  caps allow *
   -19> 2018-03-16 18:50:36.556 7f332a738700  2 mon.a@0(leader) e1 send_reply 0x55758117c080 0x5575812358c0 mon_command_ack([{"prefix": "get_command_descriptions"}]=0  v0) v1
   -18> 2018-03-16 18:50:36.556 7f332a738700  1 -- 127.0.0.1:7400/0 --> 127.0.0.1:0/417199523 -- mon_command_ack([{"prefix": "get_command_descriptions"}]=0  v0) v1 -- 0x5575812358c0 con 0
   -17> 2018-03-16 18:50:36.636 7f332a738700  1 -- 127.0.0.1:7400/0 <== client.4122 127.0.0.1:0/417199523 8 ==== mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1 ==== 150+0+0 (2277797133 0 0) 0x5575812358c0 con 0x557580a2e300
   -16> 2018-03-16 18:50:36.636 7f332a738700 20 mon.a@0(leader) e1 _ms_dispatch existing session 0x557581235d40 for client.? 127.0.0.1:0/417199523
   -15> 2018-03-16 18:50:36.636 7f332a738700 20 mon.a@0(leader) e1  caps allow *
   -14> 2018-03-16 18:50:36.636 7f332a738700  0 mon.a@0(leader) e1 handle_command mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1
   -13> 2018-03-16 18:50:36.636 7f332a738700 20 is_capable service=osd command=osd erasure-code-profile set read write on cap allow *
   -12> 2018-03-16 18:50:36.636 7f332a738700 20  allow so far , doing grant allow *
   -11> 2018-03-16 18:50:36.636 7f332a738700 20  allow all
   -10> 2018-03-16 18:50:36.636 7f332a738700 10 mon.a@0(leader) e1 _allowed_command capable
    -9> 2018-03-16 18:50:36.636 7f332a738700  0 log_channel(audit) log [INF] : from='client.? 127.0.0.1:0/417199523' entity='client.admin' cmd=[{"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"}]: dispatch
    -8> 2018-03-16 18:50:36.636 7f332a738700 10 log_client _send_to_monlog to self
    -7> 2018-03-16 18:50:36.636 7f332a738700 10 log_client  log_queue is 1 last_log 35 sent 34 num 1 unsent 1 sending 1
    -6> 2018-03-16 18:50:36.636 7f332a738700 10 log_client  will send 2018-03-16 18:50:36.640420 mon.a mon.0 127.0.0.1:7400/0 35 : audit [INF] from='client.? 127.0.0.1:0/417199523' entity='client.admin' cmd=[{"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"}]: dispatch
    -5> 2018-03-16 18:50:36.636 7f332a738700  1 -- 127.0.0.1:7400/0 --> 127.0.0.1:7400/0 -- log(1 entries from seq 35 at 2018-03-16 18:50:36.640420) v1 -- 0x55758124db80 con 0
    -4> 2018-03-16 18:50:36.636 7f332a738700 10 mon.a@0(leader).paxosservice(osdmap 1..2) dispatch 0x5575812358c0 mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1 from client.4122 127.0.0.1:0/417199523 con 0x557580a2e300
    -3> 2018-03-16 18:50:36.636 7f332a738700  5 mon.a@0(leader).paxos(paxos active c 1..19) is_readable = 1 - now=2018-03-16 18:50:36.640445 lease_expire=0.000000 has v0 lc 19
    -2> 2018-03-16 18:50:36.636 7f332a738700 10 mon.a@0(leader).osd e2 preprocess_query mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1 from client.4122 127.0.0.1:0/417199523
    -1> 2018-03-16 18:50:36.636 7f332a738700  7 mon.a@0(leader).osd e2 prepare_update mon_command({"profile": ["crush-failure-domain=osd"], "prefix": "osd erasure-code-profile set", "name": "testecprofile"} v 0) v1 from client.4122 127.0.0.1:0/417199523
     0> 2018-03-16 18:50:36.640 7f332a738700 -1 /home/dzafman/ceph/src/mon/OSDMonitor.cc: In function 'int OSDMonitor::parse_erasure_code_profile(const std::vector<std::__cxx11::basic_string<char> >&, std::map<std::__cxx11::basic_string<char>, std::__cxx11::basic_string<char> >*, std::ostream*)' thread 7f332a738700 time 2018-03-16 18:50:36.640570
/home/dzafman/ceph/src/mon/OSDMonitor.cc: 5501: FAILED assert((*erasure_code_profile_map).count("plugin"))
Actions #1

Updated by David Zafman about 6 years ago

  • Description updated (diff)
Actions #2

Updated by David Zafman about 6 years ago

../qa/run-standalone.sh ceph_objectstore_tool.py

--- ../qa/standalone/special/ceph_objectstore_tool.py ---
vstarting....DONE
Wait for health_ok...DONE
Created Replicated pool #1

ps shows this is being executed:

/src/ceph/build/bin/ceph osd erasure-code-profile set testecprofile crush-failure-domain=osd

Actions #3

Updated by David Zafman almost 6 years ago

  • Status changed from New to Can't reproduce
Actions

Also available in: Atom PDF