Project

General

Profile

Actions

Bug #38293

closed

qa/standalone/osd/osd-backfill-prio.sh failed

Added by Sage Weil about 5 years ago. Updated about 5 years ago.

Status:
Duplicate
Priority:
Urgent
Assignee:
Category:
-
Target version:
-
% Done:

0%

Source:
Tags:
Backport:
Regression:
No
Severity:
3 - minor
Reviewed:
Affected Versions:
ceph-qa-suite:
Component(RADOS):
Pull request ID:
Crash signature (v1):
Crash signature (v2):

Description

2019-02-13T05:27:20.507 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:207: TEST_backfill_priosleep 2
2019-02-13T05:27:22.509 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_prioceph pg force-backfill 3.0
2019-02-13T05:27:22.509 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_priogrep -q 'doesn'\''t require backfilling'
2019-02-13T05:27:22.970 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:207: TEST_backfill_priosleep 2
2019-02-13T05:27:24.971 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_prioceph pg force-backfill 3.0
2019-02-13T05:27:24.971 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_priogrep -q 'doesn'\''t require backfilling'
2019-02-13T05:27:25.427 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:207: TEST_backfill_priosleep 2
2019-02-13T05:27:27.428 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_prioceph pg force-backfill 3.0
2019-02-13T05:27:27.429 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_priogrep -q 'doesn'\''t require backfilling'
2019-02-13T05:27:27.887 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:207: TEST_backfill_priosleep 2
2019-02-13T05:27:29.889 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_prioceph pg force-backfill 3.0
2019-02-13T05:27:29.889 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_priogrep -q 'doesn'\''t require backfilling'
2019-02-13T05:27:30.347 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:207: TEST_backfill_priosleep 2
2019-02-13T05:27:32.348 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_prioceph pg force-backfill 3.0
2019-02-13T05:27:32.348 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:205: TEST_backfill_priogrep -q 'doesn'\''t require backfilling'
2019-02-13T05:27:32.807 INFO:tasks.workunit.client.0.smithi067.stderr:/home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-prio.sh:207: TEST_backfill_priosleep 2
2019-02-13T05:27:34.116 INFO:tasks.workunit.client.0.smithi067.stderr:Terminated

/a/sage-2019-02-13_00:42:53-rados-wip-sage2-testing-2019-02-12-1700-distro-basic-smithi/3581737

Related issues 1 (0 open1 closed)

Is duplicate of RADOS - Bug #38282: cephtool/test.sh failure in test_mon_osd_pool_setResolved02/12/2019

Actions
Actions #1

Updated by David Zafman about 5 years ago

  • Assignee set to David Zafman
Actions #2

Updated by David Zafman about 5 years ago

  • Assignee changed from David Zafman to Sage Weil

This test shouldn't loop forever so that teuthology has to time-out. I'll fix that elsewhere.

The mon received the size change for test3 pool, but there is no corresponding handle_advance_map message at the OSD.

2019-02-13 02:33:07.069 7f13ef545700  0 log_channel(audit) log [INF] : from='client.? v1:127.0.0.1:0/1343403815' entity='client.admin' cmd='[{"var": "size", "prefix": "osd pool set" 
, "pool": "test3", "val": "2"}]': finished
2019-02-13 02:33:07.069 7f13ef545700  1 -- v2:127.0.0.1:7114/0 --> v2:127.0.0.1:7114/0 -- log(1 entries from seq 437 at 2019-02-13 02:33:07.070214) v1 -- 0x556c75b80b40 con 0x556c74
61e400
2019-02-13 02:33:07.069 7f13ef545700 20 -- v2:127.0.0.1:7114/0 >> v2:127.0.0.1:7114/0 conn(0x556c7461e400 msgr2=0x556c74620b00 :-1 s=STATE_NONE l=0).send_message log(1 entries from
seq 437 at 2019-02-13 02:33:07.070214) v1 local
2019-02-13 02:33:07.069 7f13ef545700  2 mon.a@0(leader) e1 send_reply 0x556c75aca1a0 0x556c75cd4a00 mon_command_ack([{"var": "size", "prefix": "osd pool set", "pool": "test3", "val" 
: "2"}]=0 set pool 3 size to 2 v105) v1
2019-02-13 02:33:07.069 7f13ef545700  1 -- v2:127.0.0.1:7114/0 --> v1:127.0.0.1:0/1343403815 -- mon_command_ack([{"var": "size", "prefix": "osd pool set", "pool": "test3", "val": "2
"}]=0 set pool 3 size to 2 v105) v1 -- 0x556c75cd4a00 con 0x556c76480c00
2019-02-13 02:33:07.069 7f13ef545700  5 --2- v2:127.0.0.1:7114/0 >> v1:127.0.0.1:0/1343403815 conn(0x556c76480c00 0x556c75bc3200 :-1 s=READY pgs=2 cs=0 l=1).send_message enqueueing
message m=0x556c75cd4a00 type=51 mon_command_ack([{"var": "size", "prefix": "osd pool set", "pool": "test3", "val": "2"}]=0 set pool 3 size to 2 v105) v1
2019-02-13 02:33:07.069 7f13ef545700 15 --2- v2:127.0.0.1:7114/0 >> v1:127.0.0.1:0/1343403815 conn(0x556c76480c00 0x556c75bc3200 :-1 s=READY pgs=2 cs=0 l=1).send_message inline writ
e is denied, reschedule m=0x556c75cd4a00
2019-02-13 02:33:07.069 7f13ef545700 20 Event(0x556c74627480 nevent=5000 time_id=169).wakeup
2019-02-13 02:33:07.069 7f13ef545700 10 mon.a@0(leader).osd e105 update_logger
2019-02-13 02:33:07.069 7f13ef545700  0 log_channel(cluster) log [DBG] : osdmap e105: 5 total, 5 up, 4 in
2019-02-13 02:33:07.069 7f13ef545700  1 -- v2:127.0.0.1:7114/0 --> v2:127.0.0.1:7114/0 -- log(1 entries from seq 438 at 2019-02-13 02:33:07.070249) v1 -- 0x556c75b80000 con 0x556c74
61e400
2019-02-13 02:33:07.069 7f13ef545700 20 -- v2:127.0.0.1:7114/0 >> v2:127.0.0.1:7114/0 conn(0x556c7461e400 msgr2=0x556c74620b00 :-1 s=STATE_NONE l=0).send_message log(1 entries from
seq 438 at 2019-02-13 02:33:07.070249) v1 local
2019-02-13 02:33:07.069 7f13ef545700 20 queue 0x556c75da91e0 1 [0,1)
2019-02-13 02:33:07.069 7f13ef545700 20 queue 0x556c75da91e0 3 [0,1)
2019-02-13 02:33:07.069 7f13ef545700 20 queue 0x556c75da91e0 7 [0,1)
2019-02-13 02:33:07.069 7f13ef545700 10 mon.a@0(leader).osd e105 start_mapping started mapping job 0x556c75da91e0 at 2019-02-13 02:33:07.070258
2019-02-13 02:33:07.069 7f13ef545700 20 mon.a@0(leader).paxos(paxos active c 1..479) finish_round waiting_for_readable
2019-02-13 02:33:07.069 7f13ef545700 20 mon.a@0(leader).paxos(paxos active c 1..479) finish_round waiting_for_writeable
2019-02-13 02:33:07.069 7f13ef545700 10 mon.a@0(leader).paxos(paxos active c 1..479) finish_round done w/ waiters, state active
2019-02-13 02:33:07.069 7f13ef545700 10 mon.a@0(leader).paxos(paxos active c 1..479) propose_pending 480 21476 bytes
2019-02-13 02:33:07.069 7f13ef545700 10 mon.a@0(leader).paxos(paxos updating c 1..479) begin for 480 21476 bytes

After this handle_advance_map there is none that shows the size increase.

2019-02-13 02:31:20.710 7f56306be700 10 osd.1 pg_epoch: 86 pg[3.0( empty local-lis/les=36/38 n=0 ec=33/33 lis/c 36/36 les/c/f 38/38/0 36/36/33) [1,2] r=0 lpr=36 crt=0'0 mlcod 0'0 active+clean] handle_advance_map [1]/[1] -- 1/1
Actions #3

Updated by David Zafman about 5 years ago

  • Related to Bug #38282: cephtool/test.sh failure in test_mon_osd_pool_set added
Actions #4

Updated by David Zafman about 5 years ago

  • Is duplicate of Bug #38282: cephtool/test.sh failure in test_mon_osd_pool_set added
Actions #5

Updated by David Zafman about 5 years ago

  • Related to deleted (Bug #38282: cephtool/test.sh failure in test_mon_osd_pool_set)
Actions #6

Updated by David Zafman about 5 years ago

  • Status changed from 12 to Duplicate
Actions

Also available in: Atom PDF