Activity
From 03/07/2021 to 04/05/2021
04/05/2021
- 09:14 PM Backport #50018: pacific: Test Failed with: "Scrubbing terminated -- not all pgs were active and ...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40461
merged - 08:53 PM Bug #21592: LibRadosCWriteOps.CmpExt got 0 instead of -4095-1
- /a/sage-2021-04-05_18:50:52-rados-master-distro-basic-smithi/6022643
- 07:47 PM Backport #49911 (Resolved): pacific: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get...
- https://github.com/ceph/ceph/pull/40289
- 07:46 PM Bug #49427: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_missing(soi...
- @Myoungwon Oh this new failure looks very similar to the issue tracked in this ticket?...
- 03:22 PM Bug #50140 (Duplicate): test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- 03:04 PM Bug #50140: test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- a duplicate of https://tracker.ceph.com/issues/49868
- 12:13 PM Bug #50140: test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- http://pulpito.front.sepia.ceph.com/rfriedma-2021-04-01_17:51:51-rados-wip-ronenf-cscrub-class-distro-basic-smithi/60...
- 12:12 PM Bug #50140: test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- Possible fixes to consider:
- a simple fix: extending the tests min-scrub-time;
- possibly better: modify the han... - 12:09 PM Bug #50140: test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- Caused by a combination of:
- re-scrub period ("osd scrub min interval") is set in radod/thrash* to (only) 60s.
- a... - 12:06 PM Bug #50140 (Duplicate): test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- Thrashing test:
The observed error is "Exiting scrub checking -- not all pgs scrubbed".
See below for analysis.
- 03:12 PM Backport #49682: nautilus: OSD: shutdown of a OSD Host causes slow requests
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40014
merged - 03:07 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- In the log I've checked (http://pulpito.front.sepia.ceph.com/rfriedma-2021-04-01_17:51:51-rados-wip-ronenf-cscrub-cla...
- 01:40 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
- ...
04/03/2021
- 10:10 PM Backport #50131 (Resolved): pacific: monmaptool --create --add nodeA --clobber monmap aborts in e...
- https://github.com/ceph/ceph/pull/40690
- 02:25 PM Backport #50130 (Resolved): nautilus: monmaptool --create --add nodeA --clobber monmap aborts in ...
- https://github.com/ceph/ceph/pull/40700
- 02:25 PM Backport #50129 (Resolved): octopus: monmaptool --create --add nodeA --clobber monmap aborts in e...
- https://github.com/ceph/ceph/pull/40758
- 02:20 PM Backport #50125 (Resolved): nautilus: mon: Modify Paxos trim logic to be more efficient
- https://github.com/ceph/ceph/pull/41099
- 02:20 PM Backport #50124 (Resolved): pacific: mon: Modify Paxos trim logic to be more efficient
- https://github.com/ceph/ceph/pull/40691
- 02:20 PM Backport #50123 (Resolved): octopus: mon: Modify Paxos trim logic to be more efficient
- https://github.com/ceph/ceph/pull/40699
- 02:20 PM Bug #48336 (Pending Backport): monmaptool --create --add nodeA --clobber monmap aborts in entity_...
- 02:15 PM Bug #50004 (Pending Backport): mon: Modify Paxos trim logic to be more efficient
- 01:40 PM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
- ...
- 01:37 PM Bug #50119 (Can't reproduce): Invalid read of size 4 in ceph::logging::Log::dump_recent()
- the test does the 2 writes, creates the digest, deletes the pool, and submits the final op, but that op hangs
/a/s...
04/02/2021
- 11:09 PM Bug #48417: unfound EC objects in sepia's LRC after upgrade
- I think this has the same root cause as https://tracker.ceph.com/issues/48613. Accepting IO below min_size I think c...
- 10:22 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- Found it....
- 09:56 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- ...
- 09:28 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- I don't think teuthology-2021-02-03_03:31:04-rados-pacific-distro-basic-smithi/5852263 is related to can_rollback_to ...
- 08:20 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
- /a/sage-2021-04-02_14:05:32-rados-wip-sage2-testing-2021-04-01-1417-pacific-distro-basic-smithi/6017345
- 10:11 AM Backport #49991 (In Progress): nautilus: unittest_mempool.check_shard_select failed
- 10:10 AM Backport #49992 (In Progress): pacific: unittest_mempool.check_shard_select failed
- 10:08 AM Backport #49993 (In Progress): octopus: unittest_mempool.check_shard_select failed
- 07:11 AM Bug #45441 (Resolved): rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in clust...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:10 AM Bug #46323 (Resolved): thrash_cache_writeback_proxy_none: FAILED ceph_assert(version == old_value...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:09 AM Bug #48065 (Resolved): "ceph osd crush set|reweight-subtree" commands do not set weight on device...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:09 AM Bug #48566 (Resolved): nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_abort: return 1
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:09 AM Bug #48745 (Resolved): Segmentation fault in PrimaryLogPG::cancel_manifest_ops
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:08 AM Bug #48821 (Resolved): osd crash in OSD::heartbeat when dereferencing null session
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:08 AM Bug #48884 (Resolved): ceph osd df tree reporting incorrect SIZE value for rack having an empty h...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:08 AM Bug #49212 (Resolved): mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound to class 'ssd...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:06 AM Bug #49788 (Resolved): pool application metadata not propagated to the cache tier
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:06 AM Bug #49867 (Resolved): osd/scrub - a ceph_assert() in a legitimate path
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:05 AM Bug #49975 (Resolved): "make check" jenkins job fails
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 06:11 AM Backport #49978 (Resolved): pacific: "make check" jenkins job fails
- 06:11 AM Backport #49895 (Resolved): pacific: osd/scrub - a ceph_assert() in a legitimate path
- 06:10 AM Backport #49156 (Resolved): pacific: Segmentation fault in PrimaryLogPG::cancel_manifest_ops
- 06:10 AM Backport #49058 (Resolved): pacific: thrash_cache_writeback_proxy_none: FAILED ceph_assert(versio...
- 06:02 AM Backport #48596 (Resolved): octopus: nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_abor...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40278
m... - 06:01 AM Backport #49009 (Resolved): octopus: osd crash in OSD::heartbeat when dereferencing null session
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40277
m... - 06:01 AM Backport #49527 (Resolved): octopus: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40276
m... - 06:01 AM Backport #49730 (Resolved): octopus: debian ceph-common package post-inst clobbers ownership of c...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40275
m... - 06:01 AM Backport #49795 (Resolved): octopus: pool application metadata not propagated to the cache tier
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40274
m... - 06:01 AM Backport #49681 (Resolved): octopus: OSD: shutdown of a OSD Host causes slow requests
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40013
m... - 06:00 AM Backport #48985: octopus: ceph osd df tree reporting incorrect SIZE value for rack having an empt...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39970
m... - 05:57 AM Backport #49402 (Resolved): octopus: rados: Health check failed: 1/3 mons down, quorum a,c (MON_D...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40138
m... - 05:53 AM Backport #49530 (Resolved): octopus: "ceph osd crush set|reweight-subtree" commands do not set we...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39919
m...
04/01/2021
- 09:40 PM Backport #48985 (Resolved): octopus: ceph osd df tree reporting incorrect SIZE value for rack hav...
- 03:45 PM Backport #48985: octopus: ceph osd df tree reporting incorrect SIZE value for rack having an empt...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39970
merged - 09:08 PM Backport #49917 (In Progress): octopus: mon: slow ops due to osd_failure
- 08:40 PM Backport #49681 (In Progress): octopus: OSD: shutdown of a OSD Host causes slow requests
- 03:46 PM Backport #49681: octopus: OSD: shutdown of a OSD Host causes slow requests
- Mauricio Oliveira wrote:
> please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/... - 08:38 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- I'm pulling the logs to look more closely.
- 08:28 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- Looking at this more closely, I think removing that if block is still fine -- it'll be trimmed by the next IO anyway....
- 07:32 PM Bug #50106 (Can't reproduce): scrub/osd-scrub-repair.sh: corrupt_scrub_erasure: return 1
- ...
- 05:30 PM Bug #50101 (New): unhandled event in ReplicaActive
- ...
- 04:13 PM Bug #38219: rebuild-mondb hangs
- ...
- 03:50 PM Backport #48596: octopus: nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_abort: return 1
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40278
merged - 03:50 PM Backport #49009: octopus: osd crash in OSD::heartbeat when dereferencing null session
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40277
merged - 03:50 PM Bug #50100 (New): stale slow osd heartbeats health alert
- ...
- 03:49 PM Backport #49527: octopus: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound to class '...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40276
merged - 03:48 PM Backport #49730: octopus: debian ceph-common package post-inst clobbers ownership of cephadm log ...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40275
merged - 03:47 PM Backport #49795: octopus: pool application metadata not propagated to the cache tier
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40274
merged - 06:53 AM Bug #48336 (In Progress): monmaptool --create --add nodeA --clobber monmap aborts in entity_addr_...
- 12:32 AM Bug #50089 (Resolved): mon/MonMap.h: FAILED ceph_assert(m < ranks.size()) when reducing number of...
- ...
- 12:02 AM Bug #50088 (New): rados: qa: suites do not test mon removal
- We have the monthrash suite, but it merely turns monitors on and off again. It should probably also remove monitors f...
03/31/2021
- 04:10 PM Backport #50087 (Resolved): pacific: test_mon_pg: mon fails to join quorum to due election strate...
- 04:07 PM Bug #47654 (Pending Backport): test_mon_pg: mon fails to join quorum to due election strategy mis...
- Merged to master.
Pacific PR in progress at https://github.com/ceph/ceph/pull/40484 - 04:03 AM Bug #50042: rados/test.sh: api_watch_notify failures
- Looking at /a/teuthology-2021-03-27_14:59:29-upgrade:nautilus-x-pacific-distro-basic-smithi/6004568...
- 12:12 AM Bug #50042: rados/test.sh: api_watch_notify failures
- Looking at /a/sage-2021-03-28_19:04:26-rados-wip-sage2-testing-2021-03-28-0933-pacific-distro-basic-smithi/6007274
...
03/30/2021
- 11:32 PM Bug #50042: rados/test.sh: api_watch_notify failures
- Brad, as per our discussion, I am assigning this to you for now.
- 05:27 PM Bug #47654 (Fix Under Review): test_mon_pg: mon fails to join quorum to due election strategy mis...
03/29/2021
- 09:41 PM Bug #49988 (Fix Under Review): Global Recovery Event never completes
- Problem was that I did not subtract pgs that I skip because (reported_epoch_of_pg < start_epoch_of_event) from total_...
- 06:52 PM Bug #50042 (Resolved): rados/test.sh: api_watch_notify failures
- ...
- 09:01 AM Feature #48392: ceph ignores --keyring?
- I see this with v15.2.10 as well.
The problem is at the rollback stage, specifically with @osd purge-new@ where it... - 07:40 AM Bug #50032: used space in ceph df reported incorrectly for filestore OSDs
- Easily reproducible in Nautilus head indeed.
- 05:36 AM Bug #50032 (New): used space in ceph df reported incorrectly for filestore OSDs
- Hi, I noticed, that when filestore OSDs are used, used space is reported incorrectly:
filestore:
POOL ... - 06:23 AM Backport #50018 (In Progress): pacific: Test Failed with: "Scrubbing terminated -- not all pgs we...
- 04:50 AM Bug #50031 (New): osdc _throttle_op function param type of op_budget int is too small
- h2. problem
1. function calc_op_budget return type is int,when indata.length() is out of int range, it will return...
03/26/2021
- 09:59 PM Bug #49962: 'sudo ceph --cluster ceph osd crush tunables default' fails due to valgrind: Unknown ...
- Radek, can you please help take a look this issue?
- 09:53 PM Bug #50004 (Fix Under Review): mon: Modify Paxos trim logic to be more efficient
- 09:36 AM Bug #50004 (Resolved): mon: Modify Paxos trim logic to be more efficient
- When the log ingest rate is higher than the trimming rate, monitors may end up consuming a lot of space and become un...
- 08:41 PM Bug #49963 (Resolved): Crash in OSD::ms_fast_dispatch due to call to null vtable function
- 06:30 PM Backport #50018 (Resolved): pacific: Test Failed with: "Scrubbing terminated -- not all pgs were ...
- https://github.com/ceph/ceph/pull/40461
- 06:25 PM Bug #49983 (Pending Backport): Test Failed with: "Scrubbing terminated -- not all pgs were active...
- 05:53 PM Backport #49402: octopus: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40138
merged - 05:51 PM Backport #49530: octopus: "ceph osd crush set|reweight-subtree" commands do not set weight on dev...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39919
merged - 04:20 PM Backport #49640 (In Progress): nautilus: Disable and re-enable clog_to_monitors could trigger ass...
- 03:04 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- /a/yuriw-2021-03-25_20:03:40-rados-wip-yuri8-testing-2021-03-25-1042-pacific-distro-basic-smithi/5999016
- 03:03 PM Bug #45702: PGLog::read_log_and_missing: ceph_assert(miter == missing.get_items().end() || (miter...
- /a/yuriw-2021-03-25_20:03:40-rados-wip-yuri8-testing-2021-03-25-1042-pacific-distro-basic-smithi/5999051
- 12:11 PM Bug #50012 (Fix Under Review): Ceph-osd refuses to bind on an IP on the local loopback lo (again)
- the original fix was reverted by the fix of #49938. so fix in a more flexible way.
- 12:10 PM Bug #50012 (Fix Under Review): Ceph-osd refuses to bind on an IP on the local loopback lo (again)
- 09:58 AM Backport #49977 (Resolved): nautilus: "make check" jenkins job fails
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40407
m... - 09:54 AM Backport #49529 (Resolved): nautilus: "ceph osd crush set|reweight-subtree" commands do not set w...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39920
m... - 09:51 AM Backport #49976 (Resolved): octopus: "make check" jenkins job fails
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40406
m... - 07:41 AM Bug #38219: rebuild-mondb hangs
- ...
- 06:40 AM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- /kchai-2021-03-26_05:32:58-rados-wip-kefu-testing-2021-03-26-1134-distro-basic-smithi/6001105/
- 06:20 AM Backport #49993 (Resolved): octopus: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/39978
- 06:20 AM Backport #49992 (Resolved): pacific: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/40566
- 06:20 AM Backport #49991 (Resolved): nautilus: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/40567
- 06:17 AM Bug #49781 (Pending Backport): unittest_mempool.check_shard_select failed
03/25/2021
- 11:20 PM Bug #49988 (Resolved): Global Recovery Event never completes
- ...
- 10:12 PM Support #49847: OSD Fails to init after upgrading to octopus: _deferred_replay failed to decode d...
- Contrary to what I stated previously this does not seem like a software issue. The root cause was probably faulty RAM...
- 07:26 PM Bug #42884: OSDMapTest.CleanPGUpmaps failure
- https://jenkins.ceph.com/job/ceph-pull-requests/72185/consoleFull#-108728127277933967-90d1-4877-8d60-89cb08ef4eb1
- 04:44 PM Bug #49983 (Fix Under Review): Test Failed with: "Scrubbing terminated -- not all pgs were active...
- 02:42 PM Bug #49983 (Resolved): Test Failed with: "Scrubbing terminated -- not all pgs were active and cle...
- Test Run:
https://pulpito.ceph.com/nojha-2021-03-23_23:04:33-rados-wip-40323-2-distro-basic-gibba/5991116/
Failur... - 11:50 AM Backport #49977 (In Progress): nautilus: "make check" jenkins job fails
- 11:25 AM Backport #49977 (Resolved): nautilus: "make check" jenkins job fails
- https://github.com/ceph/ceph/pull/40407
- 11:45 AM Backport #49976 (In Progress): octopus: "make check" jenkins job fails
- 11:25 AM Backport #49976 (Resolved): octopus: "make check" jenkins job fails
- https://github.com/ceph/ceph/pull/40406
- 11:44 AM Backport #49978 (In Progress): pacific: "make check" jenkins job fails
- 11:25 AM Backport #49978 (Resolved): pacific: "make check" jenkins job fails
- https://github.com/ceph/ceph/pull/40405
- 11:22 AM Bug #49975 (Resolved): "make check" jenkins job fails
- ...
- 09:58 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Brad Hubbard wrote:
> I think I've found an issue related to this message in the logs but I'll need to test that the... - 05:32 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- I think I've found an issue related to this message in the logs but I'll need to test that theory....
- 04:20 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Hi Christian,
Is there no message in syslog or dmesg about the segfault at all? That seems odd. Maybe you should c... - 06:58 AM Bug #49970 (New): cmake: use GooleTest for adding gtest based tests
- see https://github.com/Kitware/CMake/blob/master/Modules/GoogleTest.cmake
for finer grained management of test, - 04:55 AM Bug #49963 (Fix Under Review): Crash in OSD::ms_fast_dispatch due to call to null vtable function
03/24/2021
- 11:43 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
- So the issue is we never should have gone this deep into the OSD::ms_fast_dispatch function.
Previously we would h... - 10:53 PM Bug #49963 (Triaged): Crash in OSD::ms_fast_dispatch due to call to null vtable function
- 10:50 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
- /a/sage-2021-03-24_06:13:24-upgrade:octopus-x-wip-sage-testing-2021-03-23-2309-distro-basic-smithi/5993446...
- 10:46 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
- m->header.type is 85...
- 10:38 PM Bug #49963 (Resolved): Crash in OSD::ms_fast_dispatch due to call to null vtable function
- /a/sage-2021-03-24_06:13:24-upgrade:octopus-x-wip-sage-testing-2021-03-23-2309-distro-basic-smithi/5993446...
- 09:09 PM Bug #49962 (Resolved): 'sudo ceph --cluster ceph osd crush tunables default' fails due to valgrin...
- ...
- 07:25 PM Bug #49961 (New): scrub/osd-recovery-scrub.sh: TEST_recovery_scrub_1 failed
- ...
- 12:21 PM Bug #49487 (Resolved): osd:scrub skip some pg
- 09:12 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Brad Hubbard wrote:
> Hi Christian,
>
> From the log this was a segfault in MonitorDBStore::get_synchronizer.
> ... - 02:51 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- So the offset into MonitorDBStore::get_synchronizer where we segfaulted was
0x3f.
If we set up a binary compatibl... - 02:44 AM Documentation #35967 (Resolved): [doc] sync documentation "OSD Config Reference" default values w...
03/23/2021
- 11:25 PM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Hi Christian,
From the log this was a segfault in MonitorDBStore::get_synchronizer.
2021-03-14 18:00:28.679 7f1... - 08:45 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Neha Ojha wrote:
> Do you happen to have a coredump for this or can you provide a copy of the mon store db?
Unfor... - 07:12 PM Bug #49781: unittest_mempool.check_shard_select failed
- https://jenkins.ceph.com/job/ceph-pull-requests/72011/consoleFull#-85362248744e9240e-b50a-4693-bac0-8a991bac86ac
- 04:55 PM Backport #49529: nautilus: "ceph osd crush set|reweight-subtree" commands do not set weight on de...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39920
merged - 12:23 AM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
- ...
03/22/2021
- 10:21 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- relevant osd.3 logs from yuriw-2021-03-19_00:00:55-rados-wip-yuri8-testing-2021-03-18-1502-pacific-distro-basic-smith...
- 09:40 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- /a/yuriw-2021-03-19_00:00:55-rados-wip-yuri8-testing-2021-03-18-1502-pacific-distro-basic-smithi/5978982
- 02:47 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Thanks, Konstantin!
- 02:46 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- @Mauricio, I was update issue backports and status.
- 02:45 PM Bug #46978 (Pending Backport): OSD: shutdown of a OSD Host causes slow requests
- 01:44 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Hi @singuliere _,
Could you please revert the backport field to include Octopus and Nautilus?
Such backports ha... - 07:40 AM Feature #49789 (Fix Under Review): common/TrackedOp: add op priority for TrackedOp
- 07:15 AM Backport #49919 (Resolved): nautilus: mon: slow ops due to osd_failure
- https://github.com/ceph/ceph/pull/41213
- 07:15 AM Backport #49918 (Resolved): pacific: mon: slow ops due to osd_failure
- https://github.com/ceph/ceph/pull/41090
- 07:15 AM Backport #49917 (Resolved): octopus: mon: slow ops due to osd_failure
- https://github.com/ceph/ceph/pull/40558
- 07:13 AM Bug #47380 (Pending Backport): mon: slow ops due to osd_failure
03/21/2021
- 06:25 PM Backport #49642 (Resolved): pacific: Disable and re-enable clog_to_monitors could trigger assertion
- 06:24 PM Backport #49008 (Resolved): pacific: osd crash in OSD::heartbeat when dereferencing null session
- 06:22 PM Backport #49728 (Resolved): pacific: debian ceph-common package post-inst clobbers ownership of c...
- 05:46 PM Bug #49883 (Rejected): librados: hang in RadosClient::wait_for_osdmap
- This turned out to be because the auth credential had wrong permissions (due to a profile name change).
- 02:33 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
- /a/sage-2021-03-20_15:11:51-rados-wip-sage2-testing-2021-03-20-0832-pacific-distro-basic-smithi/5983920
- 01:20 PM Backport #49911 (Resolved): pacific: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get...
- 01:18 PM Bug #49427 (Pending Backport): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missi...
- need this in pacific too: /a/sage-2021-03-20_15:11:51-rados-wip-sage2-testing-2021-03-20-0832-pacific-distro-basic-sm...
03/20/2021
- 01:45 PM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
- /a/sage-2021-03-20_01:11:45-rados-wip-sage-testing-2021-03-19-1647-distro-basic-smithi/5982282
with logs! - 08:06 AM Backport #48596 (In Progress): octopus: nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_a...
- 08:05 AM Backport #49009 (In Progress): octopus: osd crash in OSD::heartbeat when dereferencing null session
- 08:04 AM Backport #49527 (In Progress): octopus: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bo...
- 07:56 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Since this issue is resolved and only the pacific backport was done, I assume it means the octopus & nautilus backpor...
- 07:46 AM Backport #49730 (In Progress): octopus: debian ceph-common package post-inst clobbers ownership o...
- 07:46 AM Backport #49795 (In Progress): octopus: pool application metadata not propagated to the cache tier
03/19/2021
- 09:47 PM Bug #49809 (Need More Info): 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Do you happen to have a coredump for this or can you provide a copy of the mon store db?
- 09:13 PM Backport #49895: pacific: osd/scrub - a ceph_assert() in a legitimate path
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40226
merged - 06:08 PM Backport #49796 (Resolved): pacific: pool application metadata not propagated to the cache tier
- 06:02 PM Backport #49728 (In Progress): pacific: debian ceph-common package post-inst clobbers ownership o...
- 05:58 PM Backport #49642 (In Progress): pacific: Disable and re-enable clog_to_monitors could trigger asse...
- 05:57 PM Backport #49008 (In Progress): pacific: osd crash in OSD::heartbeat when dereferencing null session
- 05:06 PM Bug #49902 (New): monitor store.db directory has been increasing and has exceeded 450GB
- My monitor keeps growing. After it exceeds 300GB, it cannot be started because the disk is not enough. After the spac...
- 11:07 AM Backport #49320 (Resolved): octopus: thrash_cache_writeback_proxy_none: FAILED ceph_assert(versio...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39578
m... - 09:50 AM Backport #49641 (Resolved): octopus: Disable and re-enable clog_to_monitors could trigger assertion
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39935
m... - 09:48 AM Backport #49533 (Resolved): octopus: osd ok-to-stop too conservative
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39887
m...
03/18/2021
- 10:42 PM Bug #49894: set a non-zero default value for osd_client_message_cap
- Neha Ojha wrote:
> The current default of 0 doesn't help and we've tried setting it to 5000 for one of our tests htt... - 09:26 PM Bug #49894 (Resolved): set a non-zero default value for osd_client_message_cap
- The current default of 0 doesn't help and we've tried setting it to 5000 for one of our tests https://github.com/neha...
- 09:57 PM Backport #49895 (In Progress): pacific: osd/scrub - a ceph_assert() in a legitimate path
- 09:55 PM Backport #49895 (Resolved): pacific: osd/scrub - a ceph_assert() in a legitimate path
- https://github.com/ceph/ceph/pull/40226
- 09:54 PM Bug #49867 (Pending Backport): osd/scrub - a ceph_assert() in a legitimate path
- 05:57 PM Bug #49888 (Resolved): rados/singleton: radosbench.py: teuthology.exceptions.MaxWhileTries: reach...
- ...
- 04:49 PM Support #49847: OSD Fails to init after upgrading to octopus: _deferred_replay failed to decode d...
- Ended up nuking the OSD & letting it recover - this workaround "solves" the problem for me; feel free to close the is...
- 04:30 PM Backport #49641: octopus: Disable and re-enable clog_to_monitors could trigger assertion
- gerald yang wrote:
> https://github.com/ceph/ceph/pull/39935
merged - 04:27 PM Backport #49533: octopus: osd ok-to-stop too conservative
- https://github.com/ceph/ceph/pull/39887 merged
- 04:24 PM Bug #46323: thrash_cache_writeback_proxy_none: FAILED ceph_assert(version == old_value.version) i...
- https://github.com/ceph/ceph/pull/39578 merged
- 03:03 PM Bug #49883: librados: hang in RadosClient::wait_for_osdmap
- The ubuntu jobs are failing because the ceph-mgr can't find libcephsqlite.so. There were packaging changes that proba...
- 02:34 PM Bug #49883 (Rejected): librados: hang in RadosClient::wait_for_osdmap
- https://pulpito.ceph.com/pdonnell-2021-03-18_06:53:09-rados:basic-wip-pdonnell-testing-20210318.035227-distro-basic-s...
- 01:54 PM Backport #49401: pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40137
m... - 01:54 PM Backport #49401: pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40137
m... - 01:54 PM Backport #49817 (Resolved): pacific: mon: promote_standby does not update available_modules
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40132
m...
03/17/2021
- 09:19 PM Backport #49401 (Resolved): pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_D...
- 06:24 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- I think we should revert this in pacific https://github.com/ceph/ceph/pull/40195, until we can fix the test failures.
- 04:02 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- Something merged after 37f9d0a25d06a6b8529aa350110eba930fba8c9e since https://pulpito.ceph.com/yuriw-2021-03-15_23:42...
- 03:57 PM Bug #49868 (New): RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- ...
- 03:28 PM Bug #49867 (Fix Under Review): osd/scrub - a ceph_assert() in a legitimate path
- 03:15 PM Bug #49867 (Resolved): osd/scrub - a ceph_assert() in a legitimate path
- in pg_scrubber.cc, in PgScrubber::on_replica_init() there's a ceph_assert()
that might trigger in a correct path (if... - 10:14 AM Backport #49683: pacific: OSD: shutdown of a OSD Host causes slow requests
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39957
m... - 09:05 AM Bug #49781: unittest_mempool.check_shard_select failed
- Discussion on the mailing list about "mempool optimization":https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/...
03/16/2021
- 08:07 PM Support #49847 (Closed): OSD Fails to init after upgrading to octopus: _deferred_replay failed to...
- An OSD fails to start after upgrading from mimic 13.2.2 to octopus 15.2.9.
It seems like first bluestore fails at... - 03:45 PM Bug #49832 (New): Segmentation fault: in thread_name:ms_dispatch
- ...
- 03:22 PM Bug #49781: unittest_mempool.check_shard_select failed
- The test condition should not be too strict because there really is no way to predict the result. It is however good ...
- 12:56 PM Bug #49781: unittest_mempool.check_shard_select failed
- Using "pthread_self for sharding":https://github.com/ceph/ceph/blob/master/src/include/mempool.h#L261-L262 is not gre...
- 11:25 AM Bug #49781 (In Progress): unittest_mempool.check_shard_select failed
- 08:15 AM Bug #49697: prime pg temp: unexpected optimization
- ping
- 08:14 AM Bug #49787 (Resolved): test_envlibrados_for_rocksdb.sh fails on master
- 06:28 AM Backport #49682 (In Progress): nautilus: OSD: shutdown of a OSD Host causes slow requests
03/15/2021
- 10:42 PM Bug #46978 (Resolved): OSD: shutdown of a OSD Host causes slow requests
- 10:42 PM Backport #49683 (Resolved): pacific: OSD: shutdown of a OSD Host causes slow requests
- 10:41 PM Backport #49774 (Resolved): pacific: Get more parallel scrubs within osd_max_scrubs limits
- 09:56 PM Backport #49402 (In Progress): octopus: rados: Health check failed: 1/3 mons down, quorum a,c (MO...
- 09:55 PM Backport #49401 (In Progress): pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MO...
- 08:15 PM Backport #49817 (Resolved): pacific: mon: promote_standby does not update available_modules
- https://github.com/ceph/ceph/pull/40132
- 08:15 PM Backport #49816 (Resolved): octopus: mon: promote_standby does not update available_modules
- https://github.com/ceph/ceph/pull/40757
- 08:11 PM Bug #49778 (Pending Backport): mon: promote_standby does not update available_modules
- 05:26 PM Bug #49810 (Need More Info): rados/singleton: with msgr-failures/none MON_DOWN due to haven't for...
- ...
- 05:16 PM Bug #49809 (Need More Info): 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- We experienced a single mon crash (out of 3 mons) - We observed no other issues on the machine or the cluster.
I a... - 03:02 PM Bug #48793 (Resolved): out of order op
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 03:02 PM Bug #48990 (Resolved): rados/dashboard: Health check failed: Telemetry requires re-opt-in (TELEME...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 10:38 AM Bug #49781: unittest_mempool.check_shard_select failed
- master also...
- 09:38 AM Bug #49779 (Resolved): standalone: osd-recovery-scrub.sh: Recovery never started
- 09:22 AM Bug #49758 (Resolved): messages/MOSDPGNotify.h: virtual void MOSDPGNotify::encode_payload(uint64_...
- 09:10 AM Backport #49796 (Resolved): pacific: pool application metadata not propagated to the cache tier
- https://github.com/ceph/ceph/pull/40119
- 09:10 AM Backport #49795 (Resolved): octopus: pool application metadata not propagated to the cache tier
- https://github.com/ceph/ceph/pull/40274
- 09:09 AM Bug #49788 (Pending Backport): pool application metadata not propagated to the cache tier
- 01:39 AM Bug #49696: all mons crash suddenly and cann't restart unless close cephx
- Neha Ojha wrote:
> can you share a coredump from the monitor, if the issue is still reproducible?
I'm afraid not....
03/14/2021
- 11:52 AM Bug #49781: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/39978#discussion_r593341155
- 06:14 AM Feature #49789: common/TrackedOp: add op priority for TrackedOp
- PR:https://github.com/ceph/ceph/pull/40060
- 06:12 AM Feature #49789 (Fix Under Review): common/TrackedOp: add op priority for TrackedOp
- Now, we can not know a request priority by ceph daemon /var/run/ceph/ceph-osd.x.asok dump_historic_ops
if this comma... - 04:17 AM Bug #49779 (Fix Under Review): standalone: osd-recovery-scrub.sh: Recovery never started
03/13/2021
- 04:35 PM Bug #49788 (Fix Under Review): pool application metadata not propagated to the cache tier
- 04:27 PM Bug #49788 (Resolved): pool application metadata not propagated to the cache tier
- if you have a base pool with application metadata, that application is not propagated to the cache tier.
This is a... - 09:03 AM Bug #49787 (Resolved): test_envlibrados_for_rocksdb.sh fails on master
- ...
- 08:27 AM Bug #49781: unittest_mempool.check_shard_select failed
- It happened 5 days ago at https://github.com/ceph/ceph/pull/39883#issuecomment-791944956 and is related to https://gi...
- 03:33 AM Bug #49781 (Resolved): unittest_mempool.check_shard_select failed
- This test is probabilistic. Recording to see whether we find it failing more frequently.
From https://jenkins.ceph...
03/12/2021
- 09:36 PM Bug #49696 (Need More Info): all mons crash suddenly and cann't restart unless close cephx
- can you share a coredump from the monitor, if the issue is still reproducible?
- 09:31 PM Bug #49734 (Closed): [OSD]ceph osd crashes and prints Segmentation fault
- Luminous is EOL, please re-open if you see the same issue in later releases.
- 09:00 PM Backport #49775 (In Progress): nautilus: Get more parallel scrubs within osd_max_scrubs limits
- 06:20 PM Backport #49775 (Rejected): nautilus: Get more parallel scrubs within osd_max_scrubs limits
- https://github.com/ceph/ceph/pull/40142
- 08:58 PM Bug #49779 (Resolved): standalone: osd-recovery-scrub.sh: Recovery never started
In master and pacific, the TEST_recovery_scrub_2 subtest in qa/standalone/scrub/osd-recovery-scrub.sh has an interm...- 08:55 PM Backport #49776 (In Progress): octopus: Get more parallel scrubs within osd_max_scrubs limits
- 06:20 PM Backport #49776 (Rejected): octopus: Get more parallel scrubs within osd_max_scrubs limits
- https://github.com/ceph/ceph/pull/40088
- 08:52 PM Backport #49774 (In Progress): pacific: Get more parallel scrubs within osd_max_scrubs limits
- 06:20 PM Backport #49774 (Resolved): pacific: Get more parallel scrubs within osd_max_scrubs limits
- https://github.com/ceph/ceph/pull/40077
- 08:03 PM Bug #49778: mon: promote_standby does not update available_modules
- I think we probably also need a workaround so that we can upgrade from old ceph versions that have this bug...
- 08:00 PM Bug #49778 (Resolved): mon: promote_standby does not update available_modules
- originally observed during upgrade from <15.2.5 via cephadm: the cephadm migration runs immediately after upgrade and...
- 07:46 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
- ...
- 06:53 PM Bug #44595: cache tiering: Error: oid 48 copy_from 493 returned error code -2
- ...
- 06:29 PM Bug #49777 (Resolved): test_pool_min_size: 'check for active or peered' reached maximum tries (5)...
- ...
- 06:20 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
- ...
- 06:19 PM Bug #48843 (Pending Backport): Get more parallel scrubs within osd_max_scrubs limits
- 05:12 PM Bug #47181: "sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120...
- /a/yuriw-2021-03-11_19:01:40-rados-octopus-distro-basic-smithi/5956578/
- 01:59 PM Bug #48959: Primary OSD crash caused corrupted object and further crashes during backfill after s...
- We just ran into this again and had to remove the object to allow the PG to finish backfilling. The similarities betw...
- 01:38 PM Bug #49409: osd run into dead loop and tell slow request when rollback snap with using cache tier
- reopening this ticket, as its fix (https://github.com/ceph/ceph/pull/39593) was reverted as the fix of #49726
- 01:38 PM Bug #49409 (New): osd run into dead loop and tell slow request when rollback snap with using cach...
- 01:37 PM Bug #49726 (Resolved): src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_versio...
- 07:29 AM Bug #49726: src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_version64() == ve...
- created https://github.com/ceph/ceph/pull/40057 as an intermediate fix.
- 12:27 PM Bug #49427 (Resolved): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_...
- 11:50 AM Bug #48505: osdmaptool crush
- hanguang liu wrote:
> when osd map contains CRUSH_ITEM_NONE osd when i run:
> _./osdmaptool ./hkc4 --test-map-pgs-... - 11:44 AM Bug #48505: osdmaptool crush
- hanguang liu wrote:
> when osd map contains CRUSH_ITEM_NONE osd when i run:
> _./osdmaptool ./hkc4 --test-map-pgs-... - 07:26 AM Bug #49758 (Fix Under Review): messages/MOSDPGNotify.h: virtual void MOSDPGNotify::encode_payload...
- 05:37 AM Bug #49754: osd/OSD.cc: ceph_abort_msg("abort() called") during OSD::shutdown()
- ...
03/11/2021
- 11:03 PM Bug #49726: src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_version64() == ve...
- https://github.com/ceph/ceph/pull/39593#issuecomment-792503213 this is where it first showed up, most likely this PR ...
- 02:03 AM Bug #49726: src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_version64() == ve...
- /a/kchai-2021-03-09_12:22:01-rados-wip-kefu-testing-2021-03-09-1847-distro-basic-smithi/5949457
/a/ideepika-2021-03-... - 01:56 AM Bug #49726 (Resolved): src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_versio...
- ...
- 08:19 PM Backport #49054 (Resolved): pacific: pick_a_shard() always select shard 0
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39977
m... - 06:40 PM Backport #49054: pacific: pick_a_shard() always select shard 0
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39977
merged - 08:17 PM Backport #49670: pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo be...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39963
m... - 08:11 PM Backport #49565: pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39844
m... - 08:08 PM Backport #49397 (Resolved): octopus: rados/dashboard: Health check failed: Telemetry requires re-...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39704
m... - 03:59 PM Backport #49397: octopus: rados/dashboard: Health check failed: Telemetry requires re-opt-in (TEL...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39704
merged - 06:56 PM Bug #49758 (Resolved): messages/MOSDPGNotify.h: virtual void MOSDPGNotify::encode_payload(uint64_...
- ...
- 06:45 PM Bug #49754 (New): osd/OSD.cc: ceph_abort_msg("abort() called") during OSD::shutdown()
- ...
- 06:04 PM Bug #47838: mon/test_mon_osdmap_prune.sh: first_pinned != trim_to
- /a/yuriw-2021-03-10_21:08:51-rados-wip-yuri8-testing-2021-03-10-0901-pacific-distro-basic-smithi/5954442 - similar
- 01:31 PM Bug #47380: mon: slow ops due to osd_failure
- an alternative fix: https://github.com/ceph/ceph/pull/40033
- 07:11 AM Bug #49734 (Closed): [OSD]ceph osd crashes and prints Segmentation fault
- This error occurs in Mar 6th, the osd.37 was down and out with bellow log info(ceph-osd.37.log-20210306):
2021-03-... - 07:07 AM Backport #49533 (In Progress): octopus: osd ok-to-stop too conservative
- 03:30 AM Backport #49730 (Resolved): octopus: debian ceph-common package post-inst clobbers ownership of c...
- https://github.com/ceph/ceph/pull/40275
- 03:30 AM Bug #49727: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
Note that instead of a delay you can tell the OSDs to flush their pg stats. I wonder if that flushes to the mon and...- 03:16 AM Bug #49727 (Resolved): lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
This has been seen in cases where all of pool 1 PGs are scrubbed and none of pool 2's. I suggest that this is beca...- 03:30 AM Backport #49729 (Resolved): nautilus: debian ceph-common package post-inst clobbers ownership of ...
- https://github.com/ceph/ceph/pull/40698
- 03:30 AM Backport #49728 (Resolved): pacific: debian ceph-common package post-inst clobbers ownership of c...
- https://github.com/ceph/ceph/pull/40248
- 03:26 AM Backport #49145 (Resolved): pacific: out of order op
- 03:25 AM Bug #49677 (Pending Backport): debian ceph-common package post-inst clobbers ownership of cephadm...
03/10/2021
- 10:41 PM Backport #49682: nautilus: OSD: shutdown of a OSD Host causes slow requests
- please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/40014
ceph-backport.sh versi... - 10:40 PM Backport #49681: octopus: OSD: shutdown of a OSD Host causes slow requests
- please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/40013
ceph-backport.sh versi... - 04:21 PM Bug #23117: PGs stuck in "activating" after osd_max_pg_per_osd_hard_ratio has been exceeded once
- I am aware of one place where we do log withholding pg creation, the following log message in the OSD logs.
https://... - 01:08 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Hey Konstantin and Loïc,
Understood; thanks! - 07:57 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Hi Mauricio,
You are welcome to join the Stable Release team on IRC at #ceph-backports to discuss and resolve the... - 06:47 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Mauricio, just make a backport PR at GitHub, we'll attach it to tracker later.
- 08:54 AM Bug #49697 (Resolved): prime pg temp: unexpected optimization
- I encountered a problem when splitting pgs that eventually cause pg
to be inactived.
I probably think the root reas... - 07:40 AM Bug #49696 (Need More Info): all mons crash suddenly and cann't restart unless close cephx
- crash info
{
"os_version_id": "7",
"utsname_release": "4.14.0jsdx_kernel",
"os_name": "CentOS Linux... - 02:13 AM Backport #49533 (Rejected): octopus: osd ok-to-stop too conservative
- Per Sage
> I'm not sure if this is worth backporting. The primary benefit is faster upgrades, and it's the target ... - 01:24 AM Bug #47419 (Resolved): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo bench...
- 01:24 AM Backport #49670 (Resolved): pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rado...
- 12:02 AM Backport #49565 (Resolved): pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
03/09/2021
- 11:58 PM Backport #49053 (In Progress): octopus: pick_a_shard() always select shard 0
- 11:58 PM Bug #47719: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- https://github.com/ceph/ceph/pull/39844 merged
- 11:57 PM Backport #49054 (In Progress): pacific: pick_a_shard() always select shard 0
- 11:13 PM Backport #49691 (Rejected): pacific: ceph_assert(is_primary()) in PG::scrub()
- 11:10 PM Backport #49691 (Rejected): pacific: ceph_assert(is_primary()) in PG::scrub()
- 11:13 PM Bug #48712 (Resolved): ceph_assert(is_primary()) in PG::scrub()
- 11:09 PM Bug #48712 (Pending Backport): ceph_assert(is_primary()) in PG::scrub()
- 11:09 PM Bug #48712 (Resolved): ceph_assert(is_primary()) in PG::scrub()
- 11:12 PM Backport #49377 (In Progress): pacific: building libcrc32
- 10:55 PM Backport #48985 (In Progress): octopus: ceph osd df tree reporting incorrect SIZE value for rack ...
- 10:26 PM Bug #49689 (Resolved): osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch...
- ...
- 10:23 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- /a/yuriw-2021-03-08_21:03:18-rados-wip-yuri5-testing-2021-03-08-1049-pacific-distro-basic-smithi/5947439
- 10:21 PM Bug #49688 (Can't reproduce): FAILED ceph_assert(is_primary()) in submit_log_entries during Promo...
- ...
- 09:43 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- Samuel Just wrote:
> I'm...not sure what that if block is supposed to do. It was introduced as part of the initial ... - 03:21 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- I'm...not sure what that if block is supposed to do. It was introduced as part of the initial overwrites patch seque...
- 09:31 PM Backport #49670 (In Progress): pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 r...
- https://github.com/ceph/ceph/pull/39963
- 03:45 PM Backport #49670 (Resolved): pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rado...
- https://github.com/ceph/ceph/pull/39963
- 07:46 PM Backport #49683: pacific: OSD: shutdown of a OSD Host causes slow requests
- please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/39957
ceph-backport.sh versi... - 07:35 PM Backport #49683 (Resolved): pacific: OSD: shutdown of a OSD Host causes slow requests
- https://github.com/ceph/ceph/pull/39957
- 07:40 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Igor, thanks.
I'd like to / can work on submitting the backport PRs, if that's OK.
In the future, if I want to ... - 07:33 PM Bug #46978 (Pending Backport): OSD: shutdown of a OSD Host causes slow requests
- 07:25 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- The master PR has been merged.
Can someone update Status to Pending Backport, please?
Thanks! - 07:35 PM Backport #49682 (Resolved): nautilus: OSD: shutdown of a OSD Host causes slow requests
- https://github.com/ceph/ceph/pull/40014
- 07:35 PM Backport #49681 (Resolved): octopus: OSD: shutdown of a OSD Host causes slow requests
- https://github.com/ceph/ceph/pull/40013
- 05:57 PM Bug #49677 (Fix Under Review): debian ceph-common package post-inst clobbers ownership of cephadm...
- 05:54 PM Bug #49677 (Resolved): debian ceph-common package post-inst clobbers ownership of cephadm log dirs
- the debian/ubuntu ceph uid is different than the rhel/centos one used by the container. the postinst does a chown -R...
- 04:45 PM Backport #47364 (Resolved): luminous: pgs inconsistent, union_shard_errors=missing
- 03:43 PM Bug #47419 (Pending Backport): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p f...
- https://jenkins.ceph.com/job/ceph-pull-requests/70801/consoleFull#10356408840526d21-3511-427d-909c-dd086c0d1034 - thi...
- 08:32 AM Bug #48786 (Resolved): api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcount2...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 08:32 AM Bug #48984 (Resolved): lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 06:30 AM Backport #49642: pacific: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39938
- 04:11 AM Backport #49641: octopus: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39935
03/08/2021
- 05:16 PM Backport #49482: pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcou...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39773
m... - 05:14 PM Backport #49532: pacific: osd ok-to-stop too conservative
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39737
m... - 05:07 PM Backport #49529 (In Progress): nautilus: "ceph osd crush set|reweight-subtree" commands do not se...
- 05:06 PM Backport #49530 (In Progress): octopus: "ceph osd crush set|reweight-subtree" commands do not set...
- 05:05 PM Backport #49528 (Resolved): pacific: "ceph osd crush set|reweight-subtree" commands do not set we...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39736
m... - 05:02 PM Backport #49526: pacific: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound to class '...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39735
m... - 05:01 PM Backport #49404: pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39597
m... - 04:59 PM Backport #49404: pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
- https://github.com/ceph/ceph/pull/39796
https://github.com/ceph/ceph/pull/39597
(double whammy) - 01:41 PM Backport #49640: nautilus: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39912
- 11:44 AM Bug #49409 (Pending Backport): osd run into dead loop and tell slow request when rollback snap wi...
03/07/2021
- 10:02 PM Backport #49377: pacific: building libcrc32
- please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/39902
ceph-backport.sh versi... - 03:58 PM Backport #49482 (Resolved): pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/Manifes...
- 03:55 PM Backport #49642 (Resolved): pacific: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/40247
- 03:55 PM Backport #49641 (Resolved): octopus: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39935
- 03:55 PM Backport #49640 (Resolved): nautilus: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39912
- 03:54 PM Bug #48946 (Pending Backport): Disable and re-enable clog_to_monitors could trigger assertion
Also available in: Atom