Activity
From 03/16/2021 to 04/14/2021
04/14/2021
- 08:43 PM Bug #50368 (Resolved): common/PriorityCache.cc: FAILED ceph_assert(mem_avail >= 0) in radosbench_...
- ...
- 08:20 PM Bug #50042: rados/test.sh: api_watch_notify failures
- saw a failure in ceph_test_rados_api_watch_notify_pp, could be related to https://tracker.ceph.com/issues/45615
<p... - 07:32 PM Bug #50101: unhandled event in ReplicaActive
- This seems to be an unhandled event in the ReplicaActive state, couldn't find much in the logs that tell us what that...
- 04:50 PM Bug #46847: Loss of placement information on OSD reboot
- I'm working on a test reproducer here: https://github.com/ceph/ceph/pull/40849
- 07:35 AM Bug #46847: Loss of placement information on OSD reboot
- I think that this issue is the same as what I observed in https://tracker.ceph.com/issues/49104
We reproduce this ... - 04:26 PM Bug #50339: test_cls_cas failure: FAILED cls_cas.dup_get
- Revert PR https://github.com/ceph/ceph/pull/40857 based on https://github.com/ceph/ceph/pull/40811#issuecomment-81959...
- 12:38 AM Bug #50339: test_cls_cas failure: FAILED cls_cas.dup_get
- I don't think we were running it as a part of tasks/rados_cls_all before https://github.com/ceph/ceph/commit/52424874...
- 01:23 PM Bug #50355: Monitor node crashing while creating new clay pool
- Maybe related #46266.
jerasure pool creation is successfull ( without mon crash). - 01:21 PM Bug #50355 (New): Monitor node crashing while creating new clay pool
- ...
- 12:24 PM Bug #50352 (Resolved): LibRadosTwoPoolsPP.ManifestSnapRefcount failure
- ...
- 11:31 AM Bug #49104: crush weirdness: degraded PGs not marked as such, and choose_total_tries = 50 is too ...
- "Problem A" is probably a dupe of #46847
- 09:19 AM Bug #50351 (Resolved): osd: FAILED ceph_assert(recovering.count(*i)) after non-primary osd restar...
- On nautilus we have been observing an issue when an EC pg is in active+backfill_unfound+degraded state (which happens...
- 07:08 AM Bug #50346 (Resolved): OSD crash FAILED ceph_assert(!is_scrubbing())
- When I see warning PG_NOT_SCRUBBED, I set osd flag "nodeep-scrub", set config osd_max_scrubs to 2, and run:...
04/13/2021
- 11:32 PM Bug #50345 (In Progress): mon: new monitors may direct MMonJoin to a peon instead of the leader
- 10:52 PM Bug #50345 (Resolved): mon: new monitors may direct MMonJoin to a peon instead of the leader
- When monitors are joining a cluster, they may send an MMonJoin message to place themselves correctly in the map in ei...
- 10:50 PM Backport #50344 (Resolved): pacific: mon: stretch state is inconsistently-maintained on peons, pr...
- https://github.com/ceph/ceph/pull/41130
- 10:47 PM Bug #50308 (Pending Backport): mon: stretch state is inconsistently-maintained on peons, preventi...
- 06:34 AM Bug #50308 (In Progress): mon: stretch state is inconsistently-maintained on peons, preventing pr...
- 06:33 AM Bug #50308 (Resolved): mon: stretch state is inconsistently-maintained on peons, preventing prope...
- There's an issue where some in-memory monitor stretch state changes are only applied on the leader monitor, and so pe...
- 08:26 PM Bug #50342 (Resolved): test: compile errors
- 08:21 PM Bug #50342 (Fix Under Review): test: compile errors
- 06:20 PM Bug #50342: test: compile errors
- https://github.com/ceph/ceph/pull/40833
- 04:51 PM Bug #50342 (Resolved): test: compile errors
- ...
- 07:30 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
- https://pulpito.ceph.com/gregf-2021-04-13_09:22:15-rados-wip-stretch-mon-state-412-distro-basic-smithi/6043691
<pr... - 05:03 PM Bug #47617: rebuild_mondb: daemon-helper: command failed with exit status 1
- /a/sage-2021-04-12_21:28:18-rados-wip-sage4-testing-2021-04-12-1414-octopus-distro-basic-smithi/6042248...
- 02:57 PM Bug #47617: rebuild_mondb: daemon-helper: command failed with exit status 1
- /a/sage-2021-04-12_21:28:18-rados-wip-sage4-testing-2021-04-12-1414-octopus-distro-basic-smithi/6042248
- 04:26 PM Bug #50339: test_cls_cas failure: FAILED cls_cas.dup_get
- Hmm apparently I am the first to see this so far: https://sentry.ceph.com/organizations/ceph/issues/7662/events/1a7ee...
- 04:08 PM Bug #50339 (Resolved): test_cls_cas failure: FAILED cls_cas.dup_get
- This was in a wip branch of mine, but it didn't have any OSD changes:
https://pulpito.ceph.com/gregf-2021-04-13_09:2... - 09:39 AM Bug #48212: poollast_epoch_clean floor is stuck after pg merging
- We reduced the pg/pgp size of our pool to keep the 100 pgs per osd after we decommissioned a lot of osd hosts. When t...
- 12:14 AM Bug #50101: unhandled event in ReplicaActive
- ...
04/12/2021
- 08:45 PM Bug #50242: test_repair_corrupted_obj fails with assert not inconsistent
- /a/teuthology-2021-04-11_03:31:02-rados-pacific-distro-basic-smithi/6034330
- 03:22 PM Bug #50119: Invalid read of size 4 in ceph::logging::Log::dump_recent()
- Myoungwon Oh wrote:
> hm.. I think this is not related to tier_flush.
> According to /a/sage-2021-04-02_14\:05\:32... - 03:21 PM Bug #50119: Invalid read of size 4 in ceph::logging::Log::dump_recent()
- ...
- 03:18 PM Backport #49991: nautilus: unittest_mempool.check_shard_select failed
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40567
merged - 12:26 PM Bug #50299 (Resolved): PrimaryLogPG::inc_refcount_by_set leak
- ...
- 08:43 AM Feature #48182 (Resolved): osd: allow remote read by calling cls method from within cls context
04/11/2021
- 10:21 AM Backport #50153: nautilus: Reproduce https://tracker.ceph.com/issues/48417
- https://github.com/ceph/ceph/pull/40759#issuecomment-817237109 contains some advice on how to approach the backport
- 10:20 AM Backport #50152: octopus: Reproduce https://tracker.ceph.com/issues/48417
- https://github.com/ceph/ceph/pull/40759#issuecomment-817237109 contains some advice on how to approach the backport
04/10/2021
- 06:09 PM Backport #50154 (In Progress): pacific: Reproduce https://tracker.ceph.com/issues/48417
- 06:08 PM Backport #50153 (Need More Info): nautilus: Reproduce https://tracker.ceph.com/issues/48417
- not clear how this should be backported to nautilus
- 06:08 PM Backport #50152 (Need More Info): octopus: Reproduce https://tracker.ceph.com/issues/48417
- not clear how this should be backported to octopus
- 06:07 PM Bug #48417: unfound EC objects in sepia's LRC after upgrade
- Samuel Just wrote:
> I think this has the same root cause as https://tracker.ceph.com/issues/48613. Accepting IO be... - 06:02 PM Backport #50129 (In Progress): octopus: monmaptool --create --add nodeA --clobber monmap aborts i...
- 05:58 PM Backport #49816 (In Progress): octopus: mon: promote_standby does not update available_modules
- 05:52 PM Backport #49566 (In Progress): octopus: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- 01:49 AM Bug #50208: [ FAILED ] CephSQLiteTest.InsertBulk4096 [with slow ops during this test with addti...
- ...
- 12:58 AM Bug #50119: Invalid read of size 4 in ceph::logging::Log::dump_recent()
- hm.. I think this is not related to tier_flush.
According to /a/sage-2021-04-02_14\:05\:32-rados-wip-sage2-testing-...
04/09/2021
- 10:34 PM Bug #48503: scrub stat mismatch on bytes
- ...
- 10:30 PM Bug #50277 (New): nautilus: Error ETIMEDOUT: crush smoke test failed with -110: timed out during ...
- ...
- 10:06 PM Bug #50275 (New): nautilus: evicting unresponsive client
- ...
- 09:45 PM Backport #50274 (Resolved): pacific: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get...
- https://github.com/ceph/ceph/pull/44181
- 09:42 PM Bug #50192 (Pending Backport): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missi...
- 09:41 PM Bug #50208: [ FAILED ] CephSQLiteTest.InsertBulk4096 [with slow ops during this test with addti...
- Patrick, this test was using extra delays (https://github.com/ceph/ceph/pull/40593/files#diff-81a3b5f9694ef27ed306f04...
- 09:29 PM Bug #50119: Invalid read of size 4 in ceph::logging::Log::dump_recent()
- Myoungwon Oh: can you please help debug this?
- 09:21 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- Since the original patches have been reverted in pacific and master, downgrading this bug.
- 02:46 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- Neha Ojha wrote:
> https://github.com/ceph/ceph/pull/40623 - being reverted in master for the time being
merged - 08:36 PM Backport #50125 (New): nautilus: mon: Modify Paxos trim logic to be more efficient
- Nathan Cutler wrote:
> Looks like a feature. Not clear if it's worth the risk to backport this to nautilus as this l... - 04:37 PM Support #50270 (New): Is it better to use the reload instead of restart When reloading the nfs-ga...
Is it better to use the reload command instead of the restart command When reloading the nfs-ganesha(cephfs) config...- 02:45 PM Bug #49427: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_missing(soi...
- Neha Ojha wrote:
> Myoungwon Oh wrote:
> > https://github.com/ceph/ceph/pull/40606
merged - 10:20 AM Bug #49427 (Resolved): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 10:22 AM Bug #46978 (Resolved): OSD: shutdown of a OSD Host causes slow requests
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 10:18 AM Bug #49983 (Resolved): Test Failed with: "Scrubbing terminated -- not all pgs were active and cle...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 10:13 AM Backport #50018 (Resolved): pacific: Test Failed with: "Scrubbing terminated -- not all pgs were ...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40461
m... - 09:33 AM Backport #49682 (Resolved): nautilus: OSD: shutdown of a OSD Host causes slow requests
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40014
m...
04/08/2021
- 08:33 PM Bug #49962 (Triaged): 'sudo ceph --cluster ceph osd crush tunables default' fails due to valgrind...
- https://pulpito.ceph.com/yuriw-2021-04-07_17:37:43-fs-wip-yuri-testing-2021-04-07-0905-pacific-distro-basic-smithi/60...
- 08:01 PM Bug #50162: Backport to Natilus of automatic lowering min_size for repairing tasks (osd_allow_rec...
- This needs a pull request ID, or a list of master commits that are requested to be backported.
Now that Pacific is... - 07:21 PM Bug #50242: test_repair_corrupted_obj fails with assert not inconsistent
- Some observations from /a/sage-2021-04-06_20:05:38-rados-wip-sage-testing-2021-04-06-1403-distro-basic-smithi/6025317...
- 03:33 PM Bug #50242 (New): test_repair_corrupted_obj fails with assert not inconsistent
- ...
- 04:53 PM Backport #50130 (In Progress): nautilus: monmaptool --create --add nodeA --clobber monmap aborts ...
- 04:52 PM Backport #50123 (In Progress): octopus: mon: Modify Paxos trim logic to be more efficient
- 04:49 PM Backport #50125 (Need More Info): nautilus: mon: Modify Paxos trim logic to be more efficient
- Looks like a feature. Not clear if it's worth the risk to backport this to nautilus as this late stage of its lifecycle.
- 04:47 PM Backport #49919 (Need More Info): nautilus: mon: slow ops due to osd_failure
- master PR contains cleanups
for backporting to nautilus, we should consider a "more minimalistic" approach? - 04:35 PM Bug #50245 (New): TEST_recovery_scrub_2: Not enough recovery started simultaneously
- ...
- 04:20 PM Backport #49729 (In Progress): nautilus: debian ceph-common package post-inst clobbers ownership ...
- 04:18 PM Backport #49567 (In Progress): nautilus: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- 03:54 PM Backport #50124 (In Progress): pacific: mon: Modify Paxos trim logic to be more efficient
- 03:54 PM Backport #50131 (In Progress): pacific: monmaptool --create --add nodeA --clobber monmap aborts i...
- 03:43 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
- /a/sage-2021-04-06_20:05:38-rados-wip-sage-testing-2021-04-06-1403-distro-basic-smithi/6025407
- 03:08 PM Bug #50106: scrub/osd-scrub-repair.sh: corrupt_scrub_erasure: return 1
- /a/sage-2021-04-06_20:05:38-rados-wip-sage-testing-2021-04-06-1403-distro-basic-smithi/6025284
- 02:57 PM Backport #49531 (In Progress): nautilus: osd ok-to-stop too conservative
- 01:49 PM Bug #50219 (Closed): qa/standalone/erasure-code/test-erasure-eio.sh fails since pg is not in reco...
- https://pulpito.ceph.com/nojha-2021-04-07_23:49:43-rados:standalone-master-distro-basic-smithi/ - the same test passe...
- 01:23 PM Bug #50230 (Fix Under Review): mon: spawn loop after mon reinstalled
- We have tested the fix in PR 40660 and it solves our bootstrapping problem.
- 10:07 AM Bug #50230: mon: spawn loop after mon reinstalled
- Doh, ignore that fix, this is better:...
- 10:03 AM Bug #50230 (Resolved): mon: spawn loop after mon reinstalled
- This is related to #44076. (cluster is running 14.2.19 which has that fix.)
Scenario:
* mon is reinstalled (upgr... - 02:23 AM Bug #50222 (In Progress): osd: 5.2s0 deep-scrub : stat mismatch
- ...
04/07/2021
- 11:44 PM Bug #50219 (Closed): qa/standalone/erasure-code/test-erasure-eio.sh fails since pg is not in reco...
- ...
- 12:46 PM Bug #50208 (New): [ FAILED ] CephSQLiteTest.InsertBulk4096 [with slow ops during this test with...
- ...
- 11:41 AM Bug #40119: api_tier_pp hung causing a dead job
- ...
04/06/2021
- 08:54 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- https://github.com/ceph/ceph/pull/40623 - being reverted in master for the time being
- 07:17 PM Bug #49777: test_pool_min_size: 'check for active or peered' reached maximum tries (5) after wait...
- ...
- 06:52 PM Bug #49427: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_missing(soi...
- Myoungwon Oh wrote:
> https://github.com/ceph/ceph/pull/40606
Thanks for the fix, let's use https://tracker.ceph.... - 01:42 AM Bug #49427: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_missing(soi...
- https://github.com/ceph/ceph/pull/40606
- 06:52 PM Bug #50192 (Resolved): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_...
- ...
- 11:10 AM Bug #50162 (Won't Fix): Backport to Natilus of automatic lowering min_size for repairing tasks (o...
- Recently my ceph cluster (Nautilus 14.2.16 with 9 hosts, 16 4TB OSDs on each host, erasure-coding 4+2 profile, redund...
- 06:35 AM Backport #50154 (Resolved): pacific: Reproduce https://tracker.ceph.com/issues/48417
- https://github.com/ceph/ceph/pull/40759
- 06:35 AM Backport #50153 (Resolved): nautilus: Reproduce https://tracker.ceph.com/issues/48417
- https://github.com/ceph/ceph/pull/41611
- 06:35 AM Backport #50152 (Resolved): octopus: Reproduce https://tracker.ceph.com/issues/48417
- https://github.com/ceph/ceph/pull/41609
- 06:33 AM Bug #48613 (Pending Backport): Reproduce https://tracker.ceph.com/issues/48417
04/05/2021
- 09:14 PM Backport #50018: pacific: Test Failed with: "Scrubbing terminated -- not all pgs were active and ...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40461
merged - 08:53 PM Bug #21592: LibRadosCWriteOps.CmpExt got 0 instead of -4095-1
- /a/sage-2021-04-05_18:50:52-rados-master-distro-basic-smithi/6022643
- 07:47 PM Backport #49911 (Resolved): pacific: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get...
- https://github.com/ceph/ceph/pull/40289
- 07:46 PM Bug #49427: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_missing(soi...
- @Myoungwon Oh this new failure looks very similar to the issue tracked in this ticket?...
- 03:22 PM Bug #50140 (Duplicate): test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- 03:04 PM Bug #50140: test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- a duplicate of https://tracker.ceph.com/issues/49868
- 12:13 PM Bug #50140: test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- http://pulpito.front.sepia.ceph.com/rfriedma-2021-04-01_17:51:51-rados-wip-ronenf-cscrub-class-distro-basic-smithi/60...
- 12:12 PM Bug #50140: test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- Possible fixes to consider:
- a simple fix: extending the tests min-scrub-time;
- possibly better: modify the han... - 12:09 PM Bug #50140: test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- Caused by a combination of:
- re-scrub period ("osd scrub min interval") is set in radod/thrash* to (only) 60s.
- a... - 12:06 PM Bug #50140 (Duplicate): test/thrash - scrub: "not all pgs scrubbed" due to short rescrubbing period
- Thrashing test:
The observed error is "Exiting scrub checking -- not all pgs scrubbed".
See below for analysis.
- 03:12 PM Backport #49682: nautilus: OSD: shutdown of a OSD Host causes slow requests
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40014
merged - 03:07 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- In the log I've checked (http://pulpito.front.sepia.ceph.com/rfriedma-2021-04-01_17:51:51-rados-wip-ronenf-cscrub-cla...
- 01:40 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
- ...
04/03/2021
- 10:10 PM Backport #50131 (Resolved): pacific: monmaptool --create --add nodeA --clobber monmap aborts in e...
- https://github.com/ceph/ceph/pull/40690
- 02:25 PM Backport #50130 (Resolved): nautilus: monmaptool --create --add nodeA --clobber monmap aborts in ...
- https://github.com/ceph/ceph/pull/40700
- 02:25 PM Backport #50129 (Resolved): octopus: monmaptool --create --add nodeA --clobber monmap aborts in e...
- https://github.com/ceph/ceph/pull/40758
- 02:20 PM Backport #50125 (Resolved): nautilus: mon: Modify Paxos trim logic to be more efficient
- https://github.com/ceph/ceph/pull/41099
- 02:20 PM Backport #50124 (Resolved): pacific: mon: Modify Paxos trim logic to be more efficient
- https://github.com/ceph/ceph/pull/40691
- 02:20 PM Backport #50123 (Resolved): octopus: mon: Modify Paxos trim logic to be more efficient
- https://github.com/ceph/ceph/pull/40699
- 02:20 PM Bug #48336 (Pending Backport): monmaptool --create --add nodeA --clobber monmap aborts in entity_...
- 02:15 PM Bug #50004 (Pending Backport): mon: Modify Paxos trim logic to be more efficient
- 01:40 PM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
- ...
- 01:37 PM Bug #50119 (Can't reproduce): Invalid read of size 4 in ceph::logging::Log::dump_recent()
- the test does the 2 writes, creates the digest, deletes the pool, and submits the final op, but that op hangs
/a/s...
04/02/2021
- 11:09 PM Bug #48417: unfound EC objects in sepia's LRC after upgrade
- I think this has the same root cause as https://tracker.ceph.com/issues/48613. Accepting IO below min_size I think c...
- 10:22 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- Found it....
- 09:56 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- ...
- 09:28 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- I don't think teuthology-2021-02-03_03:31:04-rados-pacific-distro-basic-smithi/5852263 is related to can_rollback_to ...
- 08:20 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
- /a/sage-2021-04-02_14:05:32-rados-wip-sage2-testing-2021-04-01-1417-pacific-distro-basic-smithi/6017345
- 10:11 AM Backport #49991 (In Progress): nautilus: unittest_mempool.check_shard_select failed
- 10:10 AM Backport #49992 (In Progress): pacific: unittest_mempool.check_shard_select failed
- 10:08 AM Backport #49993 (In Progress): octopus: unittest_mempool.check_shard_select failed
- 07:11 AM Bug #45441 (Resolved): rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in clust...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:10 AM Bug #46323 (Resolved): thrash_cache_writeback_proxy_none: FAILED ceph_assert(version == old_value...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:09 AM Bug #48065 (Resolved): "ceph osd crush set|reweight-subtree" commands do not set weight on device...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:09 AM Bug #48566 (Resolved): nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_abort: return 1
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:09 AM Bug #48745 (Resolved): Segmentation fault in PrimaryLogPG::cancel_manifest_ops
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:08 AM Bug #48821 (Resolved): osd crash in OSD::heartbeat when dereferencing null session
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:08 AM Bug #48884 (Resolved): ceph osd df tree reporting incorrect SIZE value for rack having an empty h...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:08 AM Bug #49212 (Resolved): mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound to class 'ssd...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:06 AM Bug #49788 (Resolved): pool application metadata not propagated to the cache tier
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:06 AM Bug #49867 (Resolved): osd/scrub - a ceph_assert() in a legitimate path
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 07:05 AM Bug #49975 (Resolved): "make check" jenkins job fails
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 06:11 AM Backport #49978 (Resolved): pacific: "make check" jenkins job fails
- 06:11 AM Backport #49895 (Resolved): pacific: osd/scrub - a ceph_assert() in a legitimate path
- 06:10 AM Backport #49156 (Resolved): pacific: Segmentation fault in PrimaryLogPG::cancel_manifest_ops
- 06:10 AM Backport #49058 (Resolved): pacific: thrash_cache_writeback_proxy_none: FAILED ceph_assert(versio...
- 06:02 AM Backport #48596 (Resolved): octopus: nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_abor...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40278
m... - 06:01 AM Backport #49009 (Resolved): octopus: osd crash in OSD::heartbeat when dereferencing null session
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40277
m... - 06:01 AM Backport #49527 (Resolved): octopus: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40276
m... - 06:01 AM Backport #49730 (Resolved): octopus: debian ceph-common package post-inst clobbers ownership of c...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40275
m... - 06:01 AM Backport #49795 (Resolved): octopus: pool application metadata not propagated to the cache tier
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40274
m... - 06:01 AM Backport #49681 (Resolved): octopus: OSD: shutdown of a OSD Host causes slow requests
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40013
m... - 06:00 AM Backport #48985: octopus: ceph osd df tree reporting incorrect SIZE value for rack having an empt...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39970
m... - 05:57 AM Backport #49402 (Resolved): octopus: rados: Health check failed: 1/3 mons down, quorum a,c (MON_D...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40138
m... - 05:53 AM Backport #49530 (Resolved): octopus: "ceph osd crush set|reweight-subtree" commands do not set we...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39919
m...
04/01/2021
- 09:40 PM Backport #48985 (Resolved): octopus: ceph osd df tree reporting incorrect SIZE value for rack hav...
- 03:45 PM Backport #48985: octopus: ceph osd df tree reporting incorrect SIZE value for rack having an empt...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39970
merged - 09:08 PM Backport #49917 (In Progress): octopus: mon: slow ops due to osd_failure
- 08:40 PM Backport #49681 (In Progress): octopus: OSD: shutdown of a OSD Host causes slow requests
- 03:46 PM Backport #49681: octopus: OSD: shutdown of a OSD Host causes slow requests
- Mauricio Oliveira wrote:
> please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/... - 08:38 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- I'm pulling the logs to look more closely.
- 08:28 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- Looking at this more closely, I think removing that if block is still fine -- it'll be trimmed by the next IO anyway....
- 07:32 PM Bug #50106 (Can't reproduce): scrub/osd-scrub-repair.sh: corrupt_scrub_erasure: return 1
- ...
- 05:30 PM Bug #50101 (New): unhandled event in ReplicaActive
- ...
- 04:13 PM Bug #38219: rebuild-mondb hangs
- ...
- 03:50 PM Backport #48596: octopus: nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_abort: return 1
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40278
merged - 03:50 PM Backport #49009: octopus: osd crash in OSD::heartbeat when dereferencing null session
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40277
merged - 03:50 PM Bug #50100 (New): stale slow osd heartbeats health alert
- ...
- 03:49 PM Backport #49527: octopus: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound to class '...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40276
merged - 03:48 PM Backport #49730: octopus: debian ceph-common package post-inst clobbers ownership of cephadm log ...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40275
merged - 03:47 PM Backport #49795: octopus: pool application metadata not propagated to the cache tier
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40274
merged - 06:53 AM Bug #48336 (In Progress): monmaptool --create --add nodeA --clobber monmap aborts in entity_addr_...
- 12:32 AM Bug #50089 (Resolved): mon/MonMap.h: FAILED ceph_assert(m < ranks.size()) when reducing number of...
- ...
- 12:02 AM Bug #50088 (New): rados: qa: suites do not test mon removal
- We have the monthrash suite, but it merely turns monitors on and off again. It should probably also remove monitors f...
03/31/2021
- 04:10 PM Backport #50087 (Resolved): pacific: test_mon_pg: mon fails to join quorum to due election strate...
- 04:07 PM Bug #47654 (Pending Backport): test_mon_pg: mon fails to join quorum to due election strategy mis...
- Merged to master.
Pacific PR in progress at https://github.com/ceph/ceph/pull/40484 - 04:03 AM Bug #50042: rados/test.sh: api_watch_notify failures
- Looking at /a/teuthology-2021-03-27_14:59:29-upgrade:nautilus-x-pacific-distro-basic-smithi/6004568...
- 12:12 AM Bug #50042: rados/test.sh: api_watch_notify failures
- Looking at /a/sage-2021-03-28_19:04:26-rados-wip-sage2-testing-2021-03-28-0933-pacific-distro-basic-smithi/6007274
...
03/30/2021
- 11:32 PM Bug #50042: rados/test.sh: api_watch_notify failures
- Brad, as per our discussion, I am assigning this to you for now.
- 05:27 PM Bug #47654 (Fix Under Review): test_mon_pg: mon fails to join quorum to due election strategy mis...
03/29/2021
- 09:41 PM Bug #49988 (Fix Under Review): Global Recovery Event never completes
- Problem was that I did not subtract pgs that I skip because (reported_epoch_of_pg < start_epoch_of_event) from total_...
- 06:52 PM Bug #50042 (Pending Backport): rados/test.sh: api_watch_notify failures
- ...
- 09:01 AM Feature #48392: ceph ignores --keyring?
- I see this with v15.2.10 as well.
The problem is at the rollback stage, specifically with @osd purge-new@ where it... - 07:40 AM Bug #50032: used space in ceph df reported incorrectly for filestore OSDs
- Easily reproducible in Nautilus head indeed.
- 05:36 AM Bug #50032 (New): used space in ceph df reported incorrectly for filestore OSDs
- Hi, I noticed, that when filestore OSDs are used, used space is reported incorrectly:
filestore:
POOL ... - 06:23 AM Backport #50018 (In Progress): pacific: Test Failed with: "Scrubbing terminated -- not all pgs we...
- 04:50 AM Bug #50031 (New): osdc _throttle_op function param type of op_budget int is too small
- h2. problem
1. function calc_op_budget return type is int,when indata.length() is out of int range, it will return...
03/26/2021
- 09:59 PM Bug #49962: 'sudo ceph --cluster ceph osd crush tunables default' fails due to valgrind: Unknown ...
- Radek, can you please help take a look this issue?
- 09:53 PM Bug #50004 (Fix Under Review): mon: Modify Paxos trim logic to be more efficient
- 09:36 AM Bug #50004 (Resolved): mon: Modify Paxos trim logic to be more efficient
- When the log ingest rate is higher than the trimming rate, monitors may end up consuming a lot of space and become un...
- 08:41 PM Bug #49963 (Resolved): Crash in OSD::ms_fast_dispatch due to call to null vtable function
- 06:30 PM Backport #50018 (Resolved): pacific: Test Failed with: "Scrubbing terminated -- not all pgs were ...
- https://github.com/ceph/ceph/pull/40461
- 06:25 PM Bug #49983 (Pending Backport): Test Failed with: "Scrubbing terminated -- not all pgs were active...
- 05:53 PM Backport #49402: octopus: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40138
merged - 05:51 PM Backport #49530: octopus: "ceph osd crush set|reweight-subtree" commands do not set weight on dev...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39919
merged - 04:20 PM Backport #49640 (In Progress): nautilus: Disable and re-enable clog_to_monitors could trigger ass...
- 03:04 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- /a/yuriw-2021-03-25_20:03:40-rados-wip-yuri8-testing-2021-03-25-1042-pacific-distro-basic-smithi/5999016
- 03:03 PM Bug #45702: PGLog::read_log_and_missing: ceph_assert(miter == missing.get_items().end() || (miter...
- /a/yuriw-2021-03-25_20:03:40-rados-wip-yuri8-testing-2021-03-25-1042-pacific-distro-basic-smithi/5999051
- 12:11 PM Bug #50012 (Fix Under Review): Ceph-osd refuses to bind on an IP on the local loopback lo (again)
- the original fix was reverted by the fix of #49938. so fix in a more flexible way.
- 12:10 PM Bug #50012 (Fix Under Review): Ceph-osd refuses to bind on an IP on the local loopback lo (again)
- 09:58 AM Backport #49977 (Resolved): nautilus: "make check" jenkins job fails
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40407
m... - 09:54 AM Backport #49529 (Resolved): nautilus: "ceph osd crush set|reweight-subtree" commands do not set w...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39920
m... - 09:51 AM Backport #49976 (Resolved): octopus: "make check" jenkins job fails
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40406
m... - 07:41 AM Bug #38219: rebuild-mondb hangs
- ...
- 06:40 AM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- /kchai-2021-03-26_05:32:58-rados-wip-kefu-testing-2021-03-26-1134-distro-basic-smithi/6001105/
- 06:20 AM Backport #49993 (Resolved): octopus: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/39978
- 06:20 AM Backport #49992 (Resolved): pacific: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/40566
- 06:20 AM Backport #49991 (Resolved): nautilus: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/40567
- 06:17 AM Bug #49781 (Pending Backport): unittest_mempool.check_shard_select failed
03/25/2021
- 11:20 PM Bug #49988 (Resolved): Global Recovery Event never completes
- ...
- 10:12 PM Support #49847: OSD Fails to init after upgrading to octopus: _deferred_replay failed to decode d...
- Contrary to what I stated previously this does not seem like a software issue. The root cause was probably faulty RAM...
- 07:26 PM Bug #42884: OSDMapTest.CleanPGUpmaps failure
- https://jenkins.ceph.com/job/ceph-pull-requests/72185/consoleFull#-108728127277933967-90d1-4877-8d60-89cb08ef4eb1
- 04:44 PM Bug #49983 (Fix Under Review): Test Failed with: "Scrubbing terminated -- not all pgs were active...
- 02:42 PM Bug #49983 (Resolved): Test Failed with: "Scrubbing terminated -- not all pgs were active and cle...
- Test Run:
https://pulpito.ceph.com/nojha-2021-03-23_23:04:33-rados-wip-40323-2-distro-basic-gibba/5991116/
Failur... - 11:50 AM Backport #49977 (In Progress): nautilus: "make check" jenkins job fails
- 11:25 AM Backport #49977 (Resolved): nautilus: "make check" jenkins job fails
- https://github.com/ceph/ceph/pull/40407
- 11:45 AM Backport #49976 (In Progress): octopus: "make check" jenkins job fails
- 11:25 AM Backport #49976 (Resolved): octopus: "make check" jenkins job fails
- https://github.com/ceph/ceph/pull/40406
- 11:44 AM Backport #49978 (In Progress): pacific: "make check" jenkins job fails
- 11:25 AM Backport #49978 (Resolved): pacific: "make check" jenkins job fails
- https://github.com/ceph/ceph/pull/40405
- 11:22 AM Bug #49975 (Resolved): "make check" jenkins job fails
- ...
- 09:58 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Brad Hubbard wrote:
> I think I've found an issue related to this message in the logs but I'll need to test that the... - 05:32 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- I think I've found an issue related to this message in the logs but I'll need to test that theory....
- 04:20 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Hi Christian,
Is there no message in syslog or dmesg about the segfault at all? That seems odd. Maybe you should c... - 06:58 AM Bug #49970 (New): cmake: use GooleTest for adding gtest based tests
- see https://github.com/Kitware/CMake/blob/master/Modules/GoogleTest.cmake
for finer grained management of test, - 04:55 AM Bug #49963 (Fix Under Review): Crash in OSD::ms_fast_dispatch due to call to null vtable function
03/24/2021
- 11:43 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
- So the issue is we never should have gone this deep into the OSD::ms_fast_dispatch function.
Previously we would h... - 10:53 PM Bug #49963 (Triaged): Crash in OSD::ms_fast_dispatch due to call to null vtable function
- 10:50 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
- /a/sage-2021-03-24_06:13:24-upgrade:octopus-x-wip-sage-testing-2021-03-23-2309-distro-basic-smithi/5993446...
- 10:46 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
- m->header.type is 85...
- 10:38 PM Bug #49963 (Resolved): Crash in OSD::ms_fast_dispatch due to call to null vtable function
- /a/sage-2021-03-24_06:13:24-upgrade:octopus-x-wip-sage-testing-2021-03-23-2309-distro-basic-smithi/5993446...
- 09:09 PM Bug #49962 (Resolved): 'sudo ceph --cluster ceph osd crush tunables default' fails due to valgrin...
- ...
- 07:25 PM Bug #49961 (New): scrub/osd-recovery-scrub.sh: TEST_recovery_scrub_1 failed
- ...
- 12:21 PM Bug #49487 (Resolved): osd:scrub skip some pg
- 09:12 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Brad Hubbard wrote:
> Hi Christian,
>
> From the log this was a segfault in MonitorDBStore::get_synchronizer.
> ... - 02:51 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- So the offset into MonitorDBStore::get_synchronizer where we segfaulted was
0x3f.
If we set up a binary compatibl... - 02:44 AM Documentation #35967 (Resolved): [doc] sync documentation "OSD Config Reference" default values w...
03/23/2021
- 11:25 PM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Hi Christian,
From the log this was a segfault in MonitorDBStore::get_synchronizer.
2021-03-14 18:00:28.679 7f1... - 08:45 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Neha Ojha wrote:
> Do you happen to have a coredump for this or can you provide a copy of the mon store db?
Unfor... - 07:12 PM Bug #49781: unittest_mempool.check_shard_select failed
- https://jenkins.ceph.com/job/ceph-pull-requests/72011/consoleFull#-85362248744e9240e-b50a-4693-bac0-8a991bac86ac
- 04:55 PM Backport #49529: nautilus: "ceph osd crush set|reweight-subtree" commands do not set weight on de...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39920
merged - 12:23 AM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
- ...
03/22/2021
- 10:21 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- relevant osd.3 logs from yuriw-2021-03-19_00:00:55-rados-wip-yuri8-testing-2021-03-18-1502-pacific-distro-basic-smith...
- 09:40 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- /a/yuriw-2021-03-19_00:00:55-rados-wip-yuri8-testing-2021-03-18-1502-pacific-distro-basic-smithi/5978982
- 02:47 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Thanks, Konstantin!
- 02:46 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- @Mauricio, I was update issue backports and status.
- 02:45 PM Bug #46978 (Pending Backport): OSD: shutdown of a OSD Host causes slow requests
- 01:44 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Hi @singuliere _,
Could you please revert the backport field to include Octopus and Nautilus?
Such backports ha... - 07:40 AM Feature #49789 (Fix Under Review): common/TrackedOp: add op priority for TrackedOp
- 07:15 AM Backport #49919 (Resolved): nautilus: mon: slow ops due to osd_failure
- https://github.com/ceph/ceph/pull/41213
- 07:15 AM Backport #49918 (Resolved): pacific: mon: slow ops due to osd_failure
- https://github.com/ceph/ceph/pull/41090
- 07:15 AM Backport #49917 (Resolved): octopus: mon: slow ops due to osd_failure
- https://github.com/ceph/ceph/pull/40558
- 07:13 AM Bug #47380 (Pending Backport): mon: slow ops due to osd_failure
03/21/2021
- 06:25 PM Backport #49642 (Resolved): pacific: Disable and re-enable clog_to_monitors could trigger assertion
- 06:24 PM Backport #49008 (Resolved): pacific: osd crash in OSD::heartbeat when dereferencing null session
- 06:22 PM Backport #49728 (Resolved): pacific: debian ceph-common package post-inst clobbers ownership of c...
- 05:46 PM Bug #49883 (Rejected): librados: hang in RadosClient::wait_for_osdmap
- This turned out to be because the auth credential had wrong permissions (due to a profile name change).
- 02:33 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
- /a/sage-2021-03-20_15:11:51-rados-wip-sage2-testing-2021-03-20-0832-pacific-distro-basic-smithi/5983920
- 01:20 PM Backport #49911 (Resolved): pacific: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get...
- 01:18 PM Bug #49427 (Pending Backport): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missi...
- need this in pacific too: /a/sage-2021-03-20_15:11:51-rados-wip-sage2-testing-2021-03-20-0832-pacific-distro-basic-sm...
03/20/2021
- 01:45 PM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
- /a/sage-2021-03-20_01:11:45-rados-wip-sage-testing-2021-03-19-1647-distro-basic-smithi/5982282
with logs! - 08:06 AM Backport #48596 (In Progress): octopus: nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_a...
- 08:05 AM Backport #49009 (In Progress): octopus: osd crash in OSD::heartbeat when dereferencing null session
- 08:04 AM Backport #49527 (In Progress): octopus: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bo...
- 07:56 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Since this issue is resolved and only the pacific backport was done, I assume it means the octopus & nautilus backpor...
- 07:46 AM Backport #49730 (In Progress): octopus: debian ceph-common package post-inst clobbers ownership o...
- 07:46 AM Backport #49795 (In Progress): octopus: pool application metadata not propagated to the cache tier
03/19/2021
- 09:47 PM Bug #49809 (Need More Info): 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Do you happen to have a coredump for this or can you provide a copy of the mon store db?
- 09:13 PM Backport #49895: pacific: osd/scrub - a ceph_assert() in a legitimate path
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40226
merged - 06:08 PM Backport #49796 (Resolved): pacific: pool application metadata not propagated to the cache tier
- 06:02 PM Backport #49728 (In Progress): pacific: debian ceph-common package post-inst clobbers ownership o...
- 05:58 PM Backport #49642 (In Progress): pacific: Disable and re-enable clog_to_monitors could trigger asse...
- 05:57 PM Backport #49008 (In Progress): pacific: osd crash in OSD::heartbeat when dereferencing null session
- 05:06 PM Bug #49902 (New): monitor store.db directory has been increasing and has exceeded 450GB
- My monitor keeps growing. After it exceeds 300GB, it cannot be started because the disk is not enough. After the spac...
- 11:07 AM Backport #49320 (Resolved): octopus: thrash_cache_writeback_proxy_none: FAILED ceph_assert(versio...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39578
m... - 09:50 AM Backport #49641 (Resolved): octopus: Disable and re-enable clog_to_monitors could trigger assertion
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39935
m... - 09:48 AM Backport #49533 (Resolved): octopus: osd ok-to-stop too conservative
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39887
m...
03/18/2021
- 10:42 PM Bug #49894: set a non-zero default value for osd_client_message_cap
- Neha Ojha wrote:
> The current default of 0 doesn't help and we've tried setting it to 5000 for one of our tests htt... - 09:26 PM Bug #49894 (Resolved): set a non-zero default value for osd_client_message_cap
- The current default of 0 doesn't help and we've tried setting it to 5000 for one of our tests https://github.com/neha...
- 09:57 PM Backport #49895 (In Progress): pacific: osd/scrub - a ceph_assert() in a legitimate path
- 09:55 PM Backport #49895 (Resolved): pacific: osd/scrub - a ceph_assert() in a legitimate path
- https://github.com/ceph/ceph/pull/40226
- 09:54 PM Bug #49867 (Pending Backport): osd/scrub - a ceph_assert() in a legitimate path
- 05:57 PM Bug #49888 (Resolved): rados/singleton: radosbench.py: teuthology.exceptions.MaxWhileTries: reach...
- ...
- 04:49 PM Support #49847: OSD Fails to init after upgrading to octopus: _deferred_replay failed to decode d...
- Ended up nuking the OSD & letting it recover - this workaround "solves" the problem for me; feel free to close the is...
- 04:30 PM Backport #49641: octopus: Disable and re-enable clog_to_monitors could trigger assertion
- gerald yang wrote:
> https://github.com/ceph/ceph/pull/39935
merged - 04:27 PM Backport #49533: octopus: osd ok-to-stop too conservative
- https://github.com/ceph/ceph/pull/39887 merged
- 04:24 PM Bug #46323: thrash_cache_writeback_proxy_none: FAILED ceph_assert(version == old_value.version) i...
- https://github.com/ceph/ceph/pull/39578 merged
- 03:03 PM Bug #49883: librados: hang in RadosClient::wait_for_osdmap
- The ubuntu jobs are failing because the ceph-mgr can't find libcephsqlite.so. There were packaging changes that proba...
- 02:34 PM Bug #49883 (Rejected): librados: hang in RadosClient::wait_for_osdmap
- https://pulpito.ceph.com/pdonnell-2021-03-18_06:53:09-rados:basic-wip-pdonnell-testing-20210318.035227-distro-basic-s...
- 01:54 PM Backport #49401: pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40137
m... - 01:54 PM Backport #49401: pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40137
m... - 01:54 PM Backport #49817 (Resolved): pacific: mon: promote_standby does not update available_modules
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40132
m...
03/17/2021
- 09:19 PM Backport #49401 (Resolved): pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_D...
- 06:24 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- I think we should revert this in pacific https://github.com/ceph/ceph/pull/40195, until we can fix the test failures.
- 04:02 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- Something merged after 37f9d0a25d06a6b8529aa350110eba930fba8c9e since https://pulpito.ceph.com/yuriw-2021-03-15_23:42...
- 03:57 PM Bug #49868 (New): RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- ...
- 03:28 PM Bug #49867 (Fix Under Review): osd/scrub - a ceph_assert() in a legitimate path
- 03:15 PM Bug #49867 (Resolved): osd/scrub - a ceph_assert() in a legitimate path
- in pg_scrubber.cc, in PgScrubber::on_replica_init() there's a ceph_assert()
that might trigger in a correct path (if... - 10:14 AM Backport #49683: pacific: OSD: shutdown of a OSD Host causes slow requests
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39957
m... - 09:05 AM Bug #49781: unittest_mempool.check_shard_select failed
- Discussion on the mailing list about "mempool optimization":https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/...
03/16/2021
- 08:07 PM Support #49847 (Closed): OSD Fails to init after upgrading to octopus: _deferred_replay failed to...
- An OSD fails to start after upgrading from mimic 13.2.2 to octopus 15.2.9.
It seems like first bluestore fails at... - 03:45 PM Bug #49832 (New): Segmentation fault: in thread_name:ms_dispatch
- ...
- 03:22 PM Bug #49781: unittest_mempool.check_shard_select failed
- The test condition should not be too strict because there really is no way to predict the result. It is however good ...
- 12:56 PM Bug #49781: unittest_mempool.check_shard_select failed
- Using "pthread_self for sharding":https://github.com/ceph/ceph/blob/master/src/include/mempool.h#L261-L262 is not gre...
- 11:25 AM Bug #49781 (In Progress): unittest_mempool.check_shard_select failed
- 08:15 AM Bug #49697: prime pg temp: unexpected optimization
- ping
- 08:14 AM Bug #49787 (Resolved): test_envlibrados_for_rocksdb.sh fails on master
- 06:28 AM Backport #49682 (In Progress): nautilus: OSD: shutdown of a OSD Host causes slow requests
Also available in: Atom