Activity
From 03/01/2021 to 03/30/2021
03/30/2021
- 11:32 PM Bug #50042: rados/test.sh: api_watch_notify failures
- Brad, as per our discussion, I am assigning this to you for now.
- 05:27 PM Bug #47654 (Fix Under Review): test_mon_pg: mon fails to join quorum to due election strategy mis...
03/29/2021
- 09:41 PM Bug #49988 (Fix Under Review): Global Recovery Event never completes
- Problem was that I did not subtract pgs that I skip because (reported_epoch_of_pg < start_epoch_of_event) from total_...
- 06:52 PM Bug #50042 (Resolved): rados/test.sh: api_watch_notify failures
- ...
- 09:01 AM Feature #48392: ceph ignores --keyring?
- I see this with v15.2.10 as well.
The problem is at the rollback stage, specifically with @osd purge-new@ where it... - 07:40 AM Bug #50032: used space in ceph df reported incorrectly for filestore OSDs
- Easily reproducible in Nautilus head indeed.
- 05:36 AM Bug #50032 (New): used space in ceph df reported incorrectly for filestore OSDs
- Hi, I noticed, that when filestore OSDs are used, used space is reported incorrectly:
filestore:
POOL ... - 06:23 AM Backport #50018 (In Progress): pacific: Test Failed with: "Scrubbing terminated -- not all pgs we...
- 04:50 AM Bug #50031 (New): osdc _throttle_op function param type of op_budget int is too small
- h2. problem
1. function calc_op_budget return type is int,when indata.length() is out of int range, it will return...
03/26/2021
- 09:59 PM Bug #49962: 'sudo ceph --cluster ceph osd crush tunables default' fails due to valgrind: Unknown ...
- Radek, can you please help take a look this issue?
- 09:53 PM Bug #50004 (Fix Under Review): mon: Modify Paxos trim logic to be more efficient
- 09:36 AM Bug #50004 (Resolved): mon: Modify Paxos trim logic to be more efficient
- When the log ingest rate is higher than the trimming rate, monitors may end up consuming a lot of space and become un...
- 08:41 PM Bug #49963 (Resolved): Crash in OSD::ms_fast_dispatch due to call to null vtable function
- 06:30 PM Backport #50018 (Resolved): pacific: Test Failed with: "Scrubbing terminated -- not all pgs were ...
- https://github.com/ceph/ceph/pull/40461
- 06:25 PM Bug #49983 (Pending Backport): Test Failed with: "Scrubbing terminated -- not all pgs were active...
- 05:53 PM Backport #49402: octopus: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40138
merged - 05:51 PM Backport #49530: octopus: "ceph osd crush set|reweight-subtree" commands do not set weight on dev...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39919
merged - 04:20 PM Backport #49640 (In Progress): nautilus: Disable and re-enable clog_to_monitors could trigger ass...
- 03:04 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- /a/yuriw-2021-03-25_20:03:40-rados-wip-yuri8-testing-2021-03-25-1042-pacific-distro-basic-smithi/5999016
- 03:03 PM Bug #45702: PGLog::read_log_and_missing: ceph_assert(miter == missing.get_items().end() || (miter...
- /a/yuriw-2021-03-25_20:03:40-rados-wip-yuri8-testing-2021-03-25-1042-pacific-distro-basic-smithi/5999051
- 12:11 PM Bug #50012 (Fix Under Review): Ceph-osd refuses to bind on an IP on the local loopback lo (again)
- the original fix was reverted by the fix of #49938. so fix in a more flexible way.
- 12:10 PM Bug #50012 (Fix Under Review): Ceph-osd refuses to bind on an IP on the local loopback lo (again)
- 09:58 AM Backport #49977 (Resolved): nautilus: "make check" jenkins job fails
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40407
m... - 09:54 AM Backport #49529 (Resolved): nautilus: "ceph osd crush set|reweight-subtree" commands do not set w...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39920
m... - 09:51 AM Backport #49976 (Resolved): octopus: "make check" jenkins job fails
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40406
m... - 07:41 AM Bug #38219: rebuild-mondb hangs
- ...
- 06:40 AM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- /kchai-2021-03-26_05:32:58-rados-wip-kefu-testing-2021-03-26-1134-distro-basic-smithi/6001105/
- 06:20 AM Backport #49993 (Resolved): octopus: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/39978
- 06:20 AM Backport #49992 (Resolved): pacific: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/40566
- 06:20 AM Backport #49991 (Resolved): nautilus: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/40567
- 06:17 AM Bug #49781 (Pending Backport): unittest_mempool.check_shard_select failed
03/25/2021
- 11:20 PM Bug #49988 (Resolved): Global Recovery Event never completes
- ...
- 10:12 PM Support #49847: OSD Fails to init after upgrading to octopus: _deferred_replay failed to decode d...
- Contrary to what I stated previously this does not seem like a software issue. The root cause was probably faulty RAM...
- 07:26 PM Bug #42884: OSDMapTest.CleanPGUpmaps failure
- https://jenkins.ceph.com/job/ceph-pull-requests/72185/consoleFull#-108728127277933967-90d1-4877-8d60-89cb08ef4eb1
- 04:44 PM Bug #49983 (Fix Under Review): Test Failed with: "Scrubbing terminated -- not all pgs were active...
- 02:42 PM Bug #49983 (Resolved): Test Failed with: "Scrubbing terminated -- not all pgs were active and cle...
- Test Run:
https://pulpito.ceph.com/nojha-2021-03-23_23:04:33-rados-wip-40323-2-distro-basic-gibba/5991116/
Failur... - 11:50 AM Backport #49977 (In Progress): nautilus: "make check" jenkins job fails
- 11:25 AM Backport #49977 (Resolved): nautilus: "make check" jenkins job fails
- https://github.com/ceph/ceph/pull/40407
- 11:45 AM Backport #49976 (In Progress): octopus: "make check" jenkins job fails
- 11:25 AM Backport #49976 (Resolved): octopus: "make check" jenkins job fails
- https://github.com/ceph/ceph/pull/40406
- 11:44 AM Backport #49978 (In Progress): pacific: "make check" jenkins job fails
- 11:25 AM Backport #49978 (Resolved): pacific: "make check" jenkins job fails
- https://github.com/ceph/ceph/pull/40405
- 11:22 AM Bug #49975 (Resolved): "make check" jenkins job fails
- ...
- 09:58 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Brad Hubbard wrote:
> I think I've found an issue related to this message in the logs but I'll need to test that the... - 05:32 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- I think I've found an issue related to this message in the logs but I'll need to test that theory....
- 04:20 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Hi Christian,
Is there no message in syslog or dmesg about the segfault at all? That seems odd. Maybe you should c... - 06:58 AM Bug #49970 (New): cmake: use GooleTest for adding gtest based tests
- see https://github.com/Kitware/CMake/blob/master/Modules/GoogleTest.cmake
for finer grained management of test, - 04:55 AM Bug #49963 (Fix Under Review): Crash in OSD::ms_fast_dispatch due to call to null vtable function
03/24/2021
- 11:43 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
- So the issue is we never should have gone this deep into the OSD::ms_fast_dispatch function.
Previously we would h... - 10:53 PM Bug #49963 (Triaged): Crash in OSD::ms_fast_dispatch due to call to null vtable function
- 10:50 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
- /a/sage-2021-03-24_06:13:24-upgrade:octopus-x-wip-sage-testing-2021-03-23-2309-distro-basic-smithi/5993446...
- 10:46 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
- m->header.type is 85...
- 10:38 PM Bug #49963 (Resolved): Crash in OSD::ms_fast_dispatch due to call to null vtable function
- /a/sage-2021-03-24_06:13:24-upgrade:octopus-x-wip-sage-testing-2021-03-23-2309-distro-basic-smithi/5993446...
- 09:09 PM Bug #49962 (Resolved): 'sudo ceph --cluster ceph osd crush tunables default' fails due to valgrin...
- ...
- 07:25 PM Bug #49961 (New): scrub/osd-recovery-scrub.sh: TEST_recovery_scrub_1 failed
- ...
- 12:21 PM Bug #49487 (Resolved): osd:scrub skip some pg
- 09:12 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Brad Hubbard wrote:
> Hi Christian,
>
> From the log this was a segfault in MonitorDBStore::get_synchronizer.
> ... - 02:51 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- So the offset into MonitorDBStore::get_synchronizer where we segfaulted was
0x3f.
If we set up a binary compatibl... - 02:44 AM Documentation #35967 (Resolved): [doc] sync documentation "OSD Config Reference" default values w...
03/23/2021
- 11:25 PM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Hi Christian,
From the log this was a segfault in MonitorDBStore::get_synchronizer.
2021-03-14 18:00:28.679 7f1... - 08:45 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Neha Ojha wrote:
> Do you happen to have a coredump for this or can you provide a copy of the mon store db?
Unfor... - 07:12 PM Bug #49781: unittest_mempool.check_shard_select failed
- https://jenkins.ceph.com/job/ceph-pull-requests/72011/consoleFull#-85362248744e9240e-b50a-4693-bac0-8a991bac86ac
- 04:55 PM Backport #49529: nautilus: "ceph osd crush set|reweight-subtree" commands do not set weight on de...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39920
merged - 12:23 AM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
- ...
03/22/2021
- 10:21 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- relevant osd.3 logs from yuriw-2021-03-19_00:00:55-rados-wip-yuri8-testing-2021-03-18-1502-pacific-distro-basic-smith...
- 09:40 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- /a/yuriw-2021-03-19_00:00:55-rados-wip-yuri8-testing-2021-03-18-1502-pacific-distro-basic-smithi/5978982
- 02:47 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Thanks, Konstantin!
- 02:46 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- @Mauricio, I was update issue backports and status.
- 02:45 PM Bug #46978 (Pending Backport): OSD: shutdown of a OSD Host causes slow requests
- 01:44 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Hi @singuliere _,
Could you please revert the backport field to include Octopus and Nautilus?
Such backports ha... - 07:40 AM Feature #49789 (Fix Under Review): common/TrackedOp: add op priority for TrackedOp
- 07:15 AM Backport #49919 (Resolved): nautilus: mon: slow ops due to osd_failure
- https://github.com/ceph/ceph/pull/41213
- 07:15 AM Backport #49918 (Resolved): pacific: mon: slow ops due to osd_failure
- https://github.com/ceph/ceph/pull/41090
- 07:15 AM Backport #49917 (Resolved): octopus: mon: slow ops due to osd_failure
- https://github.com/ceph/ceph/pull/40558
- 07:13 AM Bug #47380 (Pending Backport): mon: slow ops due to osd_failure
03/21/2021
- 06:25 PM Backport #49642 (Resolved): pacific: Disable and re-enable clog_to_monitors could trigger assertion
- 06:24 PM Backport #49008 (Resolved): pacific: osd crash in OSD::heartbeat when dereferencing null session
- 06:22 PM Backport #49728 (Resolved): pacific: debian ceph-common package post-inst clobbers ownership of c...
- 05:46 PM Bug #49883 (Rejected): librados: hang in RadosClient::wait_for_osdmap
- This turned out to be because the auth credential had wrong permissions (due to a profile name change).
- 02:33 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
- /a/sage-2021-03-20_15:11:51-rados-wip-sage2-testing-2021-03-20-0832-pacific-distro-basic-smithi/5983920
- 01:20 PM Backport #49911 (Resolved): pacific: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get...
- 01:18 PM Bug #49427 (Pending Backport): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missi...
- need this in pacific too: /a/sage-2021-03-20_15:11:51-rados-wip-sage2-testing-2021-03-20-0832-pacific-distro-basic-sm...
03/20/2021
- 01:45 PM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
- /a/sage-2021-03-20_01:11:45-rados-wip-sage-testing-2021-03-19-1647-distro-basic-smithi/5982282
with logs! - 08:06 AM Backport #48596 (In Progress): octopus: nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_a...
- 08:05 AM Backport #49009 (In Progress): octopus: osd crash in OSD::heartbeat when dereferencing null session
- 08:04 AM Backport #49527 (In Progress): octopus: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bo...
- 07:56 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Since this issue is resolved and only the pacific backport was done, I assume it means the octopus & nautilus backpor...
- 07:46 AM Backport #49730 (In Progress): octopus: debian ceph-common package post-inst clobbers ownership o...
- 07:46 AM Backport #49795 (In Progress): octopus: pool application metadata not propagated to the cache tier
03/19/2021
- 09:47 PM Bug #49809 (Need More Info): 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- Do you happen to have a coredump for this or can you provide a copy of the mon store db?
- 09:13 PM Backport #49895: pacific: osd/scrub - a ceph_assert() in a legitimate path
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40226
merged - 06:08 PM Backport #49796 (Resolved): pacific: pool application metadata not propagated to the cache tier
- 06:02 PM Backport #49728 (In Progress): pacific: debian ceph-common package post-inst clobbers ownership o...
- 05:58 PM Backport #49642 (In Progress): pacific: Disable and re-enable clog_to_monitors could trigger asse...
- 05:57 PM Backport #49008 (In Progress): pacific: osd crash in OSD::heartbeat when dereferencing null session
- 05:06 PM Bug #49902 (New): monitor store.db directory has been increasing and has exceeded 450GB
- My monitor keeps growing. After it exceeds 300GB, it cannot be started because the disk is not enough. After the spac...
- 11:07 AM Backport #49320 (Resolved): octopus: thrash_cache_writeback_proxy_none: FAILED ceph_assert(versio...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39578
m... - 09:50 AM Backport #49641 (Resolved): octopus: Disable and re-enable clog_to_monitors could trigger assertion
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39935
m... - 09:48 AM Backport #49533 (Resolved): octopus: osd ok-to-stop too conservative
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39887
m...
03/18/2021
- 10:42 PM Bug #49894: set a non-zero default value for osd_client_message_cap
- Neha Ojha wrote:
> The current default of 0 doesn't help and we've tried setting it to 5000 for one of our tests htt... - 09:26 PM Bug #49894 (Resolved): set a non-zero default value for osd_client_message_cap
- The current default of 0 doesn't help and we've tried setting it to 5000 for one of our tests https://github.com/neha...
- 09:57 PM Backport #49895 (In Progress): pacific: osd/scrub - a ceph_assert() in a legitimate path
- 09:55 PM Backport #49895 (Resolved): pacific: osd/scrub - a ceph_assert() in a legitimate path
- https://github.com/ceph/ceph/pull/40226
- 09:54 PM Bug #49867 (Pending Backport): osd/scrub - a ceph_assert() in a legitimate path
- 05:57 PM Bug #49888 (Resolved): rados/singleton: radosbench.py: teuthology.exceptions.MaxWhileTries: reach...
- ...
- 04:49 PM Support #49847: OSD Fails to init after upgrading to octopus: _deferred_replay failed to decode d...
- Ended up nuking the OSD & letting it recover - this workaround "solves" the problem for me; feel free to close the is...
- 04:30 PM Backport #49641: octopus: Disable and re-enable clog_to_monitors could trigger assertion
- gerald yang wrote:
> https://github.com/ceph/ceph/pull/39935
merged - 04:27 PM Backport #49533: octopus: osd ok-to-stop too conservative
- https://github.com/ceph/ceph/pull/39887 merged
- 04:24 PM Bug #46323: thrash_cache_writeback_proxy_none: FAILED ceph_assert(version == old_value.version) i...
- https://github.com/ceph/ceph/pull/39578 merged
- 03:03 PM Bug #49883: librados: hang in RadosClient::wait_for_osdmap
- The ubuntu jobs are failing because the ceph-mgr can't find libcephsqlite.so. There were packaging changes that proba...
- 02:34 PM Bug #49883 (Rejected): librados: hang in RadosClient::wait_for_osdmap
- https://pulpito.ceph.com/pdonnell-2021-03-18_06:53:09-rados:basic-wip-pdonnell-testing-20210318.035227-distro-basic-s...
- 01:54 PM Backport #49401: pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40137
m... - 01:54 PM Backport #49401: pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40137
m... - 01:54 PM Backport #49817 (Resolved): pacific: mon: promote_standby does not update available_modules
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40132
m...
03/17/2021
- 09:19 PM Backport #49401 (Resolved): pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_D...
- 06:24 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- I think we should revert this in pacific https://github.com/ceph/ceph/pull/40195, until we can fix the test failures.
- 04:02 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- Something merged after 37f9d0a25d06a6b8529aa350110eba930fba8c9e since https://pulpito.ceph.com/yuriw-2021-03-15_23:42...
- 03:57 PM Bug #49868 (New): RuntimeError: Exiting scrub checking -- not all pgs scrubbed
- ...
- 03:28 PM Bug #49867 (Fix Under Review): osd/scrub - a ceph_assert() in a legitimate path
- 03:15 PM Bug #49867 (Resolved): osd/scrub - a ceph_assert() in a legitimate path
- in pg_scrubber.cc, in PgScrubber::on_replica_init() there's a ceph_assert()
that might trigger in a correct path (if... - 10:14 AM Backport #49683: pacific: OSD: shutdown of a OSD Host causes slow requests
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39957
m... - 09:05 AM Bug #49781: unittest_mempool.check_shard_select failed
- Discussion on the mailing list about "mempool optimization":https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/...
03/16/2021
- 08:07 PM Support #49847 (Closed): OSD Fails to init after upgrading to octopus: _deferred_replay failed to...
- An OSD fails to start after upgrading from mimic 13.2.2 to octopus 15.2.9.
It seems like first bluestore fails at... - 03:45 PM Bug #49832 (New): Segmentation fault: in thread_name:ms_dispatch
- ...
- 03:22 PM Bug #49781: unittest_mempool.check_shard_select failed
- The test condition should not be too strict because there really is no way to predict the result. It is however good ...
- 12:56 PM Bug #49781: unittest_mempool.check_shard_select failed
- Using "pthread_self for sharding":https://github.com/ceph/ceph/blob/master/src/include/mempool.h#L261-L262 is not gre...
- 11:25 AM Bug #49781 (In Progress): unittest_mempool.check_shard_select failed
- 08:15 AM Bug #49697: prime pg temp: unexpected optimization
- ping
- 08:14 AM Bug #49787 (Resolved): test_envlibrados_for_rocksdb.sh fails on master
- 06:28 AM Backport #49682 (In Progress): nautilus: OSD: shutdown of a OSD Host causes slow requests
03/15/2021
- 10:42 PM Bug #46978 (Resolved): OSD: shutdown of a OSD Host causes slow requests
- 10:42 PM Backport #49683 (Resolved): pacific: OSD: shutdown of a OSD Host causes slow requests
- 10:41 PM Backport #49774 (Resolved): pacific: Get more parallel scrubs within osd_max_scrubs limits
- 09:56 PM Backport #49402 (In Progress): octopus: rados: Health check failed: 1/3 mons down, quorum a,c (MO...
- 09:55 PM Backport #49401 (In Progress): pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MO...
- 08:15 PM Backport #49817 (Resolved): pacific: mon: promote_standby does not update available_modules
- https://github.com/ceph/ceph/pull/40132
- 08:15 PM Backport #49816 (Resolved): octopus: mon: promote_standby does not update available_modules
- https://github.com/ceph/ceph/pull/40757
- 08:11 PM Bug #49778 (Pending Backport): mon: promote_standby does not update available_modules
- 05:26 PM Bug #49810 (Need More Info): rados/singleton: with msgr-failures/none MON_DOWN due to haven't for...
- ...
- 05:16 PM Bug #49809 (Need More Info): 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
- We experienced a single mon crash (out of 3 mons) - We observed no other issues on the machine or the cluster.
I a... - 03:02 PM Bug #48793 (Resolved): out of order op
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 03:02 PM Bug #48990 (Resolved): rados/dashboard: Health check failed: Telemetry requires re-opt-in (TELEME...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 10:38 AM Bug #49781: unittest_mempool.check_shard_select failed
- master also...
- 09:38 AM Bug #49779 (Resolved): standalone: osd-recovery-scrub.sh: Recovery never started
- 09:22 AM Bug #49758 (Resolved): messages/MOSDPGNotify.h: virtual void MOSDPGNotify::encode_payload(uint64_...
- 09:10 AM Backport #49796 (Resolved): pacific: pool application metadata not propagated to the cache tier
- https://github.com/ceph/ceph/pull/40119
- 09:10 AM Backport #49795 (Resolved): octopus: pool application metadata not propagated to the cache tier
- https://github.com/ceph/ceph/pull/40274
- 09:09 AM Bug #49788 (Pending Backport): pool application metadata not propagated to the cache tier
- 01:39 AM Bug #49696: all mons crash suddenly and cann't restart unless close cephx
- Neha Ojha wrote:
> can you share a coredump from the monitor, if the issue is still reproducible?
I'm afraid not....
03/14/2021
- 11:52 AM Bug #49781: unittest_mempool.check_shard_select failed
- https://github.com/ceph/ceph/pull/39978#discussion_r593341155
- 06:14 AM Feature #49789: common/TrackedOp: add op priority for TrackedOp
- PR:https://github.com/ceph/ceph/pull/40060
- 06:12 AM Feature #49789 (Fix Under Review): common/TrackedOp: add op priority for TrackedOp
- Now, we can not know a request priority by ceph daemon /var/run/ceph/ceph-osd.x.asok dump_historic_ops
if this comma... - 04:17 AM Bug #49779 (Fix Under Review): standalone: osd-recovery-scrub.sh: Recovery never started
03/13/2021
- 04:35 PM Bug #49788 (Fix Under Review): pool application metadata not propagated to the cache tier
- 04:27 PM Bug #49788 (Resolved): pool application metadata not propagated to the cache tier
- if you have a base pool with application metadata, that application is not propagated to the cache tier.
This is a... - 09:03 AM Bug #49787 (Resolved): test_envlibrados_for_rocksdb.sh fails on master
- ...
- 08:27 AM Bug #49781: unittest_mempool.check_shard_select failed
- It happened 5 days ago at https://github.com/ceph/ceph/pull/39883#issuecomment-791944956 and is related to https://gi...
- 03:33 AM Bug #49781 (Resolved): unittest_mempool.check_shard_select failed
- This test is probabilistic. Recording to see whether we find it failing more frequently.
From https://jenkins.ceph...
03/12/2021
- 09:36 PM Bug #49696 (Need More Info): all mons crash suddenly and cann't restart unless close cephx
- can you share a coredump from the monitor, if the issue is still reproducible?
- 09:31 PM Bug #49734 (Closed): [OSD]ceph osd crashes and prints Segmentation fault
- Luminous is EOL, please re-open if you see the same issue in later releases.
- 09:00 PM Backport #49775 (In Progress): nautilus: Get more parallel scrubs within osd_max_scrubs limits
- 06:20 PM Backport #49775 (Rejected): nautilus: Get more parallel scrubs within osd_max_scrubs limits
- https://github.com/ceph/ceph/pull/40142
- 08:58 PM Bug #49779 (Resolved): standalone: osd-recovery-scrub.sh: Recovery never started
In master and pacific, the TEST_recovery_scrub_2 subtest in qa/standalone/scrub/osd-recovery-scrub.sh has an interm...- 08:55 PM Backport #49776 (In Progress): octopus: Get more parallel scrubs within osd_max_scrubs limits
- 06:20 PM Backport #49776 (Rejected): octopus: Get more parallel scrubs within osd_max_scrubs limits
- https://github.com/ceph/ceph/pull/40088
- 08:52 PM Backport #49774 (In Progress): pacific: Get more parallel scrubs within osd_max_scrubs limits
- 06:20 PM Backport #49774 (Resolved): pacific: Get more parallel scrubs within osd_max_scrubs limits
- https://github.com/ceph/ceph/pull/40077
- 08:03 PM Bug #49778: mon: promote_standby does not update available_modules
- I think we probably also need a workaround so that we can upgrade from old ceph versions that have this bug...
- 08:00 PM Bug #49778 (Resolved): mon: promote_standby does not update available_modules
- originally observed during upgrade from <15.2.5 via cephadm: the cephadm migration runs immediately after upgrade and...
- 07:46 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
- ...
- 06:53 PM Bug #44595: cache tiering: Error: oid 48 copy_from 493 returned error code -2
- ...
- 06:29 PM Bug #49777 (Resolved): test_pool_min_size: 'check for active or peered' reached maximum tries (5)...
- ...
- 06:20 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
- ...
- 06:19 PM Bug #48843 (Pending Backport): Get more parallel scrubs within osd_max_scrubs limits
- 05:12 PM Bug #47181: "sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120...
- /a/yuriw-2021-03-11_19:01:40-rados-octopus-distro-basic-smithi/5956578/
- 01:59 PM Bug #48959: Primary OSD crash caused corrupted object and further crashes during backfill after s...
- We just ran into this again and had to remove the object to allow the PG to finish backfilling. The similarities betw...
- 01:38 PM Bug #49409: osd run into dead loop and tell slow request when rollback snap with using cache tier
- reopening this ticket, as its fix (https://github.com/ceph/ceph/pull/39593) was reverted as the fix of #49726
- 01:38 PM Bug #49409 (New): osd run into dead loop and tell slow request when rollback snap with using cach...
- 01:37 PM Bug #49726 (Resolved): src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_versio...
- 07:29 AM Bug #49726: src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_version64() == ve...
- created https://github.com/ceph/ceph/pull/40057 as an intermediate fix.
- 12:27 PM Bug #49427 (Resolved): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_...
- 11:50 AM Bug #48505: osdmaptool crush
- hanguang liu wrote:
> when osd map contains CRUSH_ITEM_NONE osd when i run:
> _./osdmaptool ./hkc4 --test-map-pgs-... - 11:44 AM Bug #48505: osdmaptool crush
- hanguang liu wrote:
> when osd map contains CRUSH_ITEM_NONE osd when i run:
> _./osdmaptool ./hkc4 --test-map-pgs-... - 07:26 AM Bug #49758 (Fix Under Review): messages/MOSDPGNotify.h: virtual void MOSDPGNotify::encode_payload...
- 05:37 AM Bug #49754: osd/OSD.cc: ceph_abort_msg("abort() called") during OSD::shutdown()
- ...
03/11/2021
- 11:03 PM Bug #49726: src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_version64() == ve...
- https://github.com/ceph/ceph/pull/39593#issuecomment-792503213 this is where it first showed up, most likely this PR ...
- 02:03 AM Bug #49726: src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_version64() == ve...
- /a/kchai-2021-03-09_12:22:01-rados-wip-kefu-testing-2021-03-09-1847-distro-basic-smithi/5949457
/a/ideepika-2021-03-... - 01:56 AM Bug #49726 (Resolved): src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_versio...
- ...
- 08:19 PM Backport #49054 (Resolved): pacific: pick_a_shard() always select shard 0
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39977
m... - 06:40 PM Backport #49054: pacific: pick_a_shard() always select shard 0
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39977
merged - 08:17 PM Backport #49670: pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo be...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39963
m... - 08:11 PM Backport #49565: pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39844
m... - 08:08 PM Backport #49397 (Resolved): octopus: rados/dashboard: Health check failed: Telemetry requires re-...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39704
m... - 03:59 PM Backport #49397: octopus: rados/dashboard: Health check failed: Telemetry requires re-opt-in (TEL...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39704
merged - 06:56 PM Bug #49758 (Resolved): messages/MOSDPGNotify.h: virtual void MOSDPGNotify::encode_payload(uint64_...
- ...
- 06:45 PM Bug #49754 (New): osd/OSD.cc: ceph_abort_msg("abort() called") during OSD::shutdown()
- ...
- 06:04 PM Bug #47838: mon/test_mon_osdmap_prune.sh: first_pinned != trim_to
- /a/yuriw-2021-03-10_21:08:51-rados-wip-yuri8-testing-2021-03-10-0901-pacific-distro-basic-smithi/5954442 - similar
- 01:31 PM Bug #47380: mon: slow ops due to osd_failure
- an alternative fix: https://github.com/ceph/ceph/pull/40033
- 07:11 AM Bug #49734 (Closed): [OSD]ceph osd crashes and prints Segmentation fault
- This error occurs in Mar 6th, the osd.37 was down and out with bellow log info(ceph-osd.37.log-20210306):
2021-03-... - 07:07 AM Backport #49533 (In Progress): octopus: osd ok-to-stop too conservative
- 03:30 AM Backport #49730 (Resolved): octopus: debian ceph-common package post-inst clobbers ownership of c...
- https://github.com/ceph/ceph/pull/40275
- 03:30 AM Bug #49727: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
Note that instead of a delay you can tell the OSDs to flush their pg stats. I wonder if that flushes to the mon and...- 03:16 AM Bug #49727 (Resolved): lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
This has been seen in cases where all of pool 1 PGs are scrubbed and none of pool 2's. I suggest that this is beca...- 03:30 AM Backport #49729 (Resolved): nautilus: debian ceph-common package post-inst clobbers ownership of ...
- https://github.com/ceph/ceph/pull/40698
- 03:30 AM Backport #49728 (Resolved): pacific: debian ceph-common package post-inst clobbers ownership of c...
- https://github.com/ceph/ceph/pull/40248
- 03:26 AM Backport #49145 (Resolved): pacific: out of order op
- 03:25 AM Bug #49677 (Pending Backport): debian ceph-common package post-inst clobbers ownership of cephadm...
03/10/2021
- 10:41 PM Backport #49682: nautilus: OSD: shutdown of a OSD Host causes slow requests
- please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/40014
ceph-backport.sh versi... - 10:40 PM Backport #49681: octopus: OSD: shutdown of a OSD Host causes slow requests
- please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/40013
ceph-backport.sh versi... - 04:21 PM Bug #23117: PGs stuck in "activating" after osd_max_pg_per_osd_hard_ratio has been exceeded once
- I am aware of one place where we do log withholding pg creation, the following log message in the OSD logs.
https://... - 01:08 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Hey Konstantin and Loïc,
Understood; thanks! - 07:57 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Hi Mauricio,
You are welcome to join the Stable Release team on IRC at #ceph-backports to discuss and resolve the... - 06:47 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Mauricio, just make a backport PR at GitHub, we'll attach it to tracker later.
- 08:54 AM Bug #49697 (Resolved): prime pg temp: unexpected optimization
- I encountered a problem when splitting pgs that eventually cause pg
to be inactived.
I probably think the root reas... - 07:40 AM Bug #49696 (Need More Info): all mons crash suddenly and cann't restart unless close cephx
- crash info
{
"os_version_id": "7",
"utsname_release": "4.14.0jsdx_kernel",
"os_name": "CentOS Linux... - 02:13 AM Backport #49533 (Rejected): octopus: osd ok-to-stop too conservative
- Per Sage
> I'm not sure if this is worth backporting. The primary benefit is faster upgrades, and it's the target ... - 01:24 AM Bug #47419 (Resolved): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo bench...
- 01:24 AM Backport #49670 (Resolved): pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rado...
- 12:02 AM Backport #49565 (Resolved): pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
03/09/2021
- 11:58 PM Backport #49053 (In Progress): octopus: pick_a_shard() always select shard 0
- 11:58 PM Bug #47719: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- https://github.com/ceph/ceph/pull/39844 merged
- 11:57 PM Backport #49054 (In Progress): pacific: pick_a_shard() always select shard 0
- 11:13 PM Backport #49691 (Rejected): pacific: ceph_assert(is_primary()) in PG::scrub()
- 11:10 PM Backport #49691 (Rejected): pacific: ceph_assert(is_primary()) in PG::scrub()
- 11:13 PM Bug #48712 (Resolved): ceph_assert(is_primary()) in PG::scrub()
- 11:09 PM Bug #48712 (Pending Backport): ceph_assert(is_primary()) in PG::scrub()
- 11:09 PM Bug #48712 (Resolved): ceph_assert(is_primary()) in PG::scrub()
- 11:12 PM Backport #49377 (In Progress): pacific: building libcrc32
- 10:55 PM Backport #48985 (In Progress): octopus: ceph osd df tree reporting incorrect SIZE value for rack ...
- 10:26 PM Bug #49689 (Resolved): osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch...
- ...
- 10:23 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- /a/yuriw-2021-03-08_21:03:18-rados-wip-yuri5-testing-2021-03-08-1049-pacific-distro-basic-smithi/5947439
- 10:21 PM Bug #49688 (Can't reproduce): FAILED ceph_assert(is_primary()) in submit_log_entries during Promo...
- ...
- 09:43 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- Samuel Just wrote:
> I'm...not sure what that if block is supposed to do. It was introduced as part of the initial ... - 03:21 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- I'm...not sure what that if block is supposed to do. It was introduced as part of the initial overwrites patch seque...
- 09:31 PM Backport #49670 (In Progress): pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 r...
- https://github.com/ceph/ceph/pull/39963
- 03:45 PM Backport #49670 (Resolved): pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rado...
- https://github.com/ceph/ceph/pull/39963
- 07:46 PM Backport #49683: pacific: OSD: shutdown of a OSD Host causes slow requests
- please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/39957
ceph-backport.sh versi... - 07:35 PM Backport #49683 (Resolved): pacific: OSD: shutdown of a OSD Host causes slow requests
- https://github.com/ceph/ceph/pull/39957
- 07:40 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- Igor, thanks.
I'd like to / can work on submitting the backport PRs, if that's OK.
In the future, if I want to ... - 07:33 PM Bug #46978 (Pending Backport): OSD: shutdown of a OSD Host causes slow requests
- 07:25 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
- The master PR has been merged.
Can someone update Status to Pending Backport, please?
Thanks! - 07:35 PM Backport #49682 (Resolved): nautilus: OSD: shutdown of a OSD Host causes slow requests
- https://github.com/ceph/ceph/pull/40014
- 07:35 PM Backport #49681 (Resolved): octopus: OSD: shutdown of a OSD Host causes slow requests
- https://github.com/ceph/ceph/pull/40013
- 05:57 PM Bug #49677 (Fix Under Review): debian ceph-common package post-inst clobbers ownership of cephadm...
- 05:54 PM Bug #49677 (Resolved): debian ceph-common package post-inst clobbers ownership of cephadm log dirs
- the debian/ubuntu ceph uid is different than the rhel/centos one used by the container. the postinst does a chown -R...
- 04:45 PM Backport #47364 (Resolved): luminous: pgs inconsistent, union_shard_errors=missing
- 03:43 PM Bug #47419 (Pending Backport): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p f...
- https://jenkins.ceph.com/job/ceph-pull-requests/70801/consoleFull#10356408840526d21-3511-427d-909c-dd086c0d1034 - thi...
- 08:32 AM Bug #48786 (Resolved): api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcount2...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 08:32 AM Bug #48984 (Resolved): lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 06:30 AM Backport #49642: pacific: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39938
- 04:11 AM Backport #49641: octopus: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39935
03/08/2021
- 05:16 PM Backport #49482: pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcou...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39773
m... - 05:14 PM Backport #49532: pacific: osd ok-to-stop too conservative
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39737
m... - 05:07 PM Backport #49529 (In Progress): nautilus: "ceph osd crush set|reweight-subtree" commands do not se...
- 05:06 PM Backport #49530 (In Progress): octopus: "ceph osd crush set|reweight-subtree" commands do not set...
- 05:05 PM Backport #49528 (Resolved): pacific: "ceph osd crush set|reweight-subtree" commands do not set we...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39736
m... - 05:02 PM Backport #49526: pacific: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound to class '...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39735
m... - 05:01 PM Backport #49404: pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39597
m... - 04:59 PM Backport #49404: pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
- https://github.com/ceph/ceph/pull/39796
https://github.com/ceph/ceph/pull/39597
(double whammy) - 01:41 PM Backport #49640: nautilus: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39912
- 11:44 AM Bug #49409 (Pending Backport): osd run into dead loop and tell slow request when rollback snap wi...
03/07/2021
- 10:02 PM Backport #49377: pacific: building libcrc32
- please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/39902
ceph-backport.sh versi... - 03:58 PM Backport #49482 (Resolved): pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/Manifes...
- 03:55 PM Backport #49642 (Resolved): pacific: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/40247
- 03:55 PM Backport #49641 (Resolved): octopus: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39935
- 03:55 PM Backport #49640 (Resolved): nautilus: Disable and re-enable clog_to_monitors could trigger assertion
- https://github.com/ceph/ceph/pull/39912
- 03:54 PM Bug #48946 (Pending Backport): Disable and re-enable clog_to_monitors could trigger assertion
03/06/2021
- 02:58 PM Backport #49533 (In Progress): octopus: osd ok-to-stop too conservative
- https://github.com/ceph/ceph/pull/39887
- 02:43 PM Backport #49073 (Resolved): nautilus: crash in Objecter and CRUSH map lookup
- 01:16 AM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- This is where we sent the subops...
03/05/2021
- 11:10 PM Bug #49524: ceph_test_rados_delete_pools_parallel didn't start
- https://tracker.ceph.com/issues/45946 looks very similar
- 11:04 PM Bug #49525: found snap mapper error on pg 3.2s1 oid 3:4abe9991:::smithi10121515-14:e4 snaps missi...
- Ronen, can you check if this is caused due to a race between scrub and snap remove.
- 10:53 PM Bug #49403 (Duplicate): Caught signal (aborted) on mgrmap epoch 1 during librados init (rados-str...
- 07:15 PM Bug #48298: hitting mon_max_pg_per_osd right after creating OSD, then decreases slowly
- Another observation: I have nobackfill set, and I'm currently adding 8 new OSDs.
The first of the newly added OSDs... - 05:15 PM Backport #49482: pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcou...
- Myoungwon Oh wrote:
> https://github.com/ceph/ceph/pull/39773
merged - 02:39 AM Bug #47419 (Resolved): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo bench...
- Hopefully
- 01:49 AM Backport #49565 (In Progress): pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- https://github.com/ceph/ceph/pull/39844
03/04/2021
- 11:34 PM Bug #47419 (Fix Under Review): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p f...
- 11:34 PM Bug #47419 (Duplicate): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo benc...
- 04:33 PM Bug #47419: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo bench 4 write -b...
- https://jenkins.ceph.com/job/ceph-pull-requests/70513/consoleFull#10356408840526d21-3511-427d-909c-dd086c0d1034
- 11:21 PM Bug #49614 (Duplicate): src/test/smoke.sh:56: TEST_multimon: timeout 8 rados -p foo bench 4 writ...
- 11:11 PM Bug #49614: src/test/smoke.sh:56: TEST_multimon: timeout 8 rados -p foo bench 4 write -b 4096 --...
- https://jenkins.ceph.com/job/ceph-pull-requests/70513/consoleFull#-1656021838e840cee4-f4a4-4183-81dd-42855615f2c1
- 10:58 PM Bug #49614 (Duplicate): src/test/smoke.sh:56: TEST_multimon: timeout 8 rados -p foo bench 4 writ...
- ...
- 09:14 PM Bug #44631: ceph pg dump error code 124
- /ceph/teuthology-archive/pdonnell-2021-03-04_03:51:01-fs-wip-pdonnell-testing-20210303.195715-distro-basic-smithi/593...
- 05:39 PM Bug #44631: ceph pg dump error code 124
- /a/yuriw-2021-03-02_20:59:34-rados-wip-yuri7-testing-2021-03-02-1118-nautilus-distro-basic-smithi/5928174
- 09:08 PM Backport #49532 (Resolved): pacific: osd ok-to-stop too conservative
- 06:47 PM Backport #49404 (Resolved): pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
- 06:47 PM Backport #49526 (Resolved): pacific: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound...
- 06:44 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
- /a/sage-2021-03-03_16:41:22-rados-wip-sage2-testing-2021-03-03-0744-pacific-distro-basic-smithi/5930113
- 04:48 PM Bug #23117: PGs stuck in "activating" after osd_max_pg_per_osd_hard_ratio has been exceeded once
- We also his this issue last week on Ceph Version 12.2.11.
Cluster configured with a replication factor of 3, issu... - 01:21 PM Backport #48987: nautilus: ceph osd df tree reporting incorrect SIZE value for rack having an emp...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39126
m...
03/03/2021
- 10:14 PM Bug #49104: crush weirdness: degraded PGs not marked as such, and choose_total_tries = 50 is too ...
- Thanks for the analysis Neha.
Something that perhaps wasn't clear in comment 2 -- in each case where I print the `... - 06:48 PM Bug #49104 (Triaged): crush weirdness: degraded PGs not marked as such, and choose_total_tries = ...
- Thanks for the detailed logs!
Firstly, the pg dump output can sometimes be a little laggy, so I am basing my asses... - 09:53 PM Backport #48987 (Resolved): nautilus: ceph osd df tree reporting incorrect SIZE value for rack ha...
- 04:05 PM Backport #48987: nautilus: ceph osd df tree reporting incorrect SIZE value for rack having an emp...
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39126
merged - 08:47 PM Bug #20909: Error ETIMEDOUT: crush test failed with -110: timed out during smoke test (5 seconds)
- not seen in octopus and pacific so far, but pops sometimes in nautilus:...
- 08:39 PM Bug #49591 (New): no active mgr (MGR_DOWN)" in cluster log
- seen in nautilus...
- 03:37 PM Bug #49584: Ceph OSD, MDS, MGR daemon does not _only_ bind to specified address when configured t...
- After removing the specific public_addr and restarting the MDSes the situation returns to normal and the cluster reco...
- 03:22 PM Bug #49584 (New): Ceph OSD, MDS, MGR daemon does not _only_ bind to specified address when config...
- Documentation (https://docs.ceph.com/en/octopus/rados/configuration/network-config-ref/#ceph-daemons) states the foll...
- 11:32 AM Backport #49055 (Resolved): nautilus: pick_a_shard() always select shard 0
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39651
m... - 09:40 AM Bug #40029: ceph-mon: Caught signal (Aborted) in (CrushWrapper::update_choose_args(CephContext*)+...
- Florian Haas wrote:
> With thanks to Paul Emmerich in https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/threa... - 05:31 AM Bug #48417: unfound EC objects in sepia's LRC after upgrade
- https://tracker.ceph.com/issues/48613#note-13
- 12:28 AM Backport #49404 (In Progress): pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
03/02/2021
- 08:21 PM Bug #37808 (New): osd: osdmap cache weak_refs assert during shutdown
- /ceph/teuthology-archive/pdonnell-2021-03-02_17:29:53-fs:verify-wip-pdonnell-testing-20210301.234318-distro-basic-smi...
- 05:27 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
- personal ref dir: all grep reside in **/home/ideepika/pg[3.1as0.log** in teuthology server
job: /a/teuthology-2021... - 05:24 PM Bug #49572 (Duplicate): MON_DOWN: mon.c fails to join quorum after un-blacklisting mon.a
- This is the same as https://tracker.ceph.com/issues/47654...
- 04:58 PM Bug #49572 (Duplicate): MON_DOWN: mon.c fails to join quorum after un-blacklisting mon.a
- /a/sage-2021-03-01_20:24:37-rados-wip-sage-testing-2021-03-01-1118-distro-basic-smithi/5924612
it looks like the s... - 04:38 AM Backport #49482: pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcou...
- https://github.com/ceph/ceph/pull/39773
03/01/2021
- 11:25 PM Bug #49409 (Fix Under Review): osd run into dead loop and tell slow request when rollback snap wi...
- 10:16 PM Backport #49567 (Resolved): nautilus: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- https://github.com/ceph/ceph/pull/40697
- 10:15 PM Backport #49566 (Resolved): octopus: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- https://github.com/ceph/ceph/pull/40756
- 10:15 PM Backport #49565 (Resolved): pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- https://github.com/ceph/ceph/pull/39844
- 10:10 PM Bug #47719 (Pending Backport): api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
- 05:15 PM Backport #49055: nautilus: pick_a_shard() always select shard 0
- Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39651
merged - 03:43 AM Bug #49543 (New): scrub a pool which size is 1 but found stat mismatch on objects and bytes
the pg has only one primary osd:...
Also available in: Atom