Project

General

Profile

Activity

From 03/01/2021 to 03/30/2021

03/30/2021

11:32 PM Bug #50042: rados/test.sh: api_watch_notify failures
Brad, as per our discussion, I am assigning this to you for now. Neha Ojha
05:27 PM Bug #47654 (Fix Under Review): test_mon_pg: mon fails to join quorum to due election strategy mis...
Greg Farnum

03/29/2021

09:41 PM Bug #49988 (Fix Under Review): Global Recovery Event never completes
Problem was that I did not subtract pgs that I skip because (reported_epoch_of_pg < start_epoch_of_event) from total_... Kamoltat (Junior) Sirivadhna
06:52 PM Bug #50042 (Resolved): rados/test.sh: api_watch_notify failures
... Neha Ojha
09:01 AM Feature #48392: ceph ignores --keyring?
I see this with v15.2.10 as well.
The problem is at the rollback stage, specifically with @osd purge-new@ where it...
Brian Candler
07:40 AM Bug #50032: used space in ceph df reported incorrectly for filestore OSDs
Easily reproducible in Nautilus head indeed. Igor Fedotov
05:36 AM Bug #50032 (New): used space in ceph df reported incorrectly for filestore OSDs
Hi, I noticed, that when filestore OSDs are used, used space is reported incorrectly:
filestore:
POOL ...
Nikola Ciprich
06:23 AM Backport #50018 (In Progress): pacific: Test Failed with: "Scrubbing terminated -- not all pgs we...
Sridhar Seshasayee
04:50 AM Bug #50031 (New): osdc _throttle_op function param type of op_budget int is too small
h2. problem
1. function calc_op_budget return type is int,when indata.length() is out of int range, it will return...
dovefi Z

03/26/2021

09:59 PM Bug #49962: 'sudo ceph --cluster ceph osd crush tunables default' fails due to valgrind: Unknown ...
Radek, can you please help take a look this issue? Neha Ojha
09:53 PM Bug #50004 (Fix Under Review): mon: Modify Paxos trim logic to be more efficient
Neha Ojha
09:36 AM Bug #50004 (Resolved): mon: Modify Paxos trim logic to be more efficient
When the log ingest rate is higher than the trimming rate, monitors may end up consuming a lot of space and become un... Aishwarya Mathuria
08:41 PM Bug #49963 (Resolved): Crash in OSD::ms_fast_dispatch due to call to null vtable function
Sage Weil
06:30 PM Backport #50018 (Resolved): pacific: Test Failed with: "Scrubbing terminated -- not all pgs were ...
https://github.com/ceph/ceph/pull/40461 Backport Bot
06:25 PM Bug #49983 (Pending Backport): Test Failed with: "Scrubbing terminated -- not all pgs were active...
Neha Ojha
05:53 PM Backport #49402: octopus: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40138
merged
Yuri Weinstein
05:51 PM Backport #49530: octopus: "ceph osd crush set|reweight-subtree" commands do not set weight on dev...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39919
merged
Yuri Weinstein
04:20 PM Backport #49640 (In Progress): nautilus: Disable and re-enable clog_to_monitors could trigger ass...
Neha Ojha
03:04 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
/a/yuriw-2021-03-25_20:03:40-rados-wip-yuri8-testing-2021-03-25-1042-pacific-distro-basic-smithi/5999016 Neha Ojha
03:03 PM Bug #45702: PGLog::read_log_and_missing: ceph_assert(miter == missing.get_items().end() || (miter...
/a/yuriw-2021-03-25_20:03:40-rados-wip-yuri8-testing-2021-03-25-1042-pacific-distro-basic-smithi/5999051 Neha Ojha
12:11 PM Bug #50012 (Fix Under Review): Ceph-osd refuses to bind on an IP on the local loopback lo (again)
the original fix was reverted by the fix of #49938. so fix in a more flexible way. Kefu Chai
12:10 PM Bug #50012 (Fix Under Review): Ceph-osd refuses to bind on an IP on the local loopback lo (again)

Kefu Chai
09:58 AM Backport #49977 (Resolved): nautilus: "make check" jenkins job fails
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40407
m...
Loïc Dachary
09:54 AM Backport #49529 (Resolved): nautilus: "ceph osd crush set|reweight-subtree" commands do not set w...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39920
m...
Loïc Dachary
09:51 AM Backport #49976 (Resolved): octopus: "make check" jenkins job fails
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40406
m...
Loïc Dachary
07:41 AM Bug #38219: rebuild-mondb hangs
... Deepika Upadhyay
06:40 AM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
/kchai-2021-03-26_05:32:58-rados-wip-kefu-testing-2021-03-26-1134-distro-basic-smithi/6001105/ Kefu Chai
06:20 AM Backport #49993 (Resolved): octopus: unittest_mempool.check_shard_select failed
https://github.com/ceph/ceph/pull/39978 Backport Bot
06:20 AM Backport #49992 (Resolved): pacific: unittest_mempool.check_shard_select failed
https://github.com/ceph/ceph/pull/40566 Backport Bot
06:20 AM Backport #49991 (Resolved): nautilus: unittest_mempool.check_shard_select failed
https://github.com/ceph/ceph/pull/40567 Backport Bot
06:17 AM Bug #49781 (Pending Backport): unittest_mempool.check_shard_select failed
Kefu Chai

03/25/2021

11:20 PM Bug #49988 (Resolved): Global Recovery Event never completes
... Sage Weil
10:12 PM Support #49847: OSD Fails to init after upgrading to octopus: _deferred_replay failed to decode d...
Contrary to what I stated previously this does not seem like a software issue. The root cause was probably faulty RAM... Eetu Lampsijärvi
07:26 PM Bug #42884: OSDMapTest.CleanPGUpmaps failure
https://jenkins.ceph.com/job/ceph-pull-requests/72185/consoleFull#-108728127277933967-90d1-4877-8d60-89cb08ef4eb1 Neha Ojha
04:44 PM Bug #49983 (Fix Under Review): Test Failed with: "Scrubbing terminated -- not all pgs were active...
Sridhar Seshasayee
02:42 PM Bug #49983 (Resolved): Test Failed with: "Scrubbing terminated -- not all pgs were active and cle...
Test Run:
https://pulpito.ceph.com/nojha-2021-03-23_23:04:33-rados-wip-40323-2-distro-basic-gibba/5991116/
Failur...
Sridhar Seshasayee
11:50 AM Backport #49977 (In Progress): nautilus: "make check" jenkins job fails
Kefu Chai
11:25 AM Backport #49977 (Resolved): nautilus: "make check" jenkins job fails
https://github.com/ceph/ceph/pull/40407 Backport Bot
11:45 AM Backport #49976 (In Progress): octopus: "make check" jenkins job fails
Kefu Chai
11:25 AM Backport #49976 (Resolved): octopus: "make check" jenkins job fails
https://github.com/ceph/ceph/pull/40406 Backport Bot
11:44 AM Backport #49978 (In Progress): pacific: "make check" jenkins job fails
Kefu Chai
11:25 AM Backport #49978 (Resolved): pacific: "make check" jenkins job fails
https://github.com/ceph/ceph/pull/40405 Backport Bot
11:22 AM Bug #49975 (Resolved): "make check" jenkins job fails
... Kefu Chai
09:58 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
Brad Hubbard wrote:
> I think I've found an issue related to this message in the logs but I'll need to test that the...
Christian Rohmann
05:32 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
I think I've found an issue related to this message in the logs but I'll need to test that theory.... Brad Hubbard
04:20 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
Hi Christian,
Is there no message in syslog or dmesg about the segfault at all? That seems odd. Maybe you should c...
Brad Hubbard
06:58 AM Bug #49970 (New): cmake: use GooleTest for adding gtest based tests
see https://github.com/Kitware/CMake/blob/master/Modules/GoogleTest.cmake
for finer grained management of test,
Kefu Chai
04:55 AM Bug #49963 (Fix Under Review): Crash in OSD::ms_fast_dispatch due to call to null vtable function
Kefu Chai

03/24/2021

11:43 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
So the issue is we never should have gone this deep into the OSD::ms_fast_dispatch function.
Previously we would h...
Brad Hubbard
10:53 PM Bug #49963 (Triaged): Crash in OSD::ms_fast_dispatch due to call to null vtable function
Neha Ojha
10:50 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
/a/sage-2021-03-24_06:13:24-upgrade:octopus-x-wip-sage-testing-2021-03-23-2309-distro-basic-smithi/5993446... Neha Ojha
10:46 PM Bug #49963: Crash in OSD::ms_fast_dispatch due to call to null vtable function
m->header.type is 85... Brad Hubbard
10:38 PM Bug #49963 (Resolved): Crash in OSD::ms_fast_dispatch due to call to null vtable function
/a/sage-2021-03-24_06:13:24-upgrade:octopus-x-wip-sage-testing-2021-03-23-2309-distro-basic-smithi/5993446... Brad Hubbard
09:09 PM Bug #49962 (Resolved): 'sudo ceph --cluster ceph osd crush tunables default' fails due to valgrin...
... Neha Ojha
07:25 PM Bug #49961 (New): scrub/osd-recovery-scrub.sh: TEST_recovery_scrub_1 failed
... Neha Ojha
12:21 PM Bug #49487 (Resolved): osd:scrub skip some pg
Kefu Chai
09:12 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
Brad Hubbard wrote:
> Hi Christian,
>
> From the log this was a segfault in MonitorDBStore::get_synchronizer.
> ...
Christian Rohmann
02:51 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
So the offset into MonitorDBStore::get_synchronizer where we segfaulted was
0x3f.
If we set up a binary compatibl...
Brad Hubbard
02:44 AM Documentation #35967 (Resolved): [doc] sync documentation "OSD Config Reference" default values w...
Kefu Chai

03/23/2021

11:25 PM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
Hi Christian,
From the log this was a segfault in MonitorDBStore::get_synchronizer.
2021-03-14 18:00:28.679 7f1...
Brad Hubbard
08:45 AM Bug #49809: 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
Neha Ojha wrote:
> Do you happen to have a coredump for this or can you provide a copy of the mon store db?
Unfor...
Christian Rohmann
07:12 PM Bug #49781: unittest_mempool.check_shard_select failed
https://jenkins.ceph.com/job/ceph-pull-requests/72011/consoleFull#-85362248744e9240e-b50a-4693-bac0-8a991bac86ac Neha Ojha
04:55 PM Backport #49529: nautilus: "ceph osd crush set|reweight-subtree" commands do not set weight on de...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39920
merged
Yuri Weinstein
12:23 AM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
... Neha Ojha

03/22/2021

10:21 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
relevant osd.3 logs from yuriw-2021-03-19_00:00:55-rados-wip-yuri8-testing-2021-03-18-1502-pacific-distro-basic-smith... Neha Ojha
09:40 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
/a/yuriw-2021-03-19_00:00:55-rados-wip-yuri8-testing-2021-03-18-1502-pacific-distro-basic-smithi/5978982 Neha Ojha
02:47 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
Thanks, Konstantin! Mauricio Oliveira
02:46 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
@Mauricio, I was update issue backports and status. Konstantin Shalygin
02:45 PM Bug #46978 (Pending Backport): OSD: shutdown of a OSD Host causes slow requests
Konstantin Shalygin
01:44 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
Hi @singuliere _,
Could you please revert the backport field to include Octopus and Nautilus?
Such backports ha...
Mauricio Oliveira
07:40 AM Feature #49789 (Fix Under Review): common/TrackedOp: add op priority for TrackedOp
Kefu Chai
07:15 AM Backport #49919 (Resolved): nautilus: mon: slow ops due to osd_failure
https://github.com/ceph/ceph/pull/41213 Backport Bot
07:15 AM Backport #49918 (Resolved): pacific: mon: slow ops due to osd_failure
https://github.com/ceph/ceph/pull/41090 Backport Bot
07:15 AM Backport #49917 (Resolved): octopus: mon: slow ops due to osd_failure
https://github.com/ceph/ceph/pull/40558 Backport Bot
07:13 AM Bug #47380 (Pending Backport): mon: slow ops due to osd_failure
Kefu Chai

03/21/2021

06:25 PM Backport #49642 (Resolved): pacific: Disable and re-enable clog_to_monitors could trigger assertion
Sage Weil
06:24 PM Backport #49008 (Resolved): pacific: osd crash in OSD::heartbeat when dereferencing null session
Sage Weil
06:22 PM Backport #49728 (Resolved): pacific: debian ceph-common package post-inst clobbers ownership of c...
Sage Weil
05:46 PM Bug #49883 (Rejected): librados: hang in RadosClient::wait_for_osdmap
This turned out to be because the auth credential had wrong permissions (due to a profile name change). Patrick Donnelly
02:33 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
/a/sage-2021-03-20_15:11:51-rados-wip-sage2-testing-2021-03-20-0832-pacific-distro-basic-smithi/5983920
Sage Weil
01:20 PM Backport #49911 (Resolved): pacific: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get...
Backport Bot
01:18 PM Bug #49427 (Pending Backport): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missi...
need this in pacific too: /a/sage-2021-03-20_15:11:51-rados-wip-sage2-testing-2021-03-20-0832-pacific-distro-basic-sm... Sage Weil

03/20/2021

01:45 PM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
/a/sage-2021-03-20_01:11:45-rados-wip-sage-testing-2021-03-19-1647-distro-basic-smithi/5982282
with logs!
Sage Weil
08:06 AM Backport #48596 (In Progress): octopus: nautilus: qa/standalone/scrub/osd-scrub-test.sh: _scrub_a...
singuliere _
08:05 AM Backport #49009 (In Progress): octopus: osd crash in OSD::heartbeat when dereferencing null session
singuliere _
08:04 AM Backport #49527 (In Progress): octopus: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bo...
singuliere _
07:56 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
Since this issue is resolved and only the pacific backport was done, I assume it means the octopus & nautilus backpor... singuliere _
07:46 AM Backport #49730 (In Progress): octopus: debian ceph-common package post-inst clobbers ownership o...
singuliere _
07:46 AM Backport #49795 (In Progress): octopus: pool application metadata not propagated to the cache tier
singuliere _

03/19/2021

09:47 PM Bug #49809 (Need More Info): 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
Do you happen to have a coredump for this or can you provide a copy of the mon store db? Neha Ojha
09:13 PM Backport #49895: pacific: osd/scrub - a ceph_assert() in a legitimate path
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/40226
merged
Yuri Weinstein
06:08 PM Backport #49796 (Resolved): pacific: pool application metadata not propagated to the cache tier
singuliere _
06:02 PM Backport #49728 (In Progress): pacific: debian ceph-common package post-inst clobbers ownership o...
singuliere _
05:58 PM Backport #49642 (In Progress): pacific: Disable and re-enable clog_to_monitors could trigger asse...
singuliere _
05:57 PM Backport #49008 (In Progress): pacific: osd crash in OSD::heartbeat when dereferencing null session
singuliere _
05:06 PM Bug #49902 (New): monitor store.db directory has been increasing and has exceeded 450GB
My monitor keeps growing. After it exceeds 300GB, it cannot be started because the disk is not enough. After the spac... Hughen X
11:07 AM Backport #49320 (Resolved): octopus: thrash_cache_writeback_proxy_none: FAILED ceph_assert(versio...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39578
m...
Loïc Dachary
09:50 AM Backport #49641 (Resolved): octopus: Disable and re-enable clog_to_monitors could trigger assertion
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39935
m...
Loïc Dachary
09:48 AM Backport #49533 (Resolved): octopus: osd ok-to-stop too conservative
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39887
m...
Loïc Dachary

03/18/2021

10:42 PM Bug #49894: set a non-zero default value for osd_client_message_cap
Neha Ojha wrote:
> The current default of 0 doesn't help and we've tried setting it to 5000 for one of our tests htt...
Neha Ojha
09:26 PM Bug #49894 (Resolved): set a non-zero default value for osd_client_message_cap
The current default of 0 doesn't help and we've tried setting it to 5000 for one of our tests https://github.com/neha... Neha Ojha
09:57 PM Backport #49895 (In Progress): pacific: osd/scrub - a ceph_assert() in a legitimate path
Neha Ojha
09:55 PM Backport #49895 (Resolved): pacific: osd/scrub - a ceph_assert() in a legitimate path
https://github.com/ceph/ceph/pull/40226 Backport Bot
09:54 PM Bug #49867 (Pending Backport): osd/scrub - a ceph_assert() in a legitimate path
Neha Ojha
05:57 PM Bug #49888 (Resolved): rados/singleton: radosbench.py: teuthology.exceptions.MaxWhileTries: reach...
... Neha Ojha
04:49 PM Support #49847: OSD Fails to init after upgrading to octopus: _deferred_replay failed to decode d...
Ended up nuking the OSD & letting it recover - this workaround "solves" the problem for me; feel free to close the is... Eetu Lampsijärvi
04:30 PM Backport #49641: octopus: Disable and re-enable clog_to_monitors could trigger assertion
gerald yang wrote:
> https://github.com/ceph/ceph/pull/39935
merged
Yuri Weinstein
04:27 PM Backport #49533: octopus: osd ok-to-stop too conservative
https://github.com/ceph/ceph/pull/39887 merged Yuri Weinstein
04:24 PM Bug #46323: thrash_cache_writeback_proxy_none: FAILED ceph_assert(version == old_value.version) i...
https://github.com/ceph/ceph/pull/39578 merged Yuri Weinstein
03:03 PM Bug #49883: librados: hang in RadosClient::wait_for_osdmap
The ubuntu jobs are failing because the ceph-mgr can't find libcephsqlite.so. There were packaging changes that proba... Patrick Donnelly
02:34 PM Bug #49883 (Rejected): librados: hang in RadosClient::wait_for_osdmap
https://pulpito.ceph.com/pdonnell-2021-03-18_06:53:09-rados:basic-wip-pdonnell-testing-20210318.035227-distro-basic-s... Patrick Donnelly
01:54 PM Backport #49401: pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40137
m...
Nathan Cutler
01:54 PM Backport #49401: pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cl...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40137
m...
Nathan Cutler
01:54 PM Backport #49817 (Resolved): pacific: mon: promote_standby does not update available_modules
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/40132
m...
Nathan Cutler

03/17/2021

09:19 PM Backport #49401 (Resolved): pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MON_D...
Sage Weil
06:24 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
I think we should revert this in pacific https://github.com/ceph/ceph/pull/40195, until we can fix the test failures. Neha Ojha
04:02 PM Bug #49868: RuntimeError: Exiting scrub checking -- not all pgs scrubbed
Something merged after 37f9d0a25d06a6b8529aa350110eba930fba8c9e since https://pulpito.ceph.com/yuriw-2021-03-15_23:42... Neha Ojha
03:57 PM Bug #49868 (New): RuntimeError: Exiting scrub checking -- not all pgs scrubbed
... Neha Ojha
03:28 PM Bug #49867 (Fix Under Review): osd/scrub - a ceph_assert() in a legitimate path
Kefu Chai
03:15 PM Bug #49867 (Resolved): osd/scrub - a ceph_assert() in a legitimate path
in pg_scrubber.cc, in PgScrubber::on_replica_init() there's a ceph_assert()
that might trigger in a correct path (if...
Ronen Friedman
10:14 AM Backport #49683: pacific: OSD: shutdown of a OSD Host causes slow requests
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39957
m...
Nathan Cutler
09:05 AM Bug #49781: unittest_mempool.check_shard_select failed
Discussion on the mailing list about "mempool optimization":https://lists.ceph.io/hyperkitty/list/dev@ceph.io/thread/... Loïc Dachary

03/16/2021

08:07 PM Support #49847 (Closed): OSD Fails to init after upgrading to octopus: _deferred_replay failed to...
An OSD fails to start after upgrading from mimic 13.2.2 to octopus 15.2.9.
It seems like first bluestore fails at...
Eetu Lampsijärvi
03:45 PM Bug #49832 (New): Segmentation fault: in thread_name:ms_dispatch
... Deepika Upadhyay
03:22 PM Bug #49781: unittest_mempool.check_shard_select failed
The test condition should not be too strict because there really is no way to predict the result. It is however good ... Loïc Dachary
12:56 PM Bug #49781: unittest_mempool.check_shard_select failed
Using "pthread_self for sharding":https://github.com/ceph/ceph/blob/master/src/include/mempool.h#L261-L262 is not gre... Loïc Dachary
11:25 AM Bug #49781 (In Progress): unittest_mempool.check_shard_select failed
Loïc Dachary
08:15 AM Bug #49697: prime pg temp: unexpected optimization
ping fan chen
08:14 AM Bug #49787 (Resolved): test_envlibrados_for_rocksdb.sh fails on master
Kefu Chai
06:28 AM Backport #49682 (In Progress): nautilus: OSD: shutdown of a OSD Host causes slow requests
Konstantin Shalygin

03/15/2021

10:42 PM Bug #46978 (Resolved): OSD: shutdown of a OSD Host causes slow requests
Sage Weil
10:42 PM Backport #49683 (Resolved): pacific: OSD: shutdown of a OSD Host causes slow requests
Sage Weil
10:41 PM Backport #49774 (Resolved): pacific: Get more parallel scrubs within osd_max_scrubs limits
Sage Weil
09:56 PM Backport #49402 (In Progress): octopus: rados: Health check failed: 1/3 mons down, quorum a,c (MO...
Neha Ojha
09:55 PM Backport #49401 (In Progress): pacific: rados: Health check failed: 1/3 mons down, quorum a,c (MO...
Neha Ojha
08:15 PM Backport #49817 (Resolved): pacific: mon: promote_standby does not update available_modules
https://github.com/ceph/ceph/pull/40132 Backport Bot
08:15 PM Backport #49816 (Resolved): octopus: mon: promote_standby does not update available_modules
https://github.com/ceph/ceph/pull/40757 Backport Bot
08:11 PM Bug #49778 (Pending Backport): mon: promote_standby does not update available_modules
Sage Weil
05:26 PM Bug #49810 (Need More Info): rados/singleton: with msgr-failures/none MON_DOWN due to haven't for...
... Neha Ojha
05:16 PM Bug #49809 (Need More Info): 1 out of 3 mon crashed in MonitorDBStore::get_synchronizer
We experienced a single mon crash (out of 3 mons) - We observed no other issues on the machine or the cluster.
I a...
Christian Rohmann
03:02 PM Bug #48793 (Resolved): out of order op
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Nathan Cutler
03:02 PM Bug #48990 (Resolved): rados/dashboard: Health check failed: Telemetry requires re-opt-in (TELEME...
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Nathan Cutler
10:38 AM Bug #49781: unittest_mempool.check_shard_select failed
master also... Kefu Chai
09:38 AM Bug #49779 (Resolved): standalone: osd-recovery-scrub.sh: Recovery never started
Kefu Chai
09:22 AM Bug #49758 (Resolved): messages/MOSDPGNotify.h: virtual void MOSDPGNotify::encode_payload(uint64_...
Kefu Chai
09:10 AM Backport #49796 (Resolved): pacific: pool application metadata not propagated to the cache tier
https://github.com/ceph/ceph/pull/40119 Backport Bot
09:10 AM Backport #49795 (Resolved): octopus: pool application metadata not propagated to the cache tier
https://github.com/ceph/ceph/pull/40274 Backport Bot
09:09 AM Bug #49788 (Pending Backport): pool application metadata not propagated to the cache tier
Kefu Chai
01:39 AM Bug #49696: all mons crash suddenly and cann't restart unless close cephx
Neha Ojha wrote:
> can you share a coredump from the monitor, if the issue is still reproducible?
I'm afraid not....
wencong wan

03/14/2021

11:52 AM Bug #49781: unittest_mempool.check_shard_select failed
https://github.com/ceph/ceph/pull/39978#discussion_r593341155 singuliere _
06:14 AM Feature #49789: common/TrackedOp: add op priority for TrackedOp
PR:https://github.com/ceph/ceph/pull/40060 yite gu
06:12 AM Feature #49789 (Fix Under Review): common/TrackedOp: add op priority for TrackedOp
Now, we can not know a request priority by ceph daemon /var/run/ceph/ceph-osd.x.asok dump_historic_ops
if this comma...
yite gu
04:17 AM Bug #49779 (Fix Under Review): standalone: osd-recovery-scrub.sh: Recovery never started
Kefu Chai

03/13/2021

04:35 PM Bug #49788 (Fix Under Review): pool application metadata not propagated to the cache tier
Sage Weil
04:27 PM Bug #49788 (Resolved): pool application metadata not propagated to the cache tier
if you have a base pool with application metadata, that application is not propagated to the cache tier.
This is a...
Sage Weil
09:03 AM Bug #49787 (Resolved): test_envlibrados_for_rocksdb.sh fails on master
... Kefu Chai
08:27 AM Bug #49781: unittest_mempool.check_shard_select failed
It happened 5 days ago at https://github.com/ceph/ceph/pull/39883#issuecomment-791944956 and is related to https://gi... Loïc Dachary
03:33 AM Bug #49781 (Resolved): unittest_mempool.check_shard_select failed
This test is probabilistic. Recording to see whether we find it failing more frequently.
From https://jenkins.ceph...
Josh Durgin

03/12/2021

09:36 PM Bug #49696 (Need More Info): all mons crash suddenly and cann't restart unless close cephx
can you share a coredump from the monitor, if the issue is still reproducible? Neha Ojha
09:31 PM Bug #49734 (Closed): [OSD]ceph osd crashes and prints Segmentation fault
Luminous is EOL, please re-open if you see the same issue in later releases. Neha Ojha
09:00 PM Backport #49775 (In Progress): nautilus: Get more parallel scrubs within osd_max_scrubs limits
David Zafman
06:20 PM Backport #49775 (Rejected): nautilus: Get more parallel scrubs within osd_max_scrubs limits
https://github.com/ceph/ceph/pull/40142 Backport Bot
08:58 PM Bug #49779 (Resolved): standalone: osd-recovery-scrub.sh: Recovery never started

In master and pacific, the TEST_recovery_scrub_2 subtest in qa/standalone/scrub/osd-recovery-scrub.sh has an interm...
David Zafman
08:55 PM Backport #49776 (In Progress): octopus: Get more parallel scrubs within osd_max_scrubs limits
David Zafman
06:20 PM Backport #49776 (Rejected): octopus: Get more parallel scrubs within osd_max_scrubs limits
https://github.com/ceph/ceph/pull/40088 Backport Bot
08:52 PM Backport #49774 (In Progress): pacific: Get more parallel scrubs within osd_max_scrubs limits
David Zafman
06:20 PM Backport #49774 (Resolved): pacific: Get more parallel scrubs within osd_max_scrubs limits
https://github.com/ceph/ceph/pull/40077 Backport Bot
08:03 PM Bug #49778: mon: promote_standby does not update available_modules
I think we probably also need a workaround so that we can upgrade from old ceph versions that have this bug... Sage Weil
08:00 PM Bug #49778 (Resolved): mon: promote_standby does not update available_modules
originally observed during upgrade from <15.2.5 via cephadm: the cephadm migration runs immediately after upgrade and... Sage Weil
07:46 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
... Deepika Upadhyay
06:53 PM Bug #44595: cache tiering: Error: oid 48 copy_from 493 returned error code -2
... Deepika Upadhyay
06:29 PM Bug #49777 (Resolved): test_pool_min_size: 'check for active or peered' reached maximum tries (5)...
... Deepika Upadhyay
06:20 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
... Deepika Upadhyay
06:19 PM Bug #48843 (Pending Backport): Get more parallel scrubs within osd_max_scrubs limits
David Zafman
05:12 PM Bug #47181: "sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120...
/a/yuriw-2021-03-11_19:01:40-rados-octopus-distro-basic-smithi/5956578/ Neha Ojha
01:59 PM Bug #48959: Primary OSD crash caused corrupted object and further crashes during backfill after s...
We just ran into this again and had to remove the object to allow the PG to finish backfilling. The similarities betw... Tom Byrne
01:38 PM Bug #49409: osd run into dead loop and tell slow request when rollback snap with using cache tier
reopening this ticket, as its fix (https://github.com/ceph/ceph/pull/39593) was reverted as the fix of #49726 Kefu Chai
01:38 PM Bug #49409 (New): osd run into dead loop and tell slow request when rollback snap with using cach...
Kefu Chai
01:37 PM Bug #49726 (Resolved): src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_versio...
Kefu Chai
07:29 AM Bug #49726: src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_version64() == ve...
created https://github.com/ceph/ceph/pull/40057 as an intermediate fix. Kefu Chai
12:27 PM Bug #49427 (Resolved): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_...
Kefu Chai
11:50 AM Bug #48505: osdmaptool crush
hanguang liu wrote:
> when osd map contains CRUSH_ITEM_NONE osd when i run:
> _./osdmaptool ./hkc4 --test-map-pgs-...
hg liu
11:44 AM Bug #48505: osdmaptool crush
hanguang liu wrote:
> when osd map contains CRUSH_ITEM_NONE osd when i run:
> _./osdmaptool ./hkc4 --test-map-pgs-...
hg liu
07:26 AM Bug #49758 (Fix Under Review): messages/MOSDPGNotify.h: virtual void MOSDPGNotify::encode_payload...
Kefu Chai
05:37 AM Bug #49754: osd/OSD.cc: ceph_abort_msg("abort() called") during OSD::shutdown()
... Kefu Chai

03/11/2021

11:03 PM Bug #49726: src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_version64() == ve...
https://github.com/ceph/ceph/pull/39593#issuecomment-792503213 this is where it first showed up, most likely this PR ... Neha Ojha
02:03 AM Bug #49726: src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_version64() == ve...
/a/kchai-2021-03-09_12:22:01-rados-wip-kefu-testing-2021-03-09-1847-distro-basic-smithi/5949457
/a/ideepika-2021-03-...
Neha Ojha
01:56 AM Bug #49726 (Resolved): src/test/osd/RadosModel.h: FAILED ceph_assert(!version || comp->get_versio...
... Neha Ojha
08:19 PM Backport #49054 (Resolved): pacific: pick_a_shard() always select shard 0
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39977
m...
Nathan Cutler
06:40 PM Backport #49054: pacific: pick_a_shard() always select shard 0
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39977
merged
Yuri Weinstein
08:17 PM Backport #49670: pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo be...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39963
m...
Nathan Cutler
08:11 PM Backport #49565: pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39844
m...
Nathan Cutler
08:08 PM Backport #49397 (Resolved): octopus: rados/dashboard: Health check failed: Telemetry requires re-...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39704
m...
Nathan Cutler
03:59 PM Backport #49397: octopus: rados/dashboard: Health check failed: Telemetry requires re-opt-in (TEL...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39704
merged
Yuri Weinstein
06:56 PM Bug #49758 (Resolved): messages/MOSDPGNotify.h: virtual void MOSDPGNotify::encode_payload(uint64_...
... Neha Ojha
06:45 PM Bug #49754 (New): osd/OSD.cc: ceph_abort_msg("abort() called") during OSD::shutdown()
... Neha Ojha
06:04 PM Bug #47838: mon/test_mon_osdmap_prune.sh: first_pinned != trim_to
/a/yuriw-2021-03-10_21:08:51-rados-wip-yuri8-testing-2021-03-10-0901-pacific-distro-basic-smithi/5954442 - similar Neha Ojha
01:31 PM Bug #47380: mon: slow ops due to osd_failure
an alternative fix: https://github.com/ceph/ceph/pull/40033 Kefu Chai
07:11 AM Bug #49734 (Closed): [OSD]ceph osd crashes and prints Segmentation fault
This error occurs in Mar 6th, the osd.37 was down and out with bellow log info(ceph-osd.37.log-20210306):
2021-03-...
文军 丁
07:07 AM Backport #49533 (In Progress): octopus: osd ok-to-stop too conservative
Kefu Chai
03:30 AM Backport #49730 (Resolved): octopus: debian ceph-common package post-inst clobbers ownership of c...
https://github.com/ceph/ceph/pull/40275 Backport Bot
03:30 AM Bug #49727: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs

Note that instead of a delay you can tell the OSDs to flush their pg stats. I wonder if that flushes to the mon and...
David Zafman
03:16 AM Bug #49727 (Resolved): lazy_omap_stats_test: "ceph osd deep-scrub all" hangs

This has been seen in cases where all of pool 1 PGs are scrubbed and none of pool 2's. I suggest that this is beca...
David Zafman
03:30 AM Backport #49729 (Resolved): nautilus: debian ceph-common package post-inst clobbers ownership of ...
https://github.com/ceph/ceph/pull/40698 Backport Bot
03:30 AM Backport #49728 (Resolved): pacific: debian ceph-common package post-inst clobbers ownership of c...
https://github.com/ceph/ceph/pull/40248 Backport Bot
03:26 AM Backport #49145 (Resolved): pacific: out of order op
Kefu Chai
03:25 AM Bug #49677 (Pending Backport): debian ceph-common package post-inst clobbers ownership of cephadm...
Kefu Chai

03/10/2021

10:41 PM Backport #49682: nautilus: OSD: shutdown of a OSD Host causes slow requests
please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/40014
ceph-backport.sh versi...
Mauricio Oliveira
10:40 PM Backport #49681: octopus: OSD: shutdown of a OSD Host causes slow requests
please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/40013
ceph-backport.sh versi...
Mauricio Oliveira
04:21 PM Bug #23117: PGs stuck in "activating" after osd_max_pg_per_osd_hard_ratio has been exceeded once
I am aware of one place where we do log withholding pg creation, the following log message in the OSD logs.
https://...
Vikhyat Umrao
01:08 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
Hey Konstantin and Loïc,
Understood; thanks!
Mauricio Oliveira
07:57 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
Hi Mauricio,
You are welcome to join the Stable Release team on IRC at #ceph-backports to discuss and resolve the...
Loïc Dachary
06:47 AM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
Mauricio, just make a backport PR at GitHub, we'll attach it to tracker later. Konstantin Shalygin
08:54 AM Bug #49697 (Resolved): prime pg temp: unexpected optimization
I encountered a problem when splitting pgs that eventually cause pg
to be inactived.
I probably think the root reas...
fan chen
07:40 AM Bug #49696 (Need More Info): all mons crash suddenly and cann't restart unless close cephx
crash info
{
"os_version_id": "7",
"utsname_release": "4.14.0jsdx_kernel",
"os_name": "CentOS Linux...
wencong wan
02:13 AM Backport #49533 (Rejected): octopus: osd ok-to-stop too conservative
Per Sage
> I'm not sure if this is worth backporting. The primary benefit is faster upgrades, and it's the target ...
Kefu Chai
01:24 AM Bug #47419 (Resolved): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo bench...
Neha Ojha
01:24 AM Backport #49670 (Resolved): pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rado...
Neha Ojha
12:02 AM Backport #49565 (Resolved): pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
singuliere _

03/09/2021

11:58 PM Backport #49053 (In Progress): octopus: pick_a_shard() always select shard 0
singuliere _
11:58 PM Bug #47719: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
https://github.com/ceph/ceph/pull/39844 merged Yuri Weinstein
11:57 PM Backport #49054 (In Progress): pacific: pick_a_shard() always select shard 0
singuliere _
11:13 PM Backport #49691 (Rejected): pacific: ceph_assert(is_primary()) in PG::scrub()
David Zafman
11:10 PM Backport #49691 (Rejected): pacific: ceph_assert(is_primary()) in PG::scrub()
Backport Bot
11:13 PM Bug #48712 (Resolved): ceph_assert(is_primary()) in PG::scrub()
David Zafman
11:09 PM Bug #48712 (Pending Backport): ceph_assert(is_primary()) in PG::scrub()
David Zafman
11:09 PM Bug #48712 (Resolved): ceph_assert(is_primary()) in PG::scrub()
David Zafman
11:12 PM Backport #49377 (In Progress): pacific: building libcrc32
singuliere _
10:55 PM Backport #48985 (In Progress): octopus: ceph osd df tree reporting incorrect SIZE value for rack ...
Brad Hubbard
10:26 PM Bug #49689 (Resolved): osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch...
... Neha Ojha
10:23 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
/a/yuriw-2021-03-08_21:03:18-rados-wip-yuri5-testing-2021-03-08-1049-pacific-distro-basic-smithi/5947439 Neha Ojha
10:21 PM Bug #49688 (Can't reproduce): FAILED ceph_assert(is_primary()) in submit_log_entries during Promo...
... Neha Ojha
09:43 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
Samuel Just wrote:
> I'm...not sure what that if block is supposed to do. It was introduced as part of the initial ...
Neha Ojha
03:21 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
I'm...not sure what that if block is supposed to do. It was introduced as part of the initial overwrites patch seque... Samuel Just
09:31 PM Backport #49670 (In Progress): pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 r...
https://github.com/ceph/ceph/pull/39963 Neha Ojha
03:45 PM Backport #49670 (Resolved): pacific: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rado...
https://github.com/ceph/ceph/pull/39963 Backport Bot
07:46 PM Backport #49683: pacific: OSD: shutdown of a OSD Host causes slow requests
please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/39957
ceph-backport.sh versi...
Mauricio Oliveira
07:35 PM Backport #49683 (Resolved): pacific: OSD: shutdown of a OSD Host causes slow requests
https://github.com/ceph/ceph/pull/39957 Backport Bot
07:40 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
Igor, thanks.
I'd like to / can work on submitting the backport PRs, if that's OK.
In the future, if I want to ...
Mauricio Oliveira
07:33 PM Bug #46978 (Pending Backport): OSD: shutdown of a OSD Host causes slow requests
Igor Fedotov
07:25 PM Bug #46978: OSD: shutdown of a OSD Host causes slow requests
The master PR has been merged.
Can someone update Status to Pending Backport, please?
Thanks!
Mauricio Oliveira
07:35 PM Backport #49682 (Resolved): nautilus: OSD: shutdown of a OSD Host causes slow requests
https://github.com/ceph/ceph/pull/40014 Backport Bot
07:35 PM Backport #49681 (Resolved): octopus: OSD: shutdown of a OSD Host causes slow requests
https://github.com/ceph/ceph/pull/40013 Backport Bot
05:57 PM Bug #49677 (Fix Under Review): debian ceph-common package post-inst clobbers ownership of cephadm...
Sage Weil
05:54 PM Bug #49677 (Resolved): debian ceph-common package post-inst clobbers ownership of cephadm log dirs
the debian/ubuntu ceph uid is different than the rhel/centos one used by the container. the postinst does a chown -R... Sage Weil
04:45 PM Backport #47364 (Resolved): luminous: pgs inconsistent, union_shard_errors=missing
Nathan Cutler
03:43 PM Bug #47419 (Pending Backport): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p f...
https://jenkins.ceph.com/job/ceph-pull-requests/70801/consoleFull#10356408840526d21-3511-427d-909c-dd086c0d1034 - thi... Neha Ojha
08:32 AM Bug #48786 (Resolved): api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcount2...
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Nathan Cutler
08:32 AM Bug #48984 (Resolved): lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Nathan Cutler
06:30 AM Backport #49642: pacific: Disable and re-enable clog_to_monitors could trigger assertion
https://github.com/ceph/ceph/pull/39938 gerald yang
04:11 AM Backport #49641: octopus: Disable and re-enable clog_to_monitors could trigger assertion
https://github.com/ceph/ceph/pull/39935 gerald yang

03/08/2021

05:16 PM Backport #49482: pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcou...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39773
m...
Nathan Cutler
05:14 PM Backport #49532: pacific: osd ok-to-stop too conservative
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39737
m...
Nathan Cutler
05:07 PM Backport #49529 (In Progress): nautilus: "ceph osd crush set|reweight-subtree" commands do not se...
Nathan Cutler
05:06 PM Backport #49530 (In Progress): octopus: "ceph osd crush set|reweight-subtree" commands do not set...
Nathan Cutler
05:05 PM Backport #49528 (Resolved): pacific: "ceph osd crush set|reweight-subtree" commands do not set we...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39736
m...
Nathan Cutler
05:02 PM Backport #49526: pacific: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound to class '...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39735
m...
Nathan Cutler
05:01 PM Backport #49404: pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39597
m...
Nathan Cutler
04:59 PM Backport #49404: pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
https://github.com/ceph/ceph/pull/39796
https://github.com/ceph/ceph/pull/39597
(double whammy)
Nathan Cutler
01:41 PM Backport #49640: nautilus: Disable and re-enable clog_to_monitors could trigger assertion
https://github.com/ceph/ceph/pull/39912 gerald yang
11:44 AM Bug #49409 (Pending Backport): osd run into dead loop and tell slow request when rollback snap wi...
Kefu Chai

03/07/2021

10:02 PM Backport #49377: pacific: building libcrc32
please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/39902
ceph-backport.sh versi...
singuliere _
03:58 PM Backport #49482 (Resolved): pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/Manifes...
Loïc Dachary
03:55 PM Backport #49642 (Resolved): pacific: Disable and re-enable clog_to_monitors could trigger assertion
https://github.com/ceph/ceph/pull/40247 Backport Bot
03:55 PM Backport #49641 (Resolved): octopus: Disable and re-enable clog_to_monitors could trigger assertion
https://github.com/ceph/ceph/pull/39935 Backport Bot
03:55 PM Backport #49640 (Resolved): nautilus: Disable and re-enable clog_to_monitors could trigger assertion
https://github.com/ceph/ceph/pull/39912 Backport Bot
03:54 PM Bug #48946 (Pending Backport): Disable and re-enable clog_to_monitors could trigger assertion
Kefu Chai

03/06/2021

02:58 PM Backport #49533 (In Progress): octopus: osd ok-to-stop too conservative
https://github.com/ceph/ceph/pull/39887 Kefu Chai
02:43 PM Backport #49073 (Resolved): nautilus: crash in Objecter and CRUSH map lookup
Kefu Chai
01:16 AM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
This is where we sent the subops... Neha Ojha

03/05/2021

11:10 PM Bug #49524: ceph_test_rados_delete_pools_parallel didn't start
https://tracker.ceph.com/issues/45946 looks very similar Neha Ojha
11:04 PM Bug #49525: found snap mapper error on pg 3.2s1 oid 3:4abe9991:::smithi10121515-14:e4 snaps missi...
Ronen, can you check if this is caused due to a race between scrub and snap remove. Neha Ojha
10:53 PM Bug #49403 (Duplicate): Caught signal (aborted) on mgrmap epoch 1 during librados init (rados-str...
Neha Ojha
07:15 PM Bug #48298: hitting mon_max_pg_per_osd right after creating OSD, then decreases slowly
Another observation: I have nobackfill set, and I'm currently adding 8 new OSDs.
The first of the newly added OSDs...
Jonas Jelten
05:15 PM Backport #49482: pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcou...
Myoungwon Oh wrote:
> https://github.com/ceph/ceph/pull/39773
merged
Yuri Weinstein
02:39 AM Bug #47419 (Resolved): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo bench...
Hopefully Neha Ojha
01:49 AM Backport #49565 (In Progress): pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
https://github.com/ceph/ceph/pull/39844 Neha Ojha

03/04/2021

11:34 PM Bug #47419 (Fix Under Review): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p f...
Sage Weil
11:34 PM Bug #47419 (Duplicate): make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo benc...
Sage Weil
04:33 PM Bug #47419: make check: src/test/smoke.sh: TEST_multimon: timeout 8 rados -p foo bench 4 write -b...
https://jenkins.ceph.com/job/ceph-pull-requests/70513/consoleFull#10356408840526d21-3511-427d-909c-dd086c0d1034 Neha Ojha
11:21 PM Bug #49614 (Duplicate): src/test/smoke.sh:56: TEST_multimon: timeout 8 rados -p foo bench 4 writ...
Neha Ojha
11:11 PM Bug #49614: src/test/smoke.sh:56: TEST_multimon: timeout 8 rados -p foo bench 4 write -b 4096 --...
https://jenkins.ceph.com/job/ceph-pull-requests/70513/consoleFull#-1656021838e840cee4-f4a4-4183-81dd-42855615f2c1 Sage Weil
10:58 PM Bug #49614 (Duplicate): src/test/smoke.sh:56: TEST_multimon: timeout 8 rados -p foo bench 4 writ...
... Sage Weil
09:14 PM Bug #44631: ceph pg dump error code 124
/ceph/teuthology-archive/pdonnell-2021-03-04_03:51:01-fs-wip-pdonnell-testing-20210303.195715-distro-basic-smithi/593... Patrick Donnelly
05:39 PM Bug #44631: ceph pg dump error code 124
/a/yuriw-2021-03-02_20:59:34-rados-wip-yuri7-testing-2021-03-02-1118-nautilus-distro-basic-smithi/5928174 Neha Ojha
09:08 PM Backport #49532 (Resolved): pacific: osd ok-to-stop too conservative
Sage Weil
06:47 PM Backport #49404 (Resolved): pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
Sage Weil
06:47 PM Backport #49526 (Resolved): pacific: mon/crush_ops.sh fails: Error EBUSY: osd.1 has already bound...
Sage Weil
06:44 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
/a/sage-2021-03-03_16:41:22-rados-wip-sage2-testing-2021-03-03-0744-pacific-distro-basic-smithi/5930113
Sage Weil
04:48 PM Bug #23117: PGs stuck in "activating" after osd_max_pg_per_osd_hard_ratio has been exceeded once
We also his this issue last week on Ceph Version 12.2.11.
Cluster configured with a replication factor of 3, issu...
Ross Martyn
01:21 PM Backport #48987: nautilus: ceph osd df tree reporting incorrect SIZE value for rack having an emp...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39126
m...
Nathan Cutler

03/03/2021

10:14 PM Bug #49104: crush weirdness: degraded PGs not marked as such, and choose_total_tries = 50 is too ...
Thanks for the analysis Neha.
Something that perhaps wasn't clear in comment 2 -- in each case where I print the `...
Dan van der Ster
06:48 PM Bug #49104 (Triaged): crush weirdness: degraded PGs not marked as such, and choose_total_tries = ...
Thanks for the detailed logs!
Firstly, the pg dump output can sometimes be a little laggy, so I am basing my asses...
Neha Ojha
09:53 PM Backport #48987 (Resolved): nautilus: ceph osd df tree reporting incorrect SIZE value for rack ha...
Brad Hubbard
04:05 PM Backport #48987: nautilus: ceph osd df tree reporting incorrect SIZE value for rack having an emp...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39126
merged
Yuri Weinstein
08:47 PM Bug #20909: Error ETIMEDOUT: crush test failed with -110: timed out during smoke test (5 seconds)
not seen in octopus and pacific so far, but pops sometimes in nautilus:... Deepika Upadhyay
08:39 PM Bug #49591 (New): no active mgr (MGR_DOWN)" in cluster log
seen in nautilus... Deepika Upadhyay
03:37 PM Bug #49584: Ceph OSD, MDS, MGR daemon does not _only_ bind to specified address when configured t...
After removing the specific public_addr and restarting the MDSes the situation returns to normal and the cluster reco... Stefan Kooman
03:22 PM Bug #49584 (New): Ceph OSD, MDS, MGR daemon does not _only_ bind to specified address when config...
Documentation (https://docs.ceph.com/en/octopus/rados/configuration/network-config-ref/#ceph-daemons) states the foll... Stefan Kooman
11:32 AM Backport #49055 (Resolved): nautilus: pick_a_shard() always select shard 0
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39651
m...
Nathan Cutler
09:40 AM Bug #40029: ceph-mon: Caught signal (Aborted) in (CrushWrapper::update_choose_args(CephContext*)+...
Florian Haas wrote:
> With thanks to Paul Emmerich in https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/threa...
Norman Shen
05:31 AM Bug #48417: unfound EC objects in sepia's LRC after upgrade
https://tracker.ceph.com/issues/48613#note-13 Deepika Upadhyay
12:28 AM Backport #49404 (In Progress): pacific: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
David Zafman

03/02/2021

08:21 PM Bug #37808 (New): osd: osdmap cache weak_refs assert during shutdown
/ceph/teuthology-archive/pdonnell-2021-03-02_17:29:53-fs:verify-wip-pdonnell-testing-20210301.234318-distro-basic-smi... Patrick Donnelly
05:27 PM Bug #48613: Reproduce https://tracker.ceph.com/issues/48417
personal ref dir: all grep reside in **/home/ideepika/pg[3.1as0.log** in teuthology server
job: /a/teuthology-2021...
Deepika Upadhyay
05:24 PM Bug #49572 (Duplicate): MON_DOWN: mon.c fails to join quorum after un-blacklisting mon.a
This is the same as https://tracker.ceph.com/issues/47654... Neha Ojha
04:58 PM Bug #49572 (Duplicate): MON_DOWN: mon.c fails to join quorum after un-blacklisting mon.a
/a/sage-2021-03-01_20:24:37-rados-wip-sage-testing-2021-03-01-1118-distro-basic-smithi/5924612
it looks like the s...
Sage Weil
04:38 AM Backport #49482: pacific: api_tier_pp: LibRadosTwoPoolsPP.ManifestSnapRefcount/ManifestSnapRefcou...
https://github.com/ceph/ceph/pull/39773 Myoungwon Oh

03/01/2021

11:25 PM Bug #49409 (Fix Under Review): osd run into dead loop and tell slow request when rollback snap wi...
Neha Ojha
10:16 PM Backport #49567 (Resolved): nautilus: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
https://github.com/ceph/ceph/pull/40697 Backport Bot
10:15 PM Backport #49566 (Resolved): octopus: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
https://github.com/ceph/ceph/pull/40756 Backport Bot
10:15 PM Backport #49565 (Resolved): pacific: api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
https://github.com/ceph/ceph/pull/39844 Backport Bot
10:10 PM Bug #47719 (Pending Backport): api_watch_notify: LibRadosWatchNotify.AioWatchDelete2 fails
Brad Hubbard
05:15 PM Backport #49055: nautilus: pick_a_shard() always select shard 0
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39651
merged
Yuri Weinstein
03:43 AM Bug #49543 (New): scrub a pool which size is 1 but found stat mismatch on objects and bytes

the pg has only one primary osd:...
Liu Lan
 

Also available in: Atom