Project

General

Profile

Activity

From 10/26/2021 to 11/24/2021

11/24/2021

05:29 PM Bug #53308: pg-temp entries are not cleared for PGs that no longer exist
That makes sense to me, thanks Neha! Cory Snyder
05:15 PM Bug #53308 (Pending Backport): pg-temp entries are not cleared for PGs that no longer exist
Cory, I am marking this for backport to octopus and pacific, makes sense to you? Neha Ojha
05:29 PM Backport #53389 (In Progress): octopus: pg-temp entries are not cleared for PGs that no longer exist
Cory Snyder
05:20 PM Backport #53389 (Resolved): octopus: pg-temp entries are not cleared for PGs that no longer exist
https://github.com/ceph/ceph/pull/44097 Backport Bot
05:29 PM Backport #53388 (In Progress): pacific: pg-temp entries are not cleared for PGs that no longer exist
Cory Snyder
05:20 PM Backport #53388 (Resolved): pacific: pg-temp entries are not cleared for PGs that no longer exist
https://github.com/ceph/ceph/pull/44096 Backport Bot
03:50 PM Feature #51984 (Fix Under Review): [RFE] Provide warning when the 'require-osd-release' flag does...
Sridhar Seshasayee

11/23/2021

01:53 PM Bug #44286: Cache tiering shows unfound objects after OSD reboots
Update: Also happens with 16.2.5 :-( Jan-Philipp Litza
01:16 PM Bug #52948: osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
New instance seen in below pacific run:
http://pulpito.front.sepia.ceph.com/yuriw-2021-11-20_20:20:29-fs-wip-yuri6...
Kotresh Hiremath Ravishankar
10:54 AM Bug #51945: qa/workunits/mon/caps.sh: Error: Expected return 13, got 0
Seems to be the same problem in:
http://pulpito.front.sepia.ceph.com/yuriw-2021-11-20_18:00:22-rados-wip-yuri6-testi...
Ronen Friedman
07:40 AM Bug #39150: mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
/a/yuriw-2021-11-20_18:01:41-rados-wip-yuri8-testing-2021-11-20-0807-distro-basic-smithi/6516396 Aishwarya Mathuria

11/22/2021

08:29 PM Feature #21579 (Resolved): [RFE] Stop OSD's removal if the OSD's are part of inactive PGs
Vikhyat Umrao
07:11 PM Feature #51984: [RFE] Provide warning when the 'require-osd-release' flag does not match current ...
I am providing the history of PRs and commits that resulted in
the loss/removal of the checks for 'require-osd-relea...
Sridhar Seshasayee
06:45 PM Bug #53306 (Fix Under Review): ceph -s mon quorum age negative number
Sage Weil

11/20/2021

01:41 AM Bug #53349 (New): stat_sum.num_bytes of pool is incorrect when randomly writing small IOs to the ...
In a test, I found that when random writes with an IO size of 512B are performed on the rbd, The pool's stat_sum.num_... mingpo li
12:06 AM Bug #52657: MOSDPGLog::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, SERVER_NAUTILUS)'
/a/ksirivad-2021-11-19_19:14:07-rados-wip-autoscale-profile-scale-up-default-distro-basic-smithi/6514251 Neha Ojha

11/19/2021

06:23 PM Bug #53342 (In Progress): Exiting scrub checking -- not all pgs scrubbed
... Neha Ojha
04:31 PM Backport #53340 (New): pacific: osd/scrub: OSD crashes at PG removal
Backport Bot
04:30 PM Backport #53339 (Resolved): pacific: src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<cons...
https://github.com/ceph/ceph/pull/46767 Backport Bot
04:30 PM Backport #53338 (Resolved): pacific: osd/scrub: src/osd/scrub_machine.cc: 55: FAILED ceph_assert(...
Backport Bot
04:29 PM Bug #51843 (Pending Backport): osd/scrub: OSD crashes at PG removal
Neha Ojha
04:28 PM Bug #51942 (Pending Backport): src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotA...
Neha Ojha
04:27 PM Bug #52012 (Pending Backport): osd/scrub: src/osd/scrub_machine.cc: 55: FAILED ceph_assert(state_...
Neha Ojha
03:46 AM Bug #53330 (New): ceph client request connection with an old invalid key.
We have a production ceph cluster with 3 mons and 516 osds.
Ceph version: 14.2.8
CPU: Intel(R) Xeon(R) Gold 5218
...
wencong wan
01:20 AM Bug #53329 (Duplicate): Set osd_fast_shutdown_notify_mon=true by default
Neha Ojha
01:18 AM Bug #53328 (Fix Under Review): osd_fast_shutdown_notify_mon option should be true by default
Neha Ojha

11/18/2021

11:10 PM Bug #53329 (Duplicate): Set osd_fast_shutdown_notify_mon=true by default
This option was introduced in https://github.com/ceph/ceph/pull/38909, but was set false by default. There is a lot o... Neha Ojha
09:30 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
Ist Gab wrote:
> Neha Ojha wrote:
> > Set osd_delete_sleep to 2 secs and go higher if this does not help. Setting o...
Neha Ojha
09:24 PM Bug #53328: osd_fast_shutdown_notify_mon option should be true by default
Pull request ID: 44016 Satoru Takeuchi
09:14 PM Bug #53328 (Duplicate): osd_fast_shutdown_notify_mon option should be true by default
osd_fast_shutdown_notify_mon option is false by default. So users suffer
from error log flood, slow ops, and the lon...
Satoru Takeuchi
09:22 PM Bug #50608: ceph_assert(is_primary()) in PrimaryLogPG::on_local_recover
Tobias Urdin wrote:
> After upgrading osd.107 to 15.5.15 and waiting 2 hours for it to recover 3,000 objects in a si...
Neha Ojha
09:11 PM Bug #53327 (Resolved): osd: osd_fast_shutdown_notify_mon not quite right and enable osd_fast_shut...
- it should send MOSDMarkMeDead not MarkMeDown
- we must confirm that we set a flag (preparing to stop?) that makes ...
Sage Weil
08:57 PM Bug #53326 (Fix Under Review): pgs wait for read lease after osd start
Sage Weil
08:28 PM Bug #53326 (Resolved): pgs wait for read lease after osd start
- pg is healthy
- primary osd stops
- wait for things to settle
- restart primary
- pg goes into WAIT state
Th...
Sage Weil
08:08 PM Bug #51942: src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotActive*>())
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{defa... Neha Ojha
05:58 PM Bug #48298: hitting mon_max_pg_per_osd right after creating OSD, then decreases slowly
still encountering on ceph octopus 15.2.15 :(
please add the HEALTH_ERROR when the limit is hit, then one at least...
Jonas Jelten
12:51 PM Bug #53316 (New): qa: (smithi150) slow request osd_op, currently waiting for sub ops warning
The warning is seen in following teuthology run
http://pulpito.front.sepia.ceph.com/yuriw-2021-11-17_19:02:43-fs-w...
Kotresh Hiremath Ravishankar
03:05 AM Feature #52424 (Fix Under Review): [RFE] Limit slow request details to mgr log
Prashant D

11/17/2021

06:14 PM Bug #53308 (Resolved): pg-temp entries are not cleared for PGs that no longer exist
When scaling down pg_num while it was in the process of scaling up, we consistently end up with stuck pg-temp entries... Cory Snyder
04:59 PM Bug #53306 (Resolved): ceph -s mon quorum age negative number
... Sage Weil
06:24 AM Bug #52624: qa: "Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)"
Seen in this pacific run as well.
http://pulpito.front.sepia.ceph.com/yuriw-2021-11-12_00:33:28-fs-wip-yuri7-testi...
Kotresh Hiremath Ravishankar

11/16/2021

10:15 PM Bug #50659 (Fix Under Review): Segmentation fault under Pacific 16.2.1 when using a custom crush ...
Neha Ojha
03:12 PM Bug #50659: Segmentation fault under Pacific 16.2.1 when using a custom crush location hook
Thank you for this fix. It is very much appreciated. Andrew Davidoff
08:25 PM Bug #53295 (New): Leak_DefinitelyLost PrimaryLogPG::do_proxy_chunked_read()
... Neha Ojha
08:20 PM Bug #53294 (Pending Backport): rados/test.sh hangs while running LibRadosTwoPoolsPP.TierFlushDuri...
... Neha Ojha
07:14 PM Bug #52867: pick_address.cc prints: unable to find any IPv4 address in networks 'fd00:fd00:fd00:3...
Kefu Chai wrote:
> @John,
>
> per the logging message pasted at http://ix.io/3B1y
>
>
> [...]
>
> it seem...
John Fulton
06:27 PM Backport #53259 (In Progress): pacific: mon: should always display disallowed leaders when set
Greg Farnum
06:26 PM Bug #53258 (Pending Backport): mon: should always display disallowed leaders when set
Greg Farnum
06:25 PM Bug #53237 (Pending Backport): mon: stretch mode blocks kernel clients from connecting
Greg Farnum
06:24 PM Backport #53239 (In Progress): pacific: mon: stretch mode blocks kernel clients from connecting
Greg Farnum
07:21 AM Backport #52936 (Resolved): pacific: Primary OSD crash caused corrupted object and further crashe...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/43544
m...
Loïc Dachary
07:21 AM Backport #52868: stretch mode: allow users to change the tiebreaker monitor
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/43457
m...
Loïc Dachary
12:32 AM Bug #53240 (Fix Under Review): full-object read crc is mismatch, because truncate modify oi.size ...
Neha Ojha

11/15/2021

03:27 PM Bug #50608: ceph_assert(is_primary()) in PrimaryLogPG::on_local_recover
After upgrading osd.107 to 15.5.15 and waiting 2 hours for it to recover 3,000 objects in a single PG it crashed agai... Tobias Urdin
03:19 PM Bug #50608: ceph_assert(is_primary()) in PrimaryLogPG::on_local_recover
... Tobias Urdin
01:15 PM Bug #50608: ceph_assert(is_primary()) in PrimaryLogPG::on_local_recover
This is still an issue and it repeatedly hits this during recovery when upgrading the cluster where some (already upg... Tobias Urdin
02:18 AM Bug #53219: LibRadosTwoPoolsPP.ManifestRollbackRefcount failure
Calculating reference count on manifest snapshotted object requires correct refcount information. So, current unittes... Myoungwon Oh

11/14/2021

09:59 PM Bug #52901: osd/scrub: setting then clearing noscrub may lock a PG in 'scrubbing' state
A test to detect this specific bug pushed as PR 43919 Ronen Friedman
08:47 AM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
Neha Ojha wrote:
> Set osd_delete_sleep to 2 secs and go higher if this does not help. Setting osd_delete_sleep take...
Ist Gab

11/13/2021

08:01 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
Ist Gab wrote:
> Neha Ojha wrote:
> > Can you try to set a higher value of "osd delete sleep" and see if that helps...
Neha Ojha
05:50 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
Neha Ojha wrote:
> Can you try to set a higher value of "osd delete sleep" and see if that helps?
Which one speci...
Ist Gab

11/12/2021

11:11 PM Backport #53259 (Resolved): pacific: mon: should always display disallowed leaders when set
https://github.com/ceph/ceph/pull/43972 Greg Farnum
11:10 PM Bug #53258 (Resolved): mon: should always display disallowed leaders when set
I made some usability improvements in https://github.com/ceph/ceph/pull/43373, but accidentally switched things so th... Greg Farnum
11:08 PM Backport #53238 (Rejected): octopus: mon: stretch mode blocks kernel clients from connecting
Apparently I sometimes fail at sorting alphanumerically? Greg Farnum
06:57 PM Bug #51942: src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotActive*>())
Ronen, let's prioritize this. Neha Ojha
06:56 PM Bug #48909 (Duplicate): clog slow request overwhelm monitors
Neha Ojha
06:51 PM Bug #53138 (Triaged): cluster [WRN] Health check failed: Degraded data redundancy: 3/1164 objec...
This warning comes up because there are PGs recovering, probably because the test is injecting failures - we can igno... Neha Ojha
06:46 PM Bug #52969: use "ceph df" command found pool max avail increase when there are degraded objects i...
minghang zhao wrote:
> My solution is to add a function del_down_out_osd() to PGMap::get_rule_avail() to calculate t...
Neha Ojha
06:43 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
Can you try to set a higher value of "osd delete sleep" and see if that helps? Neha Ojha
06:29 PM Bug #53190: counter num_read_kb is going down
This seems possible to occur for many such counters in a distributed system like ceph, where these values are not tre... Josh Durgin
06:26 PM Bug #52901 (Resolved): osd/scrub: setting then clearing noscrub may lock a PG in 'scrubbing' state
Neha Ojha
06:20 PM Bug #52503: cli_generic.sh: slow ops when trying rand write on cache pools
Deepika Upadhyay wrote:
> /ceph/teuthology-archive/ideepika-2021-11-02_12:33:30-rbd-wip-ssd-cache-testing-distro-bas...
Neha Ojha
06:17 PM Bug #53219: LibRadosTwoPoolsPP.ManifestRollbackRefcount failure
Myoungwon Oh wrote:
> I think this is the same issue as https://tracker.ceph.com/issues/52872.
> Recovery takes alm...
Neha Ojha
07:28 AM Bug #53219: LibRadosTwoPoolsPP.ManifestRollbackRefcount failure
I think this is the same issue as https://tracker.ceph.com/issues/52872.
Recovery takes almost 8 minutes even if cur...
Myoungwon Oh
05:58 PM Bug #53251 (Closed): compiler warning about deprecated fmt::format_to()
... Casey Bodley
03:49 PM Bug #52948: osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
http://qa-proxy.ceph.com/teuthology/ideepika-2021-11-12_08:56:59-rbd-wip-deepika-testing-2021-11-12-1203-distro-basic... Deepika Upadhyay
06:31 AM Bug #53240: full-object read crc is mismatch, because truncate modify oi.size and forget to clear...
my ceph version is nautilous 14.2.5 wendong jia
04:08 AM Bug #53240: full-object read crc is mismatch, because truncate modify oi.size and forget to clear...
https://github.com/ceph/ceph/pull/43902 wendong jia
03:27 AM Bug #53240: full-object read crc is mismatch, because truncate modify oi.size and forget to clear...
The object oi.size should be 4194304, but it is actually 4063232.
The object data_digest is 0xffffffff, but read crc...
wendong jia
02:56 AM Bug #53240 (Fix Under Review): full-object read crc is mismatch, because truncate modify oi.size ...
I use 100 threads to dd on multiple files under the directory, so the same file can be truncated at any time.
When d...
wendong jia
05:53 AM Cleanup #52754: windows warnings
@Laura, they appear in windows shaman builds, anyone can take a look at the latest windows builds available here http... Deepika Upadhyay

11/11/2021

09:11 PM Bug #52867: pick_address.cc prints: unable to find any IPv4 address in networks 'fd00:fd00:fd00:3...
John Fulton wrote:
> As per comment #3 I was on the right path but I should have set an OSD setting, not a mon setti...
Neha Ojha
09:10 PM Bug #52867 (Need More Info): pick_address.cc prints: unable to find any IPv4 address in networks ...
Neha Ojha
08:40 PM Backport #53239 (Resolved): pacific: mon: stretch mode blocks kernel clients from connecting
https://github.com/ceph/ceph/pull/43971 Greg Farnum
08:40 PM Backport #53238 (Rejected): octopus: mon: stretch mode blocks kernel clients from connecting
This was reported by Red Hat at https://bugzilla.redhat.com/show_bug.cgi?id=2022190
> [66873.543382] libceph: got ...
Greg Farnum
08:30 PM Bug #53237 (Resolved): mon: stretch mode blocks kernel clients from connecting
This was reported by Red Hat at https://bugzilla.redhat.com/show_bug.cgi?id=2022190
> [66873.543382] libceph: got ...
Greg Farnum
07:48 PM Cleanup #52754: windows warnings
Deepika, the link is 404 now. Is there a way that we could preserve the Jenkins output and provide a different link? Laura Flores
03:26 PM Bug #52948: osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
Analysis of logs from JobID: 6443924
osd.3 during running of the "ceph" teuthology task didn't get initialized. As...
Sridhar Seshasayee
12:32 AM Bug #53219: LibRadosTwoPoolsPP.ManifestRollbackRefcount failure
I'll take a look Myoungwon Oh

11/10/2021

07:44 PM Bug #53219 (New): LibRadosTwoPoolsPP.ManifestRollbackRefcount failure
... Sage Weil

11/09/2021

06:21 PM Bug #50659: Segmentation fault under Pacific 16.2.1 when using a custom crush location hook
I'm also seeing this issue:... André Cruz

11/08/2021

05:45 PM Bug #52901: osd/scrub: setting then clearing noscrub may lock a PG in 'scrubbing' state
Easy to reproduce: set noscrub, then request a deep-scrub. That will get the PG's Scrubber state-machine
stuck in su...
Ronen Friedman
01:42 PM Bug #53190 (New): counter num_read_kb is going down
h3. Description of problem
An unreasonably high read metric value has been reported by monitoring (28.76TB/s).
...
Patrick Seidensal
12:36 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
Have a look the latency.png, please, all the spikes are almost outage. Ist Gab
12:27 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
Igor Fedotov wrote:
> Can we have a relevant OSD log, please. I presume suicide timeout/slow DB operations are prese...
Ist Gab
12:26 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
Neha Ojha wrote:
> Do you have any more information about this crash? How often do you see it?
I have quite a lot...
Ist Gab
10:47 AM Bug #52760: Monitor unable to rejoin the cluster
Neha Ojha wrote:
> Can you share mon logs from all the monitors with debug_mon=20 and debug_ms=1?
I will once thi...
Ruben Kerkhof

11/05/2021

09:14 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
Can we have a relevant OSD log, please. I presume suicide timeout/slow DB operations are present there. Igor Fedotov
09:12 PM Bug #53142 (Need More Info): OSD crash in PG::do_delete_work when increasing PGs
Do you have any more information about this crash? How often do you see it? Neha Ojha
09:10 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
... Neha Ojha

11/04/2021

09:17 PM Backport #53167 (Rejected): octopus: api_watch_notify: LibRadosWatchNotify.Watch3Timeout failed
Backport Bot
09:16 PM Backport #53166 (Resolved): pacific: api_watch_notify: LibRadosWatchNotify.Watch3Timeout failed
https://github.com/ceph/ceph/pull/51261 Backport Bot
09:11 PM Bug #24990 (Pending Backport): api_watch_notify: LibRadosWatchNotify.Watch3Timeout failed
Neha Ojha
09:08 PM Bug #24990: api_watch_notify: LibRadosWatchNotify.Watch3Timeout failed
https://github.com/ceph/ceph/pull/43700 merged Yuri Weinstein
01:51 AM Bug #52126 (Resolved): stretch mode: allow users to change the tiebreaker monitor
Greg Farnum
01:50 AM Backport #52868 (Resolved): stretch mode: allow users to change the tiebreaker monitor
Greg Farnum
12:38 AM Bug #52948: osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
Sridhar has started looking into this. Neha Ojha

11/03/2021

10:36 PM Backport #52936: pacific: Primary OSD crash caused corrupted object and further crashes during ba...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/43544
merged
Yuri Weinstein
10:35 PM Backport #52868: stretch mode: allow users to change the tiebreaker monitor
Greg Farnum wrote:
> https://github.com/ceph/ceph/pull/43457
merged
Yuri Weinstein
03:10 PM Bug #53142 (Need More Info): OSD crash in PG::do_delete_work when increasing PGs
I've attached the file and put the crash signature also. Ist Gab
01:40 PM Bug #53138 (Triaged): cluster [WRN] Health check failed: Degraded data redundancy: 3/1164 objec...
... Deepika Upadhyay
12:23 PM Bug #52503: cli_generic.sh: slow ops when trying rand write on cache pools
/ceph/teuthology-archive/ideepika-2021-11-02_12:33:30-rbd-wip-ssd-cache-testing-distro-basic-smithi/6477582/teutholog... Deepika Upadhyay
10:47 AM Bug #47300: mount.ceph fails to understand AAAA records from SRV record
Issue still present on 16.2.6 (ceph packages 16.2.6-1focal, kernel 5.11.0-38-generic)... Daniël Vos
09:57 AM Bug #51463: blocked requests while stopping/starting OSDs
Hi Sage,
I tested it with fast shutdown enabled (default) and disabled. In both cases I got slow ops (longer than ...
Manuel Lausch
07:15 AM Bug #52967: premerge pgs may be backfill_wait for a long time
hi, Sage, What are the conditions for production "premerge+backfill_wait"? yite gu
06:58 AM Bug #52741: pg inconsistent state is lost after the primary osd restart
yite gu wrote:
> What is the way you remove replica?
In my case it was filestore so I just remove the file on the...
Mykola Golub
03:24 AM Bug #52741: pg inconsistent state is lost after the primary osd restart
What is the way you remove replica? yite gu

11/02/2021

11:49 PM Bug #48909: clog slow request overwhelm monitors
This is being handled over https://tracker.ceph.com/issues/52424. Prashant D
10:17 PM Bug #51527 (Resolved): Ceph osd crashed due to segfault
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Loïc Dachary
10:01 PM Backport #52770 (Resolved): pacific: pg scrub stat mismatch with special objects that have hash '...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/43512
m...
Loïc Dachary
10:01 PM Backport #52620 (Resolved): pacific: partial recovery become whole object recovery after restart osd
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/43513
m...
Loïc Dachary
10:01 PM Backport #52843 (Resolved): pacific: msg/async/ProtocalV2: recv_stamp of a message is set to a wr...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/43511
m...
Loïc Dachary
10:00 PM Backport #52831 (Resolved): pacific: osd: pg may get stuck in backfill_toofull after backfill is ...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/43437
m...
Loïc Dachary
01:50 PM Bug #51463 (Need More Info): blocked requests while stopping/starting OSDs
I easily reproduced this with 'osd fast shutdown = false' (vstart default), but was unable to do so with 'osd fast sh... Sage Weil
12:22 AM Bug #52657: MOSDPGLog::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, SERVER_NAUTILUS)'
... Neha Ojha
12:21 AM Bug #52694 (Duplicate): src/messages/MOSDPGLog.h: virtual void MOSDPGLog::encode_payload(uint64_t...
Neha Ojha

11/01/2021

05:09 PM Feature #51213 (Fix Under Review): [ceph osd set noautoscale] Global on/off flag for PG autoscale...
Vikhyat Umrao

10/31/2021

10:06 AM Bug #52948: osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
... Deepika Upadhyay

10/29/2021

02:05 PM Feature #52424: [RFE] Limit slow request details to mgr log
I am working on to run qa-suite on PR. Keeping status as "In progress" for now as need to push qa suite changes if any. Prashant D

10/28/2021

11:29 PM Feature #51213: [ceph osd set noautoscale] Global on/off flag for PG autoscale feature
PR: https://github.com/ceph/ceph/pull/43716 Kamoltat (Junior) Sirivadhna
03:29 PM Backport #52845 (In Progress): pacific: osd: add scrub duration to pg dump
Cory Snyder
02:12 PM Bug #51942: src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotActive*>())
and /a/sage-2021-10-28_02:19:01-rados-wip-sage3-testing-2021-10-27-1300-distro-basic-smithi/6464056
with logs
Sage Weil
02:10 PM Bug #51942: src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotActive*>())
/a/sage-2021-10-28_02:19:01-rados-wip-sage3-testing-2021-10-27-1300-distro-basic-smithi/6464393
with osd logs
Sage Weil
02:08 PM Bug #39150: mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
/a/sage-2021-10-28_02:19:01-rados-wip-sage3-testing-2021-10-27-1300-distro-basic-smithi/6464204
with logs!
Sage Weil
02:06 PM Bug #24990 (Fix Under Review): api_watch_notify: LibRadosWatchNotify.Watch3Timeout failed
Sage Weil
02:04 PM Bug #24990: api_watch_notify: LibRadosWatchNotify.Watch3Timeout failed
/a/sage-2021-10-28_02:19:01-rados-wip-sage3-testing-2021-10-27-1300-distro-basic-smithi/6464087... Sage Weil
11:56 AM Feature #52424 (In Progress): [RFE] Limit slow request details to mgr log
Prashant D

10/27/2021

07:55 PM Feature #53050: Support blocklisting a CIDR range
Greg Farnum wrote:
> Patrick Donnelly wrote:
> > So we're going to put a huge asterisk here that the CIDR range of ...
Patrick Donnelly
05:00 AM Feature #53050: Support blocklisting a CIDR range
Patrick Donnelly wrote:
> So we're going to put a huge asterisk here that the CIDR range of machines must be hard-re...
Greg Farnum
01:39 AM Feature #53050: Support blocklisting a CIDR range
So we're going to put a huge asterisk here that the CIDR range of machines must be hard-rebooted, right? Otherwise w... Patrick Donnelly
06:46 PM Bug #53067: Fix client "version" display for kernel clients
I looked at this at one point and it was moderately irritating, but the display bug is also really confusing for user... Greg Farnum
01:10 PM Bug #53067 (New): Fix client "version" display for kernel clients
Hello
When a rhel7 client mounts a cephfs share, it appears in `ceph features` as it was a jewel client, even if t...
gustavo panizzo
10:40 AM Bug #52509: PG merge: PG stuck in premerge+peered state
We had a similar outage.
We did try to increase the number of PGs on a bucket-index-pool:...
Markus Wennrich

10/26/2021

08:30 PM Backport #52770: pacific: pg scrub stat mismatch with special objects that have hash 'ffffffff'
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/43512
merged
Yuri Weinstein
08:29 PM Backport #52620: pacific: partial recovery become whole object recovery after restart osd
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/43513
merged
Yuri Weinstein
08:28 PM Backport #52843: pacific: msg/async/ProtocalV2: recv_stamp of a message is set to a wrong value
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/43511
merged
Yuri Weinstein
08:26 PM Backport #52831: pacific: osd: pg may get stuck in backfill_toofull after backfill is interrupted...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/43437
merged
Yuri Weinstein
07:45 PM Bug #52124: Invalid read of size 8 in handle_recovery_delete()
/a/yuriw-2021-10-21_13:40:38-rados-wip-yuri2-testing-2021-10-20-1700-pacific-distro-basic-smithi/6454961/remote/smith... Neha Ojha
06:48 PM Feature #48590 (Rejected): Add ability to blocklist a cephx entity name, a set of entities by a l...
This is really impractical to do in RADOS. Closing in favor of https://tracker.ceph.com/issues/53050 Greg Farnum
06:48 PM Feature #53050 (Resolved): Support blocklisting a CIDR range
Disaster recovery use cases want to be able to fence off entire IP ranges, rather than needing to specify individual ... Greg Farnum
04:55 PM Bug #52761: OSDs announcing incorrect front_addr after upgrade to 16.2.6
Yes, I tried that, but it does not change the behavior:
>> ceph config set global public_network 10.113.0.0/16
...
Javier Cacheiro
04:35 PM Bug #52948: osd: fails to come up: "teuthology.misc:7 of 8 OSDs are up"
Neha Ojha wrote:
> This could be related to the removal of allocation metadata from rocksdb work from Gabi, I will h...
Patrick Donnelly
 

Also available in: Atom