Project

General

Profile

Activity

From 01/26/2022 to 02/24/2022

02/24/2022

10:45 PM Backport #54386: octopus: [RFE] Limit slow request details to mgr log
please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/45154
ceph-backport.sh versi...
Ponnuvel P
07:43 PM Bug #52136: Valgrind reports memory "Leak_DefinitelyLost" errors.
/a/sseshasa-2022-02-24_11:27:07-rados-wip-45118-45121-quincy-testing-distro-default-smithi/6704275/remote/smithi174/l... Laura Flores
07:19 PM Bug #53294: rados/test.sh hangs while running LibRadosTwoPoolsPP.TierFlushDuringFlush
/a/sseshasa-2022-02-24_11:27:07-rados-wip-45118-45121-quincy-testing-distro-default-smithi/6704402... Laura Flores
06:36 PM Bug #54368 (Duplicate): ModuleNotFoundError: No module named 'tasks.cephadm'
Neha Ojha
05:51 PM Backport #53644 (In Progress): pacific: Disable health warning when autoscaler is on
Christopher Hoffman
03:33 PM Backport #53551 (Resolved): pacific: [RFE] Provide warning when the 'require-osd-release' flag do...
Sridhar Seshasayee
08:56 AM Bug #54396: Setting osd_pg_max_concurrent_snap_trims to 0 prematurely clears the snaptrim queue
More context:... Daniel van der Ster
08:44 AM Bug #54396 (Fix Under Review): Setting osd_pg_max_concurrent_snap_trims to 0 prematurely clears t...
Daniel van der Ster
08:41 AM Bug #54396 (Resolved): Setting osd_pg_max_concurrent_snap_trims to 0 prematurely clears the snapt...
See https://www.spinics.net/lists/ceph-users/msg71061.html... Daniel van der Ster
08:38 AM Backport #54393 (Resolved): quincy: The built-in osd bench test shows inflated results.
https://github.com/ceph/ceph/pull/45141 Sridhar Seshasayee
08:37 AM Bug #54364 (Pending Backport): The built-in osd bench test shows inflated results.
Sridhar Seshasayee
02:45 AM Bug #51627: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_missing(soi...
The error message looks like similar before, but the cause is difference from the prior case.
Anyway, I posted the f...
Myoungwon Oh

02/23/2022

05:32 PM Bug #52124: Invalid read of size 8 in handle_recovery_delete()
Happened in a dead job.
/a/yuriw-2022-02-21_15:40:41-rados-wip-yuri4-testing-2022-02-18-0800-distro-default-smithi/6...
Laura Flores
05:16 PM Bug #51627: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_missing(soi...
Happened again. Could this be a new occurrence?
/a/yuriw-2022-02-21_15:40:41-rados-wip-yuri4-testing-2022-02-18-0800...
Laura Flores
05:00 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
/a/yuriw-2022-02-21_15:40:41-rados-wip-yuri4-testing-2022-02-18-0800-distro-default-smithi/6698327 Laura Flores
03:15 PM Backport #54386 (Resolved): octopus: [RFE] Limit slow request details to mgr log
Backport Bot

02/22/2022

09:10 PM Bug #54210 (Fix Under Review): pacific: mon/pg_autoscaler.sh: echo failed on "bash -c 'ceph osd p...
Kamoltat (Junior) Sirivadhna
09:09 PM Bug #54210: pacific: mon/pg_autoscaler.sh: echo failed on "bash -c 'ceph osd pool get a pg_num | ...
After going through sentry, I've realized that the only occurrence of this bug in master happens before the merge of
...
Kamoltat (Junior) Sirivadhna
08:39 PM Backport #54233 (In Progress): octopus: devices: mon devices appear empty when scraping SMART met...
Neha Ojha
08:39 PM Backport #54232 (In Progress): pacific: devices: mon devices appear empty when scraping SMART met...
Neha Ojha
08:14 PM Bug #54369 (New): mon/test_mon_osdmap_prune.sh: jq .osdmap_first_committed [[ 11 -eq 20 ]]
/a/yuriw-2022-02-17_23:23:56-rados-wip-yuri7-testing-2022-02-17-0852-pacific-distro-default-smithi/6692990... Kamoltat (Junior) Sirivadhna
07:31 PM Bug #54368 (Duplicate): ModuleNotFoundError: No module named 'tasks.cephadm'
/a/yuriw-2022-02-17_23:23:56-rados-wip-yuri7-testing-2022-02-17-0852-pacific-distro-default-smithi/6692894... Kamoltat (Junior) Sirivadhna
07:19 PM Bug #47589: radosbench times out "reached maximum tries (800) after waiting for 4800 seconds"
/a/yuriw-2022-02-17_23:23:56-rados-wip-yuri7-testing-2022-02-17-0852-pacific-distro-default-smithi/6692841 Kamoltat (Junior) Sirivadhna
06:29 PM Bug #52124: Invalid read of size 8 in handle_recovery_delete()
/a/yuriw-2022-02-21_18:20:15-rados-wip-yuri11-testing-2022-02-21-0831-quincy-distro-default-smithi/6699270
Happene...
Laura Flores
05:09 PM Bug #50659: Segmentation fault under Pacific 16.2.1 when using a custom crush location hook
Chris Durham wrote:
> This issue bit us in our upgrade to 16.2.7 from 15.2.15. We have a manual cluster (non-cephadm...
Chris Durham
04:11 PM Bug #50659: Segmentation fault under Pacific 16.2.1 when using a custom crush location hook
This issue bit us in our upgrade to 16.2.7 from 15.2.15. We have a manual cluster (non-cephadm). We followed the proc... Chris Durham
12:57 PM Bug #54364 (Resolved): The built-in osd bench test shows inflated results.
The built-in osd bench shows inflated results with up to 3x-4x the expected values.
Example:
Before:
{
"b...
Sridhar Seshasayee
07:13 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
We were able to narrow it down further. We can trigger the problem reliably by doing this:
- 2 clusters, multisite...
Dieter Roels

02/21/2022

05:09 PM Feature #44107 (Resolved): mon: produce stable election results when netsplits and other errors h...
Oh, this has been done for ages. Greg Farnum
12:47 PM Bug #51463: blocked requests while stopping/starting OSDs
I think we hit the same issue while upgrading our nautilus cluster to pacific.
While I did not hit this when testing...
Maximilian Stinsky
12:42 PM Backport #53339: pacific: src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotActive...
Hey I got here through the following mailing list post: https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thre... Maximilian Stinsky
01:27 AM Bug #51076: "wait_for_recovery: failed before timeout expired" during thrashosd test with EC back...
/a/yuriw-2022-02-17_22:49:55-rados-wip-yuri3-testing-2022-02-17-1256-distro-default-smithi/6692376... Laura Flores

02/20/2022

09:25 AM Bug #52901: osd/scrub: setting then clearing noscrub may lock a PG in 'scrubbing' state
Will this be backported to a stable release? André Cruz
09:23 AM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
I'm also seeing the same issue on 16.2.7, but it's been going on for almost two weeks. Already set and unset noscrub/... André Cruz
09:18 AM Backport #53339: pacific: src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotActive...
Could this be the reason I'm seeing a spam of "handle_scrub_reserve_grant: received unsolicited reservation grant" me... André Cruz

02/18/2022

01:32 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
Fortunately (or perhaps not so fortunately), in the process of dealing with this issue we performed a full restart of... David Prude
11:49 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Dieter Roels wrote:
> Hi CHristian. Are your rgws collocated with the osds of the metadata pools?
> We now notice i...
Christian Rohmann
11:31 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Hi CHristian. Are your rgws collocated with the osds of the metadata pools?
We now notice in our clusters that the...
Dieter Roels

02/17/2022

09:43 PM Bug #52124: Invalid read of size 8 in handle_recovery_delete()
/a/yuriw-2022-02-16_15:53:49-rados-wip-yuri11-testing-2022-02-15-1643-distro-default-smithi/6688846 Laura Flores
07:43 PM Bug #54210: pacific: mon/pg_autoscaler.sh: echo failed on "bash -c 'ceph osd pool get a pg_num | ...
yep this is a bug, thanks for letting me know, patch coming up. Kamoltat (Junior) Sirivadhna
06:54 PM Backport #54290 (Resolved): quincy: pybind/mgr/progress: disable pg recovery event by default
Kamoltat (Junior) Sirivadhna
05:46 PM Bug #54316 (Resolved): mon/MonCommands.h: target_size_ratio range is incorrect
Currently if we give `target_size_ratio` a value more than 1.0 using the command: `ceph osd pool create <pool-name> -... Kamoltat (Junior) Sirivadhna
05:00 PM Bug #54263: cephadm upgrade pacific to quincy autoscaler is scaling pgs from 32 -> 32768 for ceph...
Update:
From the monitor sides of things of pool creation, target_size_ratio cannot be more than 1.0 or less than ...
Kamoltat (Junior) Sirivadhna
04:52 PM Bug #51076: "wait_for_recovery: failed before timeout expired" during thrashosd test with EC back...
/a/yuriw-2022-02-16_00:25:26-rados-wip-yuri-testing-2022-02-15-1431-distro-default-smithi/6687342
Same issue with ...
Laura Flores
04:30 PM Bug #51307: LibRadosWatchNotify.Watch2Delete fails
/a/yuriw-2022-02-16_00:25:26-rados-wip-yuri-testing-2022-02-15-1431-distro-default-smithi/6687338 Laura Flores
11:54 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
We just observed 12 more scrub errors spread across 7 pgs and all on our primary (used for user access, read/write) z... Christian Rohmann
09:47 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Neha Ojha wrote:
> yite gu wrote:
> > Christian Rohmann wrote:
> > > yite gu wrote:
> > > > This is inconsistent ...
yite gu
09:51 AM Bug #54296: OSDs using too much memory
Hi Dan,
Thanks for your response.
I only adjusted osd_max_pg_log_entries and left osd_min_pg_log_entries alone. A...
Ruben Kerkhof
09:03 AM Bug #54296: OSDs using too much memory
Ruben Kerkhof wrote:
> One thing I tried was to set osd_max_pg_log_entries to 500 instead of the default of 10000, b...
Daniel van der Ster
09:33 AM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
I would just like to add that scrubs started all of the sudden and the cluster is HEALTH_OK again. Daan van Gorkum

02/16/2022

10:32 PM Bug #52657: MOSDPGLog::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, SERVER_NAUTILUS)'
/ceph/teuthology-archive/yuriw-2022-02-15_22:35:42-rados-wip-yuri8-testing-2022-02-15-1214-distro-default-smithi/6686... Neha Ojha
09:20 PM Backport #53718: pacific: mon: frequent cpu_tp had timed out messages
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44545
merged
Yuri Weinstein
08:24 PM Backport #54290: quincy: pybind/mgr/progress: disable pg recovery event by default
https://github.com/ceph/ceph/pull/45043 merged Yuri Weinstein
07:59 PM Bug #52124: Invalid read of size 8 in handle_recovery_delete()
/a/yuriw-2022-02-15_22:40:39-rados-wip-yuri7-testing-2022-02-15-1102-quincy-distro-default-smithi/6686655/remote/smit... Neha Ojha
07:07 PM Backport #53535: pacific: mon: mgrstatmonitor spams mgr with service_map
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44721
merged
Yuri Weinstein
07:06 PM Backport #53942: pacific: mon: all mon daemon always crash after rm pool
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44698
merged
Yuri Weinstein
07:00 PM Feature #54280: support truncation sequences in sparse reads
Neha mentioned taking a look at the history, so I did a bit of git archeology today. The limitation dates back to the... Jeff Layton
06:58 PM Bug #53751: "N monitors have not enabled msgr2" is always shown for new clusters
Hello. Could you please provide the output from @ceph health detail@? We suspect the warning might got replaced with ... Radoslaw Zarzynski
06:46 PM Bug #54255: utc time is used when ceph crash ls
Yaarit, was this choice intentional? Neha Ojha
06:44 PM Bug #51338 (Duplicate): osd/scrub_machine.cc: FAILED ceph_assert(state_cast&lt;const NotActive*&g...
Neha Ojha
06:44 PM Bug #51338: osd/scrub_machine.cc: FAILED ceph_assert(state_cast&lt;const NotActive*&gt;())
André Cruz wrote:
> I'm also encountering this issue on Pacific (16.2.7):
>
> [...]
>
> Any pointers?
I thi...
Neha Ojha
06:31 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
yite gu wrote:
> Christian Rohmann wrote:
> > yite gu wrote:
> > > This is inconsistent pg 7.2 from your upload fi...
Neha Ojha
06:24 PM Bug #53663 (New): Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Neha Ojha
09:50 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Dieter Roels wrote:
> After the repair the inconsistencies do not re-appear. However, we can reproduce the issue in ...
Dieter Roels
06:18 PM Bug #46847: Loss of placement information on OSD reboot
Frank Schilder wrote:
> Could somebody please set the status back to open and Affected Versions to all?
The ticke...
Neha Ojha
06:13 PM Bug #53729 (Need More Info): ceph-osd takes all memory before oom on boot
Neha Ojha
12:21 PM Bug #54296: OSDs using too much memory
Hi Igor,
See attachment.
One thing I tried was to set osd_max_pg_log_entries to 500 instead of the default of 1...
Ruben Kerkhof
12:15 PM Bug #54296: OSDs using too much memory
Hi Ruben,
please share full dump_mempools output.
Igor Fedotov
10:34 AM Bug #54296 (Resolved): OSDs using too much memory
One of our customers upgraded from Nautilus to Octopus, and now a lot of his OSDs are using way more ram than allowed... Ruben Kerkhof

02/15/2022

11:21 PM Bug #54263: cephadm upgrade pacific to quincy autoscaler is scaling pgs from 32 -> 32768 for ceph...
In summary,
the root cause of the problem is after the upgrade to quincy, cephfs meta data pool was somehow given a ...
Kamoltat (Junior) Sirivadhna
10:57 PM Bug #53855 (Fix Under Review): rados/test.sh hangs while running LibRadosTwoPoolsPP.ManifestFlush...
Laura Flores
02:07 AM Bug #53855: rados/test.sh hangs while running LibRadosTwoPoolsPP.ManifestFlushDupCount
https://github.com/ceph/ceph/pull/45035 Myoungwon Oh
07:27 PM Bug #51904: test_pool_min_size:AssertionError:wait_for_clean:failed before timeout expired due to...
/a/yuriw-2022-02-08_17:00:23-rados-wip-yuri5-testing-2022-02-08-0733-pacific-distro-default-smithi/6670539
last pg...
Laura Flores
07:15 PM Bug #50222: osd: 5.2s0 deep-scrub : stat mismatch
Looks similar, but different test.
/a/yuriw-2022-02-09_22:52:18-rados-wip-yuri5-testing-2022-02-09-1322-pacific-di...
Laura Flores
06:55 PM Bug #45702: PGLog::read_log_and_missing: ceph_assert(miter == missing.get_items().end() || (miter...
/a/yuriw-2022-02-09_22:52:18-rados-wip-yuri5-testing-2022-02-09-1322-pacific-distro-default-smithi/6672070 Laura Flores
06:47 PM Bug #53327: osd: osd_fast_shutdown_notify_mon not quite right and enable osd_fast_shutdown_notify...
Hi Nitzan,
I checked your patch on the current pacific branch.
unfortunately I still get slow ops (slow >= 5 seco...
Manuel Lausch
06:46 PM Bug #48997: rados/singleton/all/recovery-preemption: defer backfill|defer recovery not found in logs
/a/yuriw-2022-02-09_22:52:18-rados-wip-yuri5-testing-2022-02-09-1322-pacific-distro-default-smithi/6672005 Laura Flores
03:45 PM Backport #54290 (Resolved): quincy: pybind/mgr/progress: disable pg recovery event by default
Kamoltat (Junior) Sirivadhna
03:42 PM Bug #47273 (Fix Under Review): ceph report missing osdmap_clean_epochs if answered by peon
Daniel van der Ster
03:08 AM Bug #52421: test tracker
Crash signature (v1) and Crash signature (v2) are of invalid format, and are breaking the telemetry crashes bot, remo... Yaarit Hatuka

02/14/2022

11:46 PM Bug #52124: Invalid read of size 8 in handle_recovery_delete()
/a/yuriw-2022-02-08_17:00:23-rados-wip-yuri5-testing-2022-02-08-0733-pacific-distro-default-smithi/6670360 Laura Flores
11:29 PM Bug #51234: LibRadosService.StatusFormat failed, Expected: (0) != (retry), actual: 0 vs 0
Pacific:
/a/yuriw-2022-02-09_22:52:18-rados-wip-yuri5-testing-2022-02-09-1322-pacific-distro-default-smithi/6672177
Laura Flores
08:21 PM Feature #54280 (Resolved): support truncation sequences in sparse reads
I've been working on sparse read support in the kclient, and got something working today, only to notice that after t... Jeff Layton
03:39 PM Bug #51076: "wait_for_recovery: failed before timeout expired" during thrashosd test with EC back...
/a/yuriw-2022-02-11_22:59:19-rados-wip-yuri4-testing-2022-02-11-0858-distro-default-smithi/6677733
Last pg map bef...
Laura Flores
10:06 AM Bug #46847: Loss of placement information on OSD reboot
Could somebody please set the status back to open and Affected Versions to all? Frank Schilder

02/11/2022

11:01 PM Backport #52769 (Resolved): octopus: pg scrub stat mismatch with special objects that have hash '...
Igor Fedotov
10:41 PM Backport #52769: octopus: pg scrub stat mismatch with special objects that have hash 'ffffffff'
Igor Fedotov wrote:
> https://github.com/ceph/ceph/pull/44978
merged
Yuri Weinstein
10:48 PM Bug #54263: cephadm upgrade pacific to quincy autoscaler is scaling pgs from 32 -> 32768 for ceph...
The following path has MGR logs, Mon logs, Cluster logs, audit logs, and system logs.... Vikhyat Umrao
10:39 PM Bug #54263 (Resolved): cephadm upgrade pacific to quincy autoscaler is scaling pgs from 32 -> 327...
Pacific version - 16.2.7-34.el8cp
Quincy version - 17.0.0-10315-ga00e8b31
After doing some analysis it looks like...
Vikhyat Umrao
09:23 PM Bug #54262 (Closed): ERROR: test_cluster_info (tasks.cephfs.test_nfs.TestNFS)
Since the PR has not merged yet, no need to create a tracker https://github.com/ceph/ceph/pull/44911#issuecomment-103... Neha Ojha
08:48 PM Bug #54262 (Closed): ERROR: test_cluster_info (tasks.cephfs.test_nfs.TestNFS)
/a/yuriw-2022-02-11_18:38:05-rados-wip-yuri-testing-2022-02-09-1607-distro-default-smithi/6677099/... Kamoltat (Junior) Sirivadhna
09:17 PM Backport #53769 (Resolved): pacific: [ceph osd set noautoscale] Global on/off flag for PG autosca...
Kamoltat (Junior) Sirivadhna
08:52 PM Feature #51213 (Resolved): [ceph osd set noautoscale] Global on/off flag for PG autoscale feature
Kamoltat (Junior) Sirivadhna
08:37 PM Bug #50089 (Fix Under Review): mon/MonMap.h: FAILED ceph_assert(m < ranks.size()) when reducing n...
https://github.com/ceph/ceph/pull/44993 Kamoltat (Junior) Sirivadhna
12:35 PM Bug #51338: osd/scrub_machine.cc: FAILED ceph_assert(state_cast&lt;const NotActive*&gt;())
I'm also encountering this issue on Pacific (16.2.7):... André Cruz
05:42 AM Bug #54255 (New): utc time is used when ceph crash ls
ceph crash id currently uses utc time but not local time
it is a little confused when debugging issues.
liqun zhang
01:23 AM Bug #53751: "N monitors have not enabled msgr2" is always shown for new clusters
Hmm, I've just tried to get rid of... Niklas Hambuechen
12:18 AM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
Added the logs for OSD 12,23,24 part of pg 4.6b. I don't think the logs are from the beginning when the osd booted an... Daan van Gorkum

02/10/2022

06:52 PM Bug #54238: cephadm upgrade pacifc to quincy -> causing osd's FULL/cascading failure
... Vikhyat Umrao
01:26 AM Bug #54238: cephadm upgrade pacifc to quincy -> causing osd's FULL/cascading failure
The node e24-h01-000-r640 has a file - upgrade.txt from the following command:... Vikhyat Umrao
12:58 AM Bug #54238 (New): cephadm upgrade pacifc to quincy -> causing osd's FULL/cascading failure
- Upgrade was started at 2022-02-08T01:54:28... Vikhyat Umrao
05:15 PM Backport #52771 (In Progress): nautilus: pg scrub stat mismatch with special objects that have ha...
https://github.com/ceph/ceph/pull/44981 Igor Fedotov
04:19 PM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
There was (I'll have to check in which Ceph versions) a bug, where setting noscrub or nodeepscrub at the "wrong"
tim...
Ronen Friedman
02:57 PM Backport #52769 (In Progress): octopus: pg scrub stat mismatch with special objects that have has...
https://github.com/ceph/ceph/pull/44978 Igor Fedotov
02:46 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Christian Rohmann wrote:
> Dieter, could you maybe describe your test setup a little more? How many instances of R...
Dieter Roels
01:34 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Christian Rohmann wrote:
> Dieter Roels wrote:
> > All inconsistencies were on non-primary shards, so we repaired t...
Christian Rohmann
01:17 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Dieter Roels wrote:
> Not sure if this helps or not, but we are experiencing very similar issues in our clusters the...
Christian Rohmann
01:12 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Christian Rohmann wrote:
> yite gu wrote:
> > This is inconsistent pg 7.2 from your upload files. It is look like m...
yite gu
12:17 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
yite gu wrote:
> This is inconsistent pg 7.2 from your upload files. It is look like mismatch osd is 10. So, you can...
Christian Rohmann
11:20 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
yite gu wrote:
> "shards": [
> {
> "osd": 10,
> "primary": false,
> "error...
yite gu
11:15 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
"shards": [
{
"osd": 10,
"primary": false,
"errors": [

...
yite gu
10:36 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Not sure if this helps or not, but we are experiencing very similar issues in our clusters the last few days.
We a...
Dieter Roels
10:06 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
yite gu wrote:
> Can you show me that primary osd log report when happen deep-scrub error?
> I hope to know which o...
Christian Rohmann
09:25 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Can you show me that primary osd log report when happen deep-scrub error?
I hope to know which osd shard happend error
yite gu
12:00 PM Bug #53751: "N monitors have not enabled msgr2" is always shown for new clusters
I installed the cluster using the "Manual Deployment" method (https://docs.ceph.com/en/pacific/install/manual-deploym... Niklas Hambuechen
08:50 AM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
Matan Breizman wrote:
> Shu Yu wrote:
> >
> > Missing a message, PG 8.243 status
> > # ceph pg ls 8 | grep -w 8....
Shu Yu

02/09/2022

08:58 PM Bug #23117 (Fix Under Review): PGs stuck in "activating" after osd_max_pg_per_osd_hard_ratio has ...
Prashant D
06:44 PM Backport #54233: octopus: devices: mon devices appear empty when scraping SMART metrics
please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/44960
ceph-backport.sh versi...
Benoît Knecht
02:36 PM Backport #54233 (Resolved): octopus: devices: mon devices appear empty when scraping SMART metrics
https://github.com/ceph/ceph/pull/44960 Backport Bot
06:41 PM Backport #54232: pacific: devices: mon devices appear empty when scraping SMART metrics
please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/44959
ceph-backport.sh versi...
Benoît Knecht
02:36 PM Backport #54232 (Resolved): pacific: devices: mon devices appear empty when scraping SMART metrics
https://github.com/ceph/ceph/pull/44959 Backport Bot
06:39 PM Bug #52416: devices: mon devices appear empty when scraping SMART metrics
Ah, indeed! I don't think I would have been able to change the status myself though, so thanks for doing it! Benoît Knecht
02:32 PM Bug #52416 (Pending Backport): devices: mon devices appear empty when scraping SMART metrics
Thanks, Benoît,
Once the status is changed to "Pending Backport" the bot should find it.
Yaarit Hatuka
10:09 AM Bug #52416: devices: mon devices appear empty when scraping SMART metrics
I'd like to backport this to Pacific and Octopus, but the Backport Bot didn't create the corresponding tickets; what ... Benoît Knecht
04:17 PM Bug #54210: pacific: mon/pg_autoscaler.sh: echo failed on "bash -c 'ceph osd pool get a pg_num | ...
Laura Flores wrote:
> [...]
>
> Also seen in pacific:
> /a/yuriw-2022-02-05_22:51:11-rados-wip-yuri2-testing-202...
Laura Flores

02/08/2022

08:13 PM Bug #51904: test_pool_min_size:AssertionError:wait_for_clean:failed before timeout expired due to...
/a/yuriw-2022-02-05_22:51:11-rados-wip-yuri2-testing-2022-02-04-1646-pacific-distro-default-smithi/6663906
last pg...
Laura Flores
07:21 PM Bug #54210: pacific: mon/pg_autoscaler.sh: echo failed on "bash -c 'ceph osd pool get a pg_num | ...
Junior, maybe you have an idea of what's going on? Laura Flores
07:20 PM Bug #54210 (Resolved): pacific: mon/pg_autoscaler.sh: echo failed on "bash -c 'ceph osd pool get ...
... Laura Flores
03:25 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
I did run a manual deep-scrub on another inconsistent PG as well, you'll find the logs of all OSDs handling this PG i... Christian Rohmann
03:05 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Neha - I did upload the logs of a deep-scrub via ceph-post-file: 1e5ff0f8-9b76-4489-8529-ee5e6f246093
There is a lit...
Christian Rohmann
01:49 PM Bug #45457: CEPH Graylog Logging Missing "host" Field
So is this going to be backported to Pacific? Jan-Philipp Litza
03:08 AM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
Some more information on the non-default config settings (excluding MGR):... Daan van Gorkum
03:02 AM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
One more thing to add is that when I set the noscrub, nodeep-scrub flags the pgs actually don't stop scrubbing either... Daan van Gorkum
12:16 AM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
Correction. osd_recovery_sleep_hdd was set to 0.0 from the original 0.1. osd_scrub_sleep has been untouched. Daan van Gorkum
12:12 AM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
This is set to 0.0. I think we did this to speed up recovery after we did some CRUSH tuning.... Daan van Gorkum

02/07/2022

11:33 PM Bug #53924: EC PG stuckrecovery_unfound+undersized+degraded+remapped+peered
PG 7.dc4 - all osd logs. Vikhyat Umrao
11:14 PM Bug #53924: EC PG stuckrecovery_unfound+undersized+degraded+remapped+peered
- PG query:... Vikhyat Umrao
09:31 PM Bug #53924: EC PG stuckrecovery_unfound+undersized+degraded+remapped+peered
Vikhyat Umrao wrote:
> Vikhyat Umrao wrote:
> > - This was reproduced again today
>
> As this issue is random we...
Vikhyat Umrao
08:30 PM Bug #53924: EC PG stuckrecovery_unfound+undersized+degraded+remapped+peered
Vikhyat Umrao wrote:
> - This was reproduced again today
As this issue is random we did not have debug logs from ...
Vikhyat Umrao
08:27 PM Bug #53924: EC PG stuckrecovery_unfound+undersized+degraded+remapped+peered
- This was reproduced again today... Vikhyat Umrao
11:17 PM Bug #54188 (Resolved): Setting too many PGs leads error handling overflow
This happened on gibba001:... Mark Nelson
11:15 PM Bug #54166: ceph version 15.2.15, osd configuration osd_op_num_shards_ssd or osd_op_num_threads_p...
Sridhar, can you please take a look? Neha Ojha
11:12 PM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
What is osd_scrub_sleep set to?
Ronen, this sounds similar to one of the issues you were looking into, here it is ...
Neha Ojha
02:46 AM Bug #54172: ceph version 16.2.7 PG scrubs not progressing
Additional information:
I tried disabling all client I/O and after that there's zero I/O on the devices hosting th...
Daan van Gorkum
02:45 AM Bug #54172 (Resolved): ceph version 16.2.7 PG scrubs not progressing
A week ago I've upgraded a 16.2.4 cluster (3 nodes, 33 osds) to 16.2.7 using cephadm and since then we're experiencin... Daan van Gorkum
10:57 PM Bug #53751 (Need More Info): "N monitors have not enabled msgr2" is always shown for new clusters
Can you share the output of "ceph mon dump"? And how did you install this cluster? We are not seeing this issue in 16... Neha Ojha
10:39 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Massive thanks for your reply Neha, I greatly appreciate it!
Neha Ojha wrote:
> Is it possible for you to trigg...
Christian Rohmann
10:31 PM Bug #53663 (Need More Info): Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadat...
Is it possible for you to trigger a deep-scrub on one PG (with debug_osd=20,debug_ms=1), let it go into inconsistent ... Neha Ojha
03:52 PM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
The issue is still happening:
1) Find all pools with scrub errors via...
Christian Rohmann
10:23 PM Bug #54182: OSD_TOO_MANY_REPAIRS cannot be cleared in >=Octopus
We can include clear_shards_repaired in master and backport it. Neha Ojha
04:47 PM Bug #54182 (New): OSD_TOO_MANY_REPAIRS cannot be cleared in >=Octopus
The newly added warning OSD_TOO_MANY_REPAIRS (https://tracker.ceph.com/issues/41564) is raised on a certain count of ... Christian Rohmann
10:14 PM Bug #46847: Loss of placement information on OSD reboot
Can you share your ec profile and the output of "ceph osd pool ls detail"? Neha Ojha
12:35 AM Bug #46847: Loss of placement information on OSD reboot
So in even more fun news, I created the EC pool according to the instructions provided in the documentation.
It's...
Malcolm Haak
12:17 AM Bug #46847: Loss of placement information on OSD reboot
Sorry I should add some context/data
ceph version 15.2.14 (cd3bb7e87a2f62c1b862ff3fd8b1eec13391a5be) octopus (stab...
Malcolm Haak
06:58 PM Bug #54180 (In Progress): In some cases osdmaptool takes forever to complete
Neha Ojha
02:44 PM Bug #54180 (Resolved): In some cases osdmaptool takes forever to complete
with the attached file run the command:
osdmaptool osdmap.GD.bin --upmap out-file --upmap-deviation 1 --upmap-pool d...
Josh Salomon
05:52 PM Bug #53806 (Fix Under Review): unessesarily long laggy PG state
Josh Durgin

02/06/2022

11:57 PM Bug #46847: Loss of placement information on OSD reboot
Neha Ojha wrote:
> Is this issue reproducible in Octopus or later?
Yes. I hit it last night. It's minced one of m...
Malcolm Haak
12:28 PM Bug #54166 (New): ceph version 15.2.15, osd configuration osd_op_num_shards_ssd or osd_op_num_thr...
Configure osd_op_num_shards_ssd=8 or osd_op_num_threads_per_shard_ssd=8 in ceph.config, use ceph daemin osd.x config ... chao zhang

02/04/2022

06:09 PM Bug #51076: "wait_for_recovery: failed before timeout expired" during thrashosd test with EC back...
http://pulpito.front.sepia.ceph.com/lflores-2022-01-31_19:11:11-rados:thrash-erasure-code-big-master-distro-default-s... Laura Flores
02:03 PM Backport #53480 (In Progress): pacific: Segmentation fault under Pacific 16.2.1 when using a cust...
Adam Kupczyk
12:00 AM Bug #53757 (Fix Under Review): I have a rados object that data size is 0, and this object have a ...
Neha Ojha

02/03/2022

09:26 PM Backport #53974 (Resolved): quincy: BufferList.rebuild_aligned_size_and_memory failure
Ilya Dryomov
07:08 PM Backport #53974 (In Progress): quincy: BufferList.rebuild_aligned_size_and_memory failure
https://github.com/ceph/ceph/pull/44891 Casey Bodley
07:46 PM Bug #23117 (In Progress): PGs stuck in "activating" after osd_max_pg_per_osd_hard_ratio has been ...
Prashant D
07:50 AM Bug #54122 (Fix Under Review): Validate monitor ID provided with ok-to-stop similar to ok-to-rm
Nitzan Mordechai
07:49 AM Bug #54122 (Resolved): Validate monitor ID provided with ok-to-stop similar to ok-to-rm
ceph mon ok-to-stop doesn't validate the monitor ID provided. Thus returns that "quorum should be preserved " without... Nitzan Mordechai

02/02/2022

05:25 PM Backport #53551: pacific: [RFE] Provide warning when the 'require-osd-release' flag does not matc...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44259
merged
Yuri Weinstein
02:04 PM Feature #54115 (In Progress): Log pglog entry size in OSD log if it exceeds certain size limit
Even after all PGs are active+clean, we see some OSDs consuming high amount memory. From dump_mempools, osd_pglog con... Prashant D
10:32 AM Bug #51002 (Resolved): regression in ceph daemonperf command output, osd columns aren't visible a...
Igor Fedotov
10:32 AM Backport #51172 (Resolved): pacific: regression in ceph daemonperf command output, osd columns ar...
Igor Fedotov
12:02 AM Backport #51172: pacific: regression in ceph daemonperf command output, osd columns aren't visibl...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44175
merged
Yuri Weinstein
12:04 AM Backport #53702: pacific: qa/tasks/backfill_toofull.py: AssertionError: 2.0 not in backfilling
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44387
merged
Yuri Weinstein

02/01/2022

10:12 PM Bug #50192: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_missing(soi...
Myoungwon Oh wrote:
> https://github.com/ceph/ceph/pull/44181
merged
Yuri Weinstein
08:42 PM Backport #53486: pacific: LibRadosTwoPoolsPP.ManifestSnapRefcount Failure.
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44202
merged
Yuri Weinstein
08:40 PM Backport #51150: pacific: When read failed, ret can not take as data len, in FillInVerifyExtent
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44173
merged
Yuri Weinstein
08:40 PM Backport #53388: pacific: pg-temp entries are not cleared for PGs that no longer exist
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44096
merged
Yuri Weinstein
05:23 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
I will collect these logs as you've requested. As an update: I am now seeing snaptrim occurring automatically without... David Prude
04:27 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
Yes that is the case.
Can you collect the log starting at the manual repeer? The intent was to capture the logs s...
Christopher Hoffman
03:23 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
Christopher Hoffman wrote:
> 1. A single OSD will be fine, just ensure it is one exhibiting the issue.
> 2. Can you...
David Prude
02:41 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
David Prude wrote:
> Christopher Hoffman wrote:
> > Can you collect and share OSD logs (with debug_osd=20 and debug...
Christopher Hoffman
12:01 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
Christopher Hoffman wrote:
> Can you collect and share OSD logs (with debug_osd=20 and debug_ms=1) when you are enco...
David Prude
11:38 AM Bug #53751: "N monitors have not enabled msgr2" is always shown for new clusters
Neha Ojha wrote:
> Maybe you are missing the square brackets when specifying the mon_host like in https://docs.ceph....
Niklas Hambuechen
12:12 AM Bug #53751: "N monitors have not enabled msgr2" is always shown for new clusters
Maybe you are missing the square brackets when specifying the mon_host like in https://docs.ceph.com/en/pacific/rados... Neha Ojha
08:54 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Neha Ojha wrote:
> Are you using filestore or bluestore?
On bluestore
Christian Rohmann
12:20 AM Bug #53663: Random scrub errors (omap_digest_mismatch) on pgs of RADOSGW metadata pools
Are you using filestore or bluestore? Neha Ojha
02:14 AM Bug #44184: Slow / Hanging Ops after pool creation
Neha Ojha wrote:
> So on both occasions this crash was a side effect of new pool creation? Can you provide the outpu...
Ist Gab
12:28 AM Bug #53667 (Fix Under Review): osd cannot be started after being set to stop
Neha Ojha

01/31/2022

11:59 PM Bug #44184: Slow / Hanging Ops after pool creation
Ist Gab wrote:
> Neha Ojha wrote:
> > Are you still seeing this problem? Will you be able to provide debug data aro...
Neha Ojha
11:53 PM Bug #54005 (Duplicate): Why can wrong parameters be specified when creating erasure-code-profile,...
Neha Ojha
11:38 PM Bug #53294: rados/test.sh hangs while running LibRadosTwoPoolsPP.TierFlushDuringFlush
/a/benhanokh-2022-01-26_21:12:05-rados-WIP_GBH_NCB_new_alloc_map_A6-distro-basic-smithi/6642148 Laura Flores
11:35 PM Bug #53767: qa/workunits/cls/test_cls_2pc_queue.sh: killing an osd during thrashing causes timeout
/a/benhanokh-2022-01-26_21:12:05-rados-WIP_GBH_NCB_new_alloc_map_A6-distro-basic-smithi/6642122 Laura Flores
10:56 PM Bug #53767: qa/workunits/cls/test_cls_2pc_queue.sh: killing an osd during thrashing causes timeout
/a/yuriw-2022-01-27_15:09:25-rados-wip-yuri6-testing-2022-01-26-1547-distro-default-smithi/6644093... Laura Flores
11:09 PM Bug #50192: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missing().is_missing(soi...
/a/yuriw-2022-01-27_15:09:25-rados-wip-yuri6-testing-2022-01-26-1547-distro-default-smithi/6644223 Laura Flores
11:07 PM Bug #53326 (Resolved): pgs wait for read lease after osd start
Vikhyat Umrao
10:21 PM Bug #51433 (Resolved): mgr spamming with repeated set pgp_num_actual while merging
nautilus is EOL Neha Ojha
10:20 PM Backport #53876 (Resolved): pacific: pgs wait for read lease after osd start
Neha Ojha
10:12 PM Bug #51076: "wait_for_recovery: failed before timeout expired" during thrashosd test with EC back...
/a/yuriw-2022-01-27_14:57:16-rados-wip-yuri-testing-2022-01-26-1810-pacific-distro-default-smithi/6643449 Laura Flores
09:27 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
Can you collect and share OSD logs (with debug_osd=20 and debug_ms=1) when you are encountering this issue? Christopher Hoffman
03:14 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
We are also seeing this issue on *16.2.5*. We schedule cephfs snapshots via cron in a 24h7d2w rotation schedule. Over... David Prude
07:52 PM Backport #54082: pacific: mon: osd pool create <pool-name> with --bulk flag
pull request: https://github.com/ceph/ceph/pull/44847 Kamoltat (Junior) Sirivadhna
07:51 PM Backport #54082 (Resolved): pacific: mon: osd pool create <pool-name> with --bulk flag
Backporting https://github.com/ceph/ceph/pull/44241 to pacific Kamoltat (Junior) Sirivadhna
07:32 PM Bug #45318: Health check failed: 2/6 mons down, quorum b,a,c,e (MON_DOWN)" in cluster log running...
Happening in Pacific too:
/a/yuriw-2022-01-27_14:57:16-rados-wip-yuri-testing-2022-01-26-1810-pacific-distro-defau...
Laura Flores
02:38 PM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
Shu Yu wrote:
>
> Missing a message, PG 8.243 status
> # ceph pg ls 8 | grep -w 8.243
> 8.243 0 0 ...
Matan Breizman
12:18 PM Backport #53660 (Resolved): octopus: mon: "FAILED ceph_assert(session_map.sessions.empty())" when...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/44544
m...
Loïc Dachary
12:18 PM Backport #53943 (Resolved): octopus: mon: all mon daemon always crash after rm pool
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/44700
m...
Loïc Dachary
12:18 PM Backport #53534 (Resolved): octopus: mon: mgrstatmonitor spams mgr with service_map
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/44722
m...
Loïc Dachary
12:17 PM Backport #53877 (Resolved): octopus: pgs wait for read lease after osd start
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/44585
m...
Loïc Dachary
12:17 PM Backport #53701 (Resolved): octopus: qa/tasks/backfill_toofull.py: AssertionError: 2.0 not in bac...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/43438
m...
Loïc Dachary
12:17 PM Backport #52833 (Resolved): octopus: osd: pg may get stuck in backfill_toofull after backfill is ...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/43438
m...
Loïc Dachary

01/28/2022

05:17 PM Backport #54048 (Rejected): octopus: [RFE] Add health warning in ceph status for filestore OSDs
Original tracker was accidentally marked for pending backport. We are not backporting related PR to pre-quincy. Marki... Prashant D
05:16 PM Backport #54047 (Rejected): nautilus: [RFE] Add health warning in ceph status for filestore OSDs
Original tracker was accidentally marked for pending backport. We are not backporting related PR to pre-quincy. Marki... Prashant D
05:14 PM Feature #49275 (Resolved): [RFE] Add health warning in ceph status for filestore OSDs
It was accidentally marked as pending backport. We are not backporting PR for this tracker to pre-quincy. Marking it ... Prashant D
01:41 PM Feature #49275: [RFE] Add health warning in ceph status for filestore OSDs
@Dan, I think maybe this is copy-paste issue? Konstantin Shalygin
01:24 PM Feature #49275: [RFE] Add health warning in ceph status for filestore OSDs
Why is this being backported to N and O?!
Filestore is deprecated since quincy, so we should only warn in quincy a...
Daniel van der Ster
05:01 PM Backport #53978: quincy: [RFE] Limit slow request details to mgr log
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44764
merged
Yuri Weinstein
10:08 AM Bug #54050: OSD: move message to cluster log when osd hitting the pg hard limit
PR: https://github.com/ceph/ceph/pull/44821 dongdong tao
10:02 AM Bug #54050 (Closed): OSD: move message to cluster log when osd hitting the pg hard limit
OSD will print the below message if a pg creation had hit the hard limit of the max number of pgs per osd.
---
202...
dongdong tao
05:50 AM Bug #52657 (In Progress): MOSDPGLog::encode_payload(uint64_t): Assertion `HAVE_FEATURE(features, ...
Aishwarya Mathuria

01/27/2022

06:36 PM Backport #54048 (Rejected): octopus: [RFE] Add health warning in ceph status for filestore OSDs
Backport Bot
06:36 PM Backport #54047 (Rejected): nautilus: [RFE] Add health warning in ceph status for filestore OSDs
Backport Bot
06:33 PM Feature #49275 (Pending Backport): [RFE] Add health warning in ceph status for filestore OSDs
Prashant D
06:33 PM Feature #49275 (Resolved): [RFE] Add health warning in ceph status for filestore OSDs
Prashant D
03:07 PM Bug #53729: ceph-osd takes all memory before oom on boot
Gonzalo Aguilar Delgado wrote:
> Hi,
>
> Nothing a script can't do:
>
> > ceph osd pool ls | xargs -n1 -istr ...
Mark Nelson
09:33 AM Bug #53729: ceph-osd takes all memory before oom on boot
Mark Nelson wrote:
> In the mean time, Neha mentioned that you might be able to prevent the pgs from splitting by tu...
Gonzalo Aguilar Delgado
09:24 AM Bug #53729: ceph-osd takes all memory before oom on boot
Mark Nelson wrote:
> Hi Gonzalo,
>
> I'm not an expert regarding this code so please take my reply here with a gr...
Gonzalo Aguilar Delgado
02:28 PM Bug #53327 (Fix Under Review): osd: osd_fast_shutdown_notify_mon not quite right and enable osd_f...
Nitzan Mordechai
12:05 AM Backport #53660: octopus: mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44544
merged
Yuri Weinstein
12:04 AM Backport #53943: octopus: mon: all mon daemon always crash after rm pool
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44700
merged
Yuri Weinstein

01/26/2022

11:54 PM Backport #53534: octopus: mon: mgrstatmonitor spams mgr with service_map
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/44722
merged
Yuri Weinstein
11:39 PM Backport #53769: pacific: [ceph osd set noautoscale] Global on/off flag for PG autoscale feature
Kamoltat Sirivadhna wrote:
> https://github.com/ceph/ceph/pull/44540
merged
Yuri Weinstein
08:53 PM Bug #53729: ceph-osd takes all memory before oom on boot
In the mean time, Neha mentioned that you might be able to prevent the pgs from splitting by turning off the autoscal... Mark Nelson
08:34 PM Bug #53729: ceph-osd takes all memory before oom on boot
Hi Gonzalo,
I'm not an expert regarding this code so please take my reply here with a grain of salt (and others pl...
Mark Nelson
05:26 PM Bug #53729: ceph-osd takes all memory before oom on boot
How can I help to accelerate a bugfix or workaround?
If comment your investigations I can builld a docker image t...
Gonzalo Aguilar Delgado
04:16 PM Bug #53326: pgs wait for read lease after osd start
https://github.com/ceph/ceph/pull/44585 merged Yuri Weinstein
12:27 AM Bug #53326: pgs wait for read lease after osd start
https://github.com/ceph/ceph/pull/44584 merged Yuri Weinstein
04:14 PM Backport #53701: octopus: qa/tasks/backfill_toofull.py: AssertionError: 2.0 not in backfilling
Mykola Golub wrote:
> PR: https://github.com/ceph/ceph/pull/43438
merged
Yuri Weinstein
04:14 PM Backport #52833: octopus: osd: pg may get stuck in backfill_toofull after backfill is interrupted...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/43438
merged
Yuri Weinstein
12:06 PM Bug #53142: OSD crash in PG::do_delete_work when increasing PGs
>Igor Fedotov wrote:
> I doubt anyone can say what setup would be good for you without experiments in the field. M...
Ist Gab
12:04 PM Bug #44184: Slow / Hanging Ops after pool creation
Neha Ojha wrote:
> Are you still seeing this problem? Will you be able to provide debug data around this issue?
H...
Ist Gab
12:47 AM Bug #45318 (New): Health check failed: 2/6 mons down, quorum b,a,c,e (MON_DOWN)" in cluster log r...
Octopus still has this issue /a/yuriw-2022-01-24_18:01:47-rados-wip-yuri10-testing-2022-01-24-0810-octopus-distro-def... Neha Ojha
 

Also available in: Atom