Activity
From 04/12/2020 to 05/11/2020
05/11/2020
- 09:54 PM Bug #45356: nautilus: rados/upgrade/mimic-x-singleton failures due to mon_client_directed_command...
- https://github.com/ceph/ceph/pull/34884 merged
- 04:41 PM Backport #44490 (In Progress): nautilus: lz4 compressor corrupts data when buffers are unaligned
- 02:23 PM Bug #44827 (Resolved): osd: incorrect read bytes stat in SPARSE_READ
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 02:21 PM Bug #45075 (Resolved): scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed failure
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 01:51 PM Backport #45392 (Resolved): octopus: follower monitors can grow beyond memory target
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34917
m... - 12:04 PM Bug #44959 (Closed): health warning: pgs not deep-scrubbed in time although it was in time
- Aaaha, that was it. Thank you very much!
I've set the @osd deep scrub interval@ under @[osd]@ so the mgr did not g... - 11:53 AM Bug #44959: health warning: pgs not deep-scrubbed in time although it was in time
- Have you changed the values on the MGR? mgr checks that and if mgr still has defaults, it will issue warnings..
@c... - 03:28 AM Bug #45298: cram: balancer/misplaced.t fails with 'Error EAGAIN: Some objects (0.008913) are degr...
- /a/yuriw-2020-05-04_17:54:17-rados-wip-yuri5-testing-2020-05-04-1554-nautilus-distro-basic-smithi/5022793
05/10/2020
- 02:39 AM Bug #45457 (Pending Backport): CEPH Graylog Logging Missing "host" Field
- Hello,
I have tried sending CEPH logs to Graylog with the following configuration:
mon_cluster_log_to_graylog =...
05/08/2020
- 07:28 PM Backport #45392: octopus: follower monitors can grow beyond memory target
- Sridhar Seshasayee wrote:
> https://github.com/ceph/ceph/pull/34917
merged - 03:06 PM Backport #45039 (Resolved): octopus: mon: reset min_size when changing pool size
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34528
m... - 03:06 PM Backport #44836 (Resolved): octopus: librados mon_command (mgr) command hang
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34416
m... - 12:10 PM Bug #24531: Mimic MONs have slow/long running ops
- Something similar, Ceph v14.2.0....
- 09:03 AM Bug #45390: FreeBSD: osdmap decode and encode does not give the same OSDMap
- Willem Jan Withagen wrote:
> Added code to dump JSON tree for both cases, and then it seems both trees are equal.
>... - 07:28 AM Bug #45441 (Resolved): rados: Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in clust...
/a/yuriw-2020-05-05_15:20:13-rados-wip-yuri8-testing-2020-05-04-2117-octopus-distro-basic-smithi/5024888...- 05:06 AM Backport #44841 (In Progress): nautilus: nautilus: FAILED ceph_assert(head.version == 0 || e.vers...
05/07/2020
- 10:37 PM Backport #45039: octopus: mon: reset min_size when changing pool size
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34528
merged - 10:33 PM Backport #44836: octopus: librados mon_command (mgr) command hang
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34416
merged - 08:54 PM Bug #45353: FAILED ceph_assert(pg_upmap.empty())
- Haven't been able to reproduce so far post https://github.com/ceph/ceph/pull/34748
- 06:18 PM Backport #44841 (Need More Info): nautilus: nautilus: FAILED ceph_assert(head.version == 0 || e.v...
- non-trivial because of https://github.com/ceph/ceph/pull/33910/commits/d4b1cc61e6526d325fd759f98e13e5a10523f5f7
- 03:47 PM Bug #44715: common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_list.back())->ops_in_...
- http://pulpito.ceph.com/swagner-2020-05-07_09:50:39-rados-wip-swagner3-testing-2020-05-06-1727-distro-basic-smithi/50...
- 06:27 AM Bug #45424 (New): api_watch_notify_pp: [ FAILED ] LibRadosWatchNotifyECPP.WatchNotify watch_not...
- /a/yuriw-2020-05-05_15:20:13-rados-wip-yuri8-testing-2020-05-04-2117-octopus-distro-basic-smithi/5024839...
- 06:19 AM Bug #45423 (Pending Backport): api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
- /a/yuriw-2020-05-05_15:20:13-rados-wip-yuri8-testing-2020-05-04-2117-octopus-distro-basic-smithi/5024839...
05/06/2020
- 10:19 PM Bug #45390: FreeBSD: osdmap decode and encode does not give the same OSDMap
- Added code to dump JSON tree for both cases, and then it seems both trees are equal.
So the serialized OSDMap contai... - 09:06 PM Bug #45381 (Need More Info): unfound objects in erasure-coded CephFS
- Is cache tiering involved here too? Do you have any osd logs from the same time?
- 08:05 AM Backport #45392 (In Progress): octopus: follower monitors can grow beyond memory target
- 06:20 AM Backport #45392 (Resolved): octopus: follower monitors can grow beyond memory target
- https://github.com/ceph/ceph/pull/34917
- 08:02 AM Backport #45391 (In Progress): nautilus: follower monitors can grow beyond memory target
- 06:18 AM Backport #45391 (Resolved): nautilus: follower monitors can grow beyond memory target
- https://github.com/ceph/ceph/pull/34916
05/05/2020
- 08:49 PM Bug #45390 (Closed): FreeBSD: osdmap decode and encode does not give the same OSDMap
- The problems occurs both in Octopus and Master.
This is the simple version of a part of test_compression.cc:
<pre... - 07:14 PM Backport #45357: octopus: rados: Sharded OpWQ drops suicide_grace after waiting for work
- Will do, thanks for the explanation.
- 01:12 PM Backport #45357: octopus: rados: Sharded OpWQ drops suicide_grace after waiting for work
- @Dan - please leave "Target Version" empty when you stage your backports.
The name "Target Version" is a bit of a ... - 05:06 PM Bug #45388 (New): Insufficient monitor logging to diagnose downed OSDs
- We just had a case where in a Ceph Luminous cluster the monitor forced newly started OSDs to commit suicide. Communic...
- 04:30 PM Backport #44468 (Resolved): nautilus: mon: Get session_map_lock before remove_session
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34677
m... - 04:26 PM Backport #45314 (Resolved): octopus: scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34830
m... - 04:25 PM Backport #45041 (Resolved): octopus: osd: incorrect read bytes stat in SPARSE_READ
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34809
m... - 04:25 PM Backport #44842 (Resolved): octopus: nautilus: FAILED ceph_assert(head.version == 0 || e.version....
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34807
m... - 04:25 PM Backport #44685 (Resolved): octopus: osd/osd-backfill-stats.sh TEST_backfill_out2: wait_for_clean...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34806
m... - 04:49 AM Bug #45266 (Pending Backport): follower monitors can grow beyond memory target
05/04/2020
- 09:04 PM Bug #45298 (Fix Under Review): cram: balancer/misplaced.t fails with 'Error EAGAIN: Some objects ...
- 08:42 PM Backport #44468: nautilus: mon: Get session_map_lock before remove_session
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34677
merged - 08:13 PM Backport #45314: octopus: scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed failure
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34830
merged - 08:13 PM Backport #45041: octopus: osd: incorrect read bytes stat in SPARSE_READ
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34809
merged - 08:11 PM Backport #44842: octopus: nautilus: FAILED ceph_assert(head.version == 0 || e.version.version > h...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34807
merged - 08:10 PM Backport #44685: octopus: osd/osd-backfill-stats.sh TEST_backfill_out2: wait_for_clean timeout
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34806
merged - 04:18 PM Bug #45076: rados: Sharded OpWQ drops suicide_grace after waiting for work
- This issue is also present in Luminous, which is EOL now that Octopus has released.
Should I open a tracker/pr fo... - 03:58 PM Bug #45381 (Need More Info): unfound objects in erasure-coded CephFS
- Encountered something weird with cephfs today that shouldn't happen
Setup:
* Ceph 14.2.8
* 8 OSD servers, 8 SS... - 03:25 PM Bug #44286: Cache tiering shows unfound objects after OSD reboots
- this occasionally comes up on the mailing list as well. it's not reproducible on my test setup, though :(
- 03:21 PM Bug #45356 (Fix Under Review): nautilus: rados/upgrade/mimic-x-singleton failures due to mon_clie...
- 10:55 AM Feature #43324 (Fix Under Review): Make zlib windowBits configurable for compression
- 09:14 AM Bug #44715: common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_list.back())->ops_in_...
- http://pulpito.ceph.com/swagner-2020-04-30_14:11:16-rados-wip-swagner2-testing-2020-04-29-1247-distro-basic-smithi/50...
05/02/2020
- 02:15 AM Backport #45358 (In Progress): mimic: rados: Sharded OpWQ drops suicide_grace after waiting for work
- 02:15 AM Backport #45358 (New): mimic: rados: Sharded OpWQ drops suicide_grace after waiting for work
- 01:16 AM Backport #45358 (In Progress): mimic: rados: Sharded OpWQ drops suicide_grace after waiting for work
- 01:13 AM Backport #45358 (Rejected): mimic: rados: Sharded OpWQ drops suicide_grace after waiting for work
- https://github.com/ceph/ceph/pull/34883
- 02:14 AM Backport #45359 (In Progress): nautilus: rados: Sharded OpWQ drops suicide_grace after waiting fo...
- 02:13 AM Backport #45359 (New): nautilus: rados: Sharded OpWQ drops suicide_grace after waiting for work
- 01:17 AM Backport #45359 (In Progress): nautilus: rados: Sharded OpWQ drops suicide_grace after waiting fo...
- 01:13 AM Backport #45359 (Resolved): nautilus: rados: Sharded OpWQ drops suicide_grace after waiting for work
- https://github.com/ceph/ceph/pull/34882
- 02:11 AM Backport #45357 (In Progress): octopus: rados: Sharded OpWQ drops suicide_grace after waiting for...
- 02:10 AM Backport #45357 (New): octopus: rados: Sharded OpWQ drops suicide_grace after waiting for work
- 01:14 AM Backport #45357 (In Progress): octopus: rados: Sharded OpWQ drops suicide_grace after waiting for...
- 01:13 AM Backport #45357 (Resolved): octopus: rados: Sharded OpWQ drops suicide_grace after waiting for work
- https://github.com/ceph/ceph/pull/34881
05/01/2020
- 11:49 PM Bug #45076 (Pending Backport): rados: Sharded OpWQ drops suicide_grace after waiting for work
- 10:49 PM Bug #45353: FAILED ceph_assert(pg_upmap.empty())
- Damn, missed that thanks Neha. Let me run this again on current master.
- 09:06 PM Bug #45353: FAILED ceph_assert(pg_upmap.empty())
- We have removed jewel from thrash-old-clients in https://github.com/ceph/ceph/pull/34748. We should check if this fai...
- 06:56 PM Bug #45353: FAILED ceph_assert(pg_upmap.empty())
- 'rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml
1-install/jewel.yaml backoff/normal.yaml... - 05:11 AM Bug #45353 (New): FAILED ceph_assert(pg_upmap.empty())
- /a/bhubbard-2020-05-01_01:03:08-rados-wip-yuri-testing-2020-04-24-1941-master-distro-basic-smithi/5003239...
- 04:52 PM Bug #45356 (Resolved): nautilus: rados/upgrade/mimic-x-singleton failures due to mon_client_direc...
- ...
04/30/2020
- 06:55 AM Bug #45298: cram: balancer/misplaced.t fails with 'Error EAGAIN: Some objects (0.008913) are degr...
- This looks similar....
- 06:48 AM Bug #45345: tasks/rados.py fails with "psutil.NoSuchProcess: psutil.NoSuchProcess process no long...
- /a/teuthology-2020-04-26_07:01:02-rados-master-distro-basic-smithi/4985956
- 06:39 AM Bug #45345 (Can't reproduce): tasks/rados.py fails with "psutil.NoSuchProcess: psutil.NoSuchProce...
- /a/yuriw-2020-04-28_21:58:13-rados-wip-yuri-testing-2020-04-24-1941-master-distro-basic-smithi/4995279
Looking at ...
04/29/2020
- 11:53 PM Bug #45266 (Fix Under Review): follower monitors can grow beyond memory target
- 02:04 PM Bug #45266: follower monitors can grow beyond memory target
- Taking ownership of this.
-Sridhar - 11:36 PM Bug #45298 (In Progress): cram: balancer/misplaced.t fails with 'Error EAGAIN: Some objects (0.00...
- No success in reproducing this so far: http://pulpito.ceph.com/nojha-2020-04-29_18:44:55-rados:singleton-nomsgr-maste...
- 09:26 PM Bug #45240: Not able to export objects using ceph-objectstore-tool
- I don't think this is a bug in the ceph-objectstore-tool but more a case of export failing when it encounters corrupt...
- 09:17 PM Bug #45292 (Need More Info): pg autoscaler merging issue
- Can you provide pg query output for one of those PGs? Also, osd logs with debug_osd=20 will be helpful.
- 12:17 PM Backport #45314 (In Progress): octopus: scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_fai...
- 03:52 AM Bug #45318 (New): Health check failed: 2/6 mons down, quorum b,a,c,e (MON_DOWN)" in cluster log r...
- /a/teuthology-2020-04-26_02:30:03-rados-octopus-distro-basic-smithi/4984906
The MON log shows it came back up arou... - 03:15 AM Bug #44715: common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_list.back())->ops_in_...
- /a/teuthology-2020-04-26_02:30:03-rados-octopus-distro-basic-smithi/4984693
/a/teuthology-2020-04-26_02:30:03-rados-... - 03:15 AM Bug #42347: nautilus assert during osd shutdown: FAILED ceph_assert((sharded_in_flight_list.back(...
- /a/teuthology-2020-04-26_02:30:03-rados-octopus-distro-basic-smithi/4984693
04/28/2020
- 08:04 PM Bug #44076 (Resolved): mon: update + monmap update triggers spawn loop
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 08:04 PM Bug #44248 (Resolved): Receiving RemoteBackfillReserved in WaitLocalBackfillReserved can cause th...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 08:02 PM Backport #45314 (Resolved): octopus: scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed...
- https://github.com/ceph/ceph/pull/34830
- 07:28 PM Support #45270 (Resolved): after reboot osd move to localhost
- I believe this has been discussed several times on the mailing list. If your OSDs don't get reliably told what their ...
- 05:51 PM Backport #45041 (In Progress): octopus: osd: incorrect read bytes stat in SPARSE_READ
- 05:47 PM Backport #44842 (In Progress): octopus: nautilus: FAILED ceph_assert(head.version == 0 || e.versi...
- 05:46 PM Backport #44685 (In Progress): octopus: osd/osd-backfill-stats.sh TEST_backfill_out2: wait_for_cl...
- 03:44 PM Bug #45075 (Pending Backport): scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed failure
- 12:05 AM Bug #45075: scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed failure
- https://github.com/ceph/ceph/pull/34602 merged
- 09:25 AM Backport #44324: nautilus: Receiving RemoteBackfillReserved in WaitLocalBackfillReserved can caus...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34512
m... - 03:01 AM Backport #44324 (Resolved): nautilus: Receiving RemoteBackfillReserved in WaitLocalBackfillReserv...
- 09:25 AM Backport #44289: nautilus: mon: update + monmap update triggers spawn loop
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34500
m... - 03:00 AM Backport #44289 (Resolved): nautilus: mon: update + monmap update triggers spawn loop
- 02:56 AM Backport #44370 (In Progress): nautilus: msg/async: the event center is blocked by rdma construct...
- 02:05 AM Bug #45298 (Resolved): cram: balancer/misplaced.t fails with 'Error EAGAIN: Some objects (0.00891...
- /a/teuthology-2020-04-26_07:01:02-rados-master-distro-basic-smithi/4985666...
- 01:31 AM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
- /a/teuthology-2020-04-26_07:01:02-rados-master-distro-basic-smithi/4986119
04/27/2020
- 09:32 PM Backport #44324: nautilus: Receiving RemoteBackfillReserved in WaitLocalBackfillReserved can caus...
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34512
merged - 06:03 PM Bug #45292 (Need More Info): pg autoscaler merging issue
- Encountering an issue where placement groups (pgs) go into status *stuck inactive* and hang in that status. This appe...
- 11:51 AM Bug #44286: Cache tiering shows unfound objects after OSD reboots
- Issue still present on 14.2.8.
- 08:25 AM Bug #41735 (Resolved): pg_autoscaler throws HEALTH_WARN with auto_scale on for all pools
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 08:24 AM Bug #43365 (Resolved): Nautilus: Random mon crashes in failed assertion at ceph::time_detail::sig...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 03:30 AM Feature #43377 (Resolved): Make Zstandard compression level a configurable option
04/26/2020
- 02:16 PM Support #45270 (Resolved): after reboot osd move to localhost
- if my host retrieve hostname from DNS server PTR line, not set from hostnamectl set-hostname node01, i have next prob...
04/25/2020
- 06:24 PM Bug #45202: Repeatedly OSD crashes in PrimaryLogPG::hit_set_trim()
- I'm got a crash for another OSD on 4rd node, and last lines in log are not related to PG 2.f8:...
- 12:55 PM Bug #45253: Inconsistent characters allowed set for device classes
- Version is mimic 13.2.8. Sorry, forgot.
- 12:02 AM Bug #45266 (Resolved): follower monitors can grow beyond memory target
- The leader monitor periordically tells tcmalloc to release memory back to the OS, but follower monitors do not. This ...
04/24/2020
- 07:24 PM Backport #45231 (Resolved): nautilus: pg_autoscaler throws HEALTH_WARN with auto_scale on for all...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34618
m... - 06:20 PM Backport #45231 (In Progress): nautilus: pg_autoscaler throws HEALTH_WARN with auto_scale on for ...
- 06:31 PM Backport #44486 (Resolved): nautilus: Nautilus: Random mon crashes in failed assertion at ceph::t...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34542
m... - 03:07 PM Bug #45241: Error message: Mount failed with '(22) Invalid argument' when trying to import using ...
- Using --debug it turned out that the exported PGs have different fsid than the odd-directory I am trying to import. f...
- 09:28 AM Bug #45253 (New): Inconsistent characters allowed set for device classes
- I changed the device class of a number of disks yesterday successfully to "rbd.meta":...
- 02:59 AM Bug #45243 (New): nautilus: qa/standalone/scrub/osd-scrub-repair.sh fails with osd-scrub-repair.s...
- /a/yuriw-2020-04-18_19:56:53-rados-wip-yuri4-testing-2020-04-18-1756-nautilus-distro-basic-smithi/4965037...
- 02:09 AM Fix #45140: osd/tiering: flush cache pool may lead to slow write requests
- https://github.com/ceph/ceph/pull/34623
04/23/2020
- 10:48 PM Bug #45241 (New): Error message: Mount failed with '(22) Invalid argument' when trying to import ...
Hi !
When I am running the following command, I get a not so descriptive error messages:
ceph-objectstore-tool --...- 09:06 PM Bug #45240 (New): Not able to export objects using ceph-objectstore-tool
- I am trying to use the tool ceph-objectstore-tool to extract objects from offline OSD:
ceph-objectstore-tool --dat... - 03:47 PM Bug #43365: Nautilus: Random mon crashes in failed assertion at ceph::time_detail::signedspan
- For what it's worth we're still seeing it after upgrading debian to 10.3 and installing kernel "5.4.0-0.bpo.3-amd64 #...
- 01:39 PM Bug #40112 (Resolved): mon: rados/multimon tests fail with clock skew
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 01:38 PM Backport #45231 (Resolved): nautilus: pg_autoscaler throws HEALTH_WARN with auto_scale on for all...
- https://github.com/ceph/ceph/pull/34618
- 01:36 PM Bug #43889 (Resolved): expected MON_CLOCK_SKEW but got none
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 01:36 PM Documentation #43896 (Resolved): nautilus upgrade should recommend ceph-osd restarts after enabli...
- While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ...
- 01:36 PM Backport #45224 (Resolved): nautilus: LibRadosWatchNotify.WatchNotify failure
- https://github.com/ceph/ceph/pull/35049
- 12:30 PM Bug #45202: Repeatedly OSD crashes in PrimaryLogPG::hit_set_trim()
- After a hour of work both OSD's continued to crash every several seconds
- 12:08 PM Bug #45202 (New): Repeatedly OSD crashes in PrimaryLogPG::hit_set_trim()
- After a network troubles I got 1 pg in a state recovery_unfound
I tried to solve this problem using command:
<pre... - 08:20 AM Backport #45040 (Resolved): nautilus: mon: reset min_size when changing pool size
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34585
m... - 08:17 AM Backport #44908 (Resolved): mimic: mon: rados/multimon tests fail with clock skew
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34370
m... - 08:17 AM Backport #44083 (Resolved): mimic: expected MON_CLOCK_SKEW but got none
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34370
m...
04/22/2020
- 11:53 PM Bug #44062 (Pending Backport): LibRadosWatchNotify.WatchNotify failure
- 11:52 PM Bug #44062: LibRadosWatchNotify.WatchNotify failure
- Seeing this in Nautilus so setting backport.
http://pulpito.ceph.com/yuriw-2020-04-21_20:54:00-rados-wip-yuri8-tes... - 05:43 PM Bug #45191 (New): erasure-code/test-erasure-eio.sh: TEST_ec_single_recovery_error fails
- ...
- 05:15 PM Bug #45190 (New): osd dump times out
- ...
- 10:04 AM Backport #44468 (In Progress): nautilus: mon: Get session_map_lock before remove_session
04/21/2020
- 11:19 PM Backport #44908: mimic: mon: rados/multimon tests fail with clock skew
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34370
merged - 11:19 PM Backport #44083: mimic: expected MON_CLOCK_SKEW but got none
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34370
merged - 11:09 PM Backport #45040: nautilus: mon: reset min_size when changing pool size
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/34585
merged - 11:08 PM Bug #45168 (New): mimic: cephtool/test.sh: test_mon_osd_pool_set failure
- ...
- 06:46 PM Backport #45053 (Resolved): octopus: nautilus upgrade should recommend ceph-osd restarts after en...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34523
m... - 06:37 PM Backport #45054 (Resolved): nautilus: nautilus upgrade should recommend ceph-osd restarts after e...
- This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/34524
m... - 12:34 PM Bug #44715 (Fix Under Review): common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_li...
- 02:45 AM Bug #39039: mon connection reset, command not resent
- I tested this on a lab cluster after disabling cephx per https://docs.ceph.com/docs/octopus/rados/configuration/auth-...
- 02:10 AM Bug #45113 (Resolved): workunits/cls/test_cls_cmpomap.sh fails
- 02:10 AM Bug #44901: luminous: osd continue down because of the hearbeattimeout
- Slove it! It is because we deploy ceph in the docker use kolla asible.
We start some dockers by hand and miss some...
04/20/2020
- 02:24 AM Bug #39039: mon connection reset, command not resent
- This also continues to happen on octopus, I just tested on 15.2.0.
I have attached the build instructions I used t...
04/18/2020
- 08:06 AM Fix #45140: osd/tiering: flush cache pool may lead to slow write requests
- Pull request ID: 34623
- 07:26 AM Fix #45140 (New): osd/tiering: flush cache pool may lead to slow write requests
- In OSD tiering, when flushing objects from cache pool to base pool, there are two problems can lead to slow request:
...
04/17/2020
- 10:25 PM Bug #45139: osd/osd-markdown.sh: markdown_N_impl failure
- This was seen after the fix for https://tracker.ceph.com/issues/44662 merged.
- 10:24 PM Bug #45139 (New): osd/osd-markdown.sh: markdown_N_impl failure
- ...
- 09:57 PM Bug #43888: osd/osd-bench.sh 'tell osd.N bench' hang
- Still fails occasionally
/a/nojha-2020-04-10_22:42:57-rados:standalone-master-distro-basic-smithi/4943804/ - 05:02 PM Bug #41735: pg_autoscaler throws HEALTH_WARN with auto_scale on for all pools
- nautilus backport: https://github.com/ceph/ceph/pull/34618
- 04:54 PM Bug #41735 (Pending Backport): pg_autoscaler throws HEALTH_WARN with auto_scale on for all pools
- This change needs to be backported into Nautilus to fix a regression (#45135)
- 07:21 AM Bug #45113 (Fix Under Review): workunits/cls/test_cls_cmpomap.sh fails
04/16/2020
- 11:23 PM Bug #45121 (New): nautilus: osd-scrub-snaps.sh: TEST_scrub_snaps failure
- ...
- 11:09 PM Bug #45075 (Fix Under Review): scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed failure
- 07:44 PM Bug #45075 (In Progress): scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed failure
- 05:50 PM Bug #45075: scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed failure
- /a/bhubbard-2020-04-16_09:57:54-rados-wip-badone-testing-distro-basic-smithi/4957883/
- 04:15 PM Bug #45113 (Triaged): workunits/cls/test_cls_cmpomap.sh fails
- Thank you Casey! i will see if we can use the default list.
- 02:54 PM Bug #45113: workunits/cls/test_cls_cmpomap.sh fails
- i didn't realize this ran in the rados suite. it's passing in the rgw/verify suite
it looks like the rados suite... - 02:25 PM Bug #45113 (Resolved): workunits/cls/test_cls_cmpomap.sh fails
- ...
- 09:39 AM Backport #45038 (In Progress): mimic: mon: reset min_size when changing pool size
- 09:37 AM Backport #45040 (In Progress): nautilus: mon: reset min_size when changing pool size
- 08:28 AM Feature #44025 (Resolved): Make it harder to set pool replica size to 1
- 12:30 AM Bug #45076 (Fix Under Review): rados: Sharded OpWQ drops suicide_grace after waiting for work
04/15/2020
- 09:33 PM Bug #45008: [osd crash]The ceph-osd assert with rbd bench io
- Sebastian Wagner wrote:
> duplicate of 44715 ?
Looks like a dup of 42347, which was on the osd. - 03:54 PM Bug #45008: [osd crash]The ceph-osd assert with rbd bench io
- duplicate of 44715 ?
- 03:54 PM Bug #44715: common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_list.back())->ops_in_...
- http://pulpito.ceph.com/swagner-2020-04-15_09:10:55-rados-wip-swagner2-testing-2020-04-14-1813-distro-basic-smithi/
04/14/2020
- 10:25 AM Feature #45079 (New): HEALTH_WARN, if require-osd-release is < mimic and OSD wants to join the cl...
- When upgrading a cluster to octopus, users should get a warning, if require-osd-release is < mimic as this prevents o...
- 08:54 AM Bug #44715: common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_list.back())->ops_in_...
- http://pulpito.ceph.com/swagner-2020-04-09_21:46:02-rados-wip-swagner2-testing-2020-04-09-1541-distro-basic-smithi/
04/13/2020
- 10:39 PM Bug #45076 (Resolved): rados: Sharded OpWQ drops suicide_grace after waiting for work
- The Sharded OpWQ will opportunistically wait for more work when processing an empty queue. While waiting, the default...
- 07:45 PM Bug #45075 (Resolved): scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_failed failure
- ...
- 06:08 PM Backport #44486 (In Progress): nautilus: Nautilus: Random mon crashes in failed assertion at ceph...
- 02:44 PM Backport #43232: nautilus: pgs stuck in laggy state
- @Neha - Can you make a decision whether to backport this to nautilus or not? Sage wrote:
"I'm not sure whether we ...
04/12/2020
Also available in: Atom