Activity
From 03/05/2019 to 04/03/2019
04/03/2019
- 11:28 PM Bug #39099 (Resolved): Give recovery for inactive PGs a higher priority
Backfill inactive gets priority 220 and we should make sure that if we can have inactive that needs recovery only i...- 07:23 AM Bug #39087: ec_lost_unfound: a EC shard has missing object after `osd lost`
- is this `scrub error` we expect? what we should do is to find out why ceph doesn't recovery PG 2.4s0 ?
- 07:16 AM Bug #39087 (New): ec_lost_unfound: a EC shard has missing object after `osd lost`
- http://pulpito.ceph.com/kchai-2019-04-01_10:38:29-rados-wip-kefu-testing-2019-04-01-1531-distro-basic-mira/3797065/
... - 04:37 AM Feature #38616 (Resolved): Improvements to auto repair
04/02/2019
- 10:22 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- Per request on irc.
pg log:
1.cas2 on osd.2: ceph-post-file: d74a0006-c0e9-41b1-a904-7bfe41617253
1.96s3 on osd.... - 07:51 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- Output from: ceph-objectstore-tool --no-mon-config --data-path /var/lib/ceph/osd/ceph-0 --op log --pgid 1.cas0
1.c... - 06:29 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- Hi Grant, is there a way you could dump the pg log by using a command like this "ceph-objectstore-tool --no-mon-confi...
- 09:51 PM Bug #39085 (Fix Under Review): monmap created timestamp may be blank
- 09:51 PM Bug #39085: monmap created timestamp may be blank
- https://github.com/ceph/ceph/pull/27327
- 07:13 PM Bug #39085 (Resolved): monmap created timestamp may be blank
- On at least one old cluster, monmap created timestamp is empty. lab cluster:...
- 07:27 PM Bug #38219: rebuild-mondb hangs
- I reproduced this again on master, http://pulpito.ceph.com/nojha-2019-04-02_17:39:35-rados:singleton-master-distro-ba...
- 11:46 AM Bug #38219: rebuild-mondb hangs
- http://pulpito.ceph.com/kchai-2019-04-02_08:04:13-rados-wip-kefu-testing-2019-04-01-1531-distro-basic-smithi/
- 01:02 PM Bug #38124: OSD down on snaptrim.
- Hello it's been two months now is there any update about this bug?
- 12:19 PM Bug #38783: Changing mon_pg_warn_max_object_skew has no effect.
- 12:18 PM Bug #38783: Changing mon_pg_warn_max_object_skew has no effect.
- It's an mgr option. You should instead inject it to the mgr daemon.
- 05:56 AM Backport #38905 (In Progress): luminous: osd/PGLog.h: print olog_can_rollback_to before deciding ...
- https://github.com/ceph/ceph/pull/27715
- 01:52 AM Backport #38983 (Resolved): nautilus: Improvements to auto repair
- 12:12 AM Backport #38906 (In Progress): nautilus: osd/PGLog.h: print olog_can_rollback_to before deciding ...
- https://github.com/ceph/ceph/pull/27302
04/01/2019
- 10:59 PM Bug #37439: Degraded PG does not discover remapped data on originating OSD
- My proposal to fix this bug is to call @discover_all_missing@ not only if there are missing objects, but also when th...
- 09:11 PM Bug #37439: Degraded PG does not discover remapped data on originating OSD
- Hi Jonas, thanks for creating a fix for this bug. Could you please upload the latest logs from nautilus, that you hav...
- 08:58 PM Bug #37439 (Fix Under Review): Degraded PG does not discover remapped data on originating OSD
- 01:07 AM Bug #37439: Degraded PG does not discover remapped data on originating OSD
- More findings, now on Nautilus 14.2.0:
OSD.62 once was part of pg 6.65, but content on it got remapped. A restart ... - 10:46 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- Grant: I notice that the initial event outlined above is from October. Is that the very first anomalous behavior exh...
- 10:45 PM Feature #3362 (Resolved): Warn users before allowing pools to be created with more than N*<num_os...
- 09:22 PM Backport #38442: luminous: osd-markdown.sh can fail with CLI_DUP_COMMAND=1
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26616
merged - 09:22 PM Backport #38442: luminous: osd-markdown.sh can fail with CLI_DUP_COMMAND=1
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26616
merged - 04:10 PM Fix #39071 (New): monclient: initial probe is non-optimal with v2+v1
- When we are probing both v2 and v1 addrs for mons, we treat them as separate mons, which means we might be probing N ...
- 02:25 PM Feature #39066 (Fix Under Review): src/ceph-disk/tests/ceph-disk.sh is using hardcoded port
- 02:21 PM Feature #39066 (Resolved): src/ceph-disk/tests/ceph-disk.sh is using hardcoded port
- Currently it's only possible to run `...make; make tests -j8; ctest ...` on the same machine.
Please consider chan... - 01:48 PM Bug #38945 (Pending Backport): osd: leaked pg refs on shutdown
- 01:09 PM Bug #38219: rebuild-mondb hangs
- i am using following script to reproduce this issue locally, so far no luck...
- 11:51 AM Bug #39059 (Can't reproduce): assert in ceph::net::SocketMessenger::unregister_conn()
- ...
- 03:22 AM Bug #39056: localize-reads does not increment pg stats read count
- when set the flag of '--localize-reads', maybe peer_pg will complete read task, but peer_pg will not count read_num....
- 03:09 AM Bug #39056 (New): localize-reads does not increment pg stats read count
- when I mounted ceph-fuse, I setted the flag of '--localize-reads'. I found during the test that read_num count was In...
- 12:52 AM Backport #38904: mimic: osd/PGLog.h: print olog_can_rollback_to before deciding to rollback
- https://github.com/ceph/ceph/pull/27284
03/31/2019
- 07:02 PM Bug #39055 (New): OSD's crash when specific PG is trying to backfill
- Hi,
I've got a peculiar issue whereby a specific PG is trying to backfill it's objects to the other peers, but th... - 12:08 PM Bug #39054 (Closed): osd push failed because local copy is 4394'133607637
- ceph-osd.1.log:7085:2019-02-27 13:07:21.336004 7f666b5bb700 -1 log_channel(cluster) log [ERR] : 3.33 push 3:ccb8da9c:...
03/30/2019
- 07:14 PM Bug #38931: osd does not proactively remove leftover PGs
- https://github.com/ceph/ceph/pull/27205/commits/f7c5b01e181630bb15e8b923b0334eb6adfdf50a
- 06:15 PM Bug #39053 (New): changing pool crush rule may lead to IO stop
How to reproduce:
1. create some OSDs
2. change their class to, say, "xxx"
3. create replicated crush rule ref...- 01:37 PM Backport #38860 (Resolved): nautilus: upmap broken the crush rule
- 08:46 AM Bug #38784 (Pending Backport): osd: FAILED ceph_assert(attrs || !pg_log.get_missing().is_missing(...
- 08:21 AM Backport #38854 (Resolved): luminous: .mgrstat failed to decode mgrstat state; luminous dev version?
- 08:21 AM Backport #38859 (Resolved): luminous: upmap broken the crush rule
- 08:20 AM Backport #38857 (Resolved): luminous: should set EPOLLET flag on del_event()
- 08:18 AM Backport #39044 (Resolved): mimic: osd/PGLog: preserve original_crt to check rollbackability
- https://github.com/ceph/ceph/pull/27629
- 08:18 AM Backport #39043 (Resolved): nautilus: osd/PGLog: preserve original_crt to check rollbackability
- https://github.com/ceph/ceph/pull/27632
- 08:18 AM Backport #39042 (Resolved): luminous: osd/PGLog: preserve original_crt to check rollbackability
- https://github.com/ceph/ceph/pull/27715
03/29/2019
- 11:04 PM Bug #39039 (Duplicate): mon connection reset, command not resent
- ...
- 07:45 PM Backport #38854: luminous: .mgrstat failed to decode mgrstat state; luminous dev version?
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/27207
merged - 07:45 PM Backport #38859: luminous: upmap broken the crush rule
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/27224
merged - 07:44 PM Backport #38857: luminous: should set EPOLLET flag on del_event()
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/27226
merged - 07:12 AM Backport #38872 (In Progress): mimic: Rados.get_fsid() returning bytes in python3
- https://github.com/ceph/ceph/pull/27259
- 04:47 AM Backport #38858 (In Progress): mimic: upmap broken the crush rule
- https://github.com/ceph/ceph/pull/27257
- 03:04 AM Backport #38860 (In Progress): nautilus: upmap broken the crush rule
03/28/2019
- 10:23 PM Bug #39023 (Resolved): osd/PGLog: preserve original_crt to check rollbackability
- Related to the issue discovered in https://tracker.ceph.com/issues/21174#note-11.
- 07:12 PM Feature #39012 (Resolved): osd: distinguish unfound + impossible to find, vs start some down OSDs...
This may be a command that gets information from the primary of a pg listing unfound objects and where they may be ...- 06:59 PM Documentation #39011 (Resolved): Document how get_recovery_priority() and get_backfill_priority()...
Describe the get_recovery_priority() and get_backfill_priority() as it relates to these constants:...- 06:57 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- Hi Grant,
Thanks for applying the patch and updating the logs. Looks like the earlier crash on osd.2(ENOENT on cl... - 05:22 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- I am still seeing crashes with https://github.com/ceph/ceph/pull/27200 backported.
Attached are logs.
osd.2 cep... - 02:23 AM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- https://github.com/ceph/ceph/pull/27200 attempts to resolve the failure seen on osd.2
- 04:03 PM Bug #39006: ceph tell osd.xx bench help : gives wrong help
- moreover, it says that first number is a count of block, but actually it is the count of bytes for whole operation:
... - 04:01 PM Bug #39006 (Resolved): ceph tell osd.xx bench help : gives wrong help
- ```
$ ceph tell osd.11 bench help
help not valid: help doesn't represent an int
Invalid command: unused arguments... - 12:34 PM Backport #38859 (In Progress): luminous: upmap broken the crush rule
- 01:39 AM Backport #38859: luminous: upmap broken the crush rule
- https://github.com/ceph/ceph/pull/27224
- 11:10 AM Backport #38510 (Resolved): luminous: ceph CLI ability to change file ownership
- 11:09 AM Backport #38562 (Resolved): luminous: mgr deadlock
- 11:06 AM Backport #38903 (Resolved): nautilus: Minor rados related documentation fixes
- 07:50 AM Bug #38945: osd: leaked pg refs on shutdown
- please note, in luminous, we also need to stop @snap_sleep_timer@ and @scrub_sleep_timer@ into @OSDService::shutdown(...
- 07:43 AM Bug #38945 (Fix Under Review): osd: leaked pg refs on shutdown
- 06:12 AM Bug #38892: /ceph/src/tools/kvstore_tool.cc:266:1: internal compiler error: Segmentation fault
- per Brad
> If we see this again we could try temporarily adding "--param ggc-min-expand=1 --param ggc-min-heapsize... - 03:22 AM Backport #38993 (Resolved): nautilus: unable to link rocksdb library if use system rocksdb
- https://github.com/ceph/ceph/pull/27601
- 03:04 AM Bug #38992 (Resolved): unable to link rocksdb library if use system rocksdb
- 02:33 AM Backport #38750 (New): luminous: should report EINVAL in ErasureCode::parse() if m<=0
- 02:31 AM Backport #38750 (In Progress): luminous: should report EINVAL in ErasureCode::parse() if m<=0
- 02:21 AM Backport #38857 (In Progress): luminous: should set EPOLLET flag on del_event()
- https://github.com/ceph/ceph/pull/27226
- 02:00 AM Backport #38860: nautilus: upmap broken the crush rule
- https://github.com/ceph/ceph/pull/27225
03/27/2019
- 10:56 PM Bug #38839: .mgrstat failed to decode mgrstat state; luminous dev version?
- Sage, Could this have something to do with #38941 ? The timing is right.
- 05:00 PM Backport #38983 (In Progress): nautilus: Improvements to auto repair
- 04:24 PM Backport #38983 (Resolved): nautilus: Improvements to auto repair
- https://github.com/ceph/ceph/pull/27220
- 04:38 PM Bug #38784 (Fix Under Review): osd: FAILED ceph_assert(attrs || !pg_log.get_missing().is_missing(...
- 04:01 AM Bug #26971: failed to become clean before timeout expired
dzafman-2019-03-26_16:39:54-rados:thrash-wip-zafman-26971-diag-distro-basic-smithi/3776762
Another run with diag...- 03:44 AM Backport #38854 (In Progress): luminous: .mgrstat failed to decode mgrstat state; luminous dev ve...
- https://github.com/ceph/ceph/pull/27207
- 01:54 AM Bug #38945 (Resolved): osd: leaked pg refs on shutdown
- recovery_request_timer may hold some QueuePeeringEvts which PGRef,
if we dont shutdown it earlier, it potentially ca... - 01:37 AM Feature #38616: Improvements to auto repair
- Also need to backport 0fb951963ff9d03a592bad0d4442049603195e25 with this.
03/26/2019
- 11:49 PM Feature #38616 (Pending Backport): Improvements to auto repair
- 04:56 PM Backport #38510: luminous: ceph CLI ability to change file ownership
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26758
mergedReviewed-by: Sébastien Han <seb@redhat.com> - 04:49 PM Backport #38562: luminous: mgr deadlock
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26830
merged - 04:29 PM Bug #38219: rebuild-mondb hangs
- /a/sage-2019-03-26_03:52:56-rados-wip-sage-testing-2019-03-25-1934-distro-basic-smithi/3774206
- 09:38 AM Backport #38903 (In Progress): nautilus: Minor rados related documentation fixes
- 09:29 AM Backport #38901 (In Progress): mimic: Minor rados related documentation fixes
- 09:04 AM Backport #38902 (In Progress): luminous: Minor rados related documentation fixes
- 04:15 AM Feature #38940 (New): Allow marking noout by failure domain for maintainance and planned downtime.
- - Sometimes an entire host can have planned downtime for maintenance.
- Disk failures outside of the affected area ...
03/25/2019
- 10:02 PM Subtask #37731 (Resolved): upgrade/luminous-x - add "require-osd-release nautilus" and clean up
- Yes, done as a part of these.
https://github.com/ceph/ceph/pull/26389
https://github.com/ceph/ceph/pull/26455 - 07:49 PM Subtask #37731: upgrade/luminous-x - add "require-osd-release nautilus" and clean up
- @neha I think this is done, just want to confirm, pls resolve
- 09:02 PM Bug #38041 (Resolved): Fix recovery and backfill priority handling
- 09:01 PM Backport #38275 (Resolved): mimic: Fix recovery and backfill priority handling
- 06:47 PM Bug #38927 (Resolved): should print min_mon_release correctly
- 06:13 PM Bug #38357: ClsLock.TestExclusiveEphemeralStealEphemeral failed
- Similar,...
- 03:04 PM Bug #38195: osd-backfill-space.sh exposes rocksdb hang
- Seen in mimic backport testing with new osd-backfill-prio.sh test.
http://pulpito.ceph.com/dzafman-2019-03-20_19:4... - 11:14 AM Bug #38931 (Resolved): osd does not proactively remove leftover PGs
- (Context: cephfs cluster running v12.2.11)
We had an osd go nearfull this weekend. I reweighted it to move out som... - 10:58 AM Bug #38930 (Duplicate): ceph osd safe-to-destroy wrongly approves any out osd
- With v12.2.11, we found that ceph osd safe-to-destroy is wrongly reporting that all out osds are safe to destroy.
... - 10:07 AM Backport #38850: upgrade: 1 nautilus mon + 1 luminous mon can't automatically form quorum
- Agreed, my expectation would be that we can maintain quorum during the entire upgrade period. Even discounting OS upg...
03/24/2019
- 04:08 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- Yes I can still reproduce it, the cluster is still in a broken state....
- 03:37 PM Backport #38853 (Resolved): nautilus: .mgrstat failed to decode mgrstat state; luminous dev version?
- 02:31 AM Bug #38927: should print min_mon_release correctly
- > Brad Hubbard wrote:
> 15 - 15 !> 2 ?
>
>
> https://github.com/ceph/ceph/pull/27107 should fix this. - 02:30 AM Bug #38927 (Pending Backport): should print min_mon_release correctly
- 02:30 AM Bug #38927 (Resolved): should print min_mon_release correctly
dzafman-2019-03-20_19:53:02-rados-wip-zafman-testing-distro-basic-smithi/3754307
rados/upgrade/luminous-x-single...
03/23/2019
- 10:48 PM Backport #38901: mimic: Minor rados related documentation fixes
- Remove "premerge" pg state which doesn't apply in mimic.
- 09:13 PM Backport #38901 (Resolved): mimic: Minor rados related documentation fixes
- https://github.com/ceph/ceph/pull/27188
- 10:48 PM Backport #38902: luminous: Minor rados related documentation fixes
- Remove "premerge" pg state which doesn't apply in luminous.
- 09:13 PM Backport #38902 (Resolved): luminous: Minor rados related documentation fixes
- https://github.com/ceph/ceph/pull/27185
- 09:13 PM Backport #38906 (Resolved): nautilus: osd/PGLog.h: print olog_can_rollback_to before deciding to ...
- https://github.com/ceph/ceph/pull/27302
- 09:13 PM Backport #38905 (Resolved): luminous: osd/PGLog.h: print olog_can_rollback_to before deciding to ...
- https://github.com/ceph/ceph/pull/27715
- 09:13 PM Backport #38904 (Resolved): mimic: osd/PGLog.h: print olog_can_rollback_to before deciding to rol...
- https://github.com/ceph/ceph/pull/27284
- 09:13 PM Backport #38903 (Resolved): nautilus: Minor rados related documentation fixes
- https://github.com/ceph/ceph/pull/27189
- 09:13 PM Backport #38853 (In Progress): nautilus: .mgrstat failed to decode mgrstat state; luminous dev ve...
- 05:41 PM Bug #38900 (New): EC pools don't self repair on client read error
When a replicated client read fails at the primary, it will pull the object from another OSD (see rep_repair_primar...- 11:42 AM Documentation #38896 (Pending Backport): Minor rados related documentation fixes
- 12:22 AM Documentation #38896 (Resolved): Minor rados related documentation fixes
Document all pg states
Add auto repair items
"premerge" is not pg state in luminous nor mimic
03/22/2019
- 09:27 PM Bug #38845 (Resolved): mon.a@-1(probing) e1 current monmap has min_mon_release 15 (luminous) whic...
- 03:57 PM Bug #38845: mon.a@-1(probing) e1 current monmap has min_mon_release 15 (luminous) which is >2 rel...
- https://github.com/ceph/ceph/pull/27131
- 02:28 PM Bug #38845 (Fix Under Review): mon.a@-1(probing) e1 current monmap has min_mon_release 15 (lumino...
- 02:02 AM Bug #38845: mon.a@-1(probing) e1 current monmap has min_mon_release 15 (luminous) which is >2 rel...
- Brad Hubbard wrote:
> 15 - 15 !> 2 ?
https://github.com/ceph/ceph/pull/27107 should fix this. - 12:10 AM Bug #38845: mon.a@-1(probing) e1 current monmap has min_mon_release 15 (luminous) which is >2 rel...
- 15 - 15 !> 2 ?
- 09:05 PM Bug #38892: /ceph/src/tools/kvstore_tool.cc:266:1: internal compiler error: Segmentation fault
- While I was looking into this I noticed this warning in the Jenkins output....
- 04:46 PM Bug #38892 (Closed): /ceph/src/tools/kvstore_tool.cc:266:1: internal compiler error: Segmentation...
- ...
- 07:12 PM Bug #38894 (Pending Backport): osd/PGLog.h: print olog_can_rollback_to before deciding to rollback
- 05:20 PM Bug #38894 (Resolved): osd/PGLog.h: print olog_can_rollback_to before deciding to rollback
- This is important for debugging failures in merge_object_divergent_entries() before a decision to rollback is made.
- 05:16 PM Bug #38893 (Resolved): RuntimeError: expected MON_CLOCK_SKEW but got none
- ...
- 05:09 PM Cleanup #38635: bluestore: test osd_memory_target
- https://github.com/ceph/ceph/pull/27083 - Merged
Will mark Pending Backport when Part-2 merges. - 02:07 PM Bug #37766 (Resolved): rados_shutdown hang forever in ~objecter()
- 02:06 PM Backport #38398 (Resolved): mimic: rados_shutdown hang forever in ~objecter()
- 01:05 PM Backport #38881 (Resolved): nautilus: ENOENT in collection_move_rename on EC backfill target
- https://github.com/ceph/ceph/pull/27654
- 01:05 PM Backport #38880 (Resolved): luminous: ENOENT in collection_move_rename on EC backfill target
- https://github.com/ceph/ceph/pull/28110
- 01:04 PM Backport #38879 (Resolved): mimic: ENOENT in collection_move_rename on EC backfill target
- https://github.com/ceph/ceph/pull/27943
- 01:03 PM Backport #38873 (Resolved): luminous: Rados.get_fsid() returning bytes in python3
- https://github.com/ceph/ceph/pull/27674
- 01:03 PM Backport #38872 (Resolved): mimic: Rados.get_fsid() returning bytes in python3
- https://github.com/ceph/ceph/pull/27259
- 01:01 PM Backport #38860 (Resolved): nautilus: upmap broken the crush rule
- https://github.com/ceph/ceph/pull/27225
- 01:01 PM Backport #38859 (Resolved): luminous: upmap broken the crush rule
- https://github.com/ceph/ceph/pull/27224
- 01:01 PM Backport #38858 (Resolved): mimic: upmap broken the crush rule
- https://github.com/ceph/ceph/pull/27257
- 01:00 PM Backport #38857 (Resolved): luminous: should set EPOLLET flag on del_event()
- https://github.com/ceph/ceph/pull/27226
- 01:00 PM Backport #38856 (Resolved): mimic: should set EPOLLET flag on del_event()
- https://github.com/ceph/ceph/pull/29250
- 01:00 PM Backport #38854 (Resolved): luminous: .mgrstat failed to decode mgrstat state; luminous dev version?
- https://github.com/ceph/ceph/pull/27207
- 01:00 PM Backport #38853 (Resolved): nautilus: .mgrstat failed to decode mgrstat state; luminous dev version?
- https://github.com/ceph/ceph/pull/27116
- 01:00 PM Backport #38852 (Resolved): mimic: .mgrstat failed to decode mgrstat state; luminous dev version?
- https://github.com/ceph/ceph/pull/29249
- 11:05 AM Backport #38850: upgrade: 1 nautilus mon + 1 luminous mon can't automatically form quorum
- Just to clarify slightly -- I know the upgrade instructions in the Nautilus release announcement say to "upgrade moni...
- 10:19 AM Backport #38850 (Resolved): upgrade: 1 nautilus mon + 1 luminous mon can't automatically form quorum
- Seen while upgrading Luminous (12.2.10) to Nautilus (14.2.0). Three mon hosts, four osd hosts. The process was:
... - 09:30 AM Bug #38839: .mgrstat failed to decode mgrstat state; luminous dev version?
- nautilus https://github.com/ceph/ceph/pull/27116
- 09:30 AM Bug #38839 (Pending Backport): .mgrstat failed to decode mgrstat state; luminous dev version?
- 07:37 AM Bug #38826 (Pending Backport): upmap broken the crush rule
- 01:33 AM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- It is possible that the crash we are seeing on osd.2 is due to 1:537949df:::20000a2c834.00000105:head incorrectly rol...
- 01:05 AM Bug #38846: dump_pgstate_history doesn't really produce useful json output, needs an array around...
- Probably be nice if it dumped the current state stack for each pg as well.
03/21/2019
- 11:06 PM Bug #38846 (Resolved): dump_pgstate_history doesn't really produce useful json output, needs an a...
- ...
- 08:42 PM Bug #38845 (Resolved): mon.a@-1(probing) e1 current monmap has min_mon_release 15 (luminous) whic...
dzafman-2019-03-20_19:53:02-rados-wip-zafman-testing-distro-basic-smithi/3754307
rados/upgrade/luminous-x-single...- 06:05 PM Bug #38841 (New): Objects degraded higher than 100%
- 1. Working Mimic or Nautilus deployment with Bluestore (haven't tested with Filestore)
2. All OSDs up, all PGs activ... - 05:29 PM Bug #38840 (Resolved): snaps missing in mapper, should be: ca was r -2...repaired
dzafman-2019-03-20_19:53:02-rados-wip-zafman-testing-distro-basic-smithi/3754443
This looks like a cache tier ev...- 04:59 PM Bug #38839 (Fix Under Review): .mgrstat failed to decode mgrstat state; luminous dev version?
- https://github.com/ceph/ceph/pull/27101
- 04:57 PM Bug #38839 (Resolved): .mgrstat failed to decode mgrstat state; luminous dev version?
- ...
- 02:26 AM Backport #38719 (In Progress): luminous: crush: choose_args array size mis-sized when weight-sets...
- https://github.com/ceph/ceph/pull/27085
- 01:38 AM Cleanup #38635 (In Progress): bluestore: test osd_memory_target
- https://github.com/ceph/ceph/pull/27083
- 01:31 AM Backport #38720 (In Progress): mimic: crush: choose_args array size mis-sized when weight-sets ar...
- https://github.com/ceph/ceph/pull/27082
03/20/2019
- 10:50 PM Bug #26971: failed to become clean before timeout expired
I'm not sure what this means, but pg 1.0 (size 3) needs to pick another one of the 2 remaining OSDs (4 OSDs in) to ...- 12:05 PM Bug #38582: Pool storage MAX AVAIL reduction seems higher when single OSD reweight is done
- Sorry for the delay. Attaching the required.
osd 155 is the OSD mentioned in description. The one which was manually... - 11:51 AM Bug #38381 (Pending Backport): Rados.get_fsid() returning bytes in python3
- 11:40 AM Bug #38827 (In Progress): valgrind: UninitCondition in ceph::crypto::onwire::AES128GCM_OnWireRxHa...
- 11:24 AM Bug #38827: valgrind: UninitCondition in ceph::crypto::onwire::AES128GCM_OnWireRxHandler::authent...
- the test branch contains https://github.com/ceph/ceph/pull/27012
- 11:21 AM Bug #38827 (Resolved): valgrind: UninitCondition in ceph::crypto::onwire::AES128GCM_OnWireRxHandl...
- ...
- 11:27 AM Bug #38828 (Resolved): should set EPOLLET flag on del_event()
- 10:41 AM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- As requested.
osd.0: ceph-post-file: 17efe900-501c-479f-ba56-dd29fef18c58
osd.4: ceph-post-file: ff22f830-e6bc-4f... - 12:36 AM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- Hi Grant,
Looking at the logs, it seems that the first crash was seen on osd.2 on pg id 1.cas2... - 08:27 AM Bug #38826: upmap broken the crush rule
- Here is the crush rule...
- 08:24 AM Bug #38826 (Resolved): upmap broken the crush rule
- I setup a cluster and want to specify the primary osds through crush rule.
Here is the test script... - 03:14 AM Backport #38275 (In Progress): mimic: Fix recovery and backfill priority handling
- 12:43 AM Backport #38244 (Resolved): luminous: scrub warning check incorrectly uses mon scrub interval
- 12:43 AM Backport #38274 (Resolved): luminous: Fix recovery and backfill priority handling
03/19/2019
- 11:30 PM Bug #36739 (Pending Backport): ENOENT in collection_move_rename on EC backfill target
- 08:38 PM Backport #38398: mimic: rados_shutdown hang forever in ~objecter()
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26583
merged
03/18/2019
- 06:55 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- Err. I believe I mixed up two different bugs, please disregard my previous comment. I don't currently recall what I ...
- 06:52 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- For completeness: The root cause for the crashes I experienced were that I had oversized RADOS objects (2-10GB, max ...
- 02:22 PM Bug #38124: OSD down on snaptrim.
- Hello any updates about this?
- 06:35 AM Bug #38793 (New): data inconsistent
- I did some test on rbd snap, and found data inconsistent.
cluster status:...
03/17/2019
- 10:21 PM Bug #38787 (Fix Under Review): osd: cache tiering flush clone wrongly
- 02:38 AM Bug #38787 (Fix Under Review): osd: cache tiering flush clone wrongly
- because cephfs file snapcontext seq may start from 1, we find that in a never snaped fs,
the flush of file will dele... - 07:21 PM Bug #38294 (Resolved): osd/PG.cc: 6141: FAILED ceph_assert(info.history.same_interval_since != 0)...
- 10:01 AM Bug #38294 (Fix Under Review): osd/PG.cc: 6141: FAILED ceph_assert(info.history.same_interval_sin...
- https://github.com/ceph/ceph/pull/27018
- 09:57 AM Bug #38294 (In Progress): osd/PG.cc: 6141: FAILED ceph_assert(info.history.same_interval_since !=...
- /a/sage-2019-03-17_00:28:04-upgrade:luminous-x-wip-sage4-testing-2019-03-16-1713-distro-basic-smithi/3737326
pg 1.... - 12:10 AM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- ...
03/16/2019
- 11:20 PM Bug #21174: OSD crash: 903: FAILED assert(objiter->second->version > last_divergent_update)
- I have a similar issue with OSDs dropping out:...
- 06:33 PM Bug #38786 (Resolved): autoscale down can lead to max_pg_per_osd limit
- we adjust pgp_num all the way down to the target, which can make osds hit the max_pgs_per_osd if it's going too far.
...
03/15/2019
- 09:45 PM Bug #38623 (Resolved): 2.8s2 past_intervals [6539,6541) start interval does not contain the requi...
- 08:31 PM Bug #38655 (Resolved): osd: missing, size mismatch, snap mapper errors
- 06:11 PM Bug #36739: ENOENT in collection_move_rename on EC backfill target
- https://github.com/ceph/ceph/pull/26996 is a more complete fix for this issue.
- 06:06 PM Bug #38784 (Resolved): osd: FAILED ceph_assert(attrs || !pg_log.get_missing().is_missing(soid) ||...
- ...
- 05:08 PM Bug #38746 (Resolved): msgr2 leaking buffers
- https://github.com/ceph/ceph/pull/26965
- 03:20 AM Bug #38746: msgr2 leaking buffers
- hmm it happens on some osds but not others.
i added to rxbuf and txbuf lengths to the dout prefix and got this
... - 03:01 AM Bug #38746 (Resolved): msgr2 leaking buffers
- osds with bluestore consume too much ram (seeing 20GB on sepia)
to reproduce with vstart, watch bin/ceph daemon os... - 05:03 PM Bug #38783 (New): Changing mon_pg_warn_max_object_skew has no effect.
- ...
- 03:20 PM Documentation #38051 (Resolved): doc/rados/configuration: refresh osdmap section
- 03:19 PM Backport #38095 (Resolved): luminous: doc/rados/configuration: refresh osdmap section
- 12:13 PM Bug #38762 (New): Ubuntu/Debian repo has incorrect InRelease
- On Ubuntu Bionic trying to update repo package I got error:
E: Failed to fetch https://download.ceph.com/debian-mi... - 08:59 AM Backport #38751 (Resolved): mimic: should report EINVAL in ErasureCode::parse() if m<=0
- https://github.com/ceph/ceph/pull/28995
- 08:58 AM Backport #38750 (Resolved): luminous: should report EINVAL in ErasureCode::parse() if m<=0
- https://github.com/ceph/ceph/pull/28111
03/14/2019
- 04:45 PM Feature #38616: Improvements to auto repair
I don't think we need to set "failed_repair" if primary can't recover itself on a read error. We are already setti...- 02:46 PM Feature #38616 (In Progress): Improvements to auto repair
- 04:17 PM Bug #38345: mon: segv in MonOpRequest::~MonOpRequest OpHistory::cleanup
- Forgot mention, the op appears to be an MForward.
- 11:58 AM Bug #38682 (Pending Backport): should report EINVAL in ErasureCode::parse() if m<=0
- 12:53 AM Cleanup #38635: bluestore: test osd_memory_target
- Part 1: Test with a value of osd_memory_target lesser than the default, maybe half or less than that. This can be don...
03/13/2019
- 11:03 PM Bug #38345: mon: segv in MonOpRequest::~MonOpRequest OpHistory::cleanup
- ...
- 10:55 PM Bug #38345: mon: segv in MonOpRequest::~MonOpRequest OpHistory::cleanup
- ...
- 04:29 PM Bug #38724: _txc_add_transaction error (39) Directory not empty not handled on operation 21 (op 1...
- ceph-post-file: 26dab2cb-36c9-40de-8455-1379406477e8
- 04:29 PM Bug #38724 (Resolved): _txc_add_transaction error (39) Directory not empty not handled on operati...
- ...
- 03:08 PM Bug #38631 (Resolved): osd-scrub-repair.sh fails due to num_objects wrong
- 03:05 PM Bug #38678 (Resolved): Minor cleanups in tests and log output
- 01:16 PM Bug #38705 (Resolved): mgr: segv in module thread, PyArg_ParseTuple
- 12:03 PM Backport #38720 (Resolved): mimic: crush: choose_args array size mis-sized when weight-sets are e...
- https://github.com/ceph/ceph/pull/27082
- 12:03 PM Backport #38719 (Resolved): luminous: crush: choose_args array size mis-sized when weight-sets ar...
- https://github.com/ceph/ceph/pull/27085
- 11:56 AM Bug #38664 (Pending Backport): crush: choose_args array size mis-sized when weight-sets are enabled
- 11:56 AM Bug #38703 (Resolved): lazy omap stats aren't incorportaed into pg_autoscaler size value
- 11:55 AM Bug #38238: rados/test.sh: api_aio_pp doesn't seem to start
- /a/sage-2019-03-13_02:19:41-rados-wip-sage3-testing-2019-03-12-1657-distro-basic-smithi/3715202
- 11:54 AM Bug #20086: LibRadosLockECPP.LockSharedDurPP gets EEXIST
- ...
- 11:52 AM Bug #38718 (New): 'osd crush weight-set create-compat' (and other OSDMonitor commands) can leak u...
- ...
- 10:55 AM Backport #38506 (Resolved): luminous: ENOENT on setattrs (obj was recently deleted)
- 10:39 AM Bug #38258 (Resolved): filestore: fsync(2) return value not checked
- 10:38 AM Backport #38316 (Resolved): luminous: filestore: fsync(2) return value not checked
- 04:40 AM Backport #38423 (Resolved): luminous: osd/TestPGLog.cc: Verify that dup_index is being trimmed
03/12/2019
- 09:46 PM Bug #38705 (Fix Under Review): mgr: segv in module thread, PyArg_ParseTuple
- https://github.com/ceph/ceph/pull/26920
- 07:30 PM Bug #38705: mgr: segv in module thread, PyArg_ParseTuple
- appear to happen during standby. also, i see an ignored monmap message:...
- 07:28 PM Bug #38705: mgr: segv in module thread, PyArg_ParseTuple
- lots of these failures. module varies (i've seen dashboard, prometheus so far)
- 07:11 PM Bug #38705 (Resolved): mgr: segv in module thread, PyArg_ParseTuple
- ...
- 07:52 PM Backport #38506: luminous: ENOENT on setattrs (obj was recently deleted)
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26706
merged - 07:51 PM Backport #38316: luminous: filestore: fsync(2) return value not checked
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26871
merged - 06:40 PM Bug #38159: ec does not recover below min_size
- We coudl perhaps point the finger at the min_size choice:...
- 06:39 PM Bug #38159: ec does not recover below min_size
- ...
- 05:43 PM Bug #38703 (Fix Under Review): lazy omap stats aren't incorportaed into pg_autoscaler size value
- https://github.com/ceph/ceph/pull/26917
- 04:42 PM Bug #38703 (Resolved): lazy omap stats aren't incorportaed into pg_autoscaler size value
- on lab cluster,...
- 03:49 PM Bug #38579: osd: should not mark cluster_messenger when commited new osdmap
- Ok, I think I understand. This would noramlly trigger a RESETSESSION in the v1 protocol because the primary's connec...
- 02:21 PM Backport #38162 (Resolved): luminous: maybe_remove_pg_upmaps incorrectly cancels valid pending up...
- 02:11 PM Bug #37797 (Resolved): radosbench tests hit ENOSPC
- 02:11 PM Backport #38240 (Resolved): luminous: radosbench tests hit ENOSPC
- 02:09 PM Backport #38400 (Resolved): luminous: rados_shutdown hang forever in ~objecter()
- 02:01 PM Bug #38682 (Fix Under Review): should report EINVAL in ErasureCode::parse() if m<=0
- https://github.com/ceph/ceph/pull/26894
- 01:36 PM Bug #38682: should report EINVAL in ErasureCode::parse() if m<=0
- I agree that m=0 is useless--it's no better than num_rep=1... just a (much) more complicated code path and more corne...
- 06:28 AM Bug #38682: should report EINVAL in ErasureCode::parse() if m<=0
- i think we can even go further -- to prevent user from creating a profile with m=0. technically, it's correct. but pr...
- 06:24 AM Bug #38682 (Resolved): should report EINVAL in ErasureCode::parse() if m<=0
- ...
- 03:12 AM Backport #38566 (In Progress): mimic: osd_recovery_priority is not documented (but osd_recovery_o...
- https://github.com/ceph/ceph/pull/26901
03/11/2019
- 11:57 PM Bug #38631 (Duplicate): osd-scrub-repair.sh fails due to num_objects wrong
- 09:39 PM Bug #38631: osd-scrub-repair.sh fails due to num_objects wrong
- I'm going to remove create_rbd_pool because it isn't used anyway.
- 06:55 PM Bug #38631: osd-scrub-repair.sh fails due to num_objects wrong
Reopening to use for root cause fix. This tracker should also revert 10b9626ea7b.
-The commit comment is wrong,...- 11:53 PM Bug #38678 (Fix Under Review): Minor cleanups in tests and log output
- 04:57 PM Bug #38678 (Resolved): Minor cleanups in tests and log output
- 10:38 PM Bug #38655 (Fix Under Review): osd: missing, size mismatch, snap mapper errors
- https://github.com/ceph/ceph/pull/26898
- 05:24 PM Bug #38655: osd: missing, size mismatch, snap mapper errors
- The problem originates with 2.23, a merge source. It is instantiated on osd.5 with info...
- 09:10 PM Bug #38582: Pool storage MAX AVAIL reduction seems higher when single OSD reweight is done
- That does seem odd. Can you attach your crush map, "ceph osd tree", and "ceph osd dump" to this ticket?
- 07:44 PM Backport #38162: luminous: maybe_remove_pg_upmaps incorrectly cancels valid pending upmaps
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26127
merged - 07:42 PM Backport #38240: luminous: radosbench tests hit ENOSPC
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26355
merged - 07:41 PM Backport #38400: luminous: rados_shutdown hang forever in ~objecter()
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/26579
merged - 07:30 PM Bug #24613: luminous: rest/test.py fails with expected 200, got 400
- /a/yuriw-2019-03-06_22:09:13-rados-wip-yuri4-testing-2019-03-04-2231-luminous-distro-basic-smithi/3675478/
- 05:21 PM Bug #38633: /ceph/src/tools/kvstore_tool.cc:39:1: internal compiler error: Segmentation fault on ...
- https://github.com/ceph/ceph/pull/26828
- 02:00 PM Bug #38664: crush: choose_args array size mis-sized when weight-sets are enabled
- https://github.com/ceph/ceph/pull/26886
- 01:59 PM Bug #38664 (Resolved): crush: choose_args array size mis-sized when weight-sets are enabled
- simple reproducer on vstart:...
03/10/2019
- 08:58 PM Bug #38656 (New): scrub reservation leak?
- /a/sage-2019-03-10_18:54:11-rados-wip-sage2-testing-2019-03-10-1053-distro-basic-smithi/3705804
pg 1.0 scrub does ... - 04:01 PM Bug #38655 (Resolved): osd: missing, size mismatch, snap mapper errors
- ...
- 03:57 PM Bug #38238: rados/test.sh: api_aio_pp doesn't seem to start
- /a/sage-2019-03-10_01:08:05-rados-master-distro-basic-smithi/3703837
description: rados/thrash/{0-size-min-size-ov...
03/09/2019
- 07:23 PM Bug #38631 (Resolved): osd-scrub-repair.sh fails due to num_objects wrong
- 05:00 PM Bug #38633 (Resolved): /ceph/src/tools/kvstore_tool.cc:39:1: internal compiler error: Segmentatio...
- 01:53 PM Bug #38579: osd: should not mark cluster_messenger when commited new osdmap
- Sage Weil wrote:
> So in step 5, the primary hasn't seen osdmap 20, right? Only the replica has? The part I don't ... - 12:54 AM Feature #38653: Enhance health message when pool quota fills up
- 12:27 AM Backport #38316 (In Progress): luminous: filestore: fsync(2) return value not checked
- https://github.com/ceph/ceph/pull/26871
03/08/2019
- 11:34 PM Feature #38653 (In Progress): Enhance health message when pool quota fills up
- 11:00 PM Feature #38653 (New): Enhance health message when pool quota fills up
- https://bugzilla.redhat.com/show_bug.cgi?id=1481306...
- 08:46 PM Feature #22147 (In Progress): Set multiple flags in a single command line
- https://github.com/ceph/ceph/pull/26785
- 08:09 PM Backport #38646 (In Progress): mimic: OpTracker destruct assert when OSD destruct
- 02:46 PM Backport #38646 (Resolved): mimic: OpTracker destruct assert when OSD destruct
- https://github.com/ceph/ceph/pull/26862
- 03:00 PM Bug #38649 (Can't reproduce): [ERR] full status failsafe engaged, dropping updates, now -21474836...
- /a/sage-2019-03-08_07:14:13-rados-wip-sage2-testing-2019-03-07-2213-distro-basic-smithi/3682171
- 02:48 PM Bug #38377: OpTracker destruct assert when OSD destruct
- master is still being merged into nautilus AFAICT
- 04:12 AM Bug #38377 (Pending Backport): OpTracker destruct assert when OSD destruct
- 01:15 PM Bug #38579 (Need More Info): osd: should not mark cluster_messenger when commited new osdmap
- So in step 5, the primary hasn't seen osdmap 20, right? Only the replica has? The part I don't understand is that i...
- 10:37 AM Backport #38610: luminous: mon: osdmap prune
- https://github.com/ceph/ceph/pull/26834
- 10:36 AM Backport #38561 (In Progress): mimic: mgr deadlock
- https://github.com/ceph/ceph/pull/26833
- 08:19 AM Bug #38124: OSD down on snaptrim.
- Hello,
any updates regarding this bug? I would love a patch to resolve this issue ASAP. One of my monitors just... - 08:12 AM Bug #38307 (Resolved): ceph-osd fails to bind to IPv6 interface for public_network
- The PR https://github.com/ceph/ceph/pull/26692 enforces pick_addresses to fail when ms_bind_ipv4 and ms_bind_ipv6 opt...
- 04:56 AM Bug #38633 (Fix Under Review): /ceph/src/tools/kvstore_tool.cc:39:1: internal compiler error: Seg...
- 01:18 AM Bug #38633 (Resolved): /ceph/src/tools/kvstore_tool.cc:39:1: internal compiler error: Segmentatio...
- a1539b118ed6372c19f321c94e2246f4fd130a33...
- 04:26 AM Backport #38562 (In Progress): luminous: mgr deadlock
- https://github.com/ceph/ceph/pull/26830
- 04:13 AM Bug #38598 (Resolved): osdmap may include only v1 address while osd binds to v2; mon drops messages
- 04:03 AM Subtask #37732: qa/suites/rados/thrash-erasure-code*: coverage review tasks
- https://github.com/ceph/ceph/pull/26417
Addresses
- Leveldb mons no longer relevant
- Fast-read could be added t... - 03:57 AM Cleanup #38635: bluestore: test osd_memory_target
- We want to test with different values of osd_memory_target.
Also, create tests that necessarily go beyond the osd_me... - 03:50 AM Cleanup #38635 (In Progress): bluestore: test osd_memory_target
03/07/2019
- 10:26 PM Bug #38358: short pg log + cache tier ceph_test_rados out of order reply
- /a/yuriw-2019-03-07_00:04:47-rados-wip_yuri_nautilus_3.6.19-distro-basic-smithi/3675857/
- 09:36 PM Bug #38631 (Resolved): osd-scrub-repair.sh fails due to num_objects wrong
A status check for "1/1 objects unfound" is coming back as "1/2 objects unfound"
Can be reproduced easily with:
...- 03:03 PM Bug #36546 (Duplicate): common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_list.back...
- 03:02 PM Bug #38592 (Duplicate): mon,osd: src/common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_fli...
- 02:56 PM Bug #38623 (Fix Under Review): 2.8s2 past_intervals [6539,6541) start interval does not contain t...
- https://github.com/ceph/ceph/pull/26822
- 12:25 PM Bug #38623 (Resolved): 2.8s2 past_intervals [6539,6541) start interval does not contain the requi...
- ...
- 01:52 PM Bug #38579: osd: should not mark cluster_messenger when commited new osdmap
- Greg Farnum wrote:
> Do you have a reproducer for this?
>
> I get nervous when people want to remove mark_down ca... - 12:56 PM Bug #38604 (Resolved): mon logs not getting reopened after rotation
- 12:47 PM Bug #38624 (New): crush: get_rule_weight_osd_map does not handle multi-take rules
- CrushWrapper::get_rule_weight_osd_map() does not handle multi-take rules well. for example, a take 1 (primary) and t...
- 07:55 AM Backport #38565 (In Progress): mimic: Code to strip | from core pattern isn't right
- 06:43 AM Feature #38603: mon: osdmap prune
- @Nathan, I developed and tested code, I will open PR in the next couple of days soon. Please assign this to me :)
- 02:12 AM Feature #38616: Improvements to auto repair
OSD stats might have to be in meta collection- 01:29 AM Feature #38617 (Resolved): osd: Better error message when OSD count is less than osd_pool_default...
- Clearly indicate when number of OSDs is less than osd_pool_default_size, to avoid users from setting up clusters inco...
03/06/2019
- 10:38 PM Feature #38616 (Resolved): Improvements to auto repair
We should allow auto repair for bluestore pools since it has built in checksums. Currently, we are limited to er...- 10:18 PM Feature #38458: Ceph does not have command to show current osd primary-affinity
- So this is dumped as part of the osdmap output, but you want a way to see it for a particular OSD? Do we have any out...
- 10:13 PM Bug #38579: osd: should not mark cluster_messenger when commited new osdmap
- Do you have a reproducer for this?
I get nervous when people want to remove mark_down calls, as they are generally... - 07:50 PM Bug #38604 (Fix Under Review): mon logs not getting reopened after rotation
- aha, ceph-mgr and ceph-mds expliiclty set the thread name on startup.
https://github.com/ceph/ceph/pull/26797 - 07:40 PM Bug #38604: mon logs not getting reopened after rotation
- this appears to be because of /proc/$pid/stat. before,...
- 01:00 PM Bug #38604 (Resolved): mon logs not getting reopened after rotation
- ...
- 07:42 PM Bug #38219: rebuild-mondb hangs
- rados:singleton/{all/rebuild-mondb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rad...
- 06:40 PM Bug #38598 (Fix Under Review): osdmap may include only v1 address while osd binds to v2; mon drop...
- 03:10 AM Bug #38598: osdmap may include only v1 address while osd binds to v2; mon drops messages
- Proposed OSD fix:
- if we get an osdmap with require_osd_release < nautilus, and are bound to v2+v1, we rebind to ... - 03:08 AM Bug #38598 (Resolved): osdmap may include only v1 address while osd binds to v2; mon drops messages
- - osd binds to v2+v1
- osd sends osd_boot to mon
- mon adds v1 addr to osdmap only (due to require_osd_release < na... - 06:05 PM Bug #38555: scrub error on ec pg, got 6579891/0 or 7569408/6832128 bytes
- "2019-03-06 15:21:41.756014 osd.5 (osd.5) 287 : cluster [ERR] 2.2s0 scrub : stat mismatch, got 2/2 objects, 1/1 clone...
- 05:59 PM Backport #38610 (Need More Info): luminous: mon: osdmap prune
- Feature backport assumed to be non-trivial. Assigning to Joao, author of the feature, for now.
- 05:58 PM Backport #38610 (Rejected): luminous: mon: osdmap prune
- https://github.com/ceph/ceph/pull/26834
- 05:58 PM Feature #38603 (Pending Backport): mon: osdmap prune
- 10:18 AM Feature #38603 (Resolved): mon: osdmap prune
- Tracker to enable backport of this feature to luminous:
https://github.com/ceph/ceph/pull/19331 - 04:54 PM Backport #38274 (In Progress): luminous: Fix recovery and backfill priority handling
- 02:30 AM Bug #38592: mon,osd: src/common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_list.bac...
- Is this related to http://tracker.ceph.com/issues/36546?
03/05/2019
- 11:17 PM Bug #26971: failed to become clean before timeout expired
- Reproduced on master in 1 of 10 duplicate runs:
dzafman-2019-03-05_10:43:39-rados:thrash-master-distro-basic-smith... - 07:01 PM Bug #26971: failed to become clean before timeout expired
- Seen recently in luminous.
yuriw-2019-02-28_14:42:05-rados-wip-yuri4-testing-2019-02-27-2159-luminous-distro-basi... - 10:37 PM Bug #38484 (Resolved): osd: InvalidRead, PG use-after-free putting ref
- 10:09 PM Bug #38525 (Resolved): qa/standalone/osd/pg-split-merge.sh fails
- 08:48 PM Bug #38594 (New): mimic: common/Mutex.cc: 110: FAILED assert(r == 0) in powercycle
- ...
- 08:22 PM Bug #38592 (Duplicate): mon,osd: src/common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_fli...
- ...
- 06:03 PM Bug #38499 (Need More Info): ceph-mon segfaults at startup
- 03:12 PM Feature #21073 (Resolved): mgr: ceph/rgw: show hostnames and ports in ceph -s status output
- 10:51 AM Bug #38582 (New): Pool storage MAX AVAIL reduction seems higher when single OSD reweight is done
i have a 5 node ceph 11.2.0 cluster with 335 osds. Each OSD is a 4TB HDD. It has one EC 4+1 pool.
Due to high st...- 03:28 AM Backport #38511 (In Progress): mimic: ceph CLI ability to change file ownership
- https://github.com/ceph/ceph/pull/26760
- 01:52 AM Backport #38510 (In Progress): luminous: ceph CLI ability to change file ownership
- https://github.com/ceph/ceph/pull/26758
- 01:49 AM Backport #38507 (In Progress): mimic: ENOENT on setattrs (obj was recently deleted)
- https://github.com/ceph/ceph/pull/26709
- 01:36 AM Bug #38579 (Need More Info): osd: should not mark cluster_messenger when commited new osdmap
-
when we run some fault test in Luminous 12.2.10, got coredump like ...
Also available in: Atom