Activity
From 05/28/2018 to 06/26/2018
06/26/2018
- 11:32 PM Bug #23492 (In Progress): Abort in OSDMap::decode() during qa/standalone/erasure-code/test-erasur...
- 11:29 PM Feature #13507 (New): scrub APIs to read replica
- 11:28 PM Bug #24366 (Resolved): omap_digest handling still not correct
- 11:27 PM Backport #24381 (Resolved): luminous: omap_digest handling still not correct
- 11:27 PM Backport #24380 (Resolved): mimic: omap_digest handling still not correct
- 09:08 PM Bug #23352: osd: segfaults under normal operation
- Matt,
Can you provide a coredump or full backtrace? - 01:54 PM Bug #23352: osd: segfaults under normal operation
- Also confirmed on Ubuntu 18.04/Ceph 13.2.0:
ceph-mgr.log
> 2018-06-24 11:14:47.317 7ff17b0db700 -1 mgr.server s... - 02:54 AM Bug #23352: osd: segfaults under normal operation
- confirmed
ceph-mgr.log
@2018-06-20 08:46:05.528656 7fb998ff2700 -1 mgr.server send_report send_report osd,215.0x5... - 07:14 PM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
- Dexter, anyone: was there a PG split (pg_num increase) on the cluster before this happened? Or maybe a split combine...
- 07:10 PM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
- ...
- 07:06 PM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
- ...
- 06:42 PM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
- Dexter John Genterone wrote:
> Uploaded a few more logs (debug 20) here: https://storage.googleapis.com/ceph-logs/ce... - 07:11 PM Bug #24667 (Can't reproduce): osd: SIGSEGV in MMgrReport::encode_payload
- ...
- 07:07 PM Bug #24666 (New): pybind: InvalidArgumentError is missing 'errno' argument
- Instead of being derived from 'Error', the 'InvalidArgumentError' should be derived from 'OSError' which will handle ...
- 01:49 PM Bug #24664 (Resolved): osd: crash in OpTracker::unregister_inflight_op via OSD::get_health_metrics
- ...
- 09:47 AM Bug #24660 (New): admin/build-doc fails during autodoc on rados module: "AttributeError: __next__"
- I'm trying to send a doc patch and am running @admin/build-doc@ in my local environment explained in [[http://docs.ce...
06/25/2018
- 09:50 PM Bug #23352: osd: segfaults under normal operation
- Same here
2018-06-24 19:42:41.348699 7f3e53a46700 -1 mgr.server send_report send_report osd,226.0x55678069c850 sen... - 09:34 PM Bug #23352: osd: segfaults under normal operation
- Brad Hubbard wrote:
> Can anyone confirm seeing the "unknown health metric" messages in the mgr logs prior to the se... - 02:50 PM Bug #24652 (Won't Fix): OSD crashes when repairing pg
- After a deep-scrub on the primary OSD for the pg we get:...
- 01:03 PM Bug #24650 (New): mark unfound lost revert: out of order trim
- OSD crashes in a few seconds after command 'ceph pg X.XX mark_unfound_lost revert'.
-10> 2018-06-25 15:52:14.49... - 07:53 AM Bug #24645 (New): Upload to radosgw fails when there are degraded objects
- Hi,
we use Ceph RadosGW for storing and serving milions of small images. Everything is working well until recovery... - 06:46 AM Backport #24471 (In Progress): luminous: Ceph-osd crash when activate SPDK
- https://github.com/ceph/ceph/pull/22686
- 05:01 AM Backport #24472 (In Progress): mimic: Ceph-osd crash when activate SPDK
- https://github.com/ceph/ceph/pull/22684
06/22/2018
- 11:25 PM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
- Uploaded a few more logs (debug 20) here: https://storage.googleapis.com/ceph-logs/ceph-osd-logs.tar.gz
After runn... - 12:45 PM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
- Hi Sage,
We've experienced this again on a new environment we setup. Took a snippet of the logs, hope it's enough:... - 04:46 PM Bug #23622 (Resolved): qa/workunits/mon/test_mon_config_key.py fails on master
- 04:45 PM Backport #23675 (Resolved): luminous: qa/workunits/mon/test_mon_config_key.py fails on master
- 04:25 PM Backport #23675: luminous: qa/workunits/mon/test_mon_config_key.py fails on master
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21368
merged - 04:44 PM Bug #23921 (Resolved): pg-upmap cannot balance in some case
- 04:43 PM Backport #24048 (Resolved): luminous: pg-upmap cannot balance in some case
- 04:25 PM Backport #24048: luminous: pg-upmap cannot balance in some case
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/22115
merged - 04:43 PM Bug #24025 (Resolved): RocksDB compression is not supported at least on Debian.
- 04:42 PM Backport #24279 (Resolved): luminous: RocksDB compression is not supported at least on Debian.
- 04:24 PM Backport #24279: luminous: RocksDB compression is not supported at least on Debian.
- Kefu Chai wrote:
> https://github.com/ceph/ceph/pull/22215
merged - 04:40 PM Backport #24329: mimic: assert manager.get_num_active_clean() == pg_num on rados/singleton/all/ma...
- original mimic backport https://github.com/ceph/ceph/pull/22288 was merged, but deemed insufficient
- 04:38 PM Backport #24328 (Resolved): luminous: assert manager.get_num_active_clean() == pg_num on rados/si...
- 04:23 PM Bug #24321: assert manager.get_num_active_clean() == pg_num on rados/singleton/all/max-pg-per-osd...
- merged https://github.com/ceph/ceph/pull/22296
- 04:15 PM Bug #24635 (New): luminous: LibRadosTwoPoolsPP.SetRedirectRead failed
- Probably a race with the redirect code.
From http://qa-proxy.ceph.com/teuthology/yuriw-2018-06-22_03:31:56-rados-w... - 01:19 PM Bug #23352: osd: segfaults under normal operation
- Yeah. We got in log mgr before segfault ceph-osd:
> mgr.server send_report send_report osd,74.0x560276d34ed8 sent me... - 03:54 AM Bug #23352: osd: segfaults under normal operation
- 03:54 AM Bug #23352: osd: segfaults under normal operation
- In several of the crashes we are seeing lines like the following prior to the crash....
- 12:42 PM Bug #17170: mon/monclient: update "unable to obtain rotating service keys when osd init" to sugge...
- I have a bit different effect on v12.2.5, but may be related:
I have similar logs:... - 08:44 AM Backport #24351 (Resolved): luminous: slow mon ops from osd_failure
- 12:23 AM Backport #24351: luminous: slow mon ops from osd_failure
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/22568
merged - 08:44 AM Bug #23386 (Resolved): crush device class: Monitor Crash when moving Bucket into Default root
- 08:43 AM Backport #24258 (Resolved): luminous: crush device class: Monitor Crash when moving Bucket into D...
- 12:21 AM Backport #24258: luminous: crush device class: Monitor Crash when moving Bucket into Default root
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/22381
mergedReviewed-by: Nathan Cutler <ncutler@suse.com> - 08:43 AM Backport #24290 (Resolved): luminous: common: JSON output from rados bench write has typo in max_...
- 12:20 AM Backport #24290: luminous: common: JSON output from rados bench write has typo in max_latency key
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/22391
merged - 08:41 AM Backport #24356 (Resolved): luminous: osd: pg hard limit too easy to hit
- 12:18 AM Backport #24356: luminous: osd: pg hard limit too easy to hit
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/22592
merged - 08:41 AM Backport #24618 (Resolved): mimic: osd: choose_acting loop
- https://github.com/ceph/ceph/pull/22889
- 08:41 AM Backport #24617 (Resolved): mimic: ValueError: too many values to unpack due to lack of subdir
- https://github.com/ceph/ceph/pull/22888
- 12:32 AM Bug #24615 (Resolved): error message for 'unable to find any IP address' not shown
- Hi,
In my ceph.conf I have the option:...
06/21/2018
- 11:40 PM Bug #24613 (New): luminous: rest/test.py fails with expected 200, got 400
- ...
- 10:57 PM Bug #23492: Abort in OSDMap::decode() during qa/standalone/erasure-code/test-erasure-eio.sh
- possibly related lumious run: http://pulpito.ceph.com/yuriw-2018-06-11_16:27:32-rados-wip-yuri3-testing-2018-06-11-14...
- 10:15 PM Bug #23352: osd: segfaults under normal operation
- Another instance: http://pulpito.ceph.com/yuriw-2018-06-19_21:29:48-rados-wip-yuri-testing-2018-06-19-1953-luminous-d...
- 09:01 PM Bug #24487 (Pending Backport): osd: choose_acting loop
- 05:58 PM Bug #24487 (Fix Under Review): osd: choose_acting loop
- https://github.com/ceph/ceph/pull/22664
- 06:48 PM Bug #24612 (Resolved): FAILED assert(osdmap_manifest.pinned.empty()) in OSDMonitor::prune_init()
- ...
- 04:51 PM Bug #23879: test_mon_osdmap_prune.sh fails
- /a/nojha-2018-06-21_00:18:52-rados-wip-24487-distro-basic-smithi/2686362
- 09:19 AM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
- We also hit this today, happen to have osd log with --debug_osd = 20
FWIW, the cluster has an inconsistent PG and ... - 01:34 AM Bug #24601 (Resolved): FAILED assert(is_up(osd)) in OSDMap::get_inst(int)
- ...
- 12:52 AM Bug #24600 (Resolved): ValueError: too many values to unpack due to lack of subdir
- ...
06/20/2018
- 10:13 PM Bug #24422: Ceph OSDs crashing in BlueStore::queue_transactions() using EC
- Sage Weil wrote:
> Can you generate an osd log with 'debug osd = 20' for the crashing osd that leads up to the crash... - 10:13 PM Bug #24422: Ceph OSDs crashing in BlueStore::queue_transactions() using EC
- Can you generate an osd log with 'debug osd = 20' for the crashing osd that leads up to the crash?
- 09:50 PM Bug #24422 (Duplicate): Ceph OSDs crashing in BlueStore::queue_transactions() using EC
- 10:11 PM Bug #23145: OSD crashes during recovery of EC pg
- Two basic theories:
1. There is a bug that prematurely advances can_rollback_to
2. One of Peter's OSDs warped bac... - 10:05 PM Bug #23145: OSD crashes during recovery of EC pg
- Sage Weil wrote:
> Zengran Zhang wrote:
> > osd in last peering stage will call pg_log.roll_forward(at last of PG:... - 10:03 PM Bug #23145 (Need More Info): OSD crashes during recovery of EC pg
- Yong Wang, can you provide a full osd log with debug osd = 20 for the primary osd for the PG leading up to the crash...
- 09:22 PM Bug #23145: OSD crashes during recovery of EC pg
- Zengran Zhang wrote:
> osd in last peering stage will call pg_log.roll_forward(at last of PG::activate), is there p... - 01:46 AM Bug #23145: OSD crashes during recovery of EC pg
- @Sage Weil @Zengran Zhang
could you shared something about this bug recently? - 01:44 AM Bug #23145: OSD crashes during recovery of EC pg
- hi all,did it has any updates please?
- 10:02 PM Backport #24599 (In Progress): mimic: failed to load OSD map for epoch X, got 0 bytes
- 10:01 PM Backport #24599 (Resolved): mimic: failed to load OSD map for epoch X, got 0 bytes
- https://github.com/ceph/ceph/pull/22651
- 09:47 PM Bug #24448 (Won't Fix): (Filestore) ABRT report for package ceph has reached 10 occurrences
- This is likely due to filestore becoming overloaded (hence waiting on throttles) and hitting the filestore op thread ...
- 09:38 PM Bug #24511 (Duplicate): osd crushed at thread_name:safe_timer
- 09:37 PM Bug #24515: "[WRN] Health check failed: 1 slow ops, oldest one blocked for 32 sec, mon.c has slow...
- Kefu, can you take a look at this?
- 09:36 PM Bug #24531: Mimic MONs have slow/long running ops
- Joao, could you take a look at this?
- 09:34 PM Bug #24549 (Won't Fix): FileStore::read assert (ABRT report for package ceph has reached 1000 occ...
- As John described, this is not a bug in ceph but due to failing hardware or the filesystem below.
- 09:25 PM Bug #23753 (Can't reproduce): "Error ENXIO: problem getting command descriptions from osd.4" in u...
- re-open if it recurs
- 09:19 PM Bug #22624: filestore: 3180: FAILED assert(0 == "unexpected error"): error (2) No such file or di...
- 09:12 PM Bug #22085 (Can't reproduce): jewel->luminous: "[ FAILED ] LibRadosAioEC.IsSafe" in upgrade:jew...
- assuming this is the mon crush testing timeout, logs are gone so can't be sure
- 08:10 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- backport for mimic: https://github.com/ceph/ceph/pull/22651
- 08:07 PM Bug #24423 (Pending Backport): failed to load OSD map for epoch X, got 0 bytes
- 07:46 PM Bug #24597 (Resolved): FAILED assert(0 == "ERROR: source must exist") in FileStore::_collection_m...
- ...
- 06:32 PM Bug #20086: LibRadosLockECPP.LockSharedDurPP gets EEXIST
- ...
- 03:01 PM Bug #23492: Abort in OSDMap::decode() during qa/standalone/erasure-code/test-erasure-eio.sh
Now that I've looked at the code there is nothing surprising about the map handling. There is code in dequeue_op()...- 12:37 AM Bug #23492: Abort in OSDMap::decode() during qa/standalone/erasure-code/test-erasure-eio.sh
I was able to reproduce by running a loop of a single test case in qa/standalone/erasure-code/test-erasure-eio.sh
...- 01:00 PM Backport #23673 (Resolved): jewel: auth: ceph auth add does not sanity-check caps
- 12:52 PM Bug #23872 (Resolved): Deleting a pool with active watch/notify linger ops can result in seg fault
- 12:52 PM Backport #23905 (Resolved): jewel: Deleting a pool with active watch/notify linger ops can result...
- 12:21 PM Backport #24383 (In Progress): mimic: osd: stray osds in async_recovery_targets cause out of orde...
- https://github.com/ceph/ceph/pull/22642
- 08:42 AM Bug #24588 (Fix Under Review): osd: may get empty info at recovery
- -https://github.com/ceph/ceph/pull/22362-
- 01:42 AM Bug #24588 (Resolved): osd: may get empty info at recovery
- 2018-06-15 20:34:16.421720 7f89d2c24700 -1 /home/zzr/ceph.sf/src/osd/PG.cc: In function 'void PG::start_peering_inter...
- 08:40 AM Bug #24593: s390x: Ceph Monitor crashed with Caught signal (Aborted)
- I expect that only people in possession of s390x hardware will be able to debug this
I see that there is another t... - 05:33 AM Bug #24593 (New): s390x: Ceph Monitor crashed with Caught signal (Aborted)
- We are trying to setup ceph cluster on s390x platform.
ceph-mon service crashed with an error: *** Caught signal ... - 05:50 AM Feature #24591 (Fix Under Review): FileStore hasn't impl to get kv-db's statistics
- 03:22 AM Feature #24591: FileStore hasn't impl to get kv-db's statistics
- https://github.com/ceph/ceph/pull/22633
- 03:22 AM Feature #24591 (Fix Under Review): FileStore hasn't impl to get kv-db's statistics
- In BlueStore, you can see kv-db's statistics by "ceph daemon osd.X dump_objectstore_kv_stats", but FileStore hasn't i...
- 03:22 AM Feature #22147: Set multiple flags in a single command line
- I don’t think we should skip it entirely. Many of the places that implement a check like that are using a common flag...
06/19/2018
- 11:44 PM Bug #24487 (In Progress): osd: choose_acting loop
- This happens when an osd which is part of the acting set and not a part the upset, gets chosen as an async_recovery_t...
- 10:51 PM Backport #23673: jewel: auth: ceph auth add does not sanity-check caps
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21367
merged - 10:50 PM Backport #23905: jewel: Deleting a pool with active watch/notify linger ops can result in seg fault
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21754
merged - 10:49 PM Feature #22147: Set multiple flags in a single command line
- It seems fair to assume that "unset" should support this also.
Question: should settings that require --yes-i-real... - 10:40 PM Bug #24587: librados api aio tests race condition
- http://pulpito.ceph.com/yuriw-2018-06-13_14:55:30-rados-wip-yuri4-testing-2018-06-12-2037-jewel-distro-basic-smithi/2...
- 10:38 PM Bug #24587 (Resolved): librados api aio tests race condition
- Seen in a jewel integration branch with no OSD changes:
http://pulpito.ceph.com/yuriw-2018-06-12_22:32:43-rados-wi... - 09:58 PM Bug #23492: Abort in OSDMap::decode() during qa/standalone/erasure-code/test-erasure-eio.sh
- I did a run based on d9284902e1b2e292595696caf11cdead18acec96 which is a branch off of master.
http://pulpito.ceph... - 07:24 PM Backport #24584 (Resolved): luminous: osdc: wrong offset in BufferHead
- https://github.com/ceph/ceph/pull/22865
- 07:24 PM Backport #24583 (Resolved): mimic: osdc: wrong offset in BufferHead
- https://github.com/ceph/ceph/pull/22869
- 06:02 PM Bug #19971 (Resolved): osd: deletes are performed inline during pg log processing
- 06:01 PM Backport #22406 (Rejected): jewel: osd: deletes are performed inline during pg log processing
- This change was deemed too invasive at such a late stage in Jewel's life cycle.
- 06:01 PM Backport #22405 (Rejected): jewel: store longer dup op information
- This change was deemed too invasive at such a late stage in Jewel's life cycle.
- 06:00 PM Backport #22400 (Rejected): jewel: PR #16172 causing performance regression
- This change was deemed too invasive at such a late stage in Jewel's life cycle.
- 04:10 PM Bug #24484 (Pending Backport): osdc: wrong offset in BufferHead
- 11:54 AM Bug #24448: (Filestore) ABRT report for package ceph has reached 10 occurrences
- OSD killed by signal, something like OOM incidents perhaps?
- 11:53 AM Bug #24450 (Duplicate): OSD Caught signal (Aborted)
- http://tracker.ceph.com/issues/24423
- 11:51 AM Bug #24559 (Fix Under Review): building error for QAT decompress
- 02:10 AM Bug #24559 (Fix Under Review): building error for QAT decompress
- The parameter of decompress changes from 'bufferlist::iterator' to 'bufferlist::const_iterator', but chis change miss...
- 11:34 AM Bug #24549: FileStore::read assert (ABRT report for package ceph has reached 1000 occurrences)
- Presumably this is underlying FS failures tripping asserts rather than a bug (perhaps people using ZFS on centos, or ...
- 07:26 AM Backport #24355 (In Progress): mimic: osd: pg hard limit too easy to hit
- https://github.com/ceph/ceph/pull/22621
06/18/2018
- 05:51 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- 11:45 AM Bug #24549 (Won't Fix): FileStore::read assert (ABRT report for package ceph has reached 1000 occ...
- FileStore::read(coll_t, ghobject_t const&, unsigned long, unsigned long, ceph::buffer::list&, unsigned int, bool)
... - 07:11 AM Backport #24356 (In Progress): luminous: osd: pg hard limit too easy to hit
- https://github.com/ceph/ceph/pull/22592
06/16/2018
- 02:16 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- How to fix installed Mimic (upgraded from Luminous) with this fix? Is there any way to make startup OSD not requestin...
06/15/2018
- 11:40 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- I've fixed it here: https://github.com/ceph/ceph/pull/22585
- 01:36 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- Not sure if this is related, but for a few days, I'm not able to modify crushmap (like adding or removing OSD) on a l...
- 09:23 AM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- Seeing the same here with a new Mimic cluster.
I purged a few OSDs (deployment went wrong) and now they can't star... - 03:56 PM Bug #24057: cbt fails to copy results to the archive dir
- 02:48 PM Bug #24531: Mimic MONs have slow/long running ops
- ...
- 02:41 PM Bug #24531: Mimic MONs have slow/long running ops
- What's the output of "ceph versions" on this cluster?
We had issues in the lab with OSD failure reports not gettin... - 02:20 PM Bug #24531 (Resolved): Mimic MONs have slow/long running ops
- When setting up a Mimic 13.2.0 cluster I saw a message like this:...
- 08:39 AM Bug #24529 (New): monitor report empty client io rate when clock not synchronized
- we run rados bench when cluster is warn and clock is not synchronized. on the other hand, we watch io speed from resu...
- 05:08 AM Backport #24351 (In Progress): luminous: slow mon ops from osd_failure
- https://github.com/ceph/ceph/pull/22568
06/14/2018
- 10:21 PM Bug #21142 (Need More Info): OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
- 10:20 PM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
- Tim, Dexter, is this something that is reproducible in your environment? I haven't seen this one, which makes me ver...
- 07:41 PM Bug #23492: Abort in OSDMap::decode() during qa/standalone/erasure-code/test-erasure-eio.sh
This might be caused by 52dd99e3011bfc787042fe105e02c11b28867c4c which was included in https://github.com/ceph/ceph...- 07:27 PM Bug #24526: Mimic OSDs do not start after deleting some pools with size=1
- I solved this issue by monkey-patching OSD code:...
- 03:48 PM Bug #24526: Mimic OSDs do not start after deleting some pools with size=1
- P.S: This happened just after deleting some pool with size=1 - several OSDs died immediately and the latest error mes...
- 03:24 PM Bug #24526 (New): Mimic OSDs do not start after deleting some pools with size=1
- After some amount of test actions involving creating pools with size=min_size=1 and then deleting them, most OSDs fai...
- 07:06 PM Feature #24527 (New): Need a pg query that doens't include invalid peer information
Some fields in the peer info remain unchanged after a peer transitions from being the primary. This information ma...- 01:13 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- I am getting the same issue.
I also upgraded to Luminous to Mimic.
I used: ceph osd purge - 11:48 AM Backport #24198 (Resolved): luminous: mon: slow op on log message
- 11:47 AM Backport #24216 (Resolved): luminous: "process (unknown)" in ceph logs
- 11:46 AM Bug #24167 (Resolved): Module 'balancer' has failed: could not find bucket -14
- 11:46 AM Backport #24213 (Resolved): mimic: Module 'balancer' has failed: could not find bucket -14
- 11:45 AM Backport #24214 (Resolved): luminous: Module 'balancer' has failed: could not find bucket -14
- 05:54 AM Backport #24332 (In Progress): mimic: local_reserver double-reservation of backfilled pg
- https://github.com/ceph/ceph/pull/22559
06/13/2018
- 10:01 PM Backport #24198: luminous: mon: slow op on log message
- Kefu Chai wrote:
> https://github.com/ceph/ceph/pull/22109
merged - 10:00 PM Backport #24216: luminous: "process (unknown)" in ceph logs
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/22290
merged - 09:59 PM Backport #24214: luminous: Module 'balancer' has failed: could not find bucket -14
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/22308
merged - 08:13 PM Bug #24515 (New): "[WRN] Health check failed: 1 slow ops, oldest one blocked for 32 sec, mon.c ha...
- This seems to be rhel specific
Run: http://pulpito.ceph.com/yuriw-2018-06-12_21:09:43-fs-master-distro-basic-smith... - 05:19 PM Bug #23966 (Resolved): Deleting a pool with active notify linger ops can result in seg fault
- 05:19 PM Backport #24059 (Resolved): luminous: Deleting a pool with active notify linger ops can result in...
- 04:46 PM Backport #24468 (In Progress): mimic: tell ... config rm <foo> not idempotent
- 04:35 PM Backport #24245 (Resolved): luminous: Manager daemon y is unresponsive during teuthology cluster ...
- 04:34 PM Backport #24374 (Resolved): luminous: mon: auto compaction on rocksdb should kick in more often
- 12:56 PM Bug #24511 (Duplicate): osd crushed at thread_name:safe_timer
- h1. ENV
*ceph version*... - 11:29 AM Bug #23049: ceph Status shows only WARN when traffic to cluster fails
- hi,
which is the expected fix release version?
Thanks, - 10:16 AM Backport #24501 (In Progress): luminous: osd: eternal stuck PG in 'unfound_recovery'
- 10:16 AM Backport #24500 (In Progress): mimic: osd: eternal stuck PG in 'unfound_recovery'
06/12/2018
- 08:01 AM Backport #24501 (Resolved): luminous: osd: eternal stuck PG in 'unfound_recovery'
- https://github.com/ceph/ceph/pull/22546
- 08:01 AM Backport #24500 (Resolved): mimic: osd: eternal stuck PG in 'unfound_recovery'
- https://github.com/ceph/ceph/pull/22545
- 08:00 AM Backport #24495 (Resolved): luminous: osd: segv in Session::have_backoff
- https://github.com/ceph/ceph/pull/22729
- 08:00 AM Backport #24494 (Resolved): mimic: osd: segv in Session::have_backoff
- https://github.com/ceph/ceph/pull/22730
- 03:22 AM Bug #24486 (Pending Backport): osd: segv in Session::have_backoff
06/11/2018
- 09:32 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- I am going to add this test for upgrade as well, steps to recreate...
- 04:19 AM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- I have also experienced this issue while continuing the Bluestore conversion of OSDs on my Ceph cluster, after carryi...
- 02:16 PM Backport #24059: luminous: Deleting a pool with active notify linger ops can result in seg fault
- Casey Bodley wrote:
> https://github.com/ceph/ceph/pull/22143
merged - 02:33 AM Bug #24487: osd: choose_acting loop
- It looks like the "choose_async_recovery_ec candidates by cost are: 178,2(0)" line is different in the second case.. ...
- 01:45 AM Bug #24487 (Resolved): osd: choose_acting loop
- ec pg looping between [2,3,0,1] and [-,3,0,1].
osd.3 says...
06/10/2018
- 06:41 PM Bug #24486 (Fix Under Review): osd: segv in Session::have_backoff
- https://github.com/ceph/ceph/pull/22497
- 06:34 PM Bug #24486 (Resolved): osd: segv in Session::have_backoff
- ...
- 04:41 PM Bug #24485 (Resolved): LibRadosTwoPoolsPP.ManifestUnset failure
- ...
- 03:30 PM Bug #24484 (Fix Under Review): osdc: wrong offset in BufferHead
- 03:15 PM Bug #24484: osdc: wrong offset in BufferHead
- this bug will lead to an exception "buffer::end_of_buffer" which is thrown in function "buffer::list::substr_of"
Thi... - 03:08 PM Bug #24484: osdc: wrong offset in BufferHead
- PR: https://github.com/ceph/ceph/pull/22495
- 03:07 PM Bug #24484 (Resolved): osdc: wrong offset in BufferHead
- The offset of BufferHead should be "opos - bh->start()"
- 02:12 AM Backport #24329 (In Progress): mimic: assert manager.get_num_active_clean() == pg_num on rados/si...
06/09/2018
- 07:21 PM Bug #24321 (Pending Backport): assert manager.get_num_active_clean() == pg_num on rados/singleton...
- 05:56 AM Bug #24321 (Fix Under Review): assert manager.get_num_active_clean() == pg_num on rados/singleton...
- https://github.com/ceph/ceph/pull/22485
- 06:50 PM Bug #22462: mon: unknown message type 1537 in luminous->mimic upgrade tests
- Maybe i have the same issue during upgrade Jewel->Luminous http://tracker.ceph.com/issues/24481?next_issue_id=24480&p...
- 02:23 PM Bug #24373 (Pending Backport): osd: eternal stuck PG in 'unfound_recovery'
- 11:20 AM Backport #24478 (Resolved): luminous: read object attrs failed at EC recovery
- https://github.com/ceph/ceph/pull/24327
- 11:18 AM Backport #24473 (Resolved): mimic: cosbench stuck at booting cosbench driver
- https://github.com/ceph/ceph/pull/22887
- 11:18 AM Backport #24472 (Resolved): mimic: Ceph-osd crash when activate SPDK
- https://github.com/ceph/ceph/pull/22684
- 11:18 AM Backport #24471 (Resolved): luminous: Ceph-osd crash when activate SPDK
- https://github.com/ceph/ceph/pull/22686
- 11:18 AM Backport #24468 (Resolved): mimic: tell ... config rm <foo> not idempotent
- https://github.com/ceph/ceph/pull/22552
- 06:07 AM Bug #24452 (Resolved): Backfill hangs in a test case in master not mimic
06/08/2018
- 11:03 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- I can't reproduce this on any new Mimic cluster, it only happens on clusters upgraded from Luminous (which is why we ...
- 09:04 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- I'm trying to make new OSDs with ceph-volume osd create --dmcrypt --bluestore --data /dev/sdg and am getting the same...
- 07:05 PM Bug #24454 (Duplicate): failed to recover before timeout expired
- #24452
- 12:29 PM Bug #24454 (Duplicate): failed to recover before timeout expired
- tons of this on current master
http://pulpito.ceph.com/kchai-2018-06-06_04:56:43-rados-wip-kefu-testing-2018-06-06... - 07:05 PM Bug #24452 (Fix Under Review): Backfill hangs in a test case in master not mimic
- https://github.com/ceph/ceph/pull/22478
- 02:48 PM Bug #24452: Backfill hangs in a test case in master not mimic
Final messages on primary during backfill about pg 1.0....- 04:57 AM Bug #24452 (Resolved): Backfill hangs in a test case in master not mimic
../qa/run-standalone.sh "osd-backfill-stats.sh TEST_backfill_down_out" 2>&1 | tee obs.log
This test times out wa...- 02:34 PM Backport #23912: luminous: mon: High MON cpu usage when cluster is changing
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/21968
merged - 02:33 PM Backport #24245: luminous: Manager daemon y is unresponsive during teuthology cluster teardown
- Nathan Cutler wrote:
> https://github.com/ceph/ceph/pull/22331
merged - 02:31 PM Backport #24374: luminous: mon: auto compaction on rocksdb should kick in more often
- Kefu Chai wrote:
> https://github.com/ceph/ceph/pull/22360
merged - 08:18 AM Bug #23352: osd: segfaults under normal operation
- Experiencing the a safe_timer segfault with a freshly deployed cluster. No data on the cluster yet. Just an empty poo...
06/07/2018
- 03:20 PM Bug #24423: failed to load OSD map for epoch X, got 0 bytes
- We are also seeing this when creating OSDs with IDs that existed previously.
I verified that the old osd was delet... - 01:21 PM Bug #24373: osd: eternal stuck PG in 'unfound_recovery'
- https://github.com/ceph/ceph/pull/22456
- 01:14 PM Bug #24373: osd: eternal stuck PG in 'unfound_recovery'
- Okay, I see the problem. Two fixes: first, reset every pg on down->up (simpler approach), but the bigger issue is th...
- 12:58 PM Bug #24450: OSD Caught signal (Aborted)
- I have the same problem.
http://tracker.ceph.com/issues/24423 - 12:03 PM Bug #24450 (Duplicate): OSD Caught signal (Aborted)
- Hi,
I have done a rolling_upgrade to mimic with ceph-ansible. It works perfect! Now, I want to deploy new OSDs, bu... - 11:46 AM Bug #24448 (Won't Fix): (Filestore) ABRT report for package ceph has reached 10 occurrences
- https://retrace.fedoraproject.org/faf/reports/bthash/fe768f98e5fff65f0c850668c4bdae8d4da7e086/
https://retrace.fedor...
06/06/2018
- 09:11 PM Bug #24264 (Closed): ssd-primary crush rule not working as intended
- I don't think there's a good way to express that requirement in the current crush language. The rule in the docs does...
- 09:06 PM Bug #24362 (Triaged): ceph-objectstore-tool incorrectly invokes crush_location_hook
- Seems like the way to fix this is to stop ceph-objectstore-tool from trying to use the crush location hook at all.
... - 07:15 AM Bug #23145: OSD crashes during recovery of EC pg
- -3> 2018-06-06 15:00:40.462930 7fffddb25700 -1 bluestore(/var/lib/ceph/osd/ceph-12) _txc_add_transaction error (2...
- 02:45 AM Bug #23145: OSD crashes during recovery of EC pg
- @Sage Weil
@Zengran Zahng
we meet the some question, and osd crash not recover until now.
env is 12.2.5 ec 2+1 b... - 06:02 AM Backport #24293 (In Progress): jewel: mon: slow op on log message
- https://github.com/ceph/ceph/pull/22431
- 02:34 AM Bug #24373: osd: eternal stuck PG in 'unfound_recovery'
- Attached full log (download ceph-osd.3.log.gz).
Points are:... - 12:33 AM Bug #24371 (Pending Backport): Ceph-osd crash when activate SPDK
06/05/2018
- 05:34 PM Bug #24365 (Pending Backport): cosbench stuck at booting cosbench driver
- 01:33 AM Bug #24365 (Fix Under Review): cosbench stuck at booting cosbench driver
- https://github.com/ceph/ceph/pull/22405
- 04:04 PM Bug #24408 (Pending Backport): tell ... config rm <foo> not idempotent
- 11:00 AM Bug #24423 (Resolved): failed to load OSD map for epoch X, got 0 bytes
- After upgrading to Mimic I deleted a non-lvm OSD and recreated it with 'ceph-volume lvm prepare --bluestore --data /d...
- 10:37 AM Bug #24422: Ceph OSDs crashing in BlueStore::queue_transactions() using EC
- the same to https://tracker.ceph.com/issues/21475. and i already modify bluestore_deferred_throttle_bytes = 0
bluest... - 10:31 AM Bug #24422: Ceph OSDs crashing in BlueStore::queue_transactions() using EC
- 2018-06-05T17:46:28.273183+08:00 node54 ceph-osd: /work/build/rpmbuild/BUILD/infinity-3.2.5/src/os/bluestore/BlueStor...
- 10:31 AM Bug #24422: Ceph OSDs crashing in BlueStore::queue_transactions() using EC
- 鹏 张 wrote:
> ceph version: 12.2.5
> data pool use Ec module 2 + 1.
> When restart one osd,it case crash and restar... - 10:26 AM Bug #24422: Ceph OSDs crashing in BlueStore::queue_transactions() using EC
- 1.-45> 2018-06-05 17:47:56.886142 7f8972974700 -1 bluestore(/var/lib/ceph/osd/ceph-12) _txc_add_transaction error (2)...
- 10:25 AM Bug #24422 (Duplicate): Ceph OSDs crashing in BlueStore::queue_transactions() using EC
- ceph version: 12.2.5
data pool use Ec module 3 + 1.
When restart one osd,it case crash and restart more and more.
... - 04:42 AM Bug #24419 (Won't Fix): ceph-objectstore-tool unable to open mon store
- Hi,everyone;
I use luminous v12.2.5,and i try to recovery monitor database from osds,
I perform step by step acc... - 03:32 AM Backport #24291 (In Progress): jewel: common: JSON output from rados bench write has typo in max_...
- https://github.com/ceph/ceph/pull/22407
- 02:37 AM Bug #23875: Removal of snapshot with corrupt replica crashes osd
If update_snap_map() ignores the error from remove_oid() we still crash because an op from the primary related to...- 02:20 AM Backport #24292 (In Progress): mimic: common: JSON output from rados bench write has typo in max_...
- https://github.com/ceph/ceph/pull/22406
06/04/2018
- 06:32 PM Bug #24368: osd: should not restart on permanent failures
- It would, but the previous settings were there for a reason so I'm not sure if it's feasible to backport this for cep...
- 05:10 PM Bug #24371 (Fix Under Review): Ceph-osd crash when activate SPDK
- 04:00 PM Bug #24408 (Fix Under Review): tell ... config rm <foo> not idempotent
- https://github.com/ceph/ceph/pull/22395
- 03:56 PM Bug #24408 (Resolved): tell ... config rm <foo> not idempotent
- ...
- 02:56 PM Backport #24407 (In Progress): mimic: read object attrs failed at EC recovery
- 02:56 PM Backport #24407 (Resolved): mimic: read object attrs failed at EC recovery
- https://github.com/ceph/ceph/pull/22394
- 02:54 PM Bug #24406 (Resolved): read object attrs failed at EC recovery
- https://github.com/ceph/ceph/pull/22196
- 02:18 PM Backport #24290 (In Progress): luminous: common: JSON output from rados bench write has typo in m...
- https://github.com/ceph/ceph/pull/22391
- 11:53 AM Bug #24366 (Pending Backport): omap_digest handling still not correct
- 06:27 AM Bug #23352: osd: segfaults under normal operation
- Looking at the crash in http://tracker.ceph.com/issues/23352#note-14 there's a fairly glaring problem....
- 12:14 AM Bug #23352: osd: segfaults under normal operation
- Hi Kjetil,
Sure, worth a look, but AFAICT all access is protected by SafeTimers locks. - 02:08 AM Backport #24258 (In Progress): luminous: crush device class: Monitor Crash when moving Bucket int...
- https://github.com/ceph/ceph/pull/22381
06/02/2018
- 12:04 AM Bug #24365 (In Progress): cosbench stuck at booting cosbench driver
- Two things caused this issue:
1. cosbench requires openjdk-8. The cbt task does install this dependency, but we al...
06/01/2018
- 08:05 PM Bug #23352: osd: segfaults under normal operation
- Brad Hubbard wrote:
> I've confirmed that in all of the SafeTimer segfaults the 'schedule' multimap is empty, indica... - 06:01 PM Bug #24368: osd: should not restart on permanent failures
- Sounds like something that would be useful in our stable releases - Greg, do you agree?
- 05:56 PM Backport #24360 (Need More Info): luminous: osd: leaked Session on osd.7
- Do Not Backport For Now
see https://github.com/ceph/ceph/pull/22339#issuecomment-393574371 for details - 05:44 PM Backport #24383 (Resolved): mimic: osd: stray osds in async_recovery_targets cause out of order ops
- https://github.com/ceph/ceph/pull/22889
- 05:28 PM Backport #24381 (Resolved): luminous: omap_digest handling still not correct
- https://github.com/ceph/ceph/pull/22375
- 05:28 PM Backport #24380 (Resolved): mimic: omap_digest handling still not correct
- https://github.com/ceph/ceph/pull/22374
- 08:02 AM Bug #24342: Monitor's routed_requests leak
- Greg Farnum wrote:
> What version are you running? The MRoute handling is all pretty old; though we've certainly dis... - 07:16 AM Bug #24373 (Fix Under Review): osd: eternal stuck PG in 'unfound_recovery'
- 05:22 AM Bug #24373: osd: eternal stuck PG in 'unfound_recovery'
- https://github.com/ceph/ceph/pull/22358
- 04:57 AM Bug #24373 (Resolved): osd: eternal stuck PG in 'unfound_recovery'
- A PG might be eternally stuck in 'unfound_recovery' after some OSDs are marked down.
For example, the following st... - 06:12 AM Backport #24375 (In Progress): mimic: mon: auto compaction on rocksdb should kick in more often
- 06:11 AM Backport #24375 (Resolved): mimic: mon: auto compaction on rocksdb should kick in more often
- https://github.com/ceph/ceph/pull/22361
- 06:10 AM Backport #24374 (In Progress): luminous: mon: auto compaction on rocksdb should kick in more often
- 06:08 AM Backport #24374 (Resolved): luminous: mon: auto compaction on rocksdb should kick in more often
- https://github.com/ceph/ceph/pull/22360
- 06:08 AM Bug #24361 (Pending Backport): auto compaction on rocksdb should kick in more often
- 04:47 AM Bug #24371: Ceph-osd crash when activate SPDK
- This is a bug in NVMEDevice, the bug fix has been committed.
Please have a review PR https://github.com/ceph/ceph... - 02:02 AM Bug #24371: Ceph-osd crash when activate SPDK
- I'm working on the issue.
- 02:01 AM Bug #24371 (Resolved): Ceph-osd crash when activate SPDK
- Enable SPDK and configure bluestore as mentioned in http://docs.ceph.com/docs/master/rados/configuration/bluestore-co...
- 02:56 AM Feature #24363: Configure DPDK with mellanox NIC
- next, compiling pass. but all binaries can not run.
output error
EAL: VFIO_RESOURCE_LIST tailq is already registere... - 02:38 AM Feature #24363: Configure DPDK with mellanox NIC
- log details
mellanox NIC over fabric
When compiling output error.
1. lack numa and cryptopp libraries
I ... - 12:23 AM Feature #24363: Configure DPDK with mellanox NIC
- Append
NIC over optical fiber - 12:07 AM Bug #24160 (Resolved): Monitor down when large store data needs to compact triggered by ceph tell...
05/31/2018
- 11:34 PM Bug #24368 (In Progress): osd: should not restart on permanent failures
- https://github.com/ceph/ceph/pull/22349 has the simple restart interval change. Will investigate the options for cond...
- 11:25 PM Bug #24368: osd: should not restart on permanent failures
- See https://www.freedesktop.org/software/systemd/man/systemd.service.html#Restart= for the details on Restart options.
- 11:17 PM Bug #24368 (Resolved): osd: should not restart on permanent failures
- Last week at OpenStack I heard a few users report OSDs were not failing hard and fast as they should be on disk issue...
- 07:01 PM Bug #24366 (In Progress): omap_digest handling still not correct
- https://github.com/ceph/ceph/pull/22346
- 05:39 PM Bug #24366 (Resolved): omap_digest handling still not correct
When running bluestore the object info data_digest is not needed. In that case the omap_digest handling is still b...- 06:08 PM Bug #24349 (Pending Backport): osd: stray osds in async_recovery_targets cause out of order ops
- 12:51 AM Bug #24349: osd: stray osds in async_recovery_targets cause out of order ops
- https://github.com/ceph/ceph/pull/22330
- 12:46 AM Bug #24349 (Resolved): osd: stray osds in async_recovery_targets cause out of order ops
- Related to https://tracker.ceph.com/issues/23827
http://pulpito.ceph.com/yuriw-2018-05-24_17:07:20-powercycle-mast... - 05:07 PM Bug #24365 (Resolved): cosbench stuck at booting cosbench driver
- ...
- 03:54 PM Bug #24342: Monitor's routed_requests leak
- What version are you running? The MRoute handling is all pretty old; though we've certainly discovered a number of le...
- 02:17 PM Feature #24363 (New): Configure DPDK with mellanox NIC
- Hi all
Whether ceph-13.1.0 support DPDK on mellanox NIC?
I found many issues when compiling. I even though handle t... - 01:22 PM Bug #24362 (Triaged): ceph-objectstore-tool incorrectly invokes crush_location_hook
- Ceph release being used: 12.5.5 (cad919881333ac92274171586c827e01f554a70a) luminous (stable)
/etc/ceph/ceph.conf c... - 11:50 AM Backport #24359 (In Progress): mimic: osd: leaked Session on osd.7
- 07:39 AM Backport #24359 (Resolved): mimic: osd: leaked Session on osd.7
- https://github.com/ceph/ceph/pull/22339
- 09:40 AM Bug #24361 (Fix Under Review): auto compaction on rocksdb should kick in more often
- https://github.com/ceph/ceph/pull/22337
- 09:07 AM Bug #24361 (Resolved): auto compaction on rocksdb should kick in more often
- in rocksdb, by default, "max_bytes_for_level_base" is 256MB, "max_bytes_for_level_multiplier" is 10. so with this set...
- 07:39 AM Backport #24360 (Resolved): luminous: osd: leaked Session on osd.7
- https://github.com/ceph/ceph/pull/29859
- 07:38 AM Backport #24350 (In Progress): mimic: slow mon ops from osd_failure
- 07:37 AM Backport #24350 (Resolved): mimic: slow mon ops from osd_failure
- https://github.com/ceph/ceph/pull/22297
- 07:38 AM Backport #24356 (Resolved): luminous: osd: pg hard limit too easy to hit
- https://github.com/ceph/ceph/pull/22592
- 07:38 AM Backport #24355 (Resolved): mimic: osd: pg hard limit too easy to hit
- https://github.com/ceph/ceph/pull/22621
- 07:37 AM Backport #24351 (Resolved): luminous: slow mon ops from osd_failure
- https://github.com/ceph/ceph/pull/22568
- 05:31 AM Bug #20924 (Pending Backport): osd: leaked Session on osd.7
- i think https://github.com/ceph/ceph/pull/22292 indeed addresses this issue
https://github.com/ceph/ceph/pull/22384 - 04:51 AM Backport #24246 (In Progress): mimic: Manager daemon y is unresponsive during teuthology cluster ...
- https://github.com/ceph/ceph/pull/22333
- 02:55 AM Backport #24245 (In Progress): luminous: Manager daemon y is unresponsive during teuthology clust...
- https://github.com/ceph/ceph/pull/22331
05/30/2018
- 11:31 PM Bug #24160 (Fix Under Review): Monitor down when large store data needs to compact triggered by c...
- 10:45 PM Bug #23830: rados/standalone/erasure-code.yaml gets 160 byte pgmeta object
- This looks like a similar failure: http://pulpito.ceph.com/nojha-2018-05-30_20:43:02-rados-wip-async-up2-2018-05-30-d...
- 02:17 PM Bug #24342: Monitor's routed_requests leak
- It seems that this problem has been fixed by https://github.com/ceph/ceph/commit/39e06ef8f070e136e54452bdea3f6105cd79...
- 01:10 PM Bug #24342 (Closed): Monitor's routed_requests leak
- 12:09 PM Bug #24342: Monitor's routed_requests leak
- Sorry, it seems that the latest version doesn't have this problem. Really sorry. please close this.
- 09:36 AM Bug #24342: Monitor's routed_requests leak
- https://github.com/ceph/ceph/pull/22315
- 08:54 AM Bug #24342 (Closed): Monitor's routed_requests leak
- Recently, we found that, in our non-leader monitors, there are a lot of routed requests that has not been recycled, a...
- 01:58 PM Bug #24327: osd: segv in pg_log_entry_t::encode()
- Sage Weil wrote:
> This crash doesn't look familiar, and it's not clear to me what might cause segfault here. Do yo... - 01:48 PM Bug #24327 (Need More Info): osd: segv in pg_log_entry_t::encode()
- This crash doesn't look familiar, and it's not clear to me what might cause segfault here. Do you have a core file?
- 01:55 PM Bug #24339: FULL_FORCE ops are dropped if fail-safe full check fails, but not resent in scan_requ...
- Josh and I noticed this by code inspection. I'm nailing down out of space handling nits in the kernel client and wan...
- 01:46 PM Bug #24339: FULL_FORCE ops are dropped if fail-safe full check fails, but not resent in scan_requ...
- This is somewhat by design (or lack thereof)... the fail-safe check is there to prevent us from writing when we are *...
- 05:40 AM Backport #24215 (In Progress): mimic: "process (unknown)" in ceph logs
- https://github.com/ceph/ceph/pull/22311
- 03:29 AM Backport #24214 (In Progress): luminous: Module 'balancer' has failed: could not find bucket -14
- https://github.com/ceph/ceph/pull/22308
05/29/2018
- 11:01 PM Feature #23979: Limit pg log length during recovery/backfill so that we don't run out of memory.
- Initial testing is referenced here: https://github.com/ceph/ceph/pull/21508
- 10:59 PM Bug #24243 (Pending Backport): osd: pg hard limit too easy to hit
- https://github.com/ceph/ceph/pull/22187
- 10:59 PM Bug #24304 (Fix Under Review): MgrStatMonitor decode crash on 12.2.4->12.2.5 upgrade
- wrong bug
- 10:58 PM Bug #24304 (Pending Backport): MgrStatMonitor decode crash on 12.2.4->12.2.5 upgrade
- https://github.com/ceph/ceph/pull/22187
- 10:03 PM Feature #11601: osd: share cached osdmaps across osd daemons
- A vague possibility that the future seastar-based OSD may run each logical disk OSD inside a single process, which co...
- 07:38 PM Bug #24339 (New): FULL_FORCE ops are dropped if fail-safe full check fails, but not resent in sca...
- FULL_FORCE ops are dropped if fail-safe full check fails in do_op(). scan_requests() uses op->respects_full() which ...
- 06:49 PM Bug #23646 (Resolved): scrub interaction with HEAD boundaries and clones is broken
- 01:11 PM Bug #24322 (Pending Backport): slow mon ops from osd_failure
- mimic: https://github.com/ceph/ceph/pull/22297
- 12:53 PM Backport #24328 (In Progress): luminous: assert manager.get_num_active_clean() == pg_num on rados...
- 09:40 AM Backport #24328 (Resolved): luminous: assert manager.get_num_active_clean() == pg_num on rados/si...
- https://github.com/ceph/ceph/pull/22296
- 12:47 PM Backport #24329 (Resolved): mimic: assert manager.get_num_active_clean() == pg_num on rados/singl...
- 09:40 AM Backport #24329 (Resolved): mimic: assert manager.get_num_active_clean() == pg_num on rados/singl...
- https://github.com/ceph/ceph/pull/22492
- 10:02 AM Bug #22530 (Resolved): pool create cmd's expected_num_objects is not correctly interpreted
- 10:02 AM Backport #23316 (Resolved): jewel: pool create cmd's expected_num_objects is not correctly interp...
- 10:01 AM Backport #24058 (Resolved): jewel: Deleting a pool with active notify linger ops can result in se...
- 09:59 AM Backport #24244 (Resolved): jewel: osd/EC: slow/hung ops in multimds suite test
- 09:59 AM Backport #24244 (In Progress): jewel: osd/EC: slow/hung ops in multimds suite test
- 09:56 AM Backport #24294 (Resolved): mimic: control-c on ceph cli leads to segv
- 09:55 AM Backport #24294 (In Progress): mimic: control-c on ceph cli leads to segv
- 09:52 AM Backport #24256 (Resolved): mimic: osd: Assertion `!node_algorithms::inited(this->priv_value_tra...
- 09:41 AM Backport #24333 (Resolved): luminous: local_reserver double-reservation of backfilled pg
- https://github.com/ceph/ceph/pull/23493
- 09:41 AM Backport #24332 (Resolved): mimic: local_reserver double-reservation of backfilled pg
- https://github.com/ceph/ceph/pull/22559
- 08:26 AM Feature #24231: librbd/libcephfs/librgw should ignore rados_mon/osd_op_timeouts options (requires...
- libcephfs doesn't use librados, so it doesn't need any changes.
The rados_mon_op_timeout affects anything that use... - 07:55 AM Bug #20924: osd: leaked Session on osd.7
- https://github.com/ceph/ceph/pull/22292 might address this issue.
- 07:37 AM Bug #24327 (Need More Info): osd: segv in pg_log_entry_t::encode()
- The affected osd restarted itself and everything seems fine then.But what is the cause of the crash?...
- 06:37 AM Backport #24204 (In Progress): mimic: LibRadosMiscPool.PoolCreationRace segv
- https://github.com/ceph/ceph/pull/22291
- 06:20 AM Backport #24216 (In Progress): luminous: "process (unknown)" in ceph logs
- https://github.com/ceph/ceph/pull/22290
- 03:32 AM Bug #24321: assert manager.get_num_active_clean() == pg_num on rados/singleton/all/max-pg-per-osd...
- mimic: https://github.com/ceph/ceph/pull/22288
- 03:31 AM Bug #24321 (Pending Backport): assert manager.get_num_active_clean() == pg_num on rados/singleton...
05/28/2018
- 10:54 PM Feature #24176: osd: add command to drop OSD cache
- Anyone looking into this? If not, I can pick it up.
- 03:21 PM Bug #24145 (Duplicate): osdmap decode error in rados/standalone/*
- 03:19 PM Bug #23492: Abort in OSDMap::decode() during qa/standalone/erasure-code/test-erasure-eio.sh
- /a/kchai-2018-05-28_09:21:54-rados-wip-kefu-testing-2018-05-28-1113-distro-basic-smithi/2601187
on mimic branch.
... - 11:51 AM Bug #24321 (Fix Under Review): assert manager.get_num_active_clean() == pg_num on rados/singleton...
- https://github.com/ceph/ceph/pull/22275
- 05:28 AM Bug #23352: osd: segfaults under normal operation
- I've confirmed that in all of the SafeTimer segfaults the 'schedule' multimap is empty, indicating this is the last e...
- 05:16 AM Bug #23352: osd: segfaults under normal operation
- If we look at the coredump from 23585 and compare it to this message.
[117735.930255] safe_timer[52573]: segfault ... - 04:32 AM Bug #24023 (Duplicate): Segfault on OSD in 12.2.5
- Duplicate of 23352
- 04:30 AM Bug #23564 (Duplicate): OSD Segfaults
- Duplicate of 23352
- 04:28 AM Bug #23585 (Duplicate): osd: safe_timer segfault
- Duplicate of 23352
- 02:47 AM Bug #24160: Monitor down when large store data needs to compact triggered by ceph tell mon.xx com...
- PR :
https://github.com/ceph/ceph/pull/22056/
Also available in: Atom