Project

General

Profile

Activity

From 12/17/2017 to 01/15/2018

01/15/2018

07:26 PM Feature #22442: ceph daemon mon.id mon_status -> ceph daemon mon.id status
Joao, did mon_status just precede the other status commands, or was there a reason for them to be different? Greg Farnum
07:22 PM Bug #22486: ceph shows wrong MAX AVAIL with hybrid (chooseleaf firstn 1, chooseleaf firstn -1) CR...
Well, the hybrid ruleset isn't giving you as much host isolation as you're probably thinking, since it can select an ... Greg Farnum
07:11 PM Support #22664 (Closed): some random OSD are down (with a Abort signal on exception) after replac...
It's failing to create a new thread. You probably need to bump the ulimit; this is discussed in the documentation. :) Greg Farnum
07:08 PM Support #22680: mons segmentation faults New 12.2.2 cluster
This is buried in the depths of RocksDB doing IO, so the only causes I know of/can think of are
1) you've found an u...
Greg Farnum
10:39 AM Support #22680 (Resolved): mons segmentation faults New 12.2.2 cluster

Hi all,
I installed a new Luminous 12.2.2 cluster. The monitors were up at first, but quickly started failing, s...
Kenneth Waegeman
05:48 PM Backport #22387: luminous: PG stuck in recovery_unfound
Include commit 64047e1 "osd: Don't start recovery for missing until active pg state set" from https://github.com/ceph... David Zafman
11:00 AM Support #22531: OSD flapping under repair/scrub after recieve inconsistent PG LFNIndex.cc: 439: F...
Josh Durgin wrote:
> Can you provide a directory listing for pg 1.f? It seems a file that does not obey the internal...
Jan Michlik
06:12 AM Bug #22351: Couldn't init storage provider (RADOS)
Brad Hubbard wrote:
> If this is a RADOS function returning ERANGE (34) then it should be possible to find it by att...
Amine Liu
05:05 AM Bug #22351: Couldn't init storage provider (RADOS)
If this is a RADOS function returning ERANGE (34) then it should be possible to find it by attempting to start the ra... Brad Hubbard
03:26 AM Bug #20059 (Fix Under Review): miscounting degraded objects
David Zafman
02:56 AM Bug #22668: osd/ExtentCache.h: 371: FAILED assert(tid == 0)
/a//kchai-2018-01-11_06:11:31-rados-wip-kefu-testing-2018-01-11-1036-distro-basic-mira/2058373/remote/mira002/log/cep... Kefu Chai

01/14/2018

10:46 PM Bug #22672: OSDs frequently segfault in PrimaryLogPG::find_object_context() with empty clone_snap...
To (relatively) stabilise the frequently crashing OSDs, we've added an early -ENOENT return to PrimaryLogPG::find_obj... David Disseldorp
04:37 PM Bug #22677: rados/test_rados_tool.sh failure
https://github.com/ceph/ceph/pull/19946 Sage Weil

01/13/2018

03:54 PM Bug #22677 (Resolved): rados/test_rados_tool.sh failure
... Sage Weil

01/12/2018

10:43 PM Bug #22438 (Resolved): mon: leak in lttng dlopen / __tracepoints__init
Patrick Donnelly
06:29 AM Bug #22438: mon: leak in lttng dlopen / __tracepoints__init
https://github.com/ceph/teuthology/pull/1144 Kefu Chai
10:23 PM Bug #22672: OSDs frequently segfault in PrimaryLogPG::find_object_context() with empty clone_snap...
That looks like a good way to investigate. We've seen a few reports of issues with cache tier snapshots since that re... Greg Farnum
02:54 PM Bug #22672: OSDs frequently segfault in PrimaryLogPG::find_object_context() with empty clone_snap...
to detect this case during scrub, I'm currently testing the following change:
-https://github.com/ddiss/ceph/commit/...
David Disseldorp
12:55 AM Bug #22672 (Triaged): OSDs frequently segfault in PrimaryLogPG::find_object_context() with empty ...
Environment is a Luminous cache-tiered deployment with some of the hot-tier OSDs converted to bluestore. The remainin... David Disseldorp
07:38 PM Bug #22063: "RadosModel.h: 1703: FAILED assert(!version || comp->get_version64() == version)" inr...
Also in http://qa-proxy.ceph.com/teuthology/teuthology-2017-11-17_18:17:24-rados-jewel-distro-basic-smithi/1857527/te... David Zafman
07:36 PM Bug #22064: "RadosModel.h: 865: FAILED assert(0)" in rados-jewel-distro-basic-smithi
Yuri Weinstein wrote:
> Also in http://qa-proxy.ceph.com/teuthology/teuthology-2017-11-17_18:17:24-rados-jewel-distr...
David Zafman
07:18 PM Bug #22064: "RadosModel.h: 865: FAILED assert(0)" in rados-jewel-distro-basic-smithi
As 17815 has to do with when scrub is allowed to start, it wouldn't be related to this bug. David Zafman
01:03 PM Bug #22673 (Resolved): osd checks out-of-date osdmap for DESTROYED flag on start
When trying an in-place migration of a filestore to bluestore OSD, we encountered a situation where ceph-osd would re... J Mozdzen
07:45 AM Bug #22624: filestore: 3180: FAILED assert(0 == "unexpected error"): error (2) No such file or di...
i am rerunning the failed test at http://pulpito.ceph.com/kchai-2018-01-12_07:44:06-multimds-wip-pdonnell-testing-201... Kefu Chai
07:29 AM Bug #22624: filestore: 3180: FAILED assert(0 == "unexpected error"): error (2) No such file or di...
i agree it's a bug in osd. but i don't think osd should return -ENOENT in this case. as Sage pointed out, it should c... Kefu Chai
01:15 AM Bug #22351: Couldn't init storage provider (RADOS)
Abhishek Lekshmanan wrote:
> can you tell us the ceph pg num and pgp num setting in ceph.conf (or rather paste teh c...
Amine Liu

01/11/2018

09:43 PM Bug #22668 (Resolved): osd/ExtentCache.h: 371: FAILED assert(tid == 0)
... Sage Weil
06:52 PM Bug #22351: Couldn't init storage provider (RADOS)
can you tell us the ceph pg num and pgp num setting in ceph.conf (or rather paste teh ceph.conf retracting sensitive ... Abhishek Lekshmanan
04:05 PM Bug #22561: PG stuck during recovery, requires OSD restart
OSD 32 was running and actively serving client IO. Paul Emmerich
02:39 PM Support #22664 (Closed): some random OSD are down (with a Abort signal on exception) after replac...
Hello,
currently we are facing with a strange behavior, where some OSDs are got ramdomly down with a Abort signal,...
Enrico Labedzki
12:57 PM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
Recovery from non starting OSDs in this case is as following. Run OSD with debug:... Zdenek Janda
10:55 AM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
Also several osds (as you can see the ceph osd tree output) are getting dumped out of the crush map. After putting th... Michal Cila
10:44 AM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
More info on affected PG... Zdenek Janda
10:39 AM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
I have succeeded in identifying faulty PG:... Zdenek Janda
10:17 AM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
Adding last 10000 lines of strace of OSD affected by the bug.
The ABRT signal is generated right after ...
Zdenek Janda
09:45 AM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
also adding our current ceph -s/ceph osd tree state:... Josef Zelenka
09:44 AM Bug #21142: OSD crashes when loading pgs with "FAILED assert(interval.last > last)"
we are also affected by this bug. we are running luminous 12.2.2 on ubuntu 16.04, 3 node cluster, 8 HDDs per node, bl... Josef Zelenka
10:30 AM Bug #22662 (Resolved): ceph osd df json output validation reported invalid numbers (-nan) (jewel)
Hi,
we have a monitoring script which parses the 'ceph osd df -f json' output, but from time to time it will happe...
Enrico Labedzki
08:36 AM Bug #22661 (Triaged): Segmentation fault occurs when the following CLI is executed
Observation:
--------------
It is observed that when a user executes the CLI without providing the value of osd-u...
Debashis Mondal
07:34 AM Bug #22659 (In Progress): During the cache tiering configuration ,ceph-mon daemon getting crashed...
Observation:
--------------
Before setting the value of "hit_set_count" Ceph health was OK but after configuring th...
Debashis Mondal
02:54 AM Bug #22624: filestore: 3180: FAILED assert(0 == "unexpected error"): error (2) No such file or di...
OSD should reply -ENOENT for that case. should be OSD bug Zheng Yan

01/10/2018

11:38 PM Bug #22351: Couldn't init storage provider (RADOS)
Related to the ERROR: failed to initialize watch: (34) Numerical result out of range, it looks a class path issue. Th... Javier M. Mellid
11:38 PM Backport #22658 (In Progress): filestore: randomize split threshold
Josh Durgin
10:39 PM Backport #22658 (Resolved): filestore: randomize split threshold
https://github.com/ceph/ceph/pull/19906 Josh Durgin
10:16 PM Feature #15835 (Pending Backport): filestore: randomize split threshold
Josh Durgin
10:03 PM Support #22531: OSD flapping under repair/scrub after recieve inconsistent PG LFNIndex.cc: 439: F...
Can you provide a directory listing for pg 1.f? It seems a file that does not obey the internal naming rules of files... Josh Durgin
09:48 PM Bug #22561: PG stuck during recovery, requires OSD restart
Was OSD 32 running at the time? It sounds like correct behavior if OSD 32 was not reachable. It might have been marke... Josh Durgin
09:44 PM Support #22566: Some osd remain 100% CPU after upgrade jewel => luminous (v12.2.2) and some work
This is likely the singe-time startup cost of accounting for a bug in omap, where the osd has to scan the whole omap ... Josh Durgin
09:39 PM Bug #22597: "sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-0'" fails in upgrade test
IIRC we didn't have the ceph user in hammer - need to account for that in the suite if we want to keep running it at ... Josh Durgin
09:36 PM Bug #22641 (Resolved): uninit condition in PrimaryLogPG::process_copy_chunk_manifest
Josh Durgin
09:22 PM Bug #22641: uninit condition in PrimaryLogPG::process_copy_chunk_manifest
myoungwon oh wrote:
> https://github.com/ceph/ceph/pull/19874
merged
Yuri Weinstein
09:22 PM Bug #22656 (New): scrub mismatch on bytes (cache pools)
... Sage Weil
09:21 PM Bug #21557: osd.6 found snap mapper error on pg 2.0 oid 2:0e781f33:::smithi14431805-379 ... :187 ...
/a/yuriw-2018-01-09_21:50:35-rados-wip-yuri2-testing-2018-01-09-1813-distro-basic-smithi/2050823
another one.
<...
Sage Weil
09:01 PM Bug #20086: LibRadosLockECPP.LockSharedDurPP gets EEXIST
/a/yuriw-2018-01-09_21:50:35-rados-wip-yuri2-testing-2018-01-09-1813-distro-basic-smithi/2050802
Sage Weil
03:34 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
https://github.com/ceph/ceph/pull/19759 Kefu Chai
03:33 PM Bug #22539 (Pending Backport): bluestore: New OSD - Caught signal - bstore_kv_sync
Kefu Chai
02:56 PM Bug #22624: filestore: 3180: FAILED assert(0 == "unexpected error"): error (2) No such file or di...
That would be an fs bug, sure.
However, shouldn't the OSD not assert due to an object not existing?
Patrick Donnelly
02:48 PM Bug #22624: filestore: 3180: FAILED assert(0 == "unexpected error"): error (2) No such file or di...
I think the problem here is that the object doesn't exist but we're doing omap_setkeys on it.. which doesn't implicit... Sage Weil
08:57 AM Bug #22438 (Fix Under Review): mon: leak in lttng dlopen / __tracepoints__init
https://github.com/ceph/teuthology/pull/1143 Kefu Chai
08:16 AM Bug #22525 (Fix Under Review): auth: ceph auth add does not sanity-check caps
Jos Collin

01/09/2018

10:39 PM Bug #22064: "RadosModel.h: 865: FAILED assert(0)" in rados-jewel-distro-basic-smithi
Actually, I may have seen an instance of the failure in a run that did not include 17815, so please don't take what I... Nathan Cutler
05:49 PM Bug #21557: osd.6 found snap mapper error on pg 2.0 oid 2:0e781f33:::smithi14431805-379 ... :187 ...
Not 100% sure if that's the same issue but we have a customer who faces an assert in SnapMapper::get_snaps()
2018-01...
Igor Fedotov
04:02 PM Bug #22641: uninit condition in PrimaryLogPG::process_copy_chunk_manifest
https://github.com/ceph/ceph/pull/19874 Myoungwon Oh
02:43 PM Bug #22641 (Resolved): uninit condition in PrimaryLogPG::process_copy_chunk_manifest
... Sage Weil
03:54 PM Bug #22278: FreeBSD fails to build with WITH_SPDK=ON
patch merged in DPDK. waiting for SPDK to pick up the latest DPDK. Kefu Chai
03:49 PM Support #22520 (Closed): nearfull threshold is not cleared when osd really is not nearfull.
You need to change this in the osd map, not the config. "ceph osd set-nearfull-ratio" or something similar. Greg Farnum
02:59 PM Bug #22409 (Resolved): ceph_objectstore_tool: no flush before collection_empty() calls; ObjectSto...
Kefu Chai
01:52 AM Bug #22351: Couldn't init storage provider (RADOS)
Orit Wasserman wrote:
> what is your pool configuration?
all default, just a default pool 'rbd'.
Amine Liu

01/08/2018

11:54 PM Bug #22624 (Duplicate): filestore: 3180: FAILED assert(0 == "unexpected error"): error (2) No suc...
... Patrick Donnelly
12:35 PM Bug #22409 (Fix Under Review): ceph_objectstore_tool: no flush before collection_empty() calls; O...
Igor Fedotov
12:35 PM Bug #22409: ceph_objectstore_tool: no flush before collection_empty() calls; ObjectStore/StoreTes...
https://github.com/ceph/ceph/pull/19764 Igor Fedotov
08:21 AM Bug #22409: ceph_objectstore_tool: no flush before collection_empty() calls; ObjectStore/StoreTes...
sage, i am taking this ticket from you. as it's simple enough and it won't cause too much duplication of efforts.
...
Kefu Chai
07:22 AM Bug #22415 (Duplicate): 'pg dump' fails after mon rebuild
Kefu Chai

01/06/2018

01:29 AM Bug #22220: osd/ReplicatedPG.h:1667:14: internal compiler error: in force_type_die, at dwarf2out....
For DTS this should be fixed in the 7.1 release. Brad Hubbard
12:35 AM Bug #20439: PG never finishes getting created
Same thing in http://pulpito.ceph.com/yuriw-2018-01-04_20:43:14-rados-wip-yuri4-testing-2018-01-04-1750-distro-basic-... Josh Durgin

01/05/2018

03:57 PM Bug #22597 (Resolved): "sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-0'" fails in upgrade test
http://pulpito.ceph.com/kchai-2018-01-05_15:34:38-upgrade-wip-kefu-testing-2018-01-04-1836-distro-basic-mira/
<pre...
Kefu Chai
09:51 AM Bug #22525: auth: ceph auth add does not sanity-check caps
-https://github.com/ceph/ceph/pull/19794- Jing Li

01/04/2018

07:13 PM Bug #22351 (Need More Info): Couldn't init storage provider (RADOS)
what is your pool configuration? Orit Wasserman
02:42 PM Bug #22354: v12.2.2 unable to create bluestore osd using ceph-disk
So my OSDs had the default Bluestore layout the first time around, i.e. a 100MB DB/WAL (xfs) partition followed by th... Jon Heese
07:06 AM Bug #22354: v12.2.2 unable to create bluestore osd using ceph-disk
Jon Heese wrote:
> Unfortunately, `ceph-disk zap /dev/sde` does not wipe enough of the disk to avoid this issue. As...
Hua Liu
02:35 PM Bug #22266 (Pending Backport): mgr/PyModuleRegistry.cc: 139: FAILED assert(map.epoch > 0)
Kefu Chai
02:32 PM Bug #22266 (Resolved): mgr/PyModuleRegistry.cc: 139: FAILED assert(map.epoch > 0)
Sage Weil
01:23 PM Bug #22266 (Fix Under Review): mgr/PyModuleRegistry.cc: 139: FAILED assert(map.epoch > 0)
http://tracker.ceph.com/issues/22266 Kefu Chai
01:52 PM Support #22566 (New): Some osd remain 100% CPU after upgrade jewel => luminous (v12.2.2) and some...
h1. I have some OSDs that remain at 100% startup without any debug info in the logs :... David Casier
07:12 AM Support #22422: Block fsid does not match our fsid
See, [[http://tracker.ceph.com/issues/22354]] Hua Liu
01:07 AM Bug #22561 (New): PG stuck during recovery, requires OSD restart
We are sometimes encountering issues with PGs getting stuck in recovery.
For example, we ran some stress tests wit...
Paul Emmerich

01/03/2018

09:28 PM Bug #22064: "RadosModel.h: 865: FAILED assert(0)" in rados-jewel-distro-basic-smithi
So Nathan seems to have narrowed it down to https://github.com/ceph/ceph/pull/17815 - can you look at this when you'r... Josh Durgin
09:23 PM Support #22422: Block fsid does not match our fsid
It looks like you may have had a partial prepare there in the past - if you're sure it's the right disk, wipe it with... Josh Durgin
09:22 PM Bug #22438 (Resolved): mon: leak in lttng dlopen / __tracepoints__init
Josh Durgin
09:17 PM Support #22466 (Closed): PG failing to map to any OSDs
Josh Durgin
09:08 PM Support #22553: ceph-object-tool can not remove metadata pool's object
Is there possibly something wrong with that disk? Josh Durgin
03:28 PM Bug #22354: v12.2.2 unable to create bluestore osd using ceph-disk
Jon Heese wrote:
> Unfortunately, `ceph-disk zap /dev/sde` does not wipe enough of the disk to avoid this issue. As...
Curt Bruns
01:41 AM Bug #22346 (Resolved): OSD_ORPHAN issues after jewel->luminous upgrade, but orphaned osds not in ...
Not for me.
$ crushtool -d crushmap.bad -o crushmap.bad.txt
$ crushtool -d crushmap.good -o crushmap.good.txt
$ ...
Brad Hubbard

01/02/2018

09:03 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
Alright, that fixed it!
It also fixed the heavy IO issue as well as the rather large amount of consumption I was s...
Brian Woods
06:20 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
Sorry for the spam.
That broke it good!!!...
Brian Woods
06:15 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
Was able to out them all:... Brian Woods
06:14 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
I can't mark the OSDs out.... Brian Woods
03:42 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
Hard to say excatly, but I would not be surprised to see any manner of odd behaviors with a huge map like that--we ha... Sage Weil
04:28 PM Bug #22354: v12.2.2 unable to create bluestore osd using ceph-disk
Unfortunately, `ceph-disk zap /dev/sde` does not wipe enough of the disk to avoid this issue. As I mentioned above, ... Jon Heese
01:01 PM Support #22553 (New): ceph-object-tool can not remove metadata pool's object
i put an object to the rbd pool
rados -p rbd put qinli.sh
then stop osd and remove it
[root@lab71 ~]# ceph-objec...
peng zhang

12/31/2017

11:13 PM Bug #21388: inconsistent pg but repair does nothing reporting head data_digest != data_digest fro...
I'm working on fixing all my inconsistent pgs but I'm having issues with rados get... hopefully I'm just doing the co... Ryan Anstey

12/30/2017

02:30 AM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
I had no idea the ID would impact the map calculations that way (makes sense now)!!! Very good to know! And those I... Brian Woods

12/29/2017

10:34 PM Bug #22539 (In Progress): bluestore: New OSD - Caught signal - bstore_kv_sync
Brian, note that one reason why this triggered is that your osdmap is huge... because you have some osds with very la... Sage Weil

12/28/2017

11:02 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
I'm a bit lost hence trying to re-arrange things:
Let's handle the crash first.
IMO it's caused by throttle value...
Igor Fedotov

12/27/2017

04:46 AM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
A chunk from the mon log:
https://pastebin.com/MA1BStEc
Some screenshots of the IO:
https://imgur.com/a/BOKWc
...
Brian Woods
04:29 AM Bug #22544 (Resolved): objecter cannot resend split-dropped op when racing with con reset
@
if (split && con && con->has_features(CEPH_FEATUREMASK_RESEND_ON_SPLIT)) {
return RECALC_OP_TARGET_NEED_RES...
mingxin liu

12/26/2017

11:03 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
UI Lag seems to be related to heavy load to the OS SSD from the monitor services. The monitor service does a lot of I... Brian Woods
10:51 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
Edit, UI is lagging again. But its odd. SOME things lag, but GLXGears isn't. IO blocking of some sort? Adding mor... Brian Woods
10:47 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
Confirmed the line was there. Added the extra debug line, but this time when I started it is came right online (almo... Brian Woods
09:18 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
Given the object names in action it looks like that's osd map update or something that triggers the issue. Not the us... Igor Fedotov
06:41 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
Additional note, there is no data on the cluster other than the built in pools. So there is very little information ... Brian Woods
05:46 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
This will make only the fourth OSD in the cluster. Would that impact the overflowed value? What can I do to capture... Brian Woods
01:22 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
As a workaround one can try to set (temporarily until initial rebuild completes?) bluestore_throttle_bytes = 0 at the... Igor Fedotov
01:07 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
32-bit value in throttle_bytes is overflowed - see:
2017-12-25 13:18:06.783304 7f37a7a2a700 10 bluestore(/var/lib/ce...
Igor Fedotov

12/25/2017

09:27 PM Bug #22539: bluestore: New OSD - Caught signal - bstore_kv_sync
Added to ceph.conf:
debug bluestore = 20
debug osd = 20
Waited for crash, captured log, but its too large even c...
Brian Woods
08:46 PM Bug #22539 (Resolved): bluestore: New OSD - Caught signal - bstore_kv_sync
After rebuilding a demo cluster, OSD on one node can no longer be created.
Looking though the log I see this error...
Brian Woods
03:24 AM Support #22466: PG failing to map to any OSDs
When delete the osds outside of the default root , the problem's solved. Amine Liu

12/22/2017

02:15 PM Support #22531 (New): OSD flapping under repair/scrub after recieve inconsistent PG LFNIndex.cc: ...
Hi.
I have a problem when repairing PG 1.f under copy from OSD.3 on OSD.0. During the upgrade to 12.2.2, all OSDs we...
Jan Michlik
01:11 PM Bug #21262: cephfs ec data pool, many osds marked down
This looks like a Support Case rather than a Tracker Bug. Jos Collin
09:58 AM Bug #22530: pool create cmd's expected_num_objects is not correctly interpreted
fix: https://github.com/ceph/ceph/pull/19651... Honggang Yang
09:52 AM Bug #22530 (Resolved): pool create cmd's expected_num_objects is not correctly interpreted
1. disable merge... Honggang Yang
09:11 AM Bug #22354: v12.2.2 unable to create bluestore osd using ceph-disk
The problem of "ceph-disk activation issue in 12.2.2" has been catched ,It can be solved by this:
1. delete osd
2...
Hua Liu
05:33 AM Bug #21388: inconsistent pg but repair does nothing reporting head data_digest != data_digest fro...
I'm also having this issue. I'm getting new scrub errors every few days. No idea what's going on. This is something n... Ryan Anstey

12/21/2017

04:54 PM Bug #22346: OSD_ORPHAN issues after jewel->luminous upgrade, but orphaned osds not in crushmap
That did clean it up, thanks.
It is curious though that if I decompile the crushmap to text, it appears the same b...
Graham Allan
04:12 PM Feature #22528 (New): objects should not be promoted when locked
Hello,
We faced with immediate object promotion when call lock on object.
This behavior makes very hard to understa...
Aleksei Zakharov
02:57 PM Bug #21557: osd.6 found snap mapper error on pg 2.0 oid 2:0e781f33:::smithi14431805-379 ... :187 ...
/a/sage-2017-12-21_07:24:12-rados-wip-sage3-testing-2017-12-20-2253-distro-basic-smithi/1989672
but didn't have th...
Sage Weil
02:11 PM Bug #18698: BlueFS FAILED assert(0 == "allocate failed... wtf")
There is a PR#18494 addressing an issue with the symptoms similar to ones reported in comment #9 (assert during _bala... Igor Fedotov
12:50 PM Bug #22525 (Resolved): auth: ceph auth add does not sanity-check caps
When adding a keyring with "ceph auth add -i <keyring> <entity>", it does not verify that the contained capability st... Fabian Vogt
11:42 AM Support #22520: nearfull threshold is not cleared when osd really is not nearfull.
When I was delete some data from this osds, nearfull flag was also deleted.... Konstantin Shalygin
10:56 AM Support #22520 (Closed): nearfull threshold is not cleared when osd really is not nearfull.
Today one of my osd is reached nearfull ratio. mon_osd_nearfull_ratio: '.85'. I increased mon_osd_nearfull_ratio to '... Konstantin Shalygin
10:10 AM Backport #22502 (In Progress): luminous: Pool Compression type option doesn't apply to new OSD's
-https://github.com/ceph/ceph/pull/19629- Shinobu Kinjo

12/20/2017

08:59 PM Bug #22415: 'pg dump' fails after mon rebuild
/a/yuriw-2017-12-19_20:36:31-rados-wip-yuri4-testing-2017-12-19-1722-distro-basic-smithi/1980900 Sage Weil
08:58 PM Bug #22515 (Resolved): osd-config.sh fails with /usr/bin/ceph-authtool: unexpected '1000'
https://github.com/ceph/ceph/pull/19544 Sage Weil
07:56 PM Bug #22515 (Resolved): osd-config.sh fails with /usr/bin/ceph-authtool: unexpected '1000'
... Sage Weil
08:24 PM Bug #22408: objecter: sent out of order ops
/a/yuriw-2017-12-19_20:40:29-rbd-wip-yuri4-testing-2017-12-19-1722-distro-basic-smithi/1981037
rbd/basic/{base/ins...
Sage Weil
08:19 PM Bug #22369 (Resolved): out of order reply on set-chunks.yaml workload
Sage Weil
01:11 PM Bug #18698: BlueFS FAILED assert(0 == "allocate failed... wtf")
Reproduced with 12.2.2 during deep scrubbing after 7 days of workload.
36 ssds, 200G each contain 400G of rbds and...
Aleksei Gutikov
11:54 AM Backport #22502 (Resolved): luminous: Pool Compression type option doesn't apply to new OSD's
https://github.com/ceph/ceph/pull/20106 Nathan Cutler

12/19/2017

09:18 PM Bug #22486: ceph shows wrong MAX AVAIL with hybrid (chooseleaf firstn 1, chooseleaf firstn -1) CR...
Forgot to put the output in code tags, sadly I can't edit the original, so here it is again to make it more readable:... Patrick Fruh
09:14 PM Bug #22486 (New): ceph shows wrong MAX AVAIL with hybrid (chooseleaf firstn 1, chooseleaf firstn ...
I have the following configuration of OSDs:
ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS
0 hdd 5...
Patrick Fruh
03:35 PM Bug #22266: mgr/PyModuleRegistry.cc: 139: FAILED assert(map.epoch > 0)
/a/sage-2017-12-19_06:01:05-rados-wip-sage2-testing-2017-12-18-2147-distro-basic-smithi/1979661
saw this again on ...
Sage Weil
12:18 PM Bug #22445: ceph osd metadata reports wrong "back_iface"
Hmm, this could well be the first time anyone's really tested the IPv6 path here. John Spray
11:56 AM Support #22466: PG failing to map to any OSDs
More info will be needed to work out if this is a bug -- are the CRUSH rules customized? What is the topology ("ceph... John Spray
03:29 AM Bug #21218: thrash-eio + bluestore (hangs with unfound objects or read_log_and_missing assert)
/a/sage-2017-12-18_22:56:18-rados-wip-sage-testing-2017-12-18-1406-distro-basic-smithi/1976871
description: rados/si...
Sage Weil

12/18/2017

07:07 AM Bug #22468 (New): unblock backoff contend with cancel proxy write lead to out of order
1.cache primary send several proxy write to base primary
2.base pg haven't peered, backoff these op
3.base finish p...
mingxin liu
01:48 AM Support #22466 (Closed): PG failing to map to any OSDs

osdmap e88997 pg 9.d07 (9.d07) -> up [] acting []
health HEALTH_ERR
319 pgs are stuck inacti...
Amine Liu

12/17/2017

04:10 AM Backport #22450: luminous: Visibility for snap trim queue length
Nathan Cutler wrote:
> Shinobu Kinjo wrote:
> > unmerged pr can't be cherry-picked anyway...
>
> Actually, it ca...
Shinobu Kinjo
04:08 AM Backport #22450: luminous: Visibility for snap trim queue length
Shinobu Kinjo wrote:
> unmerged pr can't be cherry-picked anyway...
Actually, it can, but we definitely don't wan...
Nathan Cutler
 

Also available in: Atom