Project

General

Profile

Activity

From 07/21/2021 to 08/19/2021

08/19/2021

10:48 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
I don't have the logs right now but it prints the state of the PG so if you search for`snaptrim` in `f0208568-fbf4-48... Arthur Outhenin-Chalandre
08:48 PM Bug #52026 (New): osd: pgs went back into snaptrim state after osd restart
Thanks for providing the logs, is there a particular PG we should look at the in the logs? Neha Ojha
09:19 PM Bug #50441: cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
I assume because of MB's comment, but that seems now to be historical Dan Mick
09:17 PM Bug #50441: cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
Deepika: why is this issue in need-more-info? Looks like the original fix and pacific backport https://github.com/cep... Neha Ojha
09:12 PM Bug #48844 (Duplicate): api_watch_notify: LibRadosWatchNotify.AioWatchDelete failed
Neha Ojha
09:08 PM Bug #52261 (Need More Info): OSD takes all memory and crashes, after pg_num increase
Neha Ojha
09:08 PM Bug #52255 (Need More Info): The pgs state are degraded, but all the osds is up and there is no r...
Neha Ojha
09:08 PM Bug #52255: The pgs state are degraded, but all the osds is up and there is no recovering and bac...
can you share your osdmap? are all your osds up and in? the crushmap looks fine. Neha Ojha
08:54 PM Bug #52319: LibRadosWatchNotify.WatchNotify2 fails
Brad, are you aware of this one? Neha Ojha
03:54 AM Bug #52319 (New): LibRadosWatchNotify.WatchNotify2 fails
2021-08-17T01:34:43.023 INFO:tasks.workunit.client.0.smithi111.stdout: api_watch_notify: [ RUN ] LibRado... Aishwarya Mathuria
08:51 PM Bug #52136: Valgrind reports memory "Leak_DefinitelyLost" errors.
Adam Kupczyk wrote:
> This leak is from internals of RocksDB.
> We have no access to FileMetaData objects, we canno...
Neha Ojha
07:34 AM Bug #52136: Valgrind reports memory "Leak_DefinitelyLost" errors.
This leak is from internals of RocksDB.
We have no access to FileMetaData objects, we cannot be responsible for this...
Adam Kupczyk
08:48 PM Backport #51549: pacific: cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
Deepika Upadhyay wrote:
> https://github.com/ceph/ceph/pull/42211
merged
Yuri Weinstein
08:45 PM Bug #50659: Segmentation fault under Pacific 16.2.1 when using a custom crush location hook
Adam, can you start talking a look at this? Neha Ojha
03:24 PM Fix #52329 (Fix Under Review): src/vstart: The command "set config key osd_mclock_max_capacity_io...
Sridhar Seshasayee
02:28 PM Fix #52329 (Resolved): src/vstart: The command "set config key osd_mclock_max_capacity_iops_ssd" ...
The following was observed when bringing up a vstart cluster:... Sridhar Seshasayee
07:45 AM Backport #52322 (Resolved): pacific: LibRadosTwoPoolsPP.ManifestSnapRefcount failure
https://github.com/ceph/ceph/pull/43306 Backport Bot
07:42 AM Bug #51000 (Pending Backport): LibRadosTwoPoolsPP.ManifestSnapRefcount failure
Kefu Chai
04:47 AM Bug #51076: "wait_for_recovery: failed before timeout expired" during thrashosd test with EC back...
I see the same assertion error in this dead job - https://pulpito.ceph.com/yuriw-2021-08-16_21:15:00-rados-wip-yuri-t... Aishwarya Mathuria

08/18/2021

11:19 PM Backport #51569 (In Progress): octopus: pool last_epoch_clean floor is stuck after pg merging
Neha Ojha
09:03 PM Backport #51569: octopus: pool last_epoch_clean floor is stuck after pg merging
https://github.com/ceph/ceph/pull/42837 Steve Taylor
09:53 PM Bug #52316: qa/tasks/mon_thrash.py: _do_thrash AssertionError len(s['quorum']) == len(mons)
... Neha Ojha
07:18 PM Bug #52316 (Resolved): qa/tasks/mon_thrash.py: _do_thrash AssertionError len(s['quorum']) == len(...
2021-08-17T03:12:45.055 INFO:tasks.workunit.client.0.smithi135.stderr:2021-08-17T03:12:45.052+0000 7f27d941a700 1 --... Aishwarya Mathuria
03:50 AM Backport #52307 (Resolved): pacific: doc: clarify use of `rados rm` command
https://github.com/ceph/ceph/pull/51260 Backport Bot
03:50 AM Backport #52306 (Rejected): octopus: doc: clarify use of `rados rm` command
Backport Bot
03:47 AM Bug #52288 (Pending Backport): doc: clarify use of `rados rm` command
Kefu Chai

08/17/2021

04:40 PM Bug #52012 (Fix Under Review): osd/scrub: src/osd/scrub_machine.cc: 55: FAILED ceph_assert(state_...
Neha Ojha
01:35 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
I searched a bit through the log I sent and I don't see any traces of a pg into the snaptrim state, probably because ... Arthur Outhenin-Chalandre
07:12 AM Fix #51116: osd: Run osd bench test to override default max osd capacity for mclock.
Removed the classification of the tracker as a "Feature". This is better classified as a "Fix" with the aim of improv... Sridhar Seshasayee
04:09 AM Bug #52255: The pgs state are degraded, but all the osds is up and there is no recovering and bac...
This is my crushmap Ke Xiao

08/16/2021

11:29 PM Backport #51568: pacific: pool last_epoch_clean floor is stuck after pg merging
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/42224
merged
Yuri Weinstein
10:33 PM Bug #52288 (Resolved): doc: clarify use of `rados rm` command
The man page and the "--help" info for `rados rm ...` could be clearer. J. Eric Ivancich
08:18 PM Bug #52261: OSD takes all memory and crashes, after pg_num increase
Can you attach a 'ceph osd dump' and 'ceph pg dump', plus a log of one of the osds starting leading up to the crash w... Josh Durgin
03:22 PM Bug #52026: osd: pgs went back into snaptrim state after osd restart
I reproduced the issue by doing a `ceph pg repeer` on a pg with a non-zero snaptrimq_len. After the pg has been repee... Arthur Outhenin-Chalandre

08/15/2021

08:16 AM Bug #52261 (Need More Info): OSD takes all memory and crashes, after pg_num increase
After increasing a pool pg_num from 256 to 512, all osds are down.
On startup, they take all of the memory. After ...
Marius Leustean

08/13/2021

03:12 AM Bug #52255 (Need More Info): The pgs state are degraded, but all the osds is up and there is no r...
I removed a server yesterday, but there are 6 pgs are in stare degraded and no longer changed.
The copy size of pool...
Ke Xiao

08/11/2021

06:47 PM Bug #52233 (New): crash: void Infiniband::init(): assert(device)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=184ea175092db1eb5f584b66...
Telemetry Bot
06:47 PM Bug #52231 (New): crash: std::_Rb_tree<std::__cxx11::basic_string<char, std::char_traits<char>, s...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=1aea506f1109fd768e765158...
Telemetry Bot
06:47 PM Bug #52226 (New): crash: PosixNetworkStack::spawn_worker(unsigned int, std::function<void ()>&&)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=4bf2d022677b1bd10586cef6...
Telemetry Bot
06:47 PM Bug #52225 (New): crash: void Thread::create(const char*, size_t): assert(ret == 0)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=77077c11d9fa7cd7f8d4ccaa...
Telemetry Bot
06:47 PM Bug #52221 (Triaged): crash: void OSD::handle_osd_map(MOSDMap*): assert(p != added_maps_bl.end())

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=9ee7dc6ce5b80b3a4a423d80...
Telemetry Bot
06:47 PM Bug #52220 (New): crash: void ECUtil::HashInfo::append(uint64_t, std::map<int, ceph::buffer::v15_...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=9541d850892a4b0e1e7d3cce...
Telemetry Bot
06:47 PM Bug #52218 (Duplicate): crash: OpTracker::~OpTracker(): assert((sharded_in_flight_list.back())->o...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=886c2848ae642fafcf59efce...
Telemetry Bot
06:47 PM Bug #52217 (Duplicate): crash: OpTracker::~OpTracker(): assert((sharded_in_flight_list.back())->o...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=78325a1cfa85add67a004464...
Telemetry Bot
06:47 PM Bug #52214 (Duplicate): crash: OpTracker::~OpTracker(): assert((sharded_in_flight_list.back())->o...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=637262c9313d56f56724c439...
Telemetry Bot
06:47 PM Bug #52213 (Duplicate): crash: OpTracker::~OpTracker(): assert((sharded_in_flight_list.back())->o...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=7838cb70174ac6ee701615d8...
Telemetry Bot
06:47 PM Bug #52212 (Duplicate): crash: void PeeringState::check_past_interval_bounds() const: abort

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=cb6a35bf8176df5e9719943c...
Telemetry Bot
06:46 PM Bug #52211 (Rejected): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionRe...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=529537d03be27e8fd7c33eb3...
Telemetry Bot
06:46 PM Bug #52210 (Closed): crash: CrushWrapper::decode(ceph::buffer::v15_2_0::list::iterator_impl<true>&)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=16b0cae292a2e5aa1f4a59ae...
Telemetry Bot
06:46 PM Bug #52207 (Duplicate): crash: std::_Rb_tree<std::__cxx11::basic_string<char, std::char_traits<ch...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=e931cd074f4d4c57eafcfbec...
Telemetry Bot
06:46 PM Bug #52200 (Duplicate): crash: void OSD::handle_osd_map(MOSDMap*): assert(p != added_maps_bl.end())

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=8c8aeec7c24f8af53043cb86...
Telemetry Bot
06:46 PM Bug #52199 (Duplicate): crash: virtual Monitor::~Monitor(): assert(session_map.sessions.empty())

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=9a7de784614490d603daf107...
Telemetry Bot
06:46 PM Bug #52198 (Duplicate): crash: virtual Monitor::~Monitor(): assert(session_map.sessions.empty())

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=45aa2b2ae51cb0358e27161c...
Telemetry Bot
06:46 PM Bug #52197 (Rejected): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionRe...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=f5b9d7371888d1a4fef1a569...
Telemetry Bot
06:46 PM Bug #52195 (Duplicate): crash: /lib64/libpthread.so.0(

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=cb65b63c9b9a79458dcd7c3a...
Telemetry Bot
06:46 PM Bug #52194 (New): mon crash in rocksdb::Cleanable::~Cleanable()

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=e2287f3b36d2a97af38026b8...
Telemetry Bot
06:46 PM Bug #52193 (Rejected): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionRe...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=9f138c39ff09273c4f297dd4...
Telemetry Bot
06:46 PM Bug #52192 (Rejected): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionRe...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=3c44fd4fbb924dbf0de4d271...
Telemetry Bot
06:46 PM Bug #52191 (Rejected): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionRe...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=ef2841a5c5ebbac2def49fa6...
Telemetry Bot
06:46 PM Bug #52190 (Rejected): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionRe...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=315779c6f6febaf097208f14...
Telemetry Bot
06:45 PM Bug #52189 (Need More Info): crash in AsyncConnection::maybe_start_delay_thread()

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=f416301151f8db40b0181db8...
Telemetry Bot
06:45 PM Bug #52186 (Duplicate): crash: void OSD::handle_osd_map(MOSDMap*): assert(p != added_maps_bl.end())

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=fc7f92f74bc7bb40c5e03a81...
Telemetry Bot
06:45 PM Bug #52183 (Duplicate): crash: const entity_addrvec_t& MonMap::get_addrs(unsigned int) const: ass...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=1bfa48148eee52e245e1d06f...
Telemetry Bot
06:45 PM Bug #52180 (Duplicate): crash: void pg_missing_set<TrackChanges>::got(const hobject_t&, eversion_...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=f582692869a94580abf07e66...
Telemetry Bot
06:45 PM Bug #52178 (Duplicate): crash: virtual void AuthMonitor::update_from_paxos(bool*): assert(ret == 0)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=7edafdde10f3891aee038aa8...
Telemetry Bot
06:45 PM Bug #52176 (Duplicate): crash: std::_Rb_tree<boost::intrusive_ptr<AsyncConnection>, boost::intrus...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=c653670067a09d0d1578ea33...
Telemetry Bot
06:45 PM Bug #52174 (Triaged): crash: ceph::buffer::v15_2_0::create_aligned_in_mempool(unsigned int, unsig...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=2f0088a1a0259603ad88df29...
Telemetry Bot
06:45 PM Bug #52173 (Need More Info): crash in ProtocolV2::send_message()

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=275bfebdff86cb8d90c56459...
Telemetry Bot
06:44 PM Bug #52172 (Triaged): crash: ceph::buffer::v15_2_0::create_aligned_in_mempool(unsigned int, unsig...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=0a14d6ccc26b531fb346e8c3...
Telemetry Bot
06:44 PM Bug #52171 (Triaged): crash: virtual int RocksDBStore::get(const string&, const string&, ceph::bu...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=b8a9cbd0444778ca4112e187...
Telemetry Bot
06:44 PM Bug #52170 (Duplicate): crash: const entity_addrvec_t& MonMap::get_addrs(unsigned int) const: ass...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=96b49c839d59492286f04a76...
Telemetry Bot
06:44 PM Bug #52169 (New): crash: void SignalHandler::queue_signal_info(int, siginfo_t*, void*): assert(r ...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=88dad7f26832c8036351625c...
Telemetry Bot
06:44 PM Bug #52168 (Duplicate): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionR...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=63ba607fc458e378045e8666...
Telemetry Bot
06:44 PM Bug #52167 (Won't Fix): crash: RDMAConnectedSocketImpl::RDMAConnectedSocketImpl(ceph::common::Cep...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=07ab19a6cb27368fa09313cf...
Telemetry Bot
06:44 PM Bug #52166 (Won't Fix): crash: void Device::binding_port(ceph::common::CephContext*, int): assert...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=8c306905582a9d790b752447...
Telemetry Bot
06:44 PM Bug #52165 (Rejected): crash: void MonitorDBStore::clear(std::set<std::__cxx11::basic_string<char...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=9d5db03ed7874482b2960b7d...
Telemetry Bot
06:44 PM Bug #52164 (Duplicate): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionR...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=4d2c4ced5cb129282e81fdf9...
Telemetry Bot
06:44 PM Bug #52163 (Rejected): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionRe...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=eb081141be75e0baa7c9fe0a...
Telemetry Bot
06:44 PM Bug #52162 (Duplicate): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionR...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=cef671ee83553d16f2680c10...
Telemetry Bot
06:44 PM Bug #52161 (Rejected): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionRe...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=3ef9be60ac33158aff0fa884...
Telemetry Bot
06:44 PM Bug #52160 (Duplicate): crash: void PeeringState::check_past_interval_bounds() const: abort

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=95d3f1ffec846b1fe432b371...
Telemetry Bot
06:44 PM Bug #52159 (Duplicate): crash: void PeeringState::check_past_interval_bounds() const: abort

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=3a50bb9444331ec2b94a68f8...
Telemetry Bot
06:44 PM Bug #52158 (Need More Info): crash: ceph::common::PerfCounters::set(int, unsigned long)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=2bce05236e68a5895b61a2b3...
Telemetry Bot
06:44 PM Bug #52156 (Duplicate): crash: virtual void OSDMonitor::update_from_paxos(bool*): assert(err == 0)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=1a7da12035f04d1f6b48fbb8...
Telemetry Bot
06:44 PM Bug #52155 (Need More Info): crash: pthread_rwlock_rdlock() in queue_want_up_thru

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=ca3a3acf9c89597282538439...
Telemetry Bot
06:44 PM Bug #52154 (Won't Fix): crash: Infiniband::MemoryManager::Chunk::write(char*, unsigned int)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=93435f7389a14a7b2cf7302a...
Telemetry Bot
06:44 PM Bug #52153 (Won't Fix): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionR...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=606f52662a089d81dd216674...
Telemetry Bot
06:44 PM Bug #52152 (Duplicate): crash: pthread_getname_np()

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=1f0f8d147aa33e69a44fe0bb...
Telemetry Bot
06:43 PM Bug #52151 (New): crash: rocksdb::Cleanable::~Cleanable()

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=3c6179c451759d26d2272a88...
Telemetry Bot
06:43 PM Bug #52150 (Won't Fix): crash: bool HealthMonitor::check_member_health(): assert(store_size > 0)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=01246abff7cfc6a0f6751690...
Telemetry Bot
06:43 PM Bug #52149 (Duplicate): crash: void OSDShard::register_and_wake_split_child(PG*): assert(p != pg_...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=86176dad44ae51d3e7de7eac...
Telemetry Bot
06:43 PM Bug #52148 (Duplicate): crash: pthread_getname_np()

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=ad849d7b1a9aff5bb2b92f6f...
Telemetry Bot
06:43 PM Bug #52147 (Duplicate): crash: rocksdb::InstrumentedMutex::Lock()

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=934e8b8d53204a2de4929567...
Telemetry Bot
06:43 PM Bug #52145 (Duplicate): crash: OSDMapRef OSDService::get_map(epoch_t): assert(ret)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=3b9698aa938dbbb1fbbcfcd9...
Telemetry Bot
06:43 PM Bug #52143 (Duplicate): crash: void OSD::handle_osd_map(MOSDMap*): assert(p != added_maps_bl.end())

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=a75157c145de8e54f78f1da6...
Telemetry Bot
06:43 PM Bug #52142 (Duplicate): crash: virtual Monitor::~Monitor(): assert(session_map.sessions.empty())

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=d2e21097b6d6975cd5ebe9ff...
Telemetry Bot
06:43 PM Bug #52141 (Need More Info): crash: void OSD::load_pgs(): abort

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=2784c173c945de7507d54666...
Telemetry Bot
06:43 PM Bug #52140 (Duplicate): crash: OpTracker::~OpTracker(): assert((sharded_in_flight_list.back())->o...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=85dedc566bac7a7f47a8de6e...
Telemetry Bot
05:34 PM Bug #52136 (Resolved): Valgrind reports memory "Leak_DefinitelyLost" errors.
Valgrind reported the memory leak error in the following jobs:
/a/yuriw-2021-08-05_21:11:40-rados-wip-yuri-testing...
Sridhar Seshasayee
03:51 PM Bug #39150: mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=4d653e9c3ee37041dd2a1cf55...
Telemetry Bot
03:51 PM Bug #46266: Monitor crashed in creating pool in CrushTester::test_with_fork()

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=d6b6f43e0c31315c6493798ed...
Telemetry Bot
03:51 PM Bug #44715: common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_list.back())->ops_in_...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=58697a1c8d484e18346c670af...
Telemetry Bot
09:33 AM Bug #52129: LibRadosWatchNotify.AioWatchDelete failed
This is probably a duplicate of https://tracker.ceph.com/issues/48844. Sridhar Seshasayee

08/10/2021

09:26 PM Bug #52129 (Fix Under Review): LibRadosWatchNotify.AioWatchDelete failed
... Neha Ojha
06:52 PM Bug #52127 (New): stretch mode: disallow users from removing the tiebreaker monitor
Right now, there are no guards which prevent the user from removing the tiebreaker monitor from the monmap.
This i...
Greg Farnum
06:51 PM Bug #52126 (Resolved): stretch mode: allow users to change the tiebreaker monitor
Right now, it's impossible to change the tiebreaker monitor in stretch mode. That's an issue if the monitor needs to ... Greg Farnum
06:49 PM Bug #52125 (New): stretch mode: disallow users from changing election strategy
Right now, users can change the election strategy when in stretch mode. Uh, whoops? Greg Farnum
06:45 PM Bug #52124 (Resolved): Invalid read of size 8 in handle_recovery_delete()
... Neha Ojha
01:42 PM Bug #50659: Segmentation fault under Pacific 16.2.1 when using a custom crush location hook
Based on the progress here it seems like I'm probably the only person to have reported this. I still can't figure out... Andrew Davidoff
12:56 PM Bug #50441: cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
@Deepika finally I think this issue I mentioned last week regarding the prometheus deployment after a new cluster ins... M B
05:04 AM Bug #50441 (Need More Info): cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
M B wrote:
> Unfortunately this issue does not seem to be resolved, or at least not with Pacific 16.2.5. I installed...
Deepika Upadhyay
05:02 AM Bug #50441: cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
@Loic, sure, the PR addressing this issue was backported to pacific, spoke to Dan that octopus backport is not necess... Deepika Upadhyay

08/09/2021

09:01 PM Bug #50441: cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
Deepika, you marked this issue resolved but I can't figure out why, would you be so kind as to explain ? Thanks in ad... Loïc Dachary
09:00 PM Bug #50441 (Pending Backport): cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
Loïc Dachary
08:59 PM Backport #51549 (New): pacific: cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
Loïc Dachary
08:09 PM Backport #51840 (Resolved): pacific: osd: snaptrim logs to derr at every tick
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42482
m...
Loïc Dachary
08:05 PM Backport #51841 (Resolved): octopus: osd: snaptrim logs to derr at every tick
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42484
m...
Loïc Dachary
06:23 PM Bug #49888: rados/singleton: radosbench.py: teuthology.exceptions.MaxWhileTries: reached maximum ...
rados/singleton/{all/thrash-eio mon_election/connectivity msgr-failures/many msgr/async-v1only objectstore/bluestore-... Neha Ojha
06:07 PM Bug #45702: PGLog::read_log_and_missing: ceph_assert(miter == missing.get_items().end() || (miter...
/a/yuriw-2021-08-06_16:31:19-rados-wip-yuri-master-8.6.21-distro-basic-smithi/6324561 - no logs Neha Ojha
06:01 PM Bug #39150: mon: "FAILED ceph_assert(session_map.sessions.empty())" when out of quorum
/a/yuriw-2021-08-06_16:31:19-rados-wip-yuri-master-8.6.21-distro-basic-smithi/6324701 Neha Ojha
05:50 PM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
/a/yuriw-2021-08-06_16:31:19-rados-wip-yuri-master-8.6.21-distro-basic-smithi/6324576 Neha Ojha
05:49 PM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
/a/yuriw-2021-08-06_16:31:19-rados-wip-yuri-master-8.6.21-distro-basic-smithi/6324539 Neha Ojha
09:15 AM Bug #52026: osd: pgs went back into snaptrim state after osd restart
Here is the log of an osd that restarted and made a few pgs into the snaptrim state.
ceph-post-file: 88808267-4ec6...
Arthur Outhenin-Chalandre

08/06/2021

10:35 PM Bug #51998: PG autoscaler is wrong when pool is EC with technique=reed_sol_r6_op
I think we should improve the code and seems like you have already figured out the problem. The reason you cannot dis... Neha Ojha
10:12 PM Bug #52026 (Need More Info): osd: pgs went back into snaptrim state after osd restart
Is it possible for you share some OSD logs with debug_osd=20 from when this issue happens? Neha Ojha
10:09 PM Bug #38357: ClsLock.TestExclusiveEphemeralStealEphemeral failed
showing up more often recently Neha Ojha
10:05 PM Bug #49393 (Can't reproduce): Segmentation fault in ceph::logging::Log::entry()
Neha Ojha
10:04 PM Bug #46318: mon_recovery: quorum_status times out
Haven't seen this in recent rados runs. Neha Ojha
10:02 PM Bug #49727: lazy_omap_stats_test: "ceph osd deep-scrub all" hangs
Haven't seen this recently. Neha Ojha
10:00 PM Bug #48468: ceph-osd crash before being up again
Reducing priority for now. Neha Ojha
04:31 AM Backport #52078 (Resolved): pacific: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
https://github.com/ceph/ceph/pull/45319 Backport Bot
04:31 AM Backport #52077 (Resolved): octopus: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
https://github.com/ceph/ceph/pull/45320 Backport Bot
04:26 AM Bug #45423 (Pending Backport): api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
Kefu Chai
01:32 AM Bug #50441: cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
Can't reproduce the failure; I just started a mon-and-mgr bootstrapped cluster with no incident:... Dan Mick

08/05/2021

10:08 PM Bug #6297 (In Progress): ceph osd tell * will break when FD limit reached, messenger should close...
This has come up again so I am going to reopen this tracker so I can follow up on the resolution. Brad Hubbard

08/04/2021

03:54 PM Bug #52058: osd/scrub performance issue: multiple redundant "updates-applied" scrub events
The refactored scrub code (in Pacific and forward) changed the handling
of applied updates notifications in PrimaryL...
Ronen Friedman
03:47 PM Bug #52058 (New): osd/scrub performance issue: multiple redundant "updates-applied" scrub events
OSD logs show much too many unneeded UpdatesApplied ("updates were applied to the chunk selected for scrubbing").
...
Ronen Friedman
11:13 AM Backport #51988: pacific: osd: Add mechanism to avoid running osd benchmark on osd init when usin...
Note that only a subset of the commits from the associated parent tracker PR can be backported to pacific. More speci... Sridhar Seshasayee
11:01 AM Backport #51859 (Rejected): pacific: standalone/osd-rep-recov-eio.sh: TEST_rep_read_unfound faile...
A backport of the changes associated with the parent tracker was deemed not necessary. Sridhar Seshasayee
10:59 AM Bug #51074 (Resolved): standalone/osd-rep-recov-eio.sh: TEST_rep_read_unfound failed with "Bad da...
This doesn't need to be backported to pacific. The reason is that the mclock_scheduler will not be made default for p... Sridhar Seshasayee
08:27 AM Bug #48750: ceph config set using osd/host mask not working
I have this exact problem in 16.2.4 as well. My workaround is to set it in ceph.conf Jan-Philipp Litza
12:54 AM Bug #38357: ClsLock.TestExclusiveEphemeralStealEphemeral failed
/a/kchai-2021-08-03_15:40:41-rados-wip-kefu-testing-2021-08-03-2117-distro-basic-smithi/6309411 Kefu Chai
12:48 AM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
/a//kchai-2021-08-03_15:40:41-rados-wip-kefu-testing-2021-08-03-2117-distro-basic-smithi/6309402 Kefu Chai

08/03/2021

09:26 PM Bug #51942: src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotActive*>())
https://pulpito.ceph.com/nojha-2021-08-03_18:59:59-rados-wip-yuri-testing-2021-07-27-0830-pacific-distro-basic-smithi... Neha Ojha
07:28 PM Backport #51966 (In Progress): nautilus: set a non-zero default value for osd_client_message_cap
Neha Ojha
07:25 PM Backport #51967 (In Progress): octopus: set a non-zero default value for osd_client_message_cap
Neha Ojha
07:19 PM Backport #51830 (In Progress): pacific: set a non-zero default value for osd_client_message_cap
Neha Ojha
01:20 PM Bug #52026 (Resolved): osd: pgs went back into snaptrim state after osd restart
We are testing snapshots in CephFS. This is a 4 nodes clusters with only replicated pools. During our tests we did a ... Arthur Outhenin-Chalandre
12:39 PM Bug #44286: Cache tiering shows unfound objects after OSD reboots
Jan-Philipp Litza wrote:
> We even hit that bug twice today by rebooting two of our cache servers.
>
> What's int...
Pawel Stefanski
12:09 PM Fix #52025 (Resolved): osd: Add config option to skip running the OSD benchmark on init.
Update documentation on the steps to manually set the max osd iops capacity. Sridhar Seshasayee
10:54 AM Bug #50441: cephadm bootstrap on arm64 fails to start ceph/ceph-grafana service
Unfortunately this issue does not seem to be resolved, or at least not with Pacific 16.2.5. I installed a fresh new c... M B
05:46 AM Bug #52012: osd/scrub: src/osd/scrub_machine.cc: 55: FAILED ceph_assert(state_cast<const NotActiv...
The fix is to use (A) & (B) above as a hint to the Replica, to discard all stale scrub processes.
In the suggested f...
Ronen Friedman
05:40 AM Bug #52012 (In Progress): osd/scrub: src/osd/scrub_machine.cc: 55: FAILED ceph_assert(state_cast<...
Scenario:
- Primary reserves the replica
- Primary requests a scrub
- Replica in the process of creating the sc...
Ronen Friedman
05:34 AM Bug #52012 (Resolved): osd/scrub: src/osd/scrub_machine.cc: 55: FAILED ceph_assert(state_cast<con...
A new scrub request arriving to the replica after manual 'set noscrub' then 'unset' asserts as the replica is
still ...
Ronen Friedman
02:06 AM Bug #47025: rados/test.sh: api_watch_notify_pp LibRadosWatchNotifyECPP.WatchNotify failed
Sridhar Seshasayee wrote:
> Observed on master:
> /a/sseshasa-2021-07-14_10:37:09-rados-wip-sseshasa-testing-2021-0...
Brad Hubbard
12:23 AM Bug #45721: CommandFailedError: Command failed (workunit test rados/test_python.sh) FAIL: test_ra...
I've modified this test to only run the TestWatchNotify subtests (2) and to generate debug logging. I'll report back ... Brad Hubbard

08/02/2021

06:15 PM Bug #49888: rados/singleton: radosbench.py: teuthology.exceptions.MaxWhileTries: reached maximum ...
rados/singleton/{all/thrash-eio mon_election/classic msgr-failures/many msgr/async-v1only objectstore/bluestore-bitma... Neha Ojha
02:29 PM Feature #51984: [RFE] Provide warning when the 'require-osd-release' flag does not match current ...
Sebastian Wagner wrote:
> Thinking. cephadm sets this automatically after the upgrade finishes in https://github.com...
Vikhyat Umrao
12:58 PM Feature #51984: [RFE] Provide warning when the 'require-osd-release' flag does not match current ...
Thinking. cephadm sets this automatically after the upgrade finishes in https://github.com/ceph/ceph/blob/c50d8ebdefc... Sebastian Wagner
12:55 AM Feature #51984 (Resolved): [RFE] Provide warning when the 'require-osd-release' flag does not mat...
For more details please check:
https://bugzilla.redhat.com/show_bug.cgi?id=1988773
Vikhyat Umrao
12:54 PM Bug #50657: smart query on monitors
I also see this on mon/mgr hosts of a ceph octopus cluster:... Hannes von Haugwitz
12:18 PM Bug #51998 (New): PG autoscaler is wrong when pool is EC with technique=reed_sol_r6_op
Dear maintainer,
The PG autoscaler is wrong when trying to calculate the RATE for a pool in Erasure Coding using t...
Benjamin Mare
08:45 AM Fix #50574: qa/standalone: Modify/re-write failing standalone tests with mclock scheduler
Associating parent tracker https://tracker.ceph.com/issues/51464 to this. Sridhar Seshasayee
08:40 AM Fix #50574 (In Progress): qa/standalone: Modify/re-write failing standalone tests with mclock sch...
The PR https://github.com/ceph/ceph/pull/42133 fixes a majority of the standalone tests to work with mclock. However,... Sridhar Seshasayee
08:15 AM Backport #51988 (Resolved): pacific: osd: Add mechanism to avoid running osd benchmark on osd ini...
https://github.com/ceph/ceph/pull/41731 Backport Bot
08:12 AM Fix #51464 (Pending Backport): osd: Add mechanism to avoid running osd benchmark on osd init when...
Sridhar Seshasayee

07/30/2021

11:09 PM Bug #51858: octopus: rados/test_crash.sh failure
We should check what was tested in these batches https://sentry.ceph.com/organizations/ceph/issues/3303/events/72c13d... Neha Ojha
11:00 PM Bug #45423 (Fix Under Review): api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
Neha Ojha
12:47 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
https://github.com/ceph/ceph/pull/42564 Myoungwon Oh
10:51 PM Bug #46266 (Need More Info): Monitor crashed in creating pool in CrushTester::test_with_fork()
Neha Ojha
10:51 PM Bug #46266: Monitor crashed in creating pool in CrushTester::test_with_fork()
Can you share you crushmap? The crash is in CrushTester::test_with_fork() which could mean there is an issue with the... Neha Ojha
10:49 PM Bug #51877 (Duplicate): crash: int fork_function(int, std::ostream&, std::function<signed char()>...
Neha Ojha
10:46 PM Bug #51880 (Triaged): crash: rocksdb::ColumnFamilyData::~ColumnFamilyData() in mon
Josh Durgin
10:29 PM Bug #51887 (Triaged): crash: int Infiniband::MemoryManager::Cluster::fill(uint32_t): assert(m)
This crash has occurred in 3 clusters multiple times and because Infiniband is low priority. Neha Ojha
10:26 PM Bug #51889 (Triaged): crash: OSDMapRef OSDService::get_map(epoch_t): assert(ret)
All similar reports seem too old - prior to nautilus. These latest crashes on octopus are likely a new bug. Josh Durgin
10:15 PM Bug #44715: common/TrackedOp.cc: 163: FAILED ceph_assert((sharded_in_flight_list.back())->ops_in_...
Affected versions according to telemetry
Ceph - v15.1.0, Ceph - v15.1.1, Ceph - v15.2.1, Ceph - v15.2.10, Ceph - v...
Neha Ojha
10:10 PM Bug #51872 (Duplicate): crash: OpTracker::~OpTracker(): assert((sharded_in_flight_list.back())->o...
Neha Ojha
04:56 PM Bug #51846: rados/test.sh: LibRadosList.ListObjectsCursor did not complete.
Attempted to reproduce this issue multiple times using the original config file. This occurs rarely. After multiple a... Sridhar Seshasayee
03:24 PM Bug #51882 (Duplicate): crash: virtual Monitor::~Monitor(): assert(session_map.sessions.empty())
Sage Weil
02:51 PM Backport #51841: octopus: osd: snaptrim logs to derr at every tick
Arthur Outhenin-Chalandre wrote:
> please link this Backport tracker issue with GitHub PR https://github.com/ceph/ce...
Yuri Weinstein
02:50 PM Backport #51840: pacific: osd: snaptrim logs to derr at every tick
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/42482
merged
Yuri Weinstein
11:03 AM Bug #51644 (Resolved): don't assert on bogus CEPH_OSD_ZERO request
Kefu Chai
11:00 AM Bug #51816 (Resolved): monitor segfault on startup in container
Kefu Chai

07/29/2021

11:30 PM Backport #51840 (In Progress): pacific: osd: snaptrim logs to derr at every tick
Neha Ojha
11:25 PM Backport #51967 (Resolved): octopus: set a non-zero default value for osd_client_message_cap
https://github.com/ceph/ceph/pull/42616 Backport Bot
11:25 PM Backport #51966 (Resolved): nautilus: set a non-zero default value for osd_client_message_cap
https://github.com/ceph/ceph/pull/42617 Backport Bot
06:05 PM Bug #51942: src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotActive*>())
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{defa... Neha Ojha
01:10 AM Backport #51952 (Resolved): pacific: FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get...
https://github.com/ceph/ceph/pull/43099 Backport Bot
01:07 AM Bug #51627 (Pending Backport): FAILED ceph_assert(attrs || !recovery_state.get_pg_log().get_missi...
Neha Ojha
01:06 AM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
Myoungwon Oh wrote:
> OK, I need to time to look at this.
sure, no worries, thanks!
Neha Ojha
12:50 AM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
OK, I need to time to look at this. Myoungwon Oh

07/28/2021

09:57 PM Bug #51878 (Won't Fix): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionR...
"failed to write to db" -> probably a resource issue (ENOSPC or similar).
Sage Weil
01:47 PM Bug #51878 (Won't Fix): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionR...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=93c11009e22fb8e828c15607...
Telemetry Bot
09:55 PM Bug #51881 (Duplicate): crash: pthread_getname_np()
Sage Weil
01:47 PM Bug #51881 (Duplicate): crash: pthread_getname_np()

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=cbb64bdf7f27ebc70a80e43b...
Telemetry Bot
09:54 PM Bug #51879 (Won't Fix): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionR...
"failed to write to db" -> probably a resource issue (ENOSPC or similar). Sage Weil
01:47 PM Bug #51879 (Won't Fix): crash: int MonitorDBStore::apply_transaction(MonitorDBStore::TransactionR...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=f253a220637e71a562a88ebb...
Telemetry Bot
06:23 PM Bug #51945 (New): qa/workunits/mon/caps.sh: Error: Expected return 13, got 0
... Neha Ojha
05:59 PM Bug #51942 (Resolved): src/osd/scrub_machine.cc: FAILED ceph_assert(state_cast<const NotActive*>())
... Neha Ojha
04:56 PM Bug #51307: LibRadosWatchNotify.Watch2Delete fails
/a/yuriw-2021-07-27_17:19:39-rados-wip-yuri-testing-2021-07-27-0830-pacific-distro-basic-smithi/6297201 Neha Ojha
03:57 PM Bug #51904 (Resolved): test_pool_min_size:AssertionError:wait_for_clean:failed before timeout exp...
... Neha Ojha
03:51 PM Bug #51816: monitor segfault on startup in container
I am observing this very early when running CBT tests and when running qa/standalone tests. Bringing up ceph-mon gene... Sridhar Seshasayee
03:45 PM Bug #51816 (Fix Under Review): monitor segfault on startup in container
Sage Weil
03:37 PM Bug #51816: monitor segfault on startup in container
I assume that the "stack_sig" key is only available from the ceph crash info command right ?
The issue here, is th...
Dimitri Savineau
03:20 PM Bug #51816: monitor segfault on startup in container
"stack_sig" key is missing from the crash metadata; do you see it in any other similar crashes? Yaarit Hatuka
01:48 PM Bug #51889 (Triaged): crash: OSDMapRef OSDService::get_map(epoch_t): assert(ret)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=c2113a82e309c85fb6722417...
Telemetry Bot
01:47 PM Bug #51887 (Triaged): crash: int Infiniband::MemoryManager::Cluster::fill(uint32_t): assert(m)

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=f3fb431dbb2a3e2f5deb2623...
Telemetry Bot
01:47 PM Bug #51882 (Duplicate): crash: virtual Monitor::~Monitor(): assert(session_map.sessions.empty())

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=4d653e9c3ee37041dd2a1cf5...
Telemetry Bot
01:47 PM Bug #51880 (Triaged): crash: rocksdb::ColumnFamilyData::~ColumnFamilyData() in mon

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=4c49507ff32085966f578676...
Telemetry Bot
01:47 PM Bug #51877 (Duplicate): crash: int fork_function(int, std::ostream&, std::function<signed char()>...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=d6b6f43e0c31315c6493798e...
Telemetry Bot
01:47 PM Bug #51872: crash: OpTracker::~OpTracker(): assert((sharded_in_flight_list.back())->ops_in_flight...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=58697a1c8d484e18346c670af...
Telemetry Bot
03:51 AM Bug #51872 (Duplicate): crash: OpTracker::~OpTracker(): assert((sharded_in_flight_list.back())->o...

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=58697a1c8d484e18346c670a...
Telemetry Bot
12:51 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
/ceph/teuthology-archive/yuriw-2021-07-26_16:32:00-rados-wip-yuri5-testing-2021-07-26-0803-octopus-distro-basic-smith... Deepika Upadhyay
12:48 PM Bug #51246: error in open_pools_parallel: rados_write(0.obj) failed with error: -2
... Deepika Upadhyay
10:01 AM Bug #36304: FAILED ceph_assert(p != pg_slots.end()) in OSDShard::register_and_wake_split_child(PG*)
... Kefu Chai

07/27/2021

07:44 PM Bug #51816: monitor segfault on startup in container
Still the same issue with the latest shaman build [1]... Dimitri Savineau
09:12 AM Bug #51864 (New): Test
test Shreya Sharma
05:44 AM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
"/ceph/teuthology-archive/yuriw-2021-07-23_17:36:03-rados-wip-yuri6-testing-2021-07-23-0834-octopus-distro-basic-smit... Deepika Upadhyay
04:55 AM Backport #51859 (Rejected): pacific: standalone/osd-rep-recov-eio.sh: TEST_rep_read_unfound faile...
Backport Bot
04:54 AM Bug #51074: standalone/osd-rep-recov-eio.sh: TEST_rep_read_unfound failed with "Bad data after pr...
I assume there needs to be at least a backport to pacific and populated the Backport field accordingly. Feel free to ... Loïc Dachary
04:54 AM Bug #51858 (New): octopus: rados/test_crash.sh failure
... Deepika Upadhyay
04:54 AM Bug #49487 (Resolved): osd:scrub skip some pg
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Loïc Dachary
04:51 AM Bug #51101 (Resolved): rados/test_envlibrados_for_rocksdb.sh: cmake: symbol lookup error: cmake: ...
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Loïc Dachary
04:49 AM Bug #51451 (Resolved): Add simultaneous scrubs to rados/thrash
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Loïc Dachary
04:49 AM Fix #51580 (Resolved): qa/standalone: Add missing teardowns at the end of a subset of osd and scr...
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Loïc Dachary
04:49 AM Bug #51581 (Resolved): scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_tag failed
While running with --resolve-parent, the script "backport-create-issue" noticed that all backports of this issue are ... Loïc Dachary
04:48 AM Backport #51315 (Resolved): nautilus: osd:scrub skip some pg
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/41973
m...
Loïc Dachary
04:47 AM Backport #51582 (Resolved): octopus: osd does not proactively remove leftover PGs
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42239
m...
Loïc Dachary
04:47 AM Backport #51662 (Resolved): octopus: rados/test_envlibrados_for_rocksdb.sh: cmake: symbol lookup ...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42421
m...
Loïc Dachary
04:47 AM Backport #51496 (Resolved): octopus: mgr spamming with repeated set pgp_num_actual while merging
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42420
m...
Loïc Dachary
04:47 AM Backport #51452 (Resolved): octopus: Add simultaneous scrubs to rados/thrash
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42422
m...
Loïc Dachary
04:46 AM Backport #51314 (Resolved): octopus: osd:scrub skip some pg
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/41972
m...
Loïc Dachary
04:41 AM Backport #51766 (Resolved): pacific: scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_tag fa...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42398
m...
Loïc Dachary
04:41 AM Backport #50900 (Resolved): pacific: PGs always go into active+clean+scrubbing+deep+repair in the...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42398
m...
Loïc Dachary
04:40 AM Backport #51663 (Resolved): pacific: rados/test_envlibrados_for_rocksdb.sh: cmake: symbol lookup ...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42344
m...
Loïc Dachary
04:40 AM Backport #51584 (Resolved): pacific: osd does not proactively remove leftover PGs
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42238
m...
Loïc Dachary
04:39 AM Backport #49377 (Resolved): pacific: building libcrc32
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/39902
m...
Loïc Dachary
04:39 AM Backport #51553 (Resolved): pacific: rebuild-mondb hangs
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42411
m...
Loïc Dachary
04:35 AM Backport #51556 (Resolved): pacific: mon: return -EINVAL when handling unknown option in 'ceph os...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42229
m...
Loïc Dachary
04:34 AM Backport #51498 (Resolved): pacific: mgr spamming with repeated set pgp_num_actual while merging
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42223
m...
Loïc Dachary
04:34 AM Backport #51603: pacific: qa/standalone: Add missing teardowns at the end of a subset of osd and ...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42258
m...
Loïc Dachary
04:33 AM Backport #51570 (Resolved): pacific: CommandCrashedError: Command crashed: 'mkdir -p -- /home/ubu...
This update was made using the script "backport-resolve-issue".
backport PR https://github.com/ceph/ceph/pull/42221
m...
Loïc Dachary
02:16 AM Bug #50393: CommandCrashedError: Command crashed: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client...
octopus: http://qa-proxy.ceph.com/teuthology/yuriw-2021-07-25_15:44:02-rados-wip-yuri6-testing-2021-07-23-0834-octopu... Deepika Upadhyay

07/26/2021

07:18 PM Backport #51766: pacific: scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_tag failed
https://github.com/ceph/ceph/pull/42398 merged Yuri Weinstein
07:17 PM Backport #50900: pacific: PGs always go into active+clean+scrubbing+deep+repair in the LRC
Satoru Takeuchi wrote:
> I created a backport PR.
> https://github.com/ceph/ceph/pull/42398
merged
Yuri Weinstein
12:35 PM Bug #51847 (New): A PG in "incomplete" state may end up in a backfill loop.
This was one of the issued observed in:
/a/nojha-2021-07-22_18:00:50-rados-wip-sseshasa-testing-2021-07-14-1320-dist...
Sridhar Seshasayee
11:29 AM Bug #51846 (New): rados/test.sh: LibRadosList.ListObjectsCursor did not complete.

/a/nojha-2021-07-22_18:00:50-rados-wip-sseshasa-testing-2021-07-14-1320-distro-basic-smithi/6287310
/a/nojha-2021-...
Sridhar Seshasayee
09:02 AM Bug #51843 (Fix Under Review): osd/scrub: OSD crashes at PG removal
Ronen Friedman
06:56 AM Backport #51841: octopus: osd: snaptrim logs to derr at every tick
please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/42484
ceph-backport.sh versi...
Arthur Outhenin-Chalandre
06:55 AM Backport #51840: pacific: osd: snaptrim logs to derr at every tick
please link this Backport tracker issue with GitHub PR https://github.com/ceph/ceph/pull/42482
ceph-backport.sh versi...
Arthur Outhenin-Chalandre

07/25/2021

11:31 AM Bug #51843 (Triaged): osd/scrub: OSD crashes at PG removal
Caused by the incorrect reliance of some Scrubber sub-objects on PG::gen_prefix() for their logging.
Ronen Friedman
11:29 AM Bug #51843 (Resolved): osd/scrub: OSD crashes at PG removal
A dout() message in the scrubber code might cause an OSD crash, if issued when the PG is already removed.
BT might...
Ronen Friedman

07/24/2021

11:15 AM Backport #51841 (Resolved): octopus: osd: snaptrim logs to derr at every tick
https://github.com/ceph/ceph/pull/42484 Backport Bot
11:15 AM Backport #51840 (Resolved): pacific: osd: snaptrim logs to derr at every tick
https://github.com/ceph/ceph/pull/42482 Backport Bot
11:14 AM Bug #51799 (Pending Backport): osd: snaptrim logs to derr at every tick
Kefu Chai

07/23/2021

10:50 PM Bug #51816: monitor segfault on startup in container
I tested yesterday with the latest master build available on shaman : "ceph version 17.0.0-6285-gc011af69 (c011af6903... Dimitri Savineau
09:21 PM Bug #51816: monitor segfault on startup in container
This is related to https://github.com/ceph/ceph/pull/42014. I know there have been a few follow-on fixes for this PR,... Neha Ojha
10:10 PM Bug #51729: Upmap verification fails for multi-level crush rule
Binary osd map attached. Andras Pataki
09:38 PM Bug #51729 (Need More Info): Upmap verification fails for multi-level crush rule
Can you share copy of your binary osdmap? Neha Ojha
09:59 PM Backport #51663: pacific: rados/test_envlibrados_for_rocksdb.sh: cmake: symbol lookup error: cmak...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/42344
merged
Yuri Weinstein
09:58 PM Backport #51582: octopus: osd does not proactively remove leftover PGs
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/42239
merged
Yuri Weinstein
09:57 PM Backport #51584: pacific: osd does not proactively remove leftover PGs
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/42238
merged
Yuri Weinstein
09:54 PM Bug #51000: LibRadosTwoPoolsPP.ManifestSnapRefcount failure
Myoungwon Oh wrote:
> Ok, I'll take a look.
thanks!
Neha Ojha
09:54 PM Bug #51000 (Fix Under Review): LibRadosTwoPoolsPP.ManifestSnapRefcount failure
Neha Ojha
09:54 PM Backport #49377: pacific: building libcrc32
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/39902
merged
Yuri Weinstein
09:43 PM Bug #51083 (Need More Info): Raw space filling up faster than used space
Moving to need more info, please let us know if it happens again. Neha Ojha
09:41 PM Bug #45423: api_tier_pp: [ FAILED ] LibRadosTwoPoolsPP.HitSetWrite
Hi Myoungwon Oh, this is tiering test has been causing a lot of failures, do you think you can help investigate? Neha Ojha
09:34 PM Bug #48908 (Need More Info): EC Pool OSD crashes
Can you capture a coredump for the crash? This appears to be memory leak of some sort, what is the memory/RSS usage l... Neha Ojha
09:23 PM Bug #51799 (Fix Under Review): osd: snaptrim logs to derr at every tick
Neha Ojha
06:42 PM Backport #51662: octopus: rados/test_envlibrados_for_rocksdb.sh: cmake: symbol lookup error: cmak...
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/42421
merged
Yuri Weinstein
06:30 PM Backport #51553: pacific: rebuild-mondb hangs
Kefu Chai wrote:
> https://github.com/ceph/ceph/pull/42411
merged
Yuri Weinstein
02:40 PM Backport #51830 (Resolved): pacific: set a non-zero default value for osd_client_message_cap
https://github.com/ceph/ceph/pull/42615 Backport Bot
02:36 PM Bug #49894 (Pending Backport): set a non-zero default value for osd_client_message_cap
Neha Ojha
11:45 AM Bug #48750: ceph config set using osd/host mask not working
Do the other masks (non-host) masks work for you?
I have the same problem in octopus. class masks work, as do crus...
Dan van der Ster

07/22/2021

10:18 PM Bug #51816 (Resolved): monitor segfault on startup in container
The ceph-container project runs a demo container to validate the container build which starts few daemons and tests i... Dimitri Savineau
08:53 PM Bug #49689: osd/PeeringState.cc: ceph_abort_msg("past_interval start interval mismatch") start
/a/yuriw-2021-07-21_19:04:13-rados-wip-yuri4-testing-2021-07-21-0859-pacific-distro-basic-smithi/6285380 Neha Ojha
04:25 PM Backport #51496: octopus: mgr spamming with repeated set pgp_num_actual while merging
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/42420
merged
Yuri Weinstein
12:41 PM Bug #51799 (Resolved): osd: snaptrim logs to derr at every tick
Hi,
We reduced `osd_snap_trim_sleep` to 0.1 to speed up snap trim process but now the OSD spam the log on every ti...
Arthur Outhenin-Chalandre
05:45 AM Bug #48841 (Resolved): test_turn_off_module: wait_until_equal timed out
https://github.com/ceph/ceph/pull/41907 fixed this Kamoltat (Junior) Sirivadhna
04:04 AM Bug #51725 (Fix Under Review): make bufferlist::c_str() skip rebuild when it isn't necessary
Kefu Chai

07/21/2021

09:32 PM Backport #51452: octopus: Add simultaneous scrubs to rados/thrash
Backport Bot wrote:
> https://github.com/ceph/ceph/pull/42422
merged
Yuri Weinstein
09:28 PM Backport #51314: octopus: osd:scrub skip some pg
Mykola Golub wrote:
> https://github.com/ceph/ceph/pull/41972
merged
Yuri Weinstein
04:38 PM Backport #51315: nautilus: osd:scrub skip some pg
Mykola Golub wrote:
> https://github.com/ceph/ceph/pull/41973
merged
Yuri Weinstein
02:00 PM Backport #51766 (Resolved): pacific: scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_tag fa...
https://github.com/ceph/ceph/pull/42398 Backport Bot
01:58 PM Bug #51581 (Pending Backport): scrub/osd-scrub-repair.sh: TEST_auto_repair_bluestore_tag failed
Neha Ojha
01:55 PM Bug #51638: rados/test_envlibrados_for_rocksdb.sh: No match for argument: snappy-devel with cento...
David Galloway wrote:
> Deepika Upadhyay wrote:
> > David Galloway wrote:
> > > Oh, I see. Okay. I will fix.
> ...
Neha Ojha
03:56 AM Bug #48908: EC Pool OSD crashes

http://telemetry.front.sepia.ceph.com:4000/d/jByk5HaMz/crash-spec-x-ray?orgId=1&var-sig_v2=45667d498bde59f350e2c1c48...
Telemetry Bot
 

Also available in: Atom