Bug #42683
closedOSD Segmentation fault
0%
Description
Dear support,
I have a small ceph cluster installed with nautilus 14.2.4 meant for a test for future larger deployment. The cluster consists of 8 machines:
kernel 3.10.0-693.21.1.el7.x86_64
CentOS Linux release 7
Two of them host the OSD services (20 8TB disks each). The cluster is running fine but on one of the machines just one OSD is unable to start with this error:
ceph version 14.2.4 (75f4de193b3ea58512f204623e6c5a16e6c1e1ba) nautilus (stable)
1: (()+0xf5e0) [0x7fa7ec45c5e0]
2: (gsignal()+0x37) [0x7fa7eb46f1f7]
3: (abort()+0x148) [0x7fa7eb4708e8]
4: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x199) [0x56514f412a73]
5: (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, char const*, ...)+0) [0x56514f412bf2]
6: (BitmapAllocator::init_add_free(unsigned long, unsigned long)+0x740) [0x56514fa43870]
7: (BlueStore::_open_alloc()+0x258) [0x56514f8e9ab8]
8: (BlueStore::_open_db_and_around(bool)+0x146) [0x56514f90b306]
9: (BlueStore::_mount(bool, bool)+0x6a4) [0x56514f949c24]
10: (OSD::init()+0x3aa) [0x56514f4bcefa]
11: (main()+0x14fa) [0x56514f4171da]
12: (__libc_start_main()+0xf5) [0x7fa7eb45bc05]
13: (()+0x4b2695) [0x56514f44c695]
The relevant configuration for the OSDs is:
[osd]
osd_recovery_max_active = 40
osd_max_backfills = 64
osd_memory_target = 2684354560
I searched around the issue tracker and it seems like the error has been already reported here:
https://tracker.ceph.com/issues/39334
but I crosschecked and I appear to be running a version with the bugfix mentioned there.
Could you please help me with this?
Antonio
Files