This isn't very complete as it just indicates that something went wrong with a read. Since I presume it happens on every startup, it may help if you set "debug bluestore = 20" in the OSD's config and provide that log (perhaps with ceph-post-file if it's large). I also went through my email and see https://tracker.ceph.com/issues/24639, if you have anything in common with that deployment. (But you probably don't; an error on read generally is about bad state on disk that was created somewhere else.) -Greg
On Sun, Aug 5, 2018 at 3:19 PM Benoit Hudzia <ben...@stratoscale.com> wrote: > Hi, > > We start to see core dump occurring with luminous 12.2.7. Any idea where > this is coming from ?? We started having issues with bluestore core dumping > when we moved to 12.2.6 and hoped that 12.2.7 would have fixed it. We might > need to revert back to 12.2.5 as it seems a lot more stable. > > Pastebin link for full log: https://pastebin.com/na4E3m3N > > > Core dump : > > starting osd.7 at - osd_data /var/lib/ceph/osd/ceph-7 > /var/lib/ceph/osd/ceph-7/journal > *** Caught signal (Segmentation fault) ** > in thread 7fa8830cfd80 thread_name:ceph-osd > ceph version 12.2.7 (3ec878d1e53e1aeb47a9f619c49d9e7c0aa384d5) luminous > (stable) > 1: (()+0xa48ec1) [0x55e010afcec1] > 2: (()+0xf6d0) [0x7fa8807966d0] > 3: (BlueFS::_read(BlueFS::FileReader*, BlueFS::FileReaderBuffer*, unsigned > long, unsigned long, ceph::buffer::list*, char*)+0x452) [0x55e010ab1e72] > 4: (BlueFS::_replay(bool)+0x2ef) [0x55e010ac526f] > 5: (BlueFS::mount()+0x1d4) [0x55e010ac8fd4] > 6: (BlueStore::_open_db(bool)+0x1847) [0x55e0109e2da7] > 7: (BlueStore::_mount(bool)+0x40e) [0x55e010a1406e] > 8: (OSD::init()+0x3bd) [0x55e0105c934d] > 9: (main()+0x2d07) [0x55e0104ce947] > 10: (__libc_start_main()+0xf5) [0x7fa87f7a3445] > 11: (()+0x4b9003) [0x55e01056d003] > 2018-08-03 21:58:12.248736 7fa8830cfd80 -1 *** Caught signal (Segmentation > fault) ** > in thread 7fa8830cfd80 thread_name:ceph-osd > > ceph version 12.2.7 (3ec878d1e53e1aeb47a9f619c49d9e7c0aa384d5) luminous > (stable) > 1: (()+0xa48ec1) [0x55e010afcec1] > 2: (()+0xf6d0) [0x7fa8807966d0] > 3: (BlueFS::_read(BlueFS::FileReader*, BlueFS::FileReaderBuffer*, unsigned > long, unsigned long, ceph::buffer::list*, char*)+0x452) [0x55e010ab1e72] > 4: (BlueFS::_replay(bool)+0x2ef) [0x55e010ac526f] > 5: (BlueFS::mount()+0x1d4) [0x55e010ac8fd4] > 6: (BlueStore::_open_db(bool)+0x1847) [0x55e0109e2da7] > 7: (BlueStore::_mount(bool)+0x40e) [0x55e010a1406e] > 8: (OSD::init()+0x3bd) [0x55e0105c934d] > 9: (main()+0x2d07) [0x55e0104ce947] > 10: (__libc_start_main()+0xf5) [0x7fa87f7a3445] > 11: (()+0x4b9003) [0x55e01056d003] > NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to > interpret this. > > 0> 2018-08-03 21:58:12.248736 7fa8830cfd80 -1 *** Caught signal > (Segmentation fault) ** > in thread 7fa8830cfd80 thread_name:ceph-osd > > ceph version 12.2.7 (3ec878d1e53e1aeb47a9f619c49d9e7c0aa384d5) luminous > (stable) > 1: (()+0xa48ec1) [0x55e010afcec1] > 2: (()+0xf6d0) [0x7fa8807966d0] > 3: (BlueFS::_read(BlueFS::FileReader*, BlueFS::FileReaderBuffer*, unsigned > long, unsigned long, ceph::buffer::list*, char*)+0x452) [0x55e010ab1e72] > 4: (BlueFS::_replay(bool)+0x2ef) [0x55e010ac526f] > 5: (BlueFS::mount()+0x1d4) [0x55e010ac8fd4] > 6: (BlueStore::_open_db(bool)+0x1847) [0x55e0109e2da7] > 7: (BlueStore::_mount(bool)+0x40e) [0x55e010a1406e] > 8: (OSD::init()+0x3bd) [0x55e0105c934d] > 9: (main()+0x2d07) [0x55e0104ce947] > 10: (__libc_start_main()+0xf5) [0x7fa87f7a3445] > 11: (()+0x4b9003) [0x55e01056d003] > NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to > interpret this. > > /osd_entrypoint: line 98: 119388 Segmentation fault (core dumped) > /usr/bin/ceph-osd -f --cluster "${CEPH_CLUSTERNAME}" --id "${OSD_ID}" > --setuser root --setgroup root > > > > > -- > Dr. Benoit Hudzia > > Mobile (UK): +44 (0) 75 346 78673 > Mobile (IE): +353 (0) 89 219 3675 <+353%2089%20219%203675> > Email: ben...@stratoscale.com > > > > Web <http://www.stratoscale.com/> | Blog > <http://www.stratoscale.com/blog/> | Twitter > <https://twitter.com/Stratoscale> | Google+ > <https://plus.google.com/u/1/b/108421603458396133912/108421603458396133912/posts> > | Linkedin <https://www.linkedin.com/company/stratoscale> > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com