Thanks, I ll try to check if i can reproduce it. It's really sporadic and
occurs every 20-30 runs , I might check if it always occurs on the same
server , maybe an HW issue.

On Mon, 6 Aug 2018 at 06:12, Gregory Farnum <gfar...@redhat.com> wrote:

> This isn't very complete as it just indicates that something went wrong
> with a read. Since I presume it happens on every startup, it may help if
> you set "debug bluestore = 20" in the OSD's config and provide that log
> (perhaps with ceph-post-file if it's large).
> I also went through my email and see https://tracker.ceph.com/issues/24639,
> if you have anything in common with that deployment. (But you probably
> don't; an error on read generally is about bad state on disk that was
> created somewhere else.)
> -Greg
>
> On Sun, Aug 5, 2018 at 3:19 PM Benoit Hudzia <ben...@stratoscale.com>
> wrote:
>
>> Hi,
>>
>> We start to see core dump occurring with luminous 12.2.7. Any idea where
>> this is coming from ?? We started having issues with bluestore core dumping
>> when we moved to 12.2.6 and hoped that 12.2.7 would have fixed it. We might
>> need to revert back to 12.2.5 as it seems a lot more stable.
>>
>> Pastebin link for full log: https://pastebin.com/na4E3m3N
>>
>>
>> Core dump :
>>
>> starting osd.7 at - osd_data /var/lib/ceph/osd/ceph-7 
>> /var/lib/ceph/osd/ceph-7/journal
>> *** Caught signal (Segmentation fault) **
>>  in thread 7fa8830cfd80 thread_name:ceph-osd
>>  ceph version 12.2.7 (3ec878d1e53e1aeb47a9f619c49d9e7c0aa384d5) luminous 
>> (stable)
>>  1: (()+0xa48ec1) [0x55e010afcec1]
>>  2: (()+0xf6d0) [0x7fa8807966d0]
>>  3: (BlueFS::_read(BlueFS::FileReader*, BlueFS::FileReaderBuffer*, unsigned 
>> long, unsigned long, ceph::buffer::list*, char*)+0x452) [0x55e010ab1e72]
>>  4: (BlueFS::_replay(bool)+0x2ef) [0x55e010ac526f]
>>  5: (BlueFS::mount()+0x1d4) [0x55e010ac8fd4]
>>  6: (BlueStore::_open_db(bool)+0x1847) [0x55e0109e2da7]
>>  7: (BlueStore::_mount(bool)+0x40e) [0x55e010a1406e]
>>  8: (OSD::init()+0x3bd) [0x55e0105c934d]
>>  9: (main()+0x2d07) [0x55e0104ce947]
>>  10: (__libc_start_main()+0xf5) [0x7fa87f7a3445]
>>  11: (()+0x4b9003) [0x55e01056d003]
>> 2018-08-03 21:58:12.248736 7fa8830cfd80 -1 *** Caught signal (Segmentation 
>> fault) **
>>  in thread 7fa8830cfd80 thread_name:ceph-osd
>>
>>  ceph version 12.2.7 (3ec878d1e53e1aeb47a9f619c49d9e7c0aa384d5) luminous 
>> (stable)
>>  1: (()+0xa48ec1) [0x55e010afcec1]
>>  2: (()+0xf6d0) [0x7fa8807966d0]
>>  3: (BlueFS::_read(BlueFS::FileReader*, BlueFS::FileReaderBuffer*, unsigned 
>> long, unsigned long, ceph::buffer::list*, char*)+0x452) [0x55e010ab1e72]
>>  4: (BlueFS::_replay(bool)+0x2ef) [0x55e010ac526f]
>>  5: (BlueFS::mount()+0x1d4) [0x55e010ac8fd4]
>>  6: (BlueStore::_open_db(bool)+0x1847) [0x55e0109e2da7]
>>  7: (BlueStore::_mount(bool)+0x40e) [0x55e010a1406e]
>>  8: (OSD::init()+0x3bd) [0x55e0105c934d]
>>  9: (main()+0x2d07) [0x55e0104ce947]
>>  10: (__libc_start_main()+0xf5) [0x7fa87f7a3445]
>>  11: (()+0x4b9003) [0x55e01056d003]
>>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to 
>> interpret this.
>>
>>      0> 2018-08-03 21:58:12.248736 7fa8830cfd80 -1 *** Caught signal 
>> (Segmentation fault) **
>>  in thread 7fa8830cfd80 thread_name:ceph-osd
>>
>>  ceph version 12.2.7 (3ec878d1e53e1aeb47a9f619c49d9e7c0aa384d5) luminous 
>> (stable)
>>  1: (()+0xa48ec1) [0x55e010afcec1]
>>  2: (()+0xf6d0) [0x7fa8807966d0]
>>  3: (BlueFS::_read(BlueFS::FileReader*, BlueFS::FileReaderBuffer*, unsigned 
>> long, unsigned long, ceph::buffer::list*, char*)+0x452) [0x55e010ab1e72]
>>  4: (BlueFS::_replay(bool)+0x2ef) [0x55e010ac526f]
>>  5: (BlueFS::mount()+0x1d4) [0x55e010ac8fd4]
>>  6: (BlueStore::_open_db(bool)+0x1847) [0x55e0109e2da7]
>>  7: (BlueStore::_mount(bool)+0x40e) [0x55e010a1406e]
>>  8: (OSD::init()+0x3bd) [0x55e0105c934d]
>>  9: (main()+0x2d07) [0x55e0104ce947]
>>  10: (__libc_start_main()+0xf5) [0x7fa87f7a3445]
>>  11: (()+0x4b9003) [0x55e01056d003]
>>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to 
>> interpret this.
>>
>> /osd_entrypoint: line 98: 119388 Segmentation fault      (core dumped) 
>> /usr/bin/ceph-osd -f --cluster "${CEPH_CLUSTERNAME}" --id "${OSD_ID}" 
>> --setuser root --setgroup root
>>
>>
>>
>>
>> --
>> Dr. Benoit Hudzia
>>
>> Mobile (UK): +44 (0) 75 346 78673
>> Mobile (IE):  +353 (0) 89 219 3675 <+353%2089%20219%203675>
>> Email: ben...@stratoscale.com
>>
>>
>>
>> Web <http://www.stratoscale.com/> | Blog
>> <http://www.stratoscale.com/blog/> | Twitter
>> <https://twitter.com/Stratoscale> | Google+
>> <https://plus.google.com/u/1/b/108421603458396133912/108421603458396133912/posts>
>>  | Linkedin <https://www.linkedin.com/company/stratoscale>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>

-- 
Dr. Benoit Hudzia

Mobile (UK): +44 (0) 75 346 78673
Mobile (IE):  +353 (0) 89 219 3675
Email: ben...@stratoscale.com



Web <http://www.stratoscale.com/> | Blog <http://www.stratoscale.com/blog/>
 | Twitter <https://twitter.com/Stratoscale> | Google+
<https://plus.google.com/u/1/b/108421603458396133912/108421603458396133912/posts>
 | Linkedin <https://www.linkedin.com/company/stratoscale>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to