ceph version : master (ceph version 0.80-713-g86754cc (86754cc78ca570f19f5a68fb634d613f952a22eb)) fio version : fio-2.1.9-20-g290a
gdb backtrace #0 0x00007ffff6de5249 in AO_fetch_and_add_full (incr=1, p=0x7fff00000018) at /usr/include/atomic_ops/sysdeps/gcc/x86.h:68 #1 inc (this=0x7fff00000018) at ./include/atomic.h:98 #2 ceph::buffer::ptr::ptr (this=0x7fffecf74820, p=..., o=<optimized out>, l=0) at common/buffer.cc:575 #3 0x00007ffff6de63df in ceph::buffer::list::append (this=this@entry=0x7fffc80008e8, bp=..., off=<optimized out>, len=len@entry=0) at common/buffer.cc:1267 #4 0x00007ffff6de6a44 in ceph::buffer::list::splice (this=0x7fffe4014590, off=<optimized out>, len=64512, claim_by=0x7fffc80008e8) at common/buffer.cc:1426 #5 0x00007ffff7b89d45 in Striper::StripedReadResult::add_partial_sparse_result (this=0x7fffe4014278, cct=0x7fffe4006f50, bl=..., bl_map=..., bl_off=3670016, buffer_extents=...) at osdc/Striper.cc:291 #6 0x00007ffff7b180d8 in librbd::C_AioRead::finish (this=0x7fffe400d6a0, r=<optimized out>) at librbd/AioCompletion.cc:94 #7 0x00007ffff7b182f9 in Context::complete (this=0x7fffe400d6a0, r=<optimized out>) at ./include/Context.h:64 #8 0x00007ffff7b1840d in librbd::AioRequest::complete (this=0x7fffe4014540, r=0) at ./librbd/AioRequest.h:40 #9 0x00007ffff6d3a538 in librados::C_AioComplete::finish (this=0x7fffdc0025c0, r=<optimized out>) at ./librados/AioCompletionImpl.h:178 #10 0x00007ffff7b182f9 in Context::complete (this=0x7fffdc0025c0, r=<optimized out>) at ./include/Context.h:64 #11 0x00007ffff6dc85f0 in Finisher::finisher_thread_entry (this=0x7fffe400c7c8) at common/Finisher.cc:56 #12 0x00007ffff5f49f8e in start_thread (arg=0x7fffecf75700) at pthread_create.c:311 #13 0x00007ffff5a6fa0d in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:113 Thanks, Sushma On Tue, Jun 3, 2014 at 12:34 PM, Danny Al-Gaaf <danny.al-g...@bisect.de> wrote: > Am 03.06.2014 20:55, schrieb Sushma R: > > Haomai, > > > > I'm using the latest ceph master branch. > > > > ceph_smalliobench is a Ceph internal benchmarking tool similar to rados > > bench and the performance is more or less similar to that reported by > fio. > > > > I tried to use fio with rbd ioengine ( > > > http://telekomcloud.github.io/ceph/2014/02/26/ceph-performance-analysis_fio_rbd.html > ) > > and below are the numbers with different workloads on our setup. > > Note : fio rbd engine segfaults with randread IO pattern, only with > LevelDB > > (no issues with FileStore). With FileStore, performance of > > ceph_smalliobench and fio-rbd is similar for READs, so the numbers for > > randread for LevelDB are with ceph_smalliobench (since fio rbd > segfaults). > > Could you send me a backtrace of the segfault and some info about the > ceph and fio version you used so that I can take a look at it? > > Thanks, > > Danny > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com