From mboxrd@z Thu Jan 1 00:00:00 1970 From: Haomai Wang Subject: Re: [Annonce]The progress of KeyValueStore in Firely Date: Wed, 4 Jun 2014 13:02:39 +0800 Message-ID: References: <531123F6.5070007@bisect.de> <538E2361.2040400@bisect.de> Mime-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Return-path: In-Reply-To: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: ceph-users-bounces-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org Sender: "ceph-users" To: Sushma R Cc: "ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org" , "ceph-devel-u79uwXL29TY76Z2rM5mHXA@public.gmane.org" List-Id: ceph-devel.vger.kernel.org The fix pull request is https://github.com/ceph/ceph/pull/1912/files. Someone can help to review and merge On Wed, Jun 4, 2014 at 3:38 AM, Sushma R wrote: > ceph version : master (ceph version 0.80-713-g86754cc > (86754cc78ca570f19f5a68fb634d613f952a22eb)) > fio version : fio-2.1.9-20-g290a > > gdb backtrace > #0 0x00007ffff6de5249 in AO_fetch_and_add_full (incr=1, p=0x7fff00000018) > at /usr/include/atomic_ops/sysdeps/gcc/x86.h:68 > #1 inc (this=0x7fff00000018) at ./include/atomic.h:98 > #2 ceph::buffer::ptr::ptr (this=0x7fffecf74820, p=..., o=, > l=0) at common/buffer.cc:575 > #3 0x00007ffff6de63df in ceph::buffer::list::append > (this=this@entry=0x7fffc80008e8, bp=..., off=, > len=len@entry=0) at common/buffer.cc:1267 > #4 0x00007ffff6de6a44 in ceph::buffer::list::splice (this=0x7fffe4014590, > off=, len=64512, claim_by=0x7fffc80008e8) at > common/buffer.cc:1426 > #5 0x00007ffff7b89d45 in > Striper::StripedReadResult::add_partial_sparse_result (this=0x7fffe4014278, > cct=0x7fffe4006f50, bl=..., bl_map=..., bl_off=3670016, > buffer_extents=...) at osdc/Striper.cc:291 > #6 0x00007ffff7b180d8 in librbd::C_AioRead::finish (this=0x7fffe400d6a0, > r=) at librbd/AioCompletion.cc:94 > #7 0x00007ffff7b182f9 in Context::complete (this=0x7fffe400d6a0, > r=) at ./include/Context.h:64 > #8 0x00007ffff7b1840d in librbd::AioRequest::complete (this=0x7fffe4014540, > r=0) at ./librbd/AioRequest.h:40 > #9 0x00007ffff6d3a538 in librados::C_AioComplete::finish > (this=0x7fffdc0025c0, r=) at > ./librados/AioCompletionImpl.h:178 > #10 0x00007ffff7b182f9 in Context::complete (this=0x7fffdc0025c0, > r=) at ./include/Context.h:64 > #11 0x00007ffff6dc85f0 in Finisher::finisher_thread_entry > (this=0x7fffe400c7c8) at common/Finisher.cc:56 > #12 0x00007ffff5f49f8e in start_thread (arg=0x7fffecf75700) at > pthread_create.c:311 > #13 0x00007ffff5a6fa0d in clone () at > ../sysdeps/unix/sysv/linux/x86_64/clone.S:113 > > Thanks, > Sushma > > > On Tue, Jun 3, 2014 at 12:34 PM, Danny Al-Gaaf > wrote: >> >> Am 03.06.2014 20:55, schrieb Sushma R: >> > Haomai, >> > >> > I'm using the latest ceph master branch. >> > >> > ceph_smalliobench is a Ceph internal benchmarking tool similar to rados >> > bench and the performance is more or less similar to that reported by >> > fio. >> > >> > I tried to use fio with rbd ioengine ( >> > >> > http://telekomcloud.github.io/ceph/2014/02/26/ceph-performance-analysis_fio_rbd.html) >> > and below are the numbers with different workloads on our setup. >> > Note : fio rbd engine segfaults with randread IO pattern, only with >> > LevelDB >> > (no issues with FileStore). With FileStore, performance of >> > ceph_smalliobench and fio-rbd is similar for READs, so the numbers for >> > randread for LevelDB are with ceph_smalliobench (since fio rbd >> > segfaults). >> >> Could you send me a backtrace of the segfault and some info about the >> ceph and fio version you used so that I can take a look at it? >> >> Thanks, >> >> Danny >> > -- Best Regards, Wheat