* rbd
@ 2011-05-12 10:53 Fyodor Ustinov
2011-05-12 16:08 ` rbd Sage Weil
0 siblings, 1 reply; 7+ messages in thread
From: Fyodor Ustinov @ 2011-05-12 10:53 UTC (permalink / raw)
To: ceph-devel
Hi!
Latest (git pulled) version of 2.6 kernel. Ceph - 0.27.1
Still troubles with rbd. Now with ocfs2 no messages in syslog, but iozone
still return error:
#df -h
Filesystem Size Used Avail Use% Mounted on
/dev/sda1 237G 15G 210G 7% /
none 2.0G 164K 2.0G 1% /dev
none 2.0G 0 2.0G 0% /dev/shm
none 2.0G 52K 2.0G 1% /var/run
none 2.0G 0 2.0G 0% /var/lock
/dev/rbd0 1000G 3.2G 997G 1% /mnt
# cd /mnt/
root@stb1:/mnt# ls
lost+found
root@stb1:/mnt# iozone -a -n4g -g20g
Iozone: Performance Test of File I/O
Version $Revision: 3.373 $
Compiled for 64 bit mode.
Build: linux-AMD64
[...]
Run began: Thu May 12 13:42:45 2011
Auto Mode
Using minimum file size of 4194304 kilobytes.
Using maximum file size of 20971520 kilobytes.
Command line used: iozone -a -n4g -g20g
Output is in Kbytes/sec
Time Resolution = 0.000001 seconds.
Processor cache size set to 1024 Kbytes.
Processor cache line size set to 32 bytes.
File stride size set to 17 * record size.
random random
bkwd record stride
KB reclen write rewrite read reread read write
read rewrite read fwrite frewrite fread freread
4194304 64 146906 150779
Error in file: Found ?7f24fb70a210? Expecting ?3a3a3a3a3a3a3a3a? addr
7fe5b720e000
Error in file: Position 2481971200
Record # 37871 Record size 64 kb
where 7fe5b720e000 loop 57344
root@stb1:/mnt#
Can I help to catch this error?
WBR,
Fyodor.
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: rbd
2011-05-12 10:53 rbd Fyodor Ustinov
@ 2011-05-12 16:08 ` Sage Weil
2011-05-12 17:01 ` rbd Sage Weil
0 siblings, 1 reply; 7+ messages in thread
From: Sage Weil @ 2011-05-12 16:08 UTC (permalink / raw)
To: Fyodor Ustinov; +Cc: ceph-devel
Hi Fyodor,
> Hi!
>
> Latest (git pulled) version of 2.6 kernel. Ceph - 0.27.1
>
> Still troubles with rbd. Now with ocfs2 no messages in syslog, but iozone
> still return error:
>
> #df -h
> Filesystem Size Used Avail Use% Mounted on
> /dev/sda1 237G 15G 210G 7% /
> none 2.0G 164K 2.0G 1% /dev
> none 2.0G 0 2.0G 0% /dev/shm
> none 2.0G 52K 2.0G 1% /var/run
> none 2.0G 0 2.0G 0% /var/lock
> /dev/rbd0 1000G 3.2G 997G 1% /mnt
>
> # cd /mnt/
> root@stb1:/mnt# ls
> lost+found
> root@stb1:/mnt# iozone -a -n4g -g20g
> Iozone: Performance Test of File I/O
> Version $Revision: 3.373 $
> Compiled for 64 bit mode.
> Build: linux-AMD64
>
> [...]
>
> Run began: Thu May 12 13:42:45 2011
>
> Auto Mode
> Using minimum file size of 4194304 kilobytes.
> Using maximum file size of 20971520 kilobytes.
> Command line used: iozone -a -n4g -g20g
> Output is in Kbytes/sec
> Time Resolution = 0.000001 seconds.
> Processor cache size set to 1024 Kbytes.
> Processor cache line size set to 32 bytes.
> File stride size set to 17 * record size.
> random random
> bkwd record stride
> KB reclen write rewrite read reread read write
> read rewrite read fwrite frewrite fread freread
> 4194304 64 146906 150779
>
> Error in file: Found ?7f24fb70a210? Expecting ?3a3a3a3a3a3a3a3a? addr
> 7fe5b720e000
> Error in file: Position 2481971200
> Record # 37871 Record size 64 kb
> where 7fe5b720e000 loop 57344
> root@stb1:/mnt#
>
> Can I help to catch this error?
This one is going to be a bit difficult because there are several layers
(ocfs2, rbd client, osd) involved. It's presumably not ocfs2's fault, but
trying the same thing with ext4 or something would remove that from the
equation. I'm running iozone on ext3 now and not having any problems.
Assuming the problem is in rbd, the question is whether it's the client or
server side. That will be harder to diagnose, and will probably involve
careful use of fiemap/bmap to map the file offset to a block device
offset, which we can then translate into a rados block. That can be
pulled directly out of the object store to confirm it has the right
content...
sage
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: rbd
2011-05-12 16:08 ` rbd Sage Weil
@ 2011-05-12 17:01 ` Sage Weil
2011-05-13 15:51 ` rbd Sage Weil
0 siblings, 1 reply; 7+ messages in thread
From: Sage Weil @ 2011-05-12 17:01 UTC (permalink / raw)
To: Fyodor Ustinov; +Cc: ceph-devel
On Thu, 12 May 2011, Sage Weil wrote:
> equation. I'm running iozone on ext3 now and not having any problems.
I take it back.. I just reproduced a similar error on ext2:
random random
bkwd record stride
KB reclen write rewrite read reread read write
read rewrite read fwrite frewrite fread freread
4194304 64 3570 3040 12640 13725
Error in file: Found ?aaaaaaaaaaaaaaaa? Expecting ?3838383838383838? addr
40a00000
Error in file: Position 2813329408
Record # 42928 Record size 64 kb
where 40a00000 loop 0
I opened a tracker issue http://tracker.newdream.net/issues/1086
sage
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: rbd
2011-05-12 17:01 ` rbd Sage Weil
@ 2011-05-13 15:51 ` Sage Weil
2011-05-13 19:11 ` rbd Sage Weil
0 siblings, 1 reply; 7+ messages in thread
From: Sage Weil @ 2011-05-13 15:51 UTC (permalink / raw)
To: Fyodor Ustinov; +Cc: ceph-devel
On Thu, 12 May 2011, Sage Weil wrote:
> On Thu, 12 May 2011, Sage Weil wrote:
> > equation. I'm running iozone on ext3 now and not having any problems.
>
> I take it back.. I just reproduced a similar error on ext2:
>
> random random
> bkwd record stride
> KB reclen write rewrite read reread read write
> read rewrite read fwrite frewrite fread freread
> 4194304 64 3570 3040 12640 13725
>
> Error in file: Found ?aaaaaaaaaaaaaaaa? Expecting ?3838383838383838? addr
> 40a00000
> Error in file: Position 2813329408
> Record # 42928 Record size 64 kb
> where 40a00000 loop 0
>
> I opened a tracker issue http://tracker.newdream.net/issues/1086
Just an update: we've identified the problem but a fix is still in the
works. The bio coming down is spanning an object boundary, and we're
having trouble getting the block layer to either not do that or to handle
the bio splitting properly.
sage
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: rbd
2011-05-13 15:51 ` rbd Sage Weil
@ 2011-05-13 19:11 ` Sage Weil
2011-05-14 12:36 ` rbd Fyodor Ustinov
0 siblings, 1 reply; 7+ messages in thread
From: Sage Weil @ 2011-05-13 19:11 UTC (permalink / raw)
To: Fyodor Ustinov; +Cc: ceph-devel
Fyodor,
On Fri, 13 May 2011, Sage Weil wrote:
> On Thu, 12 May 2011, Sage Weil wrote:
> > On Thu, 12 May 2011, Sage Weil wrote:
> > > equation. I'm running iozone on ext3 now and not having any problems.
> >
> > I take it back.. I just reproduced a similar error on ext2:
> >
> > random random
> > bkwd record stride
> > KB reclen write rewrite read reread read write
> > read rewrite read fwrite frewrite fread freread
> > 4194304 64 3570 3040 12640 13725
> >
> > Error in file: Found ?aaaaaaaaaaaaaaaa? Expecting ?3838383838383838? addr
> > 40a00000
> > Error in file: Position 2813329408
> > Record # 42928 Record size 64 kb
> > where 40a00000 loop 0
> >
> > I opened a tracker issue http://tracker.newdream.net/issues/1086
>
> Just an update: we've identified the problem but a fix is still in the
> works. The bio coming down is spanning an object boundary, and we're
> having trouble getting the block layer to either not do that or to handle
> the bio splitting properly.
Can you try the req_coll branch of
git://ceph.newdream.net/git/ceph-client.git and see if that fixes things
for you? There's a memory leak fix and a fix for the split bio completion
stuff. If it checks out I want to send this to Linus quickly so it'll be
included in 2.6.39 (which is imminent).
Thanks!
sage
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: rbd
2011-05-13 19:11 ` rbd Sage Weil
@ 2011-05-14 12:36 ` Fyodor Ustinov
0 siblings, 0 replies; 7+ messages in thread
From: Fyodor Ustinov @ 2011-05-14 12:36 UTC (permalink / raw)
To: Sage Weil; +Cc: ceph-devel
On 05/13/2011 10:11 PM, Sage Weil wrote:
> Fyodor,
>
> On Fri, 13 May 2011, Sage Weil wrote:
>> On Thu, 12 May 2011, Sage Weil wrote:
>>> On Thu, 12 May 2011, Sage Weil wrote:
>>>> equation. I'm running iozone on ext3 now and not having any problems.
>>> I take it back.. I just reproduced a similar error on ext2:
>>>
>>> random random
>>> bkwd record stride
>>> KB reclen write rewrite read reread read write
>>> read rewrite read fwrite frewrite fread freread
>>> 4194304 64 3570 3040 12640 13725
>>>
>>> Error in file: Found ?aaaaaaaaaaaaaaaa? Expecting ?3838383838383838? addr
>>> 40a00000
>>> Error in file: Position 2813329408
>>> Record # 42928 Record size 64 kb
>>> where 40a00000 loop 0
>>>
>>> I opened a tracker issue http://tracker.newdream.net/issues/1086
>> Just an update: we've identified the problem but a fix is still in the
>> works. The bio coming down is spanning an object boundary, and we're
>> having trouble getting the block layer to either not do that or to handle
>> the bio splitting properly.
> Can you try the req_coll branch of
> git://ceph.newdream.net/git/ceph-client.git and see if that fixes things
> for you? There's a memory leak fix and a fix for the split bio completion
> stuff. If it checks out I want to send this to Linus quickly so it'll be
> included in 2.6.39 (which is imminent).
root@stb1:~# uname -a
Linux stb1 2.6.39-rc7-ufm-ceph #1 SMP Sat May 14 13:50:12 EEST 2011
x86_64 x86_64 x86_64 GNU/Linux
root@stb1:~# modprobe rbd
root@stb1:~# echo "77.120.112.193 name=admin rbd tmt" > /sys/bus/rbd/add
root@stb1:~# mkfs.ocfs2 -M local -T vmstore --fs-features=local,sparse
/dev/rbd0
mkfs.ocfs2 1.6.3
Overwriting existing ocfs2 partition.
Proceed (y/N): y
Filesystem Type of vmstore
Label:
Features: local sparse backup-super unwritten inline-data
strict-journal-super xattr refcount
Block size: 4096 (12 bits)
Cluster size: 1048576 (20 bits)
Volume size: 1073741824000 (1024000 clusters) (262144000 blocks)
Cluster groups: 32 (tail covers 24064 clusters, rest cover 32256 clusters)
Extent allocator size: 3221225472 (768 groups)
Journal size: 134217728
Node slots: 1
Creating bitmaps: done
Initializing superblock: done
Writing system files: done
Writing superblock: done
Writing backup superblock: 5 block(s)
Formatting Journals: done
Growing extent allocator: done
Formatting slot map: done
Formatting quota files: done
Writing lost+found: done
mkfs.ocfs2 successful
root@stb1:~# mount /dev/rbd0 /mnt
root@stb1:~# cd /mnt
root@stb1:/mnt# iozone -a -n4g -g20g
Iozone: Performance Test of File I/O
Version $Revision: 3.373 $
Compiled for 64 bit mode.
Build: linux-AMD64
Contributors:William Norcott, Don Capps, Isom Crawford, Kirby
Collins
Al Slater, Scott Rhine, Mike Wisner, Ken Goss
Steve Landherr, Brad Smith, Mark Kelly, Dr. Alain CYR,
Randy Dunlap, Mark Montague, Dan Million, Gavin
Brebner,
Jean-Marc Zucconi, Jeff Blomberg, Benny Halevy,
Dave Boone,
Erik Habbinga, Kris Strecker, Walter Wong, Joshua
Root,
Fabrice Bacchella, Zhenghua Xue, Qin Li, Darren
Sawyer.
Run began: Sat May 14 14:39:03 2011
Auto Mode
Using minimum file size of 4194304 kilobytes.
Using maximum file size of 20971520 kilobytes.
Command line used: iozone -a -n4g -g20g
Output is in Kbytes/sec
Time Resolution = 0.000001 seconds.
Processor cache size set to 1024 Kbytes.
Processor cache line size set to 32 bytes.
File stride size set to 17 * record size.
random
random bkwd record stride
KB reclen write rewrite read reread read
write read rewrite read fwrite frewrite fread freread
4194304 64 134566 135254 53718 52836 14839 78946
And now it is 20 minutes in this state.In the logs clean. Incoming
network traffic- 3Mbytes/s
But, we newer see this test before, likely an extremely low speed test
"bkwd read" this is another problem.
As you can see - trouble with "Error in file" solved. :)
WBR,
Fyodor.
^ permalink raw reply [flat|nested] 7+ messages in thread
* rbd
@ 2010-08-03 18:49 Yehuda Sadeh Weinraub
0 siblings, 0 replies; 7+ messages in thread
From: Yehuda Sadeh Weinraub @ 2010-08-03 18:49 UTC (permalink / raw)
To: Andrew Morton, Jens Axboe; +Cc: linux-kernel, linux-fsdevel, Sage Weil
Hi Andrew, Jens,
We've posted rbd a few times before. Rbd provides a block device
that stripes data over ceph's block layer (rados). Other than a few
trivial bug fixes it hasn't changed much from our last post, and the
latest version can be found on the rbd branch at
git://ceph.newdream.net/git/ceph-client.git.
First, we'd like to know if it's acceptable to include the rbd block
driver as part of the ceph module, or whether we need to restructure
ceph so that it'd be possible to add rbd as a separate module.
Second, is there any specific person's ACK or review we need in order
to send this upstream with the rest of the ceph 2.6.36 queue?
Thanks,
Yehuda
^ permalink raw reply [flat|nested] 7+ messages in thread
end of thread, other threads:[~2011-05-14 12:36 UTC | newest]
Thread overview: 7+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2011-05-12 10:53 rbd Fyodor Ustinov
2011-05-12 16:08 ` rbd Sage Weil
2011-05-12 17:01 ` rbd Sage Weil
2011-05-13 15:51 ` rbd Sage Weil
2011-05-13 19:11 ` rbd Sage Weil
2011-05-14 12:36 ` rbd Fyodor Ustinov
-- strict thread matches above, loose matches on Subject: below --
2010-08-03 18:49 rbd Yehuda Sadeh Weinraub
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.