Um...I started 2 commands in 2 separate ssh sessions:
in ssh session one:
iostat -xn 1 > stats
in ssh session two:
mkfile 10g testfile

when the mkfile was finished i did the dd command...
on the same zpool1 and zfs filesystem..that's it, really

On Fri, Jun 18, 2010 at 6:06 AM, Arne Jansen <sensi...@gmx.net> wrote:
> Curtis E. Combs Jr. wrote:
>> Sure. And hey, maybe I just need some context to know what's "normal"
>> IO for the zpool. It just...feels...slow, sometimes. It's hard to
>> explain. I attached a log of iostat -xn 1 while doing mkfile 10g
>> testfile on the zpool, as well as your dd with the bs set really high.
>> When I Ctl-C'ed the dd it said 460M/sec....like I said, maybe I just
>> need some context...
>>
>
> These iostats don't match to the creation of any large files. What are
> you doing there? Looks more like 512 byte random writes... Are you
> generating the load locally or remote?
>
>>
>> On Fri, Jun 18, 2010 at 5:36 AM, Arne Jansen <sensi...@gmx.net> wrote:
>>> artiepen wrote:
>>>> 40MB/sec is the best that it gets. Really, the average is 5. I see 4, 5, 
>>>> 2, and 6 almost 10x as many times as I see 40MB/sec. It really only bumps 
>>>> up to 40 very rarely.
>>>>
>>>> As far as random vs. sequential. Correct me if I'm wrong, but if I used dd 
>>>> to make files from /dev/zero, wouldn't that be sequential? I measure with 
>>>> zpool iostat 2 in another ssh session while making files of various sizes.
>>>>
>>>> This is a test system. I'm wondering, now, if I should just reconfigure 
>>>> with maybe 7 disks and add another spare. Seems to be the general 
>>>> consensus that bigger raid pools = worse performance. I thought the 
>>>> opposite was true...
>>> A quick test on a system with 21 1TB SATA-drives in a single
>>> RAIDZ2 group show a performance of about 400MB/s with a
>>> single dd, blocksize=1048576. Creating a 10G-file with mkfile
>>> takes 25 seconds also.
>>> So I'd say basically there is nothing wrong with the zpool
>>> configuration. Can you paste some "iostat -xn 1" output while
>>> your test is running?
>>>
>>> --Arne
>>>
>>
>>
>>
>
>



-- 
Curtis E. Combs Jr.
System Administrator Associate
University of Georgia
High Performance Computing Center
ceco...@uga.edu
Office: (706) 542-0186
Cell: (706) 206-7289
Gmail Chat: psynoph...@gmail.com
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to