Posting this back to zfs-discuss.
Roland's test case (below) is a single threaded sequential write
followed by a single threaded sequential read. His bandwidth
goes from horrible (~2MB/sec) to expected (~30MB/sec)
when prefetch is disabled. This is with relatively recent nv bits (nv110).
Roland - I'm wondering if you were tripping over
CR6732803 ZFS prefetch creates performance issues for streaming workloads.
It seems possible, but that CR is specific about multiple, concurrent IO
streams,
and your test case was only one.
I think it's more likely you were tripping over
CR6412053 zfetch needs a whole lotta love.
For both CR's the workaround is disabling prefetch
(echo "zfs_prefetch_disable/W 1" | mdb -kw)
Any other theories on this test case?
Thanks,
/jim
-------- Original Message --------
Subject: Re: [perf-discuss] ZFS performance issue - READ is slow as
hell...
Date: Tue, 31 Mar 2009 02:33:00 -0700 (PDT)
From: roland <devz...@web.de>
To: perf-disc...@opensolaris.org
Hello Jim,
i double checked again - but it`s like i told:
echo zfs_prefetch_disable/W0t1 | mdb -kw
fixes my problem.
i did a reboot and only set this single param - which immediately makes the
read troughput go up from ~2 MB/s to ~30 MB/s
I don't understand why disabling ZFS prefetch solved this
problem. The test case was a single threaded sequential write, followed
by a single threaded sequential read.
i did not even do a single write - after reboot i just did
dd if=/zfs/TESTFILE of=/dev/null
Solaris Express Community Edition snv_110 X86
FSC RX300 S2
4GB RAM
LSI Logic MegaRaid 320 Onboard SCSI Raid Controller
1x Raid1 LUN
1x Raid5 LUN (3 Disks)
(both LUN`s show same behaviour)
before:
extended device statistics
r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device
21.3 0.1 2717.6 0.1 0.7 0.0 31.8 1.7 2 4 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
16.0 0.0 2048.4 0.0 34.9 0.1 2181.8 4.8 100 3 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
28.0 0.0 3579.2 0.0 34.8 0.1 1246.2 4.9 100 5 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
45.0 0.0 5760.4 0.0 34.8 0.2 772.7 4.5 100 7 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
19.0 0.0 2431.9 0.0 34.9 0.1 1837.3 4.4 100 3 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
58.0 0.0 7421.1 0.0 34.6 0.3 597.4 5.8 100 12 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
0.0 0.0 0.0 0.0 35.0 0.0 0.0 0.0 100 0 c0t1d0
after:
extended device statistics
r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device
218.0 0.0 27842.3 0.0 0.0 0.4 0.1 1.8 1 40 c0t1d0
241.0 0.0 30848.0 0.0 0.0 0.4 0.0 1.6 0 38 c0t1d0
237.0 0.0 30340.1 0.0 0.0 0.4 0.0 1.6 0 38 c0t1d0
230.0 0.0 29434.7 0.0 0.0 0.4 0.0 1.8 0 40 c0t1d0
238.1 0.0 30471.3 0.0 0.0 0.4 0.0 1.5 0 37 c0t1d0
234.9 0.0 30001.9 0.0 0.0 0.4 0.0 1.6 1 37 c0t1d0
220.1 0.0 28171.4 0.0 0.0 0.4 0.2 1.6 5 35 c0t1d0
212.0 0.0 27137.2 0.0 0.0 0.4 0.2 1.8 4 39 c0t1d0
203.9 0.0 26103.5 0.0 0.0 0.4 0.2 1.9 5 39 c0t1d0
214.8 0.0 27489.8 0.0 0.0 0.4 0.2 1.7 5 37 c0t1d0
221.3 0.0 28327.6 0.0 0.0 0.4 0.2 1.6 5 36 c0t1d0
199.0 0.0 25407.9 0.0 0.0 0.4 0.2 2.0 4 39 c0t1d0
182.0 0.0 23297.1 0.0 0.0 0.4 0.2 2.4 4 44 c0t1d0
204.9 0.0 26230.2 0.0 0.0 0.4 0.2 1.8 5 36 c0t1d0
214.1 0.0 27399.9 0.0 0.0 0.4 0.2 1.7 5 37 c0t1d0
207.9 0.0 26611.5 0.0 0.0 0.4 0.2 1.9 4 39 c0t1d0
--
This message posted from opensolaris.org
_______________________________________________
perf-discuss mailing list
perf-disc...@opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss