Hello Matthew, Yes, my machine is 32-bit, with 1.5G of RAM.
-bash-3.00# echo ::memstat | mdb -k Page Summary Pages MB %Tot ------------ ---------------- ---------------- ---- Kernel 123249 481 32% Anon 33704 131 9% Exec and libs 7637 29 2% Page cache 1116 4 0% Free (cachelist) 222661 869 57% Free (freelist) 2685 10 1% Total 391052 1527 Physical 391051 1527 -bash-3.00# echo ::arc | mdb -k { anon = -759566176 mru = -759566136 mru_ghost = -759566096 mfu = -759566056 mfu_ghost = -759566016 size = 0x17f20c00 p = 0x160ef900 c = 0x17f16ae0 c_min = 0x4000000 c_max = 0x1da00000 hits = 0x353b misses = 0x264b deleted = 0x13bc recycle_miss = 0x31 mutex_miss = 0 evict_skip = 0 hash_elements = 0x127b hash_elements_max = 0x1a19 hash_collisions = 0x61 hash_chains = 0x4c hash_chain_max = 0x1 no_grow = 1 } now lets try: set zfs:zfs_prefetch_disable=1 bingo! r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device 609.0 0.0 77910.0 0.0 0.0 0.8 0.0 1.4 0 83 c0d0 only 1-2 % slower then dd from /dev/dsk. Do you think this is general 32-bit problem, or specific to this combination of hardware? I am using PCI/SATA Sil3114 card, and other then ZFS, performance of this interface has some limitations in Solaris. That is, single drive gives 80MB/s, but doing dd /dev/dsk/xyz simultaneously on 2 drives attached to the card gives only 46MB/s each. On Linux, however, that gives 60MB/s each, close to saturating theoretical throughput of PCI bus. Having both drives in zpool stripe gives, with prefetch disabled, close to 45MB/s each through dd from zfs file. I think that under Solaris, this card is accessed through ATA driver. There shouldn't be any issues on inside vs outside. all the reading is done on the first gig or two of the drive, as there is nothing else on them, except one 2 gig file. (well, i'm assuming simple copy onto a newly formatted zfs drive puts it at start of the drive.) Drives are completely owned by ZFS, using zpool create c0d0 c0d1 Finally, should I file a bug somewhere regarding prefetch, or is this a known issue? Many thanks. On 5/15/07, Matthew Ahrens <[EMAIL PROTECTED]> wrote:
Marko Milisavljevic wrote: > I was trying to simply test bandwidth that Solaris/ZFS (Nevada b63) can > deliver from a drive, and doing this: dd if=(raw disk) of=/dev/null gives > me around 80MB/s, while dd if=(file on ZFS) of=/dev/null gives me only > 35MB/s!?. Our experience is that ZFS gets very close to raw performance for streaming reads (assuming that there is adequate CPU and memory available). When doing reads, prefetching (and thus caching) is a critical component of performance. It may be that ZFS's prefetching or caching is misbehaving somehow. Your machine is 32-bit, right? This could be causing some caching pain... How much memory do you have? While you're running the test on ZFS, can you send the output of: echo ::memstat | mdb -k echo ::arc | mdb -k Next, try running your test with prefetch disabled, by putting set zfs:zfs_prefetch_disable=1 in /etc/system and rebooting before running your test. Send the 'iostat -xnpcz' output while this test is running. Finally, on modern drive the streaming performance can vary by up to 2x when reading the outside vs. the inside of the disk. If your pool had been used before you created your test file, it could be laid out on the inside part of the disk. Then you would be comparing raw reads of the outside of the disk vs. zfs reads of the inside of the disk. When the pool is empty, ZFS will start allocating from the outside, so you can try destroying and recreating your pool and creating the file on the fresh pool. Alternatively, create a small partition (say, 10% of the disk size) and do your tests on that to ensure that the file is not far from where your raw reads are going. Let us know how that goes. --matt
_______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss