Hi Intel X-25 M are MLC not SLC, there are very good for L2ARC.
and next, you need more RAM: ZFS can't handle 4x 80 Gb of L2ARC with only 4Gb of RAM because ZFS use memory to allocate and manage L2ARC. 2010/2/10 Felix Buenemann <felix.buenem...@googlemail.com>: > Am 09.02.10 09:58, schrieb Felix Buenemann: >> >> Am 09.02.10 02:30, schrieb Bob Friesenhahn: >>> >>> On Tue, 9 Feb 2010, Felix Buenemann wrote: >>>> >>>> Well to make things short: Using JBOD + ZFS Striped Mirrors vs. >>>> controller's RAID10, dropped the max. sequential read I/O from over >>>> 400 MByte/s to below 300 MByte/s. However random I/O and sequential >>>> writes seemed to perform >>> >>> Much of the difference is likely that your controller implements true >>> RAID10 wereas ZFS "striped" mirrors are actually load-shared mirrors. >>> Since zfs does not use true striping across vdevs, it relies on >>> sequential prefetch requests to get the sequential read rate up. >>> Sometimes zfs's prefetch is not aggressive enough. >>> >>> I have observed that there may still be considerably more read >>> performance available (to another program/thread) even while a benchmark >>> program is reading sequentially as fast as it can. >>> >>> Try running two copies of your benchmark program at once and see what >>> happens. >> >> Yes, JBOD + ZFS load-balanced mirrors does seem to work better under >> heavy load. I tried rebooting a Windows VM from NFS, which took about 43 >> sec with hot cache in both cases. But when doing this during a bonnie++ >> benchmark run, the ZFS mirrors would win big time, taking just 2:47sec >> instead of over 4min to reboot the VM. >> So I think in a real world scenario, the ZFS mirrors will win. >> >> On a sitenote however I noticed that small sequential I/O (copying a >> 150MB sourcetree to NFS), the ZFS mirrors where 50% slower than the >> controllers RAID10. > > I had a hunch that the controllers volume read ahead would interfere with > the ZFS load-shared mirrors and voilà: sequential reads jumped from 270 > MByte/s to 420 MByte/s, which checks out nicely, because writes are about > 200 MByte/s. > >> >>> Bob >> >> - Felix > > - Felix > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss