That's a very good point - in this particular case, there is no option to change the blocksize for the application.
On 3/9/10 10:42 AM, "Roch Bourbonnais" <roch.bourbonn...@sun.com> wrote: > > I think This is highlighting that there is extra CPU requirement to > manage small blocks in ZFS. > The table would probably turn over if you go to 16K zfs records and > 16K reads/writes form the application. > > Next step for you is to figure how much reads/writes IOPS do you > expect to take in the real workloads and whether or not the filesystem > portion > will represent a significant drain of CPU resource. > > -r > > > Le 8 mars 10 à 17:57, Matt Cowger a écrit : > >> Hi Everyone, >> >> It looks like I¹ve got something weird going with zfs performance on >> a ramdiskS.ZFS is performing not even a 3rd of what UFS is doing. >> >> Short version: >> >> Create 80+ GB ramdisk (ramdiskadm), system has 96GB, so we aren¹t >> swapping >> Create zpool on it (zpool create ramS.) >> Change zfs options to turn off checksumming (don¹t want it or need >> it), atime, compression, 4K block size (this is the applications >> native blocksize) etc. >> Run a simple iozone benchmark (seq. write, seq. read, rndm write, >> rndm read). >> >> Same deal for UFS, replacing the ZFS stuff with newfs stuff and >> mounting the UFS forcedirectio (no point in using a buffer cache >> memory for something that¹s already in memory) >> >> Measure IOPs performance using iozone: >> >> iozone -e -i 0 -i 1 -i 2 -n 5120 -O -q 4k -r 4k -s 5g >> >> With the ZFS filesystem I get around: >> ZFS >> (seq >> write) 42360 (seq read)31010 (random >> read)20953 (random write)32525 >> Not SOO bad, but here¹s UFS: >> UFS >> (seq >> write )42853 (seq read) 100761 (random read) >> 100471 (random write) 101141 >> >> For all tests besides the seq write, UFS utterly destroys ZFS. >> >> I¹m curious if anyone has any clever ideas on why this huge >> disparity in performance exists. At the end of the day, my >> application will run on either filesystem, it just surprises me how >> much worse ZFS performs in this (admittedly edge case) scenario. >> >> --M >> _______________________________________________ >> zfs-discuss mailing list >> zfs-discuss@opensolaris.org >> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss