Hi, I think your problem is filesystem fragmentation. When available space is less than 40% ZFS might have problems with finding free blocks. Use this script to check it:
#!/usr/sbin/dtrace -s fbt::space_map_alloc:entry { self->s = arg1; } fbt::space_map_alloc:return /arg1 != -1/ { self->s = 0; } fbt::space_map_alloc:return /self->s && (arg1 == -1)/ { @s = quantize(self->s); self->s = 0; } tick-10s { printa(@s); } Run script for few minutes. You might also have problems with space map size. This script will show you size of space map on disk: #!/bin/sh echo '::spa' | mdb -k | grep ACTIVE \ | while read pool_ptr state pool_name do echo "checking pool map size [B]: $pool_name" echo "${pool_ptr}::walk metaslab|::print -d struct metaslab ms_smo.smo_objsize" \ | mdb -k \ | nawk '{sub("^0t","",$3);sum+=$3}END{print sum}' done In memory space map takes 5 times more. All space map is loaded into memory all the time, but for example during snapshot remove all space map might be loaded, so check if you have enough RAM available on machine. Check ::kmastat in mdb. Space map uses kmem_alloc_40 ( on thumpers this is a real problem ) Workaround: 1. first you can change pool recordsize zfs set recordsize=64K POOL Maybe you wil have to use 32K or even 16K 2. You will have to disable ZIL, becuase ZIL always takes 128kB blocks. 3. Try to disable cache, tune vdev cache. Check: http://www.solarisinternals.com/wiki/index.php/ZFS_Evil_Tuning_Guide Lukas Karwacki Dnia 7-11-2007 o godz. 1:49 Walter Faleiro napisał(a): > Hi, > We have a zfs file system configured using a Sunfire 280R with a 10T > Raidweb array > > bash-3.00# zpool list > NAME SIZE USED AVAIL CAP HEALTH ALTROOT > filer 9.44T 6.97T 2.47T 73% ONLINE - > > > bash-3.00# zpool status > pool: backup > state: ONLINE > scrub: none requested > config: > > NAME STATE READ WRITE CKSUM > filer ONLINE 0 0 0 > c1t2d1 ONLINE 0 0 0 > c1t2d2 ONLINE 0 0 0 > c1t2d3 ONLINE 0 0 0 > c1t2d4 ONLINE 0 0 0 > c1t2d5 ONLINE 0 0 0 > > > the file system is shared via nfs. Off late we have seen that the file > system access slows down considerably. Running commands like find, du > on the zfs system did slow it down, but the intermittent slowdowns > cannot be explained. > > Is there a way to trace the I/O on the zfs so that we can list out > heavy read/writes to the file system to be responsible for the > slowness. > > Thanks, > --Walter > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ---------------------------------------------------- Wojna z terrorem wkracza w decydującą fazę: Robert Redford, Meryl Streep i Tom Cruise w filmie UKRYTA STRATEGIA - w kinach od 9 listopada! http://klik.wp.pl/?adr=http%3A%2F%2Fcorto.www.wp.pl%2Fas%2Fstrategia.html&sid=90 _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss