Hi...

Here's my system:

    2 Intel 3 Ghz 5160 dual-core cpu's
   10 SATA 750 GB disks running as a ZFS RAIDZ2 pool
    8 GB Memory
      SunOS 5.11 snv_79a on a separate UFS mirror
      ZFS pool version 10
      No separate ZIL or ARC cache

I ran into a problem today where the ZFS pool jammed for an extended 
period of time.  During that time, it seemed read-bound doing only read 
I/O's (as observed with "zpool iostat 1") and I saw 100% misses while 
running arcstat.pl (for "miss%", "dm%", "pm%" and "mm%").  Processes 
accessing the pool were jammed, including remote NFS mounts.  At the time, 
I was: 1) running a scrub, 2) writing 10's of MB/sec of data onto the pool 
as well as reading from the pool, 3) was deleting a large number of files 
on the pool.  I tried killing one of the jammed "rm" processes and it 
eventually died.  The # of misses seen in arcstat.pl eventually dropped 
back down to the 20-40% range ("miss%").  A while later, writes began 
occuring to the pool again and remote NFS access also freed up and overall 
system behaviour seemed to normalize.  This all occurred over the course 
of approximately an hour.

Does this kind of problem sound familiar to anyone?  Is it a ZFS problem, 
or have I hit some sort of ZFS load maximum and this is the response? 
Any suggestions for ways to avoid this are welcome...

                         Thanks...

                           Art

Arthur A. Person
Research Assistant, System Administrator
Penn State Department of Meteorology
email:  [EMAIL PROTECTED], phone:  814-863-1563
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to