Hi... Here's my system:
2 Intel 3 Ghz 5160 dual-core cpu's 10 SATA 750 GB disks running as a ZFS RAIDZ2 pool 8 GB Memory SunOS 5.11 snv_79a on a separate UFS mirror ZFS pool version 10 No separate ZIL or ARC cache I ran into a problem today where the ZFS pool jammed for an extended period of time. During that time, it seemed read-bound doing only read I/O's (as observed with "zpool iostat 1") and I saw 100% misses while running arcstat.pl (for "miss%", "dm%", "pm%" and "mm%"). Processes accessing the pool were jammed, including remote NFS mounts. At the time, I was: 1) running a scrub, 2) writing 10's of MB/sec of data onto the pool as well as reading from the pool, 3) was deleting a large number of files on the pool. I tried killing one of the jammed "rm" processes and it eventually died. The # of misses seen in arcstat.pl eventually dropped back down to the 20-40% range ("miss%"). A while later, writes began occuring to the pool again and remote NFS access also freed up and overall system behaviour seemed to normalize. This all occurred over the course of approximately an hour. Does this kind of problem sound familiar to anyone? Is it a ZFS problem, or have I hit some sort of ZFS load maximum and this is the response? Any suggestions for ways to avoid this are welcome... Thanks... Art Arthur A. Person Research Assistant, System Administrator Penn State Department of Meteorology email: [EMAIL PROTECTED], phone: 814-863-1563 _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss