This sounds like an important problem
> > Hi... > > > > Here's my system: > > > > 2 Intel 3 Ghz 5160 dual-core cpu's > > 0 SATA 750 GB disks running as a ZFS RAIDZ2 pool > > 8 GB Memory > > SunOS 5.11 snv_79a on a separate UFS mirror > > ZFS pool version 10 > > No separate ZIL or ARC cache > > ran into a problem today where the ZFS pool jammed > > for an extended > > eriod of time. During that time, it seemed > > read-bound doing only read > > I/O's (as observed with "zpool iostat 1") and I > saw > > 100% misses while > > running arcstat.pl (for "miss%", "dm%", "pm%" and > > "mm%"). Processes > > accessing the pool were jammed, including remote > NFS > > mounts. At the time, > > I was: 1) running a scrub, 2) writing 10's of > MB/sec > > of data onto the pool > > as well as reading from the pool, 3) was deleting > a > > large number of files > > on the pool. I tried killing one of the jammed > "rm" > > processes and it > > eventually died. The # of misses seen in > arcstat.pl > > eventually dropped > > back down to the 20-40% range ("miss%"). A while > > later, writes began > > occuring to the pool again and remote NFS access > also > > freed up and overall > > system behaviour seemed to normalize. This all > > occurred over the course > > of approximately an hour. > > > > Does this kind of problem sound familiar to > anyone? > > Is it a ZFS problem, > > r have I hit some sort of ZFS load maximum and > this > > is the response? > > Any suggestions for ways to avoid this are > welcome... > > > > Thanks... > > Art > > thur A. Person > > Research Assistant, System Administrator > > Penn State Department of Meteorology > > email: [EMAIL PROTECTED], phone: 814-863-1563 > > _______________________________________________ > > zfs-discuss mailing list > > zfs-discuss@opensolaris.org > > > http://mail.opensolaris.org/mailman/listinfo/zfs-discu > > > ss > > Hi Art, > > I have seen a similar problem that is happening on > several servers since a recent upgrade from b70 to > b86/b87. For no obvious reason, the servers will > stop writing to the pool for long periods of time. > Watching a "zpool iostat", I can see that 0 writes > are being done for up to a minute at a time. > Meanwhile, a large amount of small (~3K) reads are > happening. The servers behave like this for an hour > or more at a time. > > The server configuration is: > Dual-core Opteron 2212HE > 4GB ECC DDR2 RAM > 15 1TB SATA drives in a RAID-Z2 pool > 2 Supermicro SAT2-MV8 controllers > SunOS 5.11 snv_86 > UFS root and swap are on their own disk > > Have you made any progress with this problem? Has > anyone else seen this behavior? This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss