sounds like it is blocking on NFS :-)
Ask Chris for a try/buy DDRdrive X1 or whatever the latest concoction
is... it could be life change for you.
j.
On 5/8/15 11:32 AM, Joe Hetrick wrote:
Today I played a bit with set sync=disabled after watching a few f/s write
IOP's. I can't decide if I've found a particular group of users with a new
(more abusive) set of jobs;
I'm looking more and more, and I've turned sync off on a handful of filesystems
that are showing a high number of write I/O, sustained; when those systems are
bypassing the ZIL, everything is happy. The ZIL devices are never in %w, and
the pool %b coincides with spindle %b, which is almost never higher than 50 or
so; and things are streaming nicely.
Does anyone have any dtrace that I could use to poke into just what the pool is
blocking on when these others are in play? Looking at nfsv3 operations, I see
a very large number of
create
setattr
write
modify
rename
and sometimes remove
and I'm suspecting these users are doing something silly at HPC scale..
Thanks!
Joe
Hi all,
We've recently run into a situation where I'm seeing pool at 90-100 %b,
and our ZIL's at 90-100 %w, yet all of the spindles are relatively idle.
Furthermore, local I/O is normal, and testing is able to quickly and easily put
both pool and spindles in the VDEV into high activity.
The system is primarily accessed via NFS (home server for an HPC
environment). We've had users to evil things before to cause pain, but, this is
most odd, as I would only expect this behavior if we had a faulty device in the
pool with high %b (we don't) or if we had some sort of COW related issue; such as
being <15% free space or so. In this case, we are less than half full of a
108TB raidz3 pool.
latencytop shows a lot of ZFS ZIL Writer latency, but thats to be
expected given what I see above. Pool I/O with zpool iostat is normal-ish, and
as I said, simple raw writes to the pool show expected performance when done
locally.
Does anyone have any ideas?
Thanks,
Joe
--
Joe Hetrick
perl -e 'print pack(h*,a6865647279636b604269647a616e69647f627e2e65647a0)'
BOFH Excuse: doppler effect
_______________________________________________
openindiana-discuss mailing list
openindiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
openindiana-discuss mailing list
openindiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss