[EMAIL PROTECTED] said: > I have a set of threads each doing random reads to about 25% of its own, > previously written, large file ... a test run will read in about 20GB on a > server with 2GB of RAM > . . . > after several successful runs of my test application, some run of my test > will be running fine, but at some point before it finishes, I see that all IO > to the pool has stopped, and, while I still can use the system for other > things, most operations that involve the pool will also hang (e.g. a > wc on a pool based file will hang)
Bill, Unencumbered by full knowledge of the history of your project, I'll say that I think you need more RAM. I've seen this behavior on a system with 16GB RAM (and no SSD for cache), if heavy I/O goes on long enough. If larger RAM is not feasible, or you don't have a 64-bit CPU, you could try limiting the size of the ARC as well. That's not to say you're not seeing some other issue, but 2GB for heavy ZFS I/O seems a little on the small side, given my experience. Regards, Marion _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss