On Mon, 2010-05-17 at 12:54 -0400, Dan Pritts wrote: > On Mon, May 17, 2010 at 06:25:18PM +0200, Tomas Ögren wrote: > > Resilver does a whole lot of random io itself, not bulk reads.. It reads > > the filesystem tree, not "block 0, block 1, block 2..". You won't get > > 60MB/s sustained, not even close. > > Even with large, unfragmented files? > > danno > -- > Dan Pritts, Sr. Systems Engineer > Internet2 > office: +1-734-352-4953 | mobile: +1-734-834-7224
Having large, unfragmented files will certainly help keep sustained throughput. But, also, you have to consider the amount of deletions done on the pool. For instance, let's say you wrote files A, B, and C one right after another, and they're all big files. Doing a re-silver, you'd be pretty well off on getting reasonable throughput reading A, then B, then C, since they're going to be contiguous on the drive (both internally, and across the three files). However, if you have deleted B at some time, and say wrote a file D (where D < B in size) into B's old space, then, well, you seek to A, read A, seek forward to C, read C, seek back to D, etc. Thus, you'll get good throughput for resilver on these drives pretty much in just ONE case: large files with NO deletions. If you're using them for write-once/read-many/no-delete archives, then you're OK. Anything else is going to suck. :-) -- Erik Trimble Java System Support Mailstop: usca22-123 Phone: x17195 Santa Clara, CA Timezone: US/Pacific (GMT-0800) _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss