For Output ops, ZFS could setup  a 10MB I/O transfer to disk
starting at sector  X, or chunk that up  in 128K while still
assigning  the same    range    of  disk   blocks for    the
operations. Yes there will be more control information going
around, a little more  CPU  consumed, but  the disk  will be
streaming all right, I would guess.

Most heavy output load will behave this way with ZFS, random
or not. The throughput  will depend more on the availability
of  contiguous chunk of  disk blocks  that the actual record
size in use.

As for  random input, the  issue is that  ZFS does not get a
say   as to what the  application  is requesting in terms of
size  and location. Yes doing a  4M Input of contiguous disk
block will be faster than  random reading 128K chunks spread
out.  But  if the application  is  manipulating  4M objects,
those will stream out and land on contiguous disk blocks (if
available) and those should stream in as well (if our read
codepath is clever enough).

The fundamental question is really,  is there something that
ZFS  does  that causes data   that represents an application
logical unit of information (likely to  be read as one chunk)
and so that we  would like to have   contiguous on disk,  to
actually spread out everywhere on the platter.

-r




Richard Elling writes:
 > billtodd wrote:
 > > I do want to comment on the observation that "enough concurrent 128K I/O 
 > > can 
 > > saturate a disk" - the apparent implication being that one could therefore 
 > > do 
 > > no better with larger accesses, an incorrect conclusion.  Current disks 
 > > can 
 > > stream out 128 KB in 1.5 - 3 ms., while taking 5.5 - 12.5 ms. for the 
 > > average-seek-plus-partial-rotation required to get to that 128 KB in the 
 > > first 
 > > place.  Thus on a full drive serial random accesses to 128 KB chunks will 
 > > yield 
 > > only about 20% of the drive's streaming capability (by contrast, accessing 
 > > data using serial random accesses in 4 MB contiguous chunks achieves 
 > > around 
 > > 90% of a drive's streaming capability):  one can do better on disks that 
 > > support queuing if one allows queues to form, but this trades 
 > > significantly 
 > > increased average operation latency for the increase in throughput (and 
 > > said 
 > > increase still falls far short of the 90% utilization one could achieve 
 > > using 
 > > 4 MB chunking).
 > > 
 > > Enough concurrent 0.5 KB I/O can also saturate a disk, after all - but 
 > > this 
 > > says little about effective utilization.
 > 
 > I think I can summarize where we are at on this.
 > 
 > This is the classic big-{packet|block|$-line|bikini} versus
 > small-{packet|block|$-line|bikini} argument.  One size won't fit all.
 > 
 > The jury is still out on what all of this means for any given application.
 >   -- richard
 > _______________________________________________
 > zfs-discuss mailing list
 > zfs-discuss@opensolaris.org
 > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to