On Sun, 18 Jan 2009, Will Murnane wrote:
> That's easy to say, but what if there were no larger alternative?
> Suppose I have a pool composed of those 1.5TB Seagate disks, and
> Hitachi puts out some of the "same" capacity that are actually
> slightly smaller.  A drive fails in my array, I buy a Hitachi disk to
> replace it, and it doesn't work.  If I can't get a large enough drive
> to replace the missing disk with, it'd be a shame to have to destroy
> and recreate the pool on smaller media.

What do you propose that OpenSolaris should do about this?  Should 
OpenSolaris use some sort of a table of "common size" drives, or use 
an algorithm which determines certain discrete usage values based on 
declared drive sizes and a margin for error?  What should OpenSolaris 
of today do with the 20TB disk drives of tomorrow?  What should the 
margin for error of a 30TB disk drive be?  Is it ok to arbitrarily 
ignore 3/4TB of storage space?

If the "drive" is actually a huge 20TB LUN exported from a SAN RAID 
array, how should the margin for error be handled in that case?

Bob
======================================
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,    http://www.GraphicsMagick.org/

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to