I am building a 14 disk raid 6 array with 1 TB seagate AS (non-enterprise) 
drives.

So there will be 14 disks total, 2 of them will be parity, 12 TB space 
available.

My drives have a BER of 10^14

I am quite scared by my calculations - it appears that if one drive fails, and 
I do a rebuild, I will perform:

13*8*10^12 = 104000000000000

reads.  But my BER is smaller:

10^14 = 100000000000000

So I am (theoretically) guaranteed to lose another drive on raid rebuild.  Then 
the calculation for _that_ rebuild is:

12*8*10^12 = 96000000000000

So no longer guaranteed, but 96% isn't good.

I have looked all over, and these seem to be the accepted calculations - which 
means if I ever have to rebuild, I'm toast.

But here is the question - the part I am having trouble understanding:

The 13*8*10^12 operations required for the first rebuild .... isn't that the 
number for _the entire array_ ?  Any given 1 TB disk only has 10^12 bits on it 
_total_.  So why would I ever do more than 10^12 operations on the disk ?

It seems very odd to me that a raid controller would have to access any given 
bit more than once to do a rebuild ... and the total number of bits on a drive 
is 10^12, which is far below the 10^14 BER number.

So I guess my question is - why are we all doing this calculation, wherein we 
apply the total operations across an entire array rebuild to a single drives 
BER number ?

Thanks.
 
 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to