On Mar 2, 2010, at 11:09 AM, Bob Friesenhahn wrote:

> On Tue, 2 Mar 2010, Brian Kolaci wrote:
>> 
>> What is probability of corruption with ZFS in Solaris 10 U6 and up in a SAN 
>> environment?  Have people successfully recovered?
> 
> The probability of corruption in a "SAN environment" depends entirely on your 
> SAN environment.  With proper design, the probability should be "zero".  If 
> it is non-zero then there must be a design defect in your SAN hardware which 
> is liable to do harm at any time (aka "snake in the grass").

The only problem here is that they've already had a few corruptions.  They were 
on U3 however, but I realize a lot of changes went in on U6 and I'm not aware 
of any since then.  Unfortunately I wan't told about them until after they 
rebuilt the pools, and the devices were reused.  All of the corruptions I'm 
aware of were OS disk images on ZFS in the control domain.  One I believe 
happened when they filled the pool because of a snapshot in place, and another 
when the control domain was patched & rebooted while it had guests still 
running.  But I don't know the extent of the corruption whether it was 
pool-wide or just files and/or metadata.  I don't know if anything could have 
been recovered.  I know my only experience with a corrupted pool (which I'm 
still working on) there were 2 files and one dataset found bad.  And a recent 
one was due to someone accidentally adding the same devices to two different 
pools.  I had him quickly copy as much data off as he could (got 5 out of 7
  disk images copied) before it paniced.

So if there is corruption, can it be safely isolated so as to not affect other 
datasets or LDoms?  Or would it be likely to take down the whole pool?

> 
>> In this environment the redundancy is performed at the hardware level, not 
>> at the host.  So there's no chance of self-healing here.  Yes, its been 
>> discussed, but because of the legacy storage environment thats shared with 
>> other non-ZFS systems, they require redundancy at the hardware level, and 
>> they won't budge on that and won't do additional redundancy at the ZFS level.
> 
> That is unfortunate.  In this case, probabilty of failure increases from 
> "zero".

Yes, and need to expect the SA's to do the unexpected.

> 
>> So given the environment, would it be better for lots of small pools, or a 
>> large shared pool?
> 
> I think that a larger shared pool will be more satisfying and less wasteful 
> of resources.  However, a large pool written to a single huge SAN LUN suffers 
> from concurrency issues.  ZFS loses the ability to intelligently schedule I/O 
> for individual disks and instead must use the strategy to post a lot of (up 
> to 35) simultaneous I/Os and hope for the best.

This is what I had written in my document too.  But all the SAN LUNs are only 
33GB (yes, even when they need over a TB there are lots of 33GB LUNs pooled).  
They're trying to get very efficient with storage, which is why I started down 
this path.  But I'm also trying to measure the risk, if any, for moving to a 
larger pool rather than a bunch of smaller per-virtual-host pools.  Regardless 
of the approach, the LUNs are all 33GB and thats the granularity of all 
allocations.


_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to