On Jun 17, 2013, at 1:36 PM, Sebastian Gabler <sequoiamo...@gmx.net> wrote:

> Dear Bill, Peter, Richard, and Saso.
> 
> Thanks for the great comments.
> 
> Now, changing to reverse gear, isn't it more likely to loose data by having a 
> pool that spans across mutiple HBAs than if you connect all drives to a 
> single HBA?

That has an easy answer: yes. But you were asking about data availability, 
which is not
the same as data loss.

> I mean, unless you make sure that there are never any more drives served by 
> one HBA alone (single-ported SATA drives) in a leaf VDEV than can be 
> tolerated by the provided redundancy, a VDEV in the pool could become 
> unavailable upon HBA failure, ultimately leading to loss of the whole pool?

In ZFS, if a pool's redundancy cannot be assured, then the pool goes into the 
FAULTED
state and I/O is suspended until the fault is cleared. In many cases, there is 
no data loss, even
though the data is unavailable while the pool is FAULTED.

In general, HBAs do not have long-lived persistent storage. The data passes 
through them to
the disks. ZFS can recover from the typical failure modes where data is lost 
in-flight without a
commitment to media (see the many posts on cache flush behaviour)

> That is, given that the failure of the HBA would not lead to an immediate 
> crash of the host which would make it identical to the previous scenario. I'd 
> claim that such failures are probably not handled, and so the consequences 
> are not predictable.

I have direct experience where an HBA failed for the root pool. I ran for about 
an hour before I
got bored and rebooted. The HBA failed POST and was replaced. But thanks to ZFS 
awesomeness,
I didn't lose data. But since the only available data to the OS was the data in 
ARC, there wasn't 
much I could do if it required reading a file from disk (eg /usr/bin/ls)

> Similar scenarios are feasible if one disk shelf dies completely, and the 
> pool spans across more than one.

Yes, there are well-known techniques for managing diversity for availability.

> I have personally seen a single vdev in a pool to go down by drive 
> incompatibility, and when I had to decide to give up the pool or try 
> recovery, I got the impression from iostat that there probably had been 
> transactions to the remaining VDEVs, making the recovery forensics. Not sure 
> if this was indeed accurate, but then I was jumping to the conclusion that an 
> immediate, hard crash would have been preferable over a slow melt-down. 
> Prejudice or fact?

I am not familiar with your case, so cannot render an opinion. In my career 
I've seen very few
protected ZFS pool losses that were due to single disk failures. For SVM, LVM, 
and RAID 
controllers... the record is nowhere near as good.
 -- richard

--

richard.ell...@richardelling.com
+1-760-896-4422



_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss

Reply via email to