On Thu, 2010-06-03 at 12:22 -0400, Dennis Clarke wrote: > > If you're clever, you'll also try to make sure each side of the mirror > > is on a different controller, and if you have enough controllers > > available, you'll also try to balance the controllers across stripes. > > Something like this ? > > # zpool status fibre0 > pool: fibre0 > state: ONLINE > status: The pool is formatted using an older on-disk format. The pool can > still be used, but some features are unavailable. > action: Upgrade the pool using 'zpool upgrade'. Once this is done, the > pool will no longer be accessible on older software versions. > scrub: none requested > config: > > NAME STATE READ WRITE CKSUM > fibre0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c2t16d0 ONLINE 0 0 0 > c5t0d0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c5t1d0 ONLINE 0 0 0 > c2t17d0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c5t2d0 ONLINE 0 0 0 > c2t18d0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c2t20d0 ONLINE 0 0 0 > c5t4d0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c2t21d0 ONLINE 0 0 0 > c5t6d0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c2t19d0 ONLINE 0 0 0 > c5t5d0 ONLINE 0 0 0 > spares > c2t22d0 AVAIL > > errors: No known data errors
That looks like a good configuration to me! > > However, unlike the bad old days of SVM ( DiskSuite or Solstice Disksuite > or Online Disk Suite etc ) I have no idea what algorithm is used to pick > the hot spare in the event of a failure. I mean, if I had more than one > hotspare there of course. Also, I think the weird order of controllers is > a user mistake on my part. Some of them have c5 listed first and others > have c2 listed first. I don't know if that matters at all however. I don't think the order matters. It certainly won't make a difference for write, since you have to use both sides of the mirror. It *could* make a difference for reading... but I suspect that zfs will try to sufficiently balance things out that any difference in ordering will be lost in the noise. The hotspare replacement shouldn't matter all that much... except when you're using them to upgrade to bigger drives. (Then just use a single hot spare to force the selection.) > > I can add mirrors on the fly but I can not ( yet ) remove them. I would > imagine that the algorithm to remove data from vdevs would be fairly > gnarly. Indeed. However, you shouldn't need to remove vdevs. With redundancy, you can resilver to a hot spare, and then remove the drive, but ultimately, you can't condense the data onto fewer drives. You have to accept when you configure your array that, modulo hot spares, your pool will always consume the same number of spindles. > > The item that I find somewhat confusing is how to apply multi-path fibre > devices to a stripe of mirrors. Consider these : > > > This is not disk redundency but rather fibre path redundency. When I drop > these guys into a ZPool it looks like this : > > NAME STATE READ WRITE CKSUM > fp0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c4t20000004CFBFD4BDd0s0 ONLINE 0 0 0 > c4t20000004CFA4D3A1d0s0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c4t20000004CFA4D2D9d0s0 ONLINE 0 0 0 > c4t20000004CFA4D2C7d0s0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c4t20000004CFA4D655d0s0 ONLINE 0 0 0 > c4t20000004CF9B63D0d0s0 ONLINE 0 0 0 The above configuration looks good to me. > > So the manner in which any given IO transaction gets to the zfs filesystem > just gets ever more complicated and convoluted and it makes me wonder if I > am tossing away performance to get higher levels of safety. If you're using multipathing, then you get path load balancing automatically, and you can pretty much ignore the controller balancing issue, as long as you use the mpxio (scsi_vhci) path. mpxio should take care of ensuring that I/O is balanced across ports for you; you just need to make sure that you are balancing *spindles* properly. -- Garrett _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss