On Mar 16, 2013, at 7:01 PM, Andrew Werchowiecki <andrew.werchowie...@xpanse.com.au> wrote:
> It's a home set up, the performance penalty from splitting the cache devices > is non-existant, and that work around sounds like some pretty crazy amount of > overhead where I could instead just have a mirrored slog. > > I'm less concerned about wasted space, more concerned about amount of SAS > ports I have available. > > I understand that p0 refers to the whole disk... in the logs I pasted in I'm > not attempting to mount p0. I'm trying to work out why I'm getting an error > attempting to mount p2, after p1 has successfully mounted. Further, this has > been done before on other systems in the same hardware configuration in the > exact same fashion, and I've gone over the steps trying to make sure I > haven't missed something but can't see a fault. You can have only one Solaris partition at a time. Ian already shared the answer, "Create one 100% Solaris partition and then use format to create two slices." -- richard > > I'm not keen on using Solaris slices because I don't have an understanding of > what that does to the pool's OS interoperability. > ________________________________________ > From: Edward Ned Harvey (opensolarisisdeadlongliveopensolaris) > [opensolarisisdeadlongliveopensola...@nedharvey.com] > Sent: Friday, 15 March 2013 8:44 PM > To: Andrew Werchowiecki; zfs-discuss@opensolaris.org > Subject: RE: partioned cache devices > >> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- >> boun...@opensolaris.org] On Behalf Of Andrew Werchowiecki >> >> muslimwookie@Pyzee:~$ sudo zpool add aggr0 cache c25t10d1p2 >> Password: >> cannot open '/dev/dsk/c25t10d1p2': I/O error >> muslimwookie@Pyzee:~$ >> >> I have two SSDs in the system, I've created an 8gb partition on each drive >> for >> use as a mirrored write cache. I also have the remainder of the drive >> partitioned for use as the read only cache. However, when attempting to add >> it I get the error above. > > Sounds like you're probably running into confusion about how to partition the > drive. If you create fdisk partitions, they will be accessible as p0, p1, > p2, but I think p0 unconditionally refers to the whole drive, so the first > partition is p1, and the second is p2. > > If you create one big solaris fdisk parititon and then slice it via > "partition" where s2 is typically the encompassing slice, and people usually > use s1 and s2 and s6 for actual slices, then they will be accessible via s1, > s2, s6 > > Generally speaking, it's unadvisable to split the slog/cache devices anyway. > Because: > > If you're splitting it, evidently you're focusing on the wasted space. > Buying an expensive 128G device where you couldn't possibly ever use more > than 4G or 8G in the slog. But that's not what you should be focusing on. > You should be focusing on the speed (that's why you bought it in the first > place.) The slog is write-only, and the cache is a mixture of read/write, > where it should be hopefully doing more reads than writes. But regardless of > your actual success with the cache device, your cache device will be busy > most of the time, and competing against the slog. > > You have a mirror, you say. You should probably drop both the cache & log. > Use one whole device for the cache, use one whole device for the log. The > only risk you'll run is: > > Since a slog is write-only (except during mount, typically at boot) it's > possible to have a failure mode where you think you're writing to the log, > but the first time you go back and read, you discover an error, and discover > the device has gone bad. In other words, without ever doing any reads, you > might not notice when/if the device goes bad. Fortunately, there's an easy > workaround. You could periodically (say, once a month) script the removal of > your log device, create a junk pool, write a bunch of data to it, scrub it > (thus verifying it was written correctly) and in the absence of any scrub > errors, destroy the junk pool and re-add the device as a slog to the main > pool. > > I've never heard of anyone actually being that paranoid, and I've never heard > of anyone actually experiencing the aforementioned possible undetected device > failure mode. So this is all mostly theoretical. > > Mirroring the slog device really isn't necessary in the modern age. > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- ZFS and performance consulting http://www.RichardElling.com _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss