Hi guys, after reading the mailings yesterday i noticed someone was after upgrading to zfs v21 (deduplication) i'm after the same, i installed osol-dev-127 earlier which comes with v19 and then followed the instructions on http://pkg.opensolaris.org/dev/en/index.shtml to bring my system up to date, however, the system is reporting no updates are available and stays at zfs v19, any ideas?
On 17 Nov 2009, at 10:28, zfs-discuss-requ...@opensolaris.org wrote: > Send zfs-discuss mailing list submissions to > zfs-discuss@opensolaris.org > > To subscribe or unsubscribe via the World Wide Web, visit > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > or, via email, send a message with subject or body 'help' to > zfs-discuss-requ...@opensolaris.org > > You can reach the person managing the list at > zfs-discuss-ow...@opensolaris.org > > When replying, please edit your Subject line so it is more specific > than "Re: Contents of zfs-discuss digest..." > > > Today's Topics: > > 1. Re: permanent files error, unable to access pool > (Cindy Swearingen) > 2. Re: hung pool on iscsi (Tim Cook) > 3. Re: hung pool on iscsi (Richard Elling) > 4. Re: hung pool on iscsi (Jacob Ritorto) > 5. Re: permanent files error, unable to access pool > (daniel.rodriguez.delg...@gmail.com) > 6. building zpools on device aliases (sean walmsley) > 7. Re: Best config for different sized disks (Erik Trimble) > 8. [Fwd: [zfs-auto-snapshot] Heads up: SUNWzfs-auto-snapshot > obsoletion in snv 128] (Tim Foster) > > > ---------------------------------------------------------------------- > > Message: 1 > Date: Mon, 16 Nov 2009 15:46:26 -0700 > From: Cindy Swearingen <cindy.swearin...@sun.com> > To: "daniel.rodriguez.delg...@gmail.com" > <daniel.rodriguez.delg...@gmail.com> > Cc: zfs-discuss@opensolaris.org > Subject: Re: [zfs-discuss] permanent files error, unable to access > pool > Message-ID: <4b01d642.4010...@sun.com> > Content-Type: text/plain; CHARSET=US-ASCII; format=flowed > > Hi Daniel, > > In some cases, when I/O is suspended, permanent errors are logged and > you need to run a zpool scrub to clear the errors. > > Are you saying that a zpool scrub cleared the errors that were > displayed in the zpool status output? Or, did you also use zpool > clear? > > Metadata is duplicated even in a one-device pool but recovery > must depend on the severity of metadata errors. > > Thanks, > > Cindy > > On 11/16/09 13:18, daniel.rodriguez.delg...@gmail.com wrote: >> Thanks Cindy, >> >> In fact, after some research, I ran into the scrub suggestion and it worked >> perfectly. Now I think that the automated message in >> http://www.sun.com/msg/ZFS-8000-8A should mention something about scrub as a >> worthy attempt. >> >> It was related to an external usb disk. I guess I am happy it happened now >> before I invested in getting a couple of other external disks as mirrors of >> the existing one. I guess I am better off installing an extra internal disk. >> >> is this something common on usb disks? would it get improved in later >> versions of osol or it is somewhat of an incompatibility/unfriendliness of >> zfs with external usb disks? > > > ------------------------------ > > Message: 2 > Date: Mon, 16 Nov 2009 17:04:48 -0600 > From: Tim Cook <t...@cook.ms> > To: Martin Vool <mardi...@gmail.com> > Cc: zfs-discuss@opensolaris.org > Subject: Re: [zfs-discuss] hung pool on iscsi > Message-ID: > <fbef46640911161504y6e193b32obe3d7bebaf2d9...@mail.gmail.com> > Content-Type: text/plain; charset="iso-8859-1" > > On Mon, Nov 16, 2009 at 4:00 PM, Martin Vool <mardi...@gmail.com> wrote: > >> I already got my files back acctuay and the disc contains already new >> pools, so i have no idea how it was set. >> >> I have to make a virtualbox installation and test it. >> Can you please tell me how-to set the failmode? >> >> >> > > http://prefetch.net/blog/index.php/2008/03/01/configuring-zfs-to-gracefully-deal-with-failures/ > -------------- next part -------------- > An HTML attachment was scrubbed... > URL: > <http://mail.opensolaris.org/pipermail/zfs-discuss/attachments/20091116/f51a8a6e/attachment-0001.html> > > ------------------------------ > > Message: 3 > Date: Mon, 16 Nov 2009 15:13:49 -0800 > From: Richard Elling <richard.ell...@gmail.com> > To: Martin Vool <mardi...@gmail.com> > Cc: zfs-discuss@opensolaris.org > Subject: Re: [zfs-discuss] hung pool on iscsi > Message-ID: <7adfc8e2-3be0-48e6-8d5a-506d975a2...@gmail.com> > Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes > > > On Nov 16, 2009, at 2:00 PM, Martin Vool wrote: > >> I already got my files back acctuay and the disc contains already >> new pools, so i have no idea how it was set. >> >> I have to make a virtualbox installation and test it. > > Don't forget to change VirtualBox's default cache flush setting. > http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide#OpenSolaris.2FZFS.2FVirtual_Box_Recommendations > -- richard > > > > > ------------------------------ > > Message: 4 > Date: Mon, 16 Nov 2009 18:22:17 -0500 > From: Jacob Ritorto <jacob.rito...@gmail.com> > To: zfs-discuss@opensolaris.org > Subject: Re: [zfs-discuss] hung pool on iscsi > Message-ID: > <1f3f8f1d0911161522u55db284bw5668cbe48f321...@mail.gmail.com> > Content-Type: text/plain; charset=ISO-8859-1 > > On Mon, Nov 16, 2009 at 4:49 PM, Tim Cook <t...@cook.ms> wrote: >> Is your failmode set to wait? > > Yes. This box has like ten prod zones and ten corresponding zpools > that initiate to iscsi targets on the filers. We can't panic the > whole box just because one {zone/zpool/iscsi target} fails. Are there > undocumented commands to reset a specific zpool or something? > > thx > jake > > > ------------------------------ > > Message: 5 > Date: Mon, 16 Nov 2009 15:53:50 PST > From: "daniel.rodriguez.delg...@gmail.com" > <daniel.rodriguez.delg...@gmail.com> > To: zfs-discuss@opensolaris.org > Subject: Re: [zfs-discuss] permanent files error, unable to access > pool > Message-ID: <317370742.281258415660371.javamail.tweb...@sf-app1> > Content-Type: text/plain; charset=UTF-8 > > to be the best of my recollection, I only needed to run zfs scrub, reboot and > the disk became operational again.... > > the irony was that the error message was asking me to recover from backup, > but the disk involved was my backup of my working pool..... > -- > This message posted from opensolaris.org > > > ------------------------------ > > Message: 6 > Date: Mon, 16 Nov 2009 18:17:39 PST > From: sean walmsley <s...@fpp.nuclearsafetysolutions.com> > To: zfs-discuss@opensolaris.org > Subject: [zfs-discuss] building zpools on device aliases > Message-ID: <1647062936.311258424289885.javamail.tweb...@sf-app1> > Content-Type: text/plain; charset=UTF-8 > > We have a number of Sun J4200 SAS JBOD arrays which we have multipathed using > Sun's MPxIO facility. While this is great for reliability, it results in the > /dev/dsk device IDs changing from cXtYd0 to something virtually unreadable > like "c4t5000C5000B21AC63d0s3". > > Since the entries in /dev/{rdsk,dsk} are simply symbolic links anyway, would > there be any problem with adding "alias" links to /devices there and building > our zpools on them? We've tried this and it seems to work fine producing a > zpool status similar to the following: > > ... > NAME STATE READ WRITE CKSUM > vol01 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > top00 ONLINE 0 0 0 > bot00 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > top01 ONLINE 0 0 0 > bot01 ONLINE 0 0 0 > ... > > Here our aliases are "topnn" and "botnn" to denote the disks in the top and > bottom JBODs. > > The obvious question is "what happens if the alias link disappears?". We've > tested this, and ZFS seems to handle it quite nicely by finding the "normal" > /dev/dsk link and simply working with that (although it's more difficult to > get ZFS to use the alias again once it is recreated). > > If anyone can think of anything really nasty that we've missed, we'd > appreciate knowing about it. Alternatively, if there is a better supported > means of having ZFS display human-readable device ids we're all ears :-) > > Perhaps an MPxIO RFE for "vanity" device names would be in order? > -- > This message posted from opensolaris.org > > > ------------------------------ > > Message: 7 > Date: Tue, 17 Nov 2009 00:51:05 -0800 > From: Erik Trimble <erik.trim...@sun.com> > To: Tim Cook <t...@cook.ms> > Cc: Les Pritchard <les.pritch...@gmail.com>, > zfs-discuss@opensolaris.org > Subject: Re: [zfs-discuss] Best config for different sized disks > Message-ID: <4b0263f9.1060...@sun.com> > Content-Type: text/plain; CHARSET=US-ASCII; format=flowed > > Tim Cook wrote: >> >> >> On Mon, Nov 16, 2009 at 12:09 PM, Bob Friesenhahn >> <bfrie...@simple.dallas.tx.us <mailto:bfrie...@simple.dallas.tx.us>> >> wrote: >> >> On Sun, 15 Nov 2009, Tim Cook wrote: >> >> >> Once again I question why you're wasting your time with >> raid-z. You might as well just stripe across all the drives. >> You're taking a performance penalty for a setup that >> essentially has 0 redundancy. You lose a 500gb drive, you >> lose everything. >> >> >> Why do you say that this user will lose everything? The two >> concatenated/striped devices on the local system are no different >> than if they were concatenated on SAN array and made available as >> one LUN. If one of those two drives fails, then it would have the >> same effect as if one larger drive failed. >> >> Bob >> >> >> Can I blame it on too many beers? I was thinking losing half of one >> drive, rather than an entire vdev would just cause "weirdness" in the >> pool, rather than a clean failure. I suppose without experimentation >> there's no way to really no, in theory though, zfs should be able to >> handle it. >> >> --Tim >> > Back to the original question: the "concat using SVM" method works OK > if the disk you have are all integer multiples of each other (that is, > this worked because he had 2 500GB drives to make a 1TB drive out of). > It certainly seems the best method - both for performance and maximum > disk space - that I can think of. However, it won't work well in other > cases: i.e. a couple of 250GB drives, and a couple of 1.5TB drives. > > In cases of serious mis-match between the drive sizes, especially when > there's not a real good way to concat to get a metadrive big enough to > match others, I'd recommend going for multiple zpools, and slicing up > the bigger drives to allow for RAIDZ-ing with the smaller ones "natively". > > E.g. > > let's say you have 3 250GB drives, and 3 1.5TB drives. You could > partition the 1.5TB drives into 250GB and 1.25TB partitions, and then > RAIDZ the 3 250GB drives together, plus the 250GB partitions as one > zpool, then the 1.25TB partitions as another zpool. > > You'll have some problems with contending I/O if you try to write to > both zpools at once, but it's the best way I can think of to maximize > space and at the same time maximize performance for single-pool I/O. > > I think it would be a serious performance mistake to combine the two > pools as vdevs in a single pool, though it's perfectly possible. > > I.e. > (preferred) > zpool create smalltank raidz c0t0d0 c0t1d0 c0t2d0 c1t0d0s0 c1t1d0s0 c1t2d0s0 > zpool create largetank raidz c1t0d0s1 c1t1d0s1 c1t2d0s1 > > instead of > > zpool create supertank raidz c0t0d0 c0t1d0 c0t2d0 c1t0d0s0 c1t1d0s0 > c1t2d0s0 raidz c1t0d0s1 c1t1d0s1 c1t2d0s1 > > > > -- > Erik Trimble > Java System Support > Mailstop: usca22-123 > Phone: x17195 > Santa Clara, CA > Timezone: US/Pacific (GMT-0800) > > > > ------------------------------ > > Message: 8 > Date: Tue, 17 Nov 2009 10:27:38 +0000 > From: Tim Foster <tim.fos...@sun.com> > To: zfs-discuss <zfs-discuss@opensolaris.org> > Subject: [zfs-discuss] [Fwd: [zfs-auto-snapshot] Heads up: > SUNWzfs-auto-snapshot obsoletion in snv 128] > Message-ID: <1258453658.2502.28.ca...@igyo> > Content-Type: text/plain; CHARSET=US-ASCII > > Hi all, > > Just forwarding Niall's heads-up message about the impending removal of > the existing zfs-auto-snapshot implementation in nv_128 > > I've not been involved in the rewrite, but what I've read about the new > code, it'll be a lot more efficient than the old ksh-based code, and > will fix many of the problems with the old solution. > > The new code currently lacks support for the 'zfs/backup' functionality > - which allowed you to specify a command to tell the service to generate > incremental or full send streams at each time interval, along with a > commandline to process that stream (sshing to a remote server and doing > a zfs recv, for example) > > If you do use that functionality, it'd be good to drop a mail to the > thread[1] on the zfs-auto-snapshot alias. > > > It's been a wild ride, but my work on zfs-auto-snapshot is done I > think :-) > > cheers, > tim > > [1] > http://mail.opensolaris.org/pipermail/zfs-auto-snapshot/2009-November/thread.html#199 > > -------- Forwarded Message -------- > From: Niall Power <niall.po...@sun.com> > To: zfs-auto-snaps...@opensolaris.org > Subject: [zfs-auto-snapshot] Heads up: SUNWzfs-auto-snapshot obsoletion > in snv 128 > Date: Mon, 16 Nov 2009 18:26:28 +1100 > > This is a heads up for user of Tim's zfs-auto-snapshot auto snapshot > service currently delivered > in Solaris Nevada and OpenSolaris. As of build 128 the zfs-auto-snapshot > scripts are being replaced > by a rewritten daemon (time-sliderd). > Time-sliderd will continue to support the existing SMF > auto-snapshot:<schedule> service instances > as it's configuration mechanism so for most users there should be no > significant differences noticeable. > Some of the options will no longer be supported however since they are > either obsolete or too > specifically tied to the zfs-auto-snapshot implementation to make them > portable. > > Stuff that will work: > - Default out of the box schedules (frequent, hourly, daily, weekly, > monthly) > - Custom schedules > > SMF properties that will be supported: > - interval > - period > - keep > > SMF properties that won't be supported > - "offset" The time-slider daemon schedules snapshots based on relative > times rather that absolute times which makes it more suitable for systems > that do not have constant 24/7 uptime so this feature isn't so relevant > anymore (it only got implemented recently in zfs-auto-snapshot also) > > - "label" Dropping it allows naming shemes to be simplified and > enforces uniqueness when SMF tries to import an auto-snapshot instance > > - backup/backup-save-cmd/backup-lock > We plan to implement an rsync based backup mechanism that allows backups > to be browsed like a proper filesystem instead of having binary > snapshot blobs > that are explicitly classified as unstable/volatile by zfs(1) > > For those who want to use time-slider without going through the GUI, simply > enable/configure (or create) the auto-snapshot instances you need then > enable > the time-slider SMF service. time-slider will pick up the enabled > auto-snapshot > instances and schedule snapshots for them. > > For folks who prefer to continue using zfs-auto-snapshot, you will need to > remove SUNWgnome-time-slider and install the existing zfs-auto-snapshot > packages instead. > > Comments/questions welcome ;) > > Cheers, > Niall. > > _______________________________________________ > zfs-auto-snapshot mailing list > zfs-auto-snaps...@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-auto-snapshot > > > > > ------------------------------ > > _______________________________________________ > zfs-discuss mailing list > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > > End of zfs-discuss Digest, Vol 49, Issue 78 > ******************************************* _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss