On Jun 28, 2010, at 10:03 AM, zfs-discuss-requ...@opensolaris.org wrote:

> Send zfs-discuss mailing list submissions to
>       zfs-discuss@opensolaris.org
> 
> To subscribe or unsubscribe via the World Wide Web, visit
>       http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> or, via email, send a message with subject or body 'help' to
>       zfs-discuss-requ...@opensolaris.org
> 
> You can reach the person managing the list at
>       zfs-discuss-ow...@opensolaris.org
> 
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of zfs-discuss digest..."
> 
> 
> Today's Topics:
> 
>   1. Re: ZFS bug - should I be worried about this? (Gabriele Bulfon)
>   2. Re: ZFS bug - should I be worried about this? (Victor Latushkin)
>   3. Re: OCZ Vertex 2 Pro performance numbers (Frank Cusack)
>   4. Re: ZFS bug - should I be worried about this? (Garrett D'Amore)
>   5. Announce: zfsdump (Tristram Scott)
>   6. Re: Announce: zfsdump (Brian Kolaci)
>   7. Re: zpool import hangs indefinitely (retry post in parts; too
>      long?) (Andrew Jones)
>   8. Re: Announce: zfsdump (Tristram Scott)
>   9. Re: Announce: zfsdump (Brian Kolaci)
> 
> 
> ----------------------------------------------------------------------
> 
> Message: 1
> Date: Mon, 28 Jun 2010 05:16:00 PDT
> From: Gabriele Bulfon <gbul...@sonicle.com>
> To: zfs-discuss@opensolaris.org
> Subject: Re: [zfs-discuss] ZFS bug - should I be worried about this?
> Message-ID: <593812734.121277727391600.javamail.tweb...@sf-app1>
> Content-Type: text/plain; charset=UTF-8
> 
> Yes...they're still running...but being aware that a power failure causing an 
> unexpected poweroff may make the pool unreadable is a pain....
> 
> Yes. Patches should be available.
> Or adoption may be lowering a lot...
> -- 
> This message posted from opensolaris.org
> 
> 
> ------------------------------
> 
> Message: 2
> Date: Mon, 28 Jun 2010 18:14:12 +0400
> From: Victor Latushkin <victor.latush...@sun.com>
> To: Gabriele Bulfon <gbul...@sonicle.com>
> Cc: zfs-discuss@opensolaris.org
> Subject: Re: [zfs-discuss] ZFS bug - should I be worried about this?
> Message-ID: <4c28ae34.1030...@sun.com>
> Content-Type: text/plain; CHARSET=US-ASCII; format=flowed
> 
> On 28.06.10 16:16, Gabriele Bulfon wrote:
>> Yes...they're still running...but being aware that a power failure causing an
>> unexpected poweroff may make the pool unreadable is a pain....
> 
> Pool integrity is not affected by this issue.
> 
> 
> 
> ------------------------------
> 
> Message: 3
> Date: Mon, 28 Jun 2010 07:26:45 -0700
> From: Frank Cusack <frank+lists/z...@linetwo.net>
> To: 'OpenSolaris ZFS discuss' <zfs-discuss@opensolaris.org>
> Subject: Re: [zfs-discuss] OCZ Vertex 2 Pro performance numbers
> Message-ID: <5f1b59775f3ffc0e1781f...@cusack.local>
> Content-Type: text/plain; charset=us-ascii; format=flowed
> 
> On 6/26/10 9:47 AM -0400 David Magda wrote:
>> Crickey. Who's the genius who thinks of these URLs?
> 
> SEOs
> 
> 
> ------------------------------
> 
> Message: 4
> Date: Mon, 28 Jun 2010 08:17:21 -0700
> From: "Garrett D'Amore" <garr...@nexenta.com>
> To: Gabriele Bulfon <gbul...@sonicle.com>
> Cc: zfs-discuss@opensolaris.org
> Subject: Re: [zfs-discuss] ZFS bug - should I be worried about this?
> Message-ID: <1277738241.5596.4325.ca...@velocity>
> Content-Type: text/plain; charset="UTF-8"
> 
> On Mon, 2010-06-28 at 05:16 -0700, Gabriele Bulfon wrote:
>> Yes...they're still running...but being aware that a power failure causing 
>> an unexpected poweroff may make the pool unreadable is a pain....
>> 
>> Yes. Patches should be available.
>> Or adoption may be lowering a lot...
> 
> 
> I don't have access to the information, but if this problem is the same
> one I think it is, then the pool does not become unreadable.  Rather,
> its state after such an event represents a *consistent* state from some
> point of time *earlier* than that confirmed fsync() (or a write on a
> file opened with O_SYNC or O_DSYNC).
> 
> For most users, this is not a critical failing.  For users using
> databases or requiring transactional integrity for data stored on ZFS,
> then yes, this is a very nasty problem indeed.
> 
> I suspect that this is the problem I reported earlier in my blog
> (http://gdamore.blogspot.com) about certain kernels having O_SYNC and
> O_DSYNC problems.  I can't confirm this though, because I don't have
> access to the SunSolve database to read the report.
> 
> (This is something I'll have to check into fixing... it seems like my
> employer ought to have access to that information...)
> 
>       - Garrett
> 
> 
> 
> ------------------------------
> 
> Message: 5
> Date: Mon, 28 Jun 2010 08:26:02 PDT
> From: Tristram Scott <tristram.sc...@quantmodels.co.uk>
> To: zfs-discuss@opensolaris.org
> Subject: [zfs-discuss] Announce: zfsdump
> Message-ID: <311835455.361277738793747.javamail.tweb...@sf-app1>
> Content-Type: text/plain; charset=UTF-8
> 
> For quite some time I have been using zfs send -R fsn...@snapname | dd 
> of=/dev/rmt/1ln to make a tape backup of my zfs file system.  A few weeks 
> back the size of the file system grew to larger than would fit on a single 
> DAT72 tape, and I once again searched for a simple solution to allow dumping 
> of a zfs file system to multiple tapes.  Once again I was disappointed...
> 
> I expect there are plenty of other ways this could have been handled, but 
> none leapt out at me.  I didn't want to pay large sums of cash for a 
> commercial backup product, and I didn't see that Amanda would be an easy 
> thing to fit into my existing scripts.  In particular, (and I could well be 
> reading this incorrectly) it seems that the commercial products, Amanda, 
> star, all are dumping the zfs file system file by file (with or without 
> ACLs).  I found none which would allow me to dump the file system and its 
> snapshots, unless I used zfs send to a scratch disk, and dumped to tape from 
> there.  But, of course, that assumes I have a scratch disk large enough.
> 
> So, I have implemented zfsdump as a ksh script.  The method is as follows:
> 1. Make a bunch of fifos.
> 2. Pipe the stream from zfs send to split, with split writing to the fifos 
> (in sequence).
> 3. Use dd to copy from the fifos to tape(s).
> 
> When the first tape is complete, zfsdump returns.  One then calls it again, 
> specifying that the second tape is to be used, and so on.
> 
>> From the man page:
> 
>     Example 1.  Dump the @Tues snapshot of the  tank  filesystem
>     to  the  non-rewinding,  non-compressing  tape,  with a 36GB
>     capacity:
> 
>          zfsdump -z t...@tues -a "-R" -f /dev/rmt/1ln  -s  36864 -t 0
> 
>     For the second tape:
> 
>          zfsdump -z t...@tues -a "-R" -f /dev/rmt/1ln  -s  36864 -t 1
> 
> If you would like to try it out, download the package from:
> http://www.quantmodels.co.uk/zfsdump/
> 
> I have packaged it up, so do the usual pkgadd stuff to install.
> 
> Please, though, [b]try this out with caution[/b].  Build a few test file 
> systems, and see that it works for you. 
> [b]It comes without warranty of any kind.[/b]
> 
> 
> Tristram
> -- 
> This message posted from opensolaris.org
> 
> 
> ------------------------------
> 
> Message: 6
> Date: Mon, 28 Jun 2010 11:51:07 -0400
> From: Brian Kolaci <brian.kol...@sun.com>
> To: Tristram Scott <tristram.sc...@quantmodels.co.uk>
> Cc: ZFS Discussions <zfs-discuss@opensolaris.org>
> Subject: Re: [zfs-discuss] Announce: zfsdump
> Message-ID: <4c28c4eb.2010...@sun.com>
> Content-Type: text/plain; CHARSET=US-ASCII; format=flowed
> 
> I use Bacula which works very well (much better than Amanda did).
> You may be able to customize it to do direct zfs send/receive, however I find 
> that although they are great for copying file systems to other machines, they 
> are inadequate for backups unless you always intend to restore the whole file 
> system.  Most people want to restore a file or directory tree of files, not a 
> whole file system.  In the past 25 years of backups and restores, I've never 
> had to restore a whole file system.  I get requests for a few files, or 
> somebody's mailbox or somebody's http document root.
> You can directly install it from CSW (or blastwave).
> 
> On 6/28/2010 11:26 AM, Tristram Scott wrote:
>> For quite some time I have been using zfs send -R fsn...@snapname | dd 
>> of=/dev/rmt/1ln to make a tape backup of my zfs file system.  A few weeks 
>> back the size of the file system grew to larger than would fit on a single 
>> DAT72 tape, and I once again searched for a simple solution to allow dumping 
>> of a zfs file system to multiple tapes.  Once again I was disappointed...
>> 
>> I expect there are plenty of other ways this could have been handled, but 
>> none leapt out at me.  I didn't want to pay large sums of cash for a 
>> commercial backup product, and I didn't see that Amanda would be an easy 
>> thing to fit into my existing scripts.  In particular, (and I could well be 
>> reading this incorrectly) it seems that the commercial products, Amanda, 
>> star, all are dumping the zfs file system file by file (with or without 
>> ACLs).  I found none which would allow me to dump the file system and its 
>> snapshots, unless I used zfs send to a scratch disk, and dumped to tape from 
>> there.  But, of course, that assumes I have a scratch disk large enough.
>> 
>> So, I have implemented zfsdump as a ksh script.  The method is as follows:
>> 1. Make a bunch of fifos.
>> 2. Pipe the stream from zfs send to split, with split writing to the fifos 
>> (in sequence).
>> 3. Use dd to copy from the fifos to tape(s).
>> 
>> When the first tape is complete, zfsdump returns.  One then calls it again, 
>> specifying that the second tape is to be used, and so on.
>> 
>> From the man page:
>> 
>>      Example 1.  Dump the @Tues snapshot of the  tank  filesystem
>>      to  the  non-rewinding,  non-compressing  tape,  with a 36GB
>>      capacity:
>> 
>>           zfsdump -z t...@tues -a "-R" -f /dev/rmt/1ln  -s  36864 -t 0
>> 
>>      For the second tape:
>> 
>>           zfsdump -z t...@tues -a "-R" -f /dev/rmt/1ln  -s  36864 -t 1
>> 
>> If you would like to try it out, download the package from:
>> http://www.quantmodels.co.uk/zfsdump/
>> 
>> I have packaged it up, so do the usual pkgadd stuff to install.
>> 
>> Please, though, [b]try this out with caution[/b].  Build a few test file 
>> systems, and see that it works for you.
>> [b]It comes without warranty of any kind.[/b]
>> 
>> 
>> Tristram
> 
> 
> 
> ------------------------------
> 
> Message: 7
> Date: Mon, 28 Jun 2010 08:59:21 PDT
> From: Andrew Jones <andrewnjo...@gmail.com>
> To: zfs-discuss@opensolaris.org
> Subject: Re: [zfs-discuss] zpool import hangs indefinitely (retry post
>       in parts; too long?)
> Message-ID: <185781013.401277740792036.javamail.tweb...@sf-app1>
> Content-Type: text/plain; charset=UTF-8
> 
> Now at 36 hours since zdb process start and:
> 
> 
> PID USERNAME  SIZE   RSS STATE  PRI NICE      TIME  CPU PROCESS/NLWP
>   827 root     4936M 4931M sleep   59    0   0:50:47 0.2% zdb/209
> 
> Idling at 0.2% processor for nearly the past 24 hours... feels very stuck. 
> Thoughts on how to determine where and why?
> -- 
> This message posted from opensolaris.org
> 
> 
> ------------------------------
> 
> Message: 8
> Date: Mon, 28 Jun 2010 09:26:12 PDT
> From: Tristram Scott <tristram.sc...@quantmodels.co.uk>
> To: zfs-discuss@opensolaris.org
> Subject: Re: [zfs-discuss] Announce: zfsdump
> Message-ID: <1916741256.441277742403278.javamail.tweb...@sf-app1>
> Content-Type: text/plain; charset=UTF-8
> 
>> I use Bacula which works very well (much better than
>> Amanda did).
>> You may be able to customize it to do direct zfs
>> send/receive, however I find that although they are
>> great for copying file systems to other machines,
>> they are inadequate for backups unless you always
>> intend to restore the whole file system.  Most people
>> want to restore a file or directory tree of files,
>> not a whole file system.  In the past 25 years of
>> backups and restores, I've never had to restore a
>> whole file system.  I get requests for a few files,
>> or somebody's mailbox or somebody's http document
>> root.
>> You can directly install it from CSW (or blastwave).
> 
> Thanks for your comments, Brian.  I should look at Bacula in more detail.
> 
> As for full restore versus ad hoc requests for files I just deleted, my 
> experience is mostly similar to yours, although I have had need for full 
> system restore more than once.
> 
> For the restore of a few files here and there, I believe this is now well 
> handled with zfs snapshots.  I have always found these requests to be down to 
> human actions.  The need for full system restore has (almost) always been 
> hardware failure. 
> 
> If the file was there an hour ago, or yesterday, or last week, or last month, 
> then we have it in a snapshot.
> 
> If the disk died horribly during a power outage (grrr!) then it would be very 
> nice to be able to restore not only the full file system, but also the 
> snapshots too.  The only way I know of achieving that is by using zfs send 
> etc.  
> 
>> 
>> On 6/28/2010 11:26 AM, Tristram Scott wrote:
> [snip]
> 
>>> 
>>> Tristram
>> 
>> _______________________________________________
>> zfs-discuss mailing list
>> zfs-discuss@opensolaris.org
>> http://mail.opensolaris.org/mailman/listinfo/zfs-discu
>> ss
>> 
> -- 
> This message posted from opensolaris.org
> 
> 
> ------------------------------
> 
> Message: 9
> Date: Mon, 28 Jun 2010 13:00:06 -0400
> From: Brian Kolaci <brian.kol...@sun.com>
> To: Tristram Scott <tristram.sc...@quantmodels.co.uk>
> Cc: zfs-discuss@opensolaris.org
> Subject: Re: [zfs-discuss] Announce: zfsdump
> Message-ID: <eb3a216e-dfd0-40fa-932c-65a8aee4f...@sun.com>
> Content-Type: text/plain; CHARSET=US-ASCII
> 
> 
> On Jun 28, 2010, at 12:26 PM, Tristram Scott wrote:
> 
>>> I use Bacula which works very well (much better than
>>> Amanda did).
>>> You may be able to customize it to do direct zfs
>>> send/receive, however I find that although they are
>>> great for copying file systems to other machines,
>>> they are inadequate for backups unless you always
>>> intend to restore the whole file system.  Most people
>>> want to restore a file or directory tree of files,
>>> not a whole file system.  In the past 25 years of
>>> backups and restores, I've never had to restore a
>>> whole file system.  I get requests for a few files,
>>> or somebody's mailbox or somebody's http document
>>> root.
>>> You can directly install it from CSW (or blastwave).
>> 
>> Thanks for your comments, Brian.  I should look at Bacula in more detail.
>> 
>> As for full restore versus ad hoc requests for files I just deleted, my 
>> experience is mostly similar to yours, although I have had need for full 
>> system restore more than once.
>> 
>> For the restore of a few files here and there, I believe this is now well 
>> handled with zfs snapshots.  I have always found these requests to be down 
>> to human actions.  The need for full system restore has (almost) always been 
>> hardware failure. 
>> 
>> If the file was there an hour ago, or yesterday, or last week, or last 
>> month, then we have it in a snapshot.
>> 
>> If the disk died horribly during a power outage (grrr!) then it would be 
>> very nice to be able to restore not only the full file system, but also the 
>> snapshots too.  The only way I know of achieving that is by using zfs send 
>> etc.  
>> 
> 
> I like snapshots when I'm making a major change to the system or for cloning. 
>  So to me, snapshots are good for transaction based operations.  Such as 
> stopping & flushing a database, take a snapshot, then resume the database.  
> Then you can back up the snapshot with Bacula and destroy the snapshot when 
> the backup is complete.  I have Bacula configured with a pre-backup and 
> post-backup scripts to do just that.  When you do the restore, it will create 
> something that "looks" like a snapshot from the file system perspective, but 
> isn't really one.
> 
> But if you're looking for a copy of a file from a specific date, Bacula 
> retains that.  In fact you specify the retention period you want and you'll 
> have access to any/all individual files on a per date basis.  You can retain 
> the files for months or years if you like, and you specify that in the Bacula 
> config file as to how long you want to keep the tapes around.  So it really 
> comes down to your use-case.
> 
> 
> 
> 
> ------------------------------
> 
> _______________________________________________
> zfs-discuss mailing list
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> 
> 
> End of zfs-discuss Digest, Vol 56, Issue 126
> ********************************************

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to