again i say (eventually) some "zfs sendndmp" type of mechanism seems the right
way to go here *shrug*
-=dave
> Date: Mon, 5 Nov 2007 05:54:15 -0800> From: [EMAIL PROTECTED]> To:
> zfs-discuss@opensolaris.org> Subject: Re: [zfs-discuss] HAMMER> > Peter
> Tribble wrote: > > I'm not worried about the compression effect. Where I see
> problems is> > backing up million/tens of millions of files in a single > >
> dataset. Backing up> > each file is essentially a random read (and this isn't
> helped by raidz> > which gives you a single disks worth of random read I/O
> per vdev). I> > would love to see better ways of backing up huge numbers of
> files.> > It's worth correcting this point... the RAIDZ behavior you mention
> only> occurs if the read size is not aligned to the dataset's block size.
> The> checksum verifier must read the entire stripe to validate the data, but>
> it does that in parallel across the stripe's vdevs. The whole block is> then
> available for delivery to the application.> > Although, backing up
> millions/tens of millions of files in a single> backup dataset is a bad idea
> anyway. The metadata searches will kill> you, no matter what backend
> filesystem is supporting it.> > "zfs send" is the faster way of backing up
> huge numbers of files. But> you pay the price in restore time. (But that's
> the normal tradeoff)> > --Joe>
> _______________________________________________> zfs-discuss mailing list>
> zfs-discuss@opensolaris.org>
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss