Brian McCann wrote:
>      Hi all...I'm having a conceptual problem I can't get around and
> was hoping someone can change my focus here.  I've been backing up
> roughly 6-8 million small files (roughly 2-4k each) using dump, but
> restores take forever due to the huge number of files and directories.
>  Luckily, I haven't had to restore for an emergency yet...but if I
> need to, I'm kinda stuck.  I've looked at distributed file systems
> like CODA, but the number of files I have to deal with will make it
> choke.  Can anyone offer any suggestions?  I've pondered running
> rsync, but am very worried about how long that will take...

Do the files change a lot, or is it more like a few files added/changed
every day, and the bulk don't change?

If it's the latter, you could maybe get best performance from something
like Subversion (a CVS derivative).

Though I suspect rsync would also do well in that case.

If a ton of those files are changing all the time, try doing a test on
creating a tarball and then backing up the tarball.  That may be a simple
managable solution.  There are probably other more complex solutions of
which I am ignorant :-)

-- 
Like Music?
http://l-i-e.com/artists.htm

_______________________________________________
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"

Reply via email to