On Wed, Dec 21, 2005 at 06:36:23PM -0800, Alvin Oga wrote:

> if you don't trust find|tar ... you have major problems with the machine's
> reliability and these brand new commands nobody used for 30 yrs :-)
> 
> using any other "favorite backup programs" will suffer the same fate of
> losing "huge amounts of data", and more importantly, is there a way to
> recover the lost data and/or alternative apps that doesn't have the "bug"
> or just simply fix the hardware ..
> 
> - there is nothing sw can to fix flaky hardware .... and unreliable
>   hardware cannot be used as a means to invalidate "methodology"
> 
>       - good methdologies would already acocunt for the various hundred
>       ways that it can fail in the first place

That's exactly what I'm saying: your tar | gpg methodology has not accounted
for the chance of a few flipped bits, because if it had, it wouldn't lead to
massive data loss, which it does.  Compressing/encrypting after archiving is
inferior to compressing/encrypting before archiving when considering
robustness.  I just can't comprehend how you could dispute that.

> > I'm open to hearing any advantages of tar over afio for backups, because I
> > don't know of even one.
> 
> :-)
> 
> i will bet any amount of $$$ and data .. that find | tar is better than
> the average "backup specific apps" that meets all my backup requirements
> 
> my backup specs
>       - it will NOT corrupt my prev backups, say going back 5 years
>       - it is fast and is live with the simple change of an ip#
>       and untar as needed depending on the purpose of that tar files
>       - confidential data is encrypted and root read only
> 
>       - i can restore to any random data and random time at any
>       time somebody says "prove that it can be done"
> 
>       - it can support 20Terabytes of data in a 4U chassis ... and
>       obvisously, that data is also backed up ... i keep at leaast 
>       3 copies of everything in various state of readiness
> 
>       - it doesn't costs more than the bare costs of the hw in both
>       labor to write or test the "program" and methodology
> 
>       - it must survive a failure of 2 successive full backups
>       ( ie have a work around backup failures )
> 
>       - bare metal restore should be done in a matter of few minutes
>       except that "restore" of 10TB sized data will take a FEW seconds
> 
>       - backup system must also be flexible and extensible and
>       can support 180degree methodology changes
>       ( managers are known to change directions ya know and budgets
>         come and go randomly )
> 
>       - and it obviously has to be searchable
> 
>       - some people like gui's... but i think gui's is for windoze kids
>       
>       - more detailed specs... and semi endless list of major points
> 
>       - find | tar meets all those specs above ...
> 
>       and trivially scriptable and anybody can maintain it since
>       it's not wirtten in martian code, even if it might loook like it
>       after a few dozen people add their $0.01 to it

afio is no more of a "backup specific app" than tar is.  afio has had no more
code changes than tar in the last 5 years.  I'm guessing you don't know
anything about it based on your comments.

I still don't see anything in that list that tar has but afio doesn't.  I *do*
know one thing that afio has that tar doesn't: much greater robustness in the
case of corruption.  Whether you "trust" your hardware or not, it doesn't make
sense to me to choose a less robust solution over a more robust solution.


-- 
To UNSUBSCRIBE, email to [EMAIL PROTECTED] 
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]

Reply via email to