On 2/29/08, Kent West <[EMAIL PROTECTED]> wrote:
>
> I have a small server on which I need to backup the /home partition.
>
> I have a Barracuda Terastation Pro backup server sitting right next to
> it, connected via Ethernet.
>
> The problem is that the Terastation Pro only offers three connection
> methods: Windows Fileshare (Samba/smb/cifs), Apple Filesharing (AFS),
> and FTP).
>
> I came into the Linux world about the time that FTP was being deprecated
> in favor of SFTP and its variants, so I have a real skittishness of
> using plain FTP. AFS is irrelevant for me. And Samba, whereas slightly
> distasteful, would be okay, except for two problems:
>
> 1. file permissions are not preserved when doing something like rsync, and
> 2. tarballs get truncated at an apparent 2GB limit when using tar.
>
> I don't need anything fancy; just simple and reliable. I had put enough
> effort into learning tar and rsync to make them work (I thought! (is
> this 2GB limit when tarring over smb documented anywhere?)), but then I
> kept running into these show-stoppers as above. I've been very
> frustrated that over the past year of off-and-on "I'm going to get
> serious now and find a solution" I've been unable to find something
> simple(!!!) and reliable. (And by "simple", I mean "easy-to-comprehend
> in two-minutes", not "easy to implement after having mastered every
> command-line switch available".) So I've decided to finally give in and
> ask the big guns on this list.
>
> Before I spend any more effort trying to setup/learn some other system
> (Bacula, Amanda, whatever), do you folks want to give me any suggestions
> as to the best way to proceed?
>
> I want something:
>
> * simple
> * that will back up 10 - 40 GB of /home partition
> * preferably making a full backup every week or so with incrementals
> every day, tossing out old fulls/incrementals as new ones are made
> * that will work over SMB or FTP securely enough that I can stomach it
> * that preserves directory structure/permissions
> * that doesn't run into arbitrary limits like 2GB (or works around them)
> * is automatic, so once set up, I don't have to think about it
> * does not require X or web server installation/tweaking to configure
> * does not require any sort of server piece (other than perhaps an
> /etc/init.d "service" installed as part of a quick and easy "aptitude
> install ..."
> * does not require fancy back-end stuff, like MySQL
>
> I know some of you experts see a solution immediately in using tar or
> rsync, and are thinking, "Well, if Kent had just done his research he'd
> know that if he'd XXX, then YYY...", but that's just it; I'm not a
> full-time researcher of how tar and rsync and Bacula works, and thus I'm
> throwing myself on the mercy of the community for a workable solution.
>
> (I suspect there may be a lot of people like me who knows we need to be
> doing backups but can't find a 2nd-grade-easy system to accomplish the
> task.)
>
> Thanks for any suggestions/help!
>
> --
> Kent
>
>
>
> --
> To UNSUBSCRIBE, email to [EMAIL PROTECTED]
> with a subject of "unsubscribe". Trouble? Contact
> [EMAIL PROTECTED]
>
>
Hey,
  If all is working except for the 2GB file limit (documented or otherwise)
you can just use 'split' to break the archive into smaller parts:
eg:
  `split -b 2000m backup.tar.gz backup.tar.gz.`
or pipe tar straight to it:
   `tar ${your-args} | split -b 2000m - backup.tar.gz.`

  You can then join them by using 'cat':
  `cat backup.tar.gz.* > backup.tar.gz`

cheers,
Owen.

-- If it ain't broke, fix it till it is.

Reply via email to