I want to back up some data over a potentially unreliable link. The
ongoing incremental data to back up will be fairly small and the job
will run as a perpetual incremental using VirtualFull to synthesize a
full backup once a week or so. The initial data though will be 10GB and
will take upwards of 4 days to complete at 1mbit/second, which would
saturate the uplink - and I'd hate to throw away 3 days of backup just
because the link dropped.

If bacula had an option like "job limit = 200MB" where the limit
represents the amount of compressed data, and is 'soft' (eg if it is
180MB through a job and starts backing up a 50MB file, the job will
complete after that file, not the instant it hits 200MB), it would be as
if bacula was seeing the missed files as new files each time the job was
run until eventually I get the whole 10GB. It would also allow me to
only run the job during the night when saturating the uplink doesn't
matter.

Would this be that hard to implement? I think it could be as simple as:

1. new option in job definition (not fileset, I think)
2. code to pass the option from the director to the fd
3. code to not back up a file if the limit has been reached, but still
mark previously backed up files as seen and detect deleted files (eg
you'd still have to do a complete iteration of all the files, just not
actually back them up)

And would anyone else find this useful? A "job limit = 6 hours" might
also work.

Thanks

James


------------------------------------------------------------------------------
All of the data generated in your IT infrastructure is seriously valuable.
Why? It contains a definitive record of application performance, security 
threats, fraudulent activity, and more. Splunk takes this data and makes 
sense of it. IT sense. And common sense.
http://p.sf.net/sfu/splunk-d2d-c2
_______________________________________________
Bacula-devel mailing list
Bacula-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-devel

Reply via email to