I want to back up some data over a potentially unreliable link. The ongoing incremental data to back up will be fairly small and the job will run as a perpetual incremental using VirtualFull to synthesize a full backup once a week or so. The initial data though will be 10GB and will take upwards of 4 days to complete at 1mbit/second, which would saturate the uplink - and I'd hate to throw away 3 days of backup just because the link dropped.
If bacula had an option like "job limit = 200MB" where the limit represents the amount of compressed data, and is 'soft' (eg if it is 180MB through a job and starts backing up a 50MB file, the job will complete after that file, not the instant it hits 200MB), it would be as if bacula was seeing the missed files as new files each time the job was run until eventually I get the whole 10GB. It would also allow me to only run the job during the night when saturating the uplink doesn't matter. Would this be that hard to implement? I think it could be as simple as: 1. new option in job definition (not fileset, I think) 2. code to pass the option from the director to the fd 3. code to not back up a file if the limit has been reached, but still mark previously backed up files as seen and detect deleted files (eg you'd still have to do a complete iteration of all the files, just not actually back them up) And would anyone else find this useful? A "job limit = 6 hours" might also work. Thanks James ------------------------------------------------------------------------------ All of the data generated in your IT infrastructure is seriously valuable. Why? It contains a definitive record of application performance, security threats, fraudulent activity, and more. Splunk takes this data and makes sense of it. IT sense. And common sense. http://p.sf.net/sfu/splunk-d2d-c2 _______________________________________________ Bacula-devel mailing list Bacula-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-devel