On 6/27/2015 1:37 AM, Andrew Noonan wrote: > On Fri, Jun 26, 2015 at 2:17 PM, Ana Emília M. Arruda > <emiliaarr...@gmail.com> wrote: >> Are you going to generate a .tar of about 250TB every day? Which will >> be the nature of your restores? You´re going to need always the >> restore of the whole data set or occasionally you will need to >> restore a small set of files? > After the backfill, I expect the daily backups to be in the hundreds > of gigs a day range. During the backfill, I'll want to maximize > writes to catch up as soon as I can. In general, we already have a > copy of the data, but given how it syncs, this doesn't protect against > an accidental "rm" getting synced downstream, so I'd imagine other > then for testing purposes, restores will be for accidental deletes or > for some sort of massive disaster situation. This is why I figure the > backfill should be close to the size of the tape, though I'm not 100%.
Keep in mind that one file spanning N tapes is N times more likely to fail. The loss of a single tape results in the loss of the file. For this reason, I prefer to break large data sets like that into N different backup jobs so that a job's full backup fits on a single tape, if at all possible. Tapes do sometimes fail after they have been successfully written. ------------------------------------------------------------------------------ Monitor 25 network devices or servers for free with OpManager! OpManager is web-based network management software that monitors network devices and physical & virtual servers, alerts via email & sms for fault. Monitor 25 devices for free with no restriction. Download now http://ad.doubleclick.net/ddm/clk/292181274;119417398;o _______________________________________________ Bacula-users mailing list Bacula-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-users