-----Original Message-----

>I suggest running smaller jobs. I don't mean to sound trite, but thatreally is 
>the solution. Given that the alternative is non-trivial, thesensible choice 
>is, I'm afraid, cancel the job.

I'm already kicking off 20+ jobs for a single system already. This does not 
work when we're talking over the 100TB/nearly 200TB mark. And when these errors 
happen it does not matter how many jobs you have as /all/ outstanding jobs fail 
when you have concurancy (in this case all jobs that were qued and were not 
even writing to the same tape were canceled). > This sounds like a 
configuration issue. Queued jobs should not be cancelled when a previous job 
cancels.

Not queued, concurent jobs (all are active at the same time but only one writes 
at a time from it's spool file) This was done to avoid the 
write|spool|write|spool loop for a serial job against a large system cutting 
backup times in half.






------------------------------------------------------------------------------
All of the data generated in your IT infrastructure is seriously valuable.
Why? It contains a definitive record of application performance, security 
threats, fraudulent activity, and more. Splunk takes this data and makes 
sense of it. IT sense. And common sense.
http://p.sf.net/sfu/splunk-d2d-c2
_______________________________________________
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

Reply via email to