On Fri, Jun 19, 2009 at 6:29 PM, Dirk Bartley<bartle...@chartermi.net> wrote: > Greetings > > I'm trying to understand Maximum Concurrent Jobs. > > Here is an example output from status dir > > > > Running Jobs: > Console connected at 19-Jun-09 17:51 > JobId Level Name Status > ====================================================================== > 9334 Full Alum3JobMain.2009-06-19_17.01.00_47 is running > 9336 Full Alum2RootJob.2009-06-19_17.01.00_49 is waiting on Storage > Alum2FileStorage > 9337 Full Alum1Job.2009-06-19_17.01.00_50 is waiting on Storage > Alum2FileStorage > 9338 Full Alum3root.2009-06-19_17.01.00_51 is waiting on Storage > Alum2FileStorage > 9339 Full Alum3sharedlvm.2009-06-19_17.01.00_52 is waiting on > Storage Alum2FileStorage > 9340 Full DatabaseJob.2009-06-19_17.01.00_53 is waiting execution > 9341 Full SubversionJob.2009-06-19_17.01.00_54 is waiting for > higher priority jobs to finish > ==== > > > These jobs are all being written to filestorage. Then when they are all > done the low priority jobs will trigger a run after script that will > start them all being copied to tape, but that is later. > > So it looks like only one job is running because I have a maximum > concurrent directive relating to storage limiting the number of jobs > running concurrently. > > Reading the storage daemons config documentation, there is only one and > it is on the storage daemon itself, not on the Device resource. It is > set to 50. > > On this storage daemon I have 2 tape devices and one Filestorage device > on this storage Daemon. > > I can see from the director configuration, there is a maximum concurrent > jobs directive for just about every resource. The director resource is > set to 5. All jobs are set to 5 with the exception of the copy jobs. > All client resources are set to 10. Each storage resource in the > directors config is set to a Max of 10. > > So I'm not understanding why I can't have more jobs running at the same > time. The pool resource for the File-xxx pool sets Maximum Volume Jobs > = 1 so that each job will have it's own volume (or file on the hard > drive) that it is writing to. > > Because I am writing to file first, I am not setting up spooling. > > Any possible assistance in helping me understand this is apprecieated. > I guess I could be happy that everything is working, I'ts just that when > someone asks me why I can't get more jobs running concurrently, I want > to be able to answer. Because I know one person will ask. >
Are the priorities the same for all jobs but the SubversionJob? John ------------------------------------------------------------------------------ Are you an open source citizen? Join us for the Open Source Bridge conference! Portland, OR, June 17-19. Two days of sessions, one day of unconference: $250. Need another reason to go? 24-hour hacker lounge. Register today! http://ad.doubleclick.net/clk;215844324;13503038;v?http://opensourcebridge.org _______________________________________________ Bacula-users mailing list Bacula-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-users