During the 38 minutes that it takes the FD to do it's set up the CPU is running at nearly 100% for the FD process. After the FD begins sending data to the SD the CPU use drops to around 10%, which is normal for a backup on that server. The transfer rate is also about normal. That server has a mix of very large db files and many small files - I expect the backup to take at least 12 hours based on prior experience when using a more "normal" fileset specification based on directory names rather than individual files.
---- Alan Davis Senior Architect Ruckus Network, Inc. 703.464.6578 (o) 410.365.7175 (m) [EMAIL PROTECTED] alancdavis AIM > -----Original Message----- > From: Kern Sibbald [mailto:[EMAIL PROTECTED] > Sent: Tuesday, January 30, 2007 12:51 PM > To: Alan Davis > Cc: bacula-users@lists.sourceforge.net > Subject: Re: [Bacula-users] Experience with extremely large fileset > include lists? > > On Tuesday 30 January 2007 17:39, Alan Davis wrote: > > I've modified the timeout in stored/job.c to allow the SD to wait 90 > > minutes instead of 30, recompiled and installed the modified SD. > > > > The test job takes about 38 minutes for the FD to process the fileset, > > with the FD memory used : > > > > Heap: bytes=24,593,473 max_bytes=25,570,460 bufs=296,047 > > max_bufs=298,836 > > Yes, 24MB is really large memory utilization. > > > > > The SD waited for the FD to connect and is running the job as expected. > > I'll be interested to hear the results of running the job. I suspect that > it > will be catastrophically slow. > > > > > > > > > ---- > > Alan Davis > > Senior Architect > > Ruckus Network, Inc. > > 703.464.6578 (o) > > 410.365.7175 (m) > > [EMAIL PROTECTED] > > alancdavis AIM > > > > > -----Original Message----- > > > From: Alan Davis [mailto:[EMAIL PROTECTED] > > > Sent: Tuesday, January 30, 2007 10:44 AM > > > To: 'Kern Sibbald'; 'bacula-users@lists.sourceforge.net' > > > Subject: RE: [Bacula-users] Experience with extremely large fileset > > > include lists? > > > > > > Returning to the original thread... > > > > > > Just to make sure I'm being clear - my FileSet specification is: > > > > > > FileSet { > > > Name = "u2LgFileList" > > > Include { > > > Options { > > > signature = MD5 > > > } > > > File = </local/etc/u2LgFileList.list > > > > > > } > > > } > > > > > > The file /local/etc/u2LgFileList.list has 29K+ entries in it. > > > Note that this is /not/ an exclude list - it's explicitly listing the > > > files to be backed up. > > > > > > The FD takes about 40 minutes to read in the file list. > > > The SD times out in 30 minutes waiting for the FD. > > > > > > From my reading of the manual there are directives that set the time > > > > that > > > > > the Director will wait for an FD to respond "FD Connect Timeout" and > > > > the > > > > > time that the FD will wait for the SD to respond "SD Connect Timeout" > > > > as > > > > > well as the "Heartbeat Interval" that will keep the connection open > > > > during > > > > > long backups. > > > > > > I've not found a directive to modify the length of time that the SD > > > > will > > > > > wait for the FD to begin transferring data. > > > > > > This is the error message from the failed backup message. Note that > > > > the > > > > > authorization rejection is /not/ the problem - a test backup that > > > succeeded was used to verify proper authorization and communication > > > between FD and SD. > > > > > > 29-Jan 16:33 athos-dir: Start Backup JobId 112, > > > > Job=u2FullBackupJob.2007- > > > > > 01-29_16.33.15 > > > 29-Jan 17:21 u2-fd: u2FullBackupJob.2007-01-29_16.33.15 Fatal error: > > > Authorization key rejected by Storage daemon. > > > Please see > > > > http://www.bacula.org/rel-manual/faq.html#AuthorizationErrors > > > > > for help. > > > 29-Jan 17:21 u2-fd: u2FullBackupJob.2007-01-29_16.33.15 Fatal error: > > > Failed to authenticate Storage daemon. > > > 29-Jan 17:21 athos-dir: u2FullBackupJob.2007-01-29_16.33.15 Fatal > > > > error: > > > Socket error on Storage command: ERR=No data available 29-Jan 17:21 > > > > athos- > > > > > dir: u2FullBackupJob.2007-01-29_16.33.15 Error: Bacula 1.39.22 > > > > (09Sep06): > > > 29-Jan-2007 17:21:22 > > > > > > I think that the timeout is being specified in this code from > > > > stored/job.c > > > > > gettimeofday(&tv, &tz); > > > timeout.tv_nsec = tv.tv_usec * 1000; > > > timeout.tv_sec = tv.tv_sec + 30 * 60; /* wait 30 minutes */ > > > > > > Dmsg1(100, "%s waiting on FD to contact SD\n", jcr->Job); > > > /* > > > * Wait for the File daemon to contact us to start the Job, > > > * when he does, we will be released, unless the 30 minutes > > > * expires. > > > */ > > > P(mutex); > > > for ( ;!job_canceled(jcr); ) { > > > errstat = pthread_cond_timedwait(&jcr->job_start_wait, &mutex, > > > &timeout); > > > if (errstat == 0 || errstat == ETIMEDOUT) { > > > break; > > > } > > > } > > > V(mutex); > > > > > > > > > > > > ---- > > > Alan Davis > > > Senior Architect > > > Ruckus Network, Inc. > > > 703.464.6578 (o) > > > 410.365.7175 (m) > > > [EMAIL PROTECTED] > > > alancdavis AIM > > > > > > > -----Original Message----- > > > > From: Kern Sibbald [mailto:[EMAIL PROTECTED] > > > > Sent: Monday, January 29, 2007 4:06 PM > > > > To: bacula-users@lists.sourceforge.net > > > > Cc: Alan Davis > > > > Subject: Re: [Bacula-users] Experience with extremely large fileset > > > > include lists? > > > > > > > > Hello, > > > > > > > > On Monday 29 January 2007 21:19, Alan Davis wrote: > > > > > Kern, > > > > > > > > > > Thanks for the fast response. To clarify a bit - the file list > > > > that I > > > > > > > would be using would be individual files, not directories. There > > > > would > > > > > > > be no exclude list as only the files that I need backed up would > > > > be > > > > > > > listed. > > > > > > > > Yes, my answer was based on that assumption. > > > > > > > > > I have about 30TB of data files spread over several hundred > > > > > > directories. > > > > > > > > A true incremental backup will spend large amounts of time > > > > determining > > > > > > > what files have been changed or added. The information about the > > > > > modified or new files is stored in a db as a side-effect of > > > > processing > > > > > > > the files for release to production so building a file list is > > > > > > trivial. > > > > > > > > The only problem would be the FD's capability of handling a file > > > > list > > > > > of > > > > > > > > 10K+ entries. > > > > > > > > All I can say is to try it, but I won't be surprised if it chews up > > > > a > > > > > lot > > > > > > > of > > > > CPU. > > > > > > > > However, doing an equivalent of an incremental backup by means of an > > > > exclusion > > > > list doesn't seem possible to me. > > > > > > > > Bacula is really quite fast in traversing a very large filesystem > > > > during > > > > > > an > > > > incremental backup. > > > > > > > > > Thanks. > > > > > > > > > > ---- > > > > > Alan Davis > > > > > Senior Architect > > > > > Ruckus Network, Inc. > > > > > 703.464.6578 (o) > > > > > 410.365.7175 (m) > > > > > [EMAIL PROTECTED] > > > > > alancdavis AIM > > > > > > > > > > > -----Original Message----- > > > > > > From: Kern Sibbald [mailto:[EMAIL PROTECTED] > > > > > > Sent: Monday, January 29, 2007 2:47 PM > > > > > > To: bacula-users@lists.sourceforge.net > > > > > > Cc: Alan Davis > > > > > > Subject: Re: [Bacula-users] Experience with extremely large > > > > fileset > > > > > > > > include lists? > > > > > > > > > > > > On Monday 29 January 2007 18:17, Alan Davis wrote: > > > > > > > I understand that one of the projects is to incorporate > > > > features > > > > > > > that > > > > > > > > > > > > will make very large exclude lists feasible, but does anyone > > > > have > > > > > > > > > experience, good or bad, with very large include lists in a > > > > > > fileset? > > > > > > > > > > I'm looking at the possibility of building a backup list from > > > > a db > > > > > > > query > > > > > > > > > > > > that has the potential to return tens of thousands of files > > > > stored > > > > > > > in > > > > > > > > > > > > hundreds of directories. > > > > > > > > > > > > For each file in the directories you specify (normally your > > > > whole > > > > > > > > filesystem), > > > > > > Bacula will do a linear search through the exclude list. Thus > > > > it > > > > > > > could be > > > > > > > > > > > extremely CPU intensive. For a large list (more than 1000 > > > > files) I > > > > > > > > believe > > > > > > it (the list) needs to be put into a hash tree, which is code > > > > that > > > > > > > does > > > > > > > > > > > not > > > > > > exist. > > > > > > > > > > > > > Thanks > > > > > > > > > > > > > > > > > > > > > > > > > > > > ---- > > > > > > > > > > > > > > Alan Davis > > > > > > > > > > > > > > Senior Architect > > > > > > > > > > > > > > Ruckus Network, Inc. > > > > > > > > > > > > > > 703.464.6578 (o) > > > > > > > > > > > > > > 410.365.7175 (m) > > > > > > > > > > > > > > [EMAIL PROTECTED] > > > > > > > > > > > > > > alancdavis AIM > > > > ---------------------------------------------------------------------- > > > > > -- > > > > > > > - > > > > > > > > > Take Surveys. Earn Cash. Influence the Future of IT > > > > > Join SourceForge.net's Techsay panel and you'll get the chance to > > > > > > share > > > > > > > > your opinions on IT & business topics through brief surveys - and > > > > earn > > > > > > cash > > > > http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDE > > V > > > > > > > _______________________________________________ > > > > > Bacula-users mailing list > > > > > Bacula-users@lists.sourceforge.net > > > > > https://lists.sourceforge.net/lists/listinfo/bacula-users ------------------------------------------------------------------------- Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT & business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV _______________________________________________ Bacula-users mailing list Bacula-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-users