In the message dated: Mon, 08 Jan 2007 18:19:37 +0100, The pithy ruminations from Kern Sibbald on <Re: [Bacula-users] "resume" failed job> were: => On Monday 08 January 2007 14:22, Aaron Knister wrote: => > Thanks for the sarcasm. NOT. I come here for support, not to be => > ridiculed. And i'll have you know that the restore from that backup => > set did work. => => Yes, I was sarcastic, but life is (at least I am) like that. => => I certainly didn't mean to ridicule you, but rather to warn you that IMO, you
Well, the original poster isn't the only one who thought your response was very strong. I've had similar failures--due to client problems--in the midst of multi-TB backups, and was extremely interested to hear about what might have been a solution to resume the backup without starting from the beginning. => were doing something that is *very* unlikely to work. Perhaps I was wrong -- => I guess you fell into the 1% uncertainty that I had. I have my doubts, => anyway, good luck. I appreciate the warning, and probably won't try the same technique of mucking with the backup status value in the database. I think that this is a very interesting thread, in terms of making bacula more robust. I can see many applications for the idea of resuming an interrupted backup (the failure that the the original poster described, desktop backups where users may shut off machines, support for mobile users over unreliable network links, etc.). => => My advice to other users, remains the same: If a Job fails, the File records => will most likely not have been inserted, and in such a case, marking the job => as successfully terminated will most likely result in restore failure (or => screw up of some sort) later down the road because those File records are => critical for most restore operations. Hmmm.... I don't know anything about the internal structure of bacula, or much about databases, but it seems to me that this is a serious weakness. Would it be possible for baclula to function more like a journaling filesystems in terms of keeping consistency? Would it be possible to use the existing algorithms in bacula to insert the File records into a temporary db table--and also write File records to the temporary table in anticipation that the write to storage will succeed, and then [move/ copy/insert] those records into the real table only when the write to the storage media has been acknowledged? Would this scheme make it possible to: read the real table corresponding to the media read the media read the temporary table reconcile the differences (ie., any filesets written to media, with entries in the temporary table, but lacking records in the permanent table would have the File records updated from the temporary table) Mark "exposing my db ignorance daily" Bergman => => > => > => > On Jan 8, 2007, at 4:29 AM, Kern Sibbald wrote: => > => > > On Sunday 07 January 2007 23:03, Aaron Knister wrote: => > >> I solved this problem myself. I'm not sure how elegant the solution => > >> is, however. => > >> => > >> Using myphpadmin I changed the "JobStatus" field in the respective => > >> jobid's mysql entry from "f" to "T". I then re-ran the job and it => > >> picked up more or less where it left off. => > > => > > Yes, well congratulations. I give you 99% probability of having => > > created a set => > > of backups that cannot be restored. => > > => > > I *strongly* recommend that other users don't try manually => > > modifying the DB => > > unless you understand *all* the little details of how job records => > > are put => > > into the DB. => > > => > >> => > >> -Aaron => > >> => > >> On Jan 5, 2007, at 10:29 PM, Aaron Knister wrote: => > >> => > >>> Hi, => > >>> I recently had a backup job fail mid way. It was backing up 5 => > >>> terabytes of data, and had written 3tb off to tape. The job stopped => > >>> because there were no more writable volumes in the particular volume => > >>> pool to which the job was assigned. I cleared up a volume however => > >>> the => > >>> job did not resume and after a while errored out. I would like to => > >>> know if i can salvage the 3 terabytes that was already written to => > >>> tape and just continue from that point. => > >>> => > >>> Many thanks. => > >>> => > >>> -Aaron => > >>> ---- Mark Bergman [EMAIL PROTECTED] System Administrator Section of Biomedical Image Analysis 215-662-7310 Department of Radiology, University of Pennsylvania http://pgpkeys.pca.dfn.de:11371/pks/lookup?search=mark.bergman%40.uphs.upenn.edu The information contained in this e-mail message is intended only for the personal and confidential use of the recipient(s) named above. If the reader of this message is not the intended recipient or an agent responsible for delivering it to the intended recipient, you are hereby notified that you have received this document in error and that any review, dissemination, distribution, or copying of this message is strictly prohibited. If you have received this communication in error, please notify us immediately by e-mail, and delete the original message. ------------------------------------------------------------------------- Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT & business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV _______________________________________________ Bacula-users mailing list Bacula-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-users