> > I mean that my goal is to save any new file, once and permanently. > If monday I have file1 on my nas, I want it to be saved on tape. > Tuesday I add file2 : I want it to be saved on tape. > Wednesday, file1 is deleted from NAS: it's a mistake, and I still want to > keep file1 forever on tape (and be able to restore it ). > Every file that has existed once on my NAS must be saved permanently on a > tape. > Okay I understand it like this: You have done one full backup at the beginning. After that, you are doing incremental backups every night to save every new file. If the tape is full it gets packed away as archive and never gets rewritten? Right? Your primary goal is to save your current data and archive "deleted" files forever?
I don't use tapes, but I think if you do incremental backups and you want to restore something you need to insert a huge part of the tapes because bacula needs to read them.(I'm not sure about that.) If bacula should to this, you will have a huge problem if you want to restore a file in lets say 10years. And being honest I really don't like the idea of doing incremental backups endlessly without differential- and full-backups in between(I wrote more about that later) > Let me show you my (simplified) configuration : > > I mounted ( nfs ) my first NAS on, say, /mnt/NAS1/ > My file set is : FileSet { > Name = "NAS1" > File = /mnt/NAS1 > } > > My job is Job { > Name = "BackupNAS1" > JobDefs = "DefaultJob" > Level = Incremental > FileSet="NAS1" > #Accurate = yes # Not clear what I should do here. activate to yes seemed to > add many unwanted files - probably moved/renamed files ? > Pool = BACKUP1 > Storage = ScalarI3-BACKUP1 # this is my tape library > Schedule = NAS1Daily #run every day > > } > > with > JobDefs { > Name = "DefaultJob" > Type = Backup > Level = Incremental > Client = lto8-fd > FileSet = "Test File Set" > Messages = Standard > SpoolAttributes = yes > Priority = 10 > Write Bootstrap = "/var/lib/bacula/%c.bsr" > } > > My pool is : > Pool { > Name = BACKUP1 > Pool Type = Backup > Recycle = no > AutoPrune = no > Volume Retention = 100 years > Job Retention = 100 years > Maximum Volume Bytes = 0 > Maximum Volumes = 1000 > Storage = ScalarI3-BACKUP1 > Next Pool = BACKUP1 > } To your .conf: -under JobDefs-DefaultJob :you declare "FileSet = "Test File Set"" and in your jobdef you declare "FileSet="NAS1"" if that's your standard fileset, set it like this or try to ommit it in the jobdef. It is a little bit confusing. -you use the "Next Pool"-Ressource in your Pool. Documentation states: it belongs under Schedule>Run>Next Pool. Either way it describes a migrating job. I think that's not what you want to do? If I would be in your place, I would do it differently(assumed i got your primary goal right that you want to save your current data and archive "deleted" files forever?): -first I would set nfs user permissions, if nfs or samba doesn't to the trick I would straight head to nextcloud(also Open-Source with a pricing plan for companies) Why?: -> you can set permissions that your users can't delete their files and are forced to move it in a archive-folder with a good naming-convention, when they want to get rid of it(maybe you can automate it and the files go in an archive-folder, when your users hit the delete button and not to the bin. Should they do mistakes, it's up to you, to figure out the right file(might not be that clean). -> having a good naming-convention and some sort of documentation makes it 1million times easier to find the right file in the future. I think you have two major goals: 1. keeping the productiveData(the data your users are currently using) save and 2. archiving old files, your users don't need anymore. To achieve the first goal I would go ahead and implement a backup-strategy with three pools(one incremental-, one differential-, one full-pool) and rotating tapes(rewriting them after a given time). Should one of your NASs fail on you, you will be able to restore the files by yourself fast, keeping offline-time short, and therefore blackout-costs small. To achieve the second goal I would go ahead and search for companies that are specialized in securing data for a long time(I've read about them in a book). The first idea I have had, was using a tape like a "special harddrive". Collect the files your users don't need anymore somewhere and write them once to a tape and label it by hand. If something happens to this tape, the data will be gone. I don't like this idea. I wouldn't do that. Probably the best idea would be to call a data-securing-company which do that job for you. Either way I wouldn't keep the productiveData-tapes and the archiv-tapes at the same spot(that would be another pro for the data-securing-company), because it violates the 3-2-1-backup-rule(everything will be gone when disaster strikes(flood, fire, hurricane....)). If you don't know about the 3-2-1-backup-rule please look it up on the internet(this rule discusses good backups in more detail). > I'm not sure I fully understand here : you say "since the volume-use duration > is set to short." . But I believe it's exactly the contrary here : my > volume-use duration is set to 100 years !? isn't it ?. Yes, is exactly the contrary. I'm not sure but that shouldn't be a problem. If you want to write to it indefinitely you can specify it as 0(the default) as specified in the documentation(chapter: configuring the director). > > In the bacula-dir.conf you specified the director ressource type "Messages" > > there is a option called "append" > > A part of my bacula-dir.conf: > > # WARNING! the following will create a file that you must cycle from > > # time to time as it will grow indefinitely. However, it will > > # also keep all your messages if they scroll off the console. > > append = "/var/log/bacula/bacula.log" = all, !skipped > > console = all, !skipped > > > > At the end "all, !skipped" are the types or classes of messages which go > > into it. They are described in more detail in the "Messages > > Resource"-Chapter: > > https://www.bacula.org/11.0.x-manuals/en/main/Messages_Resource.html > > > > If I type the "messages"-command in the bconsole the output is in my case > > in both cases the same. > > > > This is regarding logs, right ? Doesn't seem to apply to me here. I'm dealing > with big video files being unnecessarily saved 10, 15 or 20 times on > tapes.... > Or maybe I missed something here ? In your last email, you asked "Specifically, how do you go about identifying exactly which volumes / jobid are to be "deactivated" and how do you do that? You know the day where everything came to a halt. Knowing this, you can look through your logs which jobs ran on that day. For every Job there is a longer list with one tag named "Volume name(s)". Under this tag are the volumes listed, that got used in that job. Sorry for making it not clearer. Maybe there is someone who has experiences with a similar backup-job or such data-securing-companies and can help you better. Kind regards Sebastian ------------------------------------------------------------------------------------------------- FreeMail powered by mail.de - MEHR SICHERHEIT, SERIOSITÄT UND KOMFORT _______________________________________________ Bacula-users mailing list Bacula-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-users