> 
> I mean that my goal is to save any new file, once and permanently. 
> If monday I have file1 on my nas, I want it to be saved on tape. 
> Tuesday I add file2 : I want it to be saved on tape. 
> Wednesday, file1 is deleted from NAS: it's a mistake, and I still want to 
> keep file1 forever on tape (and be able to restore it ). 
> Every file that has existed once on my NAS must be saved permanently on a 
> tape.  
> 
Okay I understand it like this:
You have done one full backup at the beginning. After that, you are doing 
incremental backups every night to save every new file. If the tape is full it 
gets packed away as archive and never gets rewritten? Right?
Your primary goal is to save your current data and archive "deleted" files 
forever?

I don't use tapes, but I think if you do incremental backups and you want to 
restore something you need to insert a huge part of the tapes because bacula 
needs to read them.(I'm not sure about that.)
If bacula should to this, you will have a huge problem if you want to restore a 
file in lets say 10years.

And being honest I really don't like the idea of doing incremental backups 
endlessly without differential- and full-backups in between(I wrote more about 
that later) 


> Let me show you my (simplified)  configuration : 
> 
> I mounted ( nfs ) my first NAS on, say, /mnt/NAS1/  
> My file set is :  FileSet {
> Name = "NAS1"
> File = /mnt/NAS1
> } 
> 
> My job is  Job {
> Name = "BackupNAS1"
> JobDefs = "DefaultJob"
> Level = Incremental
> FileSet="NAS1"
> #Accurate = yes # Not clear what I should do here. activate to yes seemed to 
> add many unwanted files - probably moved/renamed files ? 
> Pool = BACKUP1
> Storage = ScalarI3-BACKUP1 # this is my tape library
> Schedule = NAS1Daily #run every day
> 
> } 
> 
> with  
> JobDefs {
> Name = "DefaultJob"
> Type = Backup
> Level = Incremental
> Client = lto8-fd
> FileSet = "Test File Set"
> Messages = Standard
> SpoolAttributes = yes
> Priority = 10
> Write Bootstrap = "/var/lib/bacula/%c.bsr"
> } 
> 
> My pool is :  
> Pool {
> Name = BACKUP1
> Pool Type = Backup
> Recycle = no
> AutoPrune = no
> Volume Retention = 100 years
> Job Retention = 100 years
> Maximum Volume Bytes = 0
> Maximum Volumes = 1000 
> Storage = ScalarI3-BACKUP1
> Next Pool = BACKUP1
> } 


To your .conf:
-under JobDefs-DefaultJob :you declare  "FileSet = "Test File Set"" and in your 
jobdef you declare "FileSet="NAS1"" if that's your standard fileset, set it 
like this or try to ommit it in the jobdef. It is a little bit confusing.
-you use the "Next Pool"-Ressource in your Pool. Documentation states: it 
belongs under Schedule>Run>Next Pool. Either way it describes a migrating job. 
I think that's not what you want to do?


If I would be in your place, I would do it differently(assumed i got your 
primary goal right that you want to save your current data and archive 
"deleted" files forever?):
-first I would set nfs user permissions, if nfs or samba doesn't to the trick I 
would straight head to nextcloud(also Open-Source with a pricing plan for 
companies)
Why?: -> you can set permissions that your users can't delete their files and 
are forced to move it in a archive-folder with a good naming-convention, when 
they want to get rid of it(maybe you can automate it and the files go in an 
archive-folder, when your users hit the delete button and not to the bin. 
Should they do mistakes, it's up to you, to figure out the right file(might not 
be that clean).
-> having a good naming-convention and some sort of documentation makes it 
1million times easier to find the right file in the future.

I think you have two major goals: 1. keeping the productiveData(the data your 
users are currently using) save and 2. archiving old files, your users don't 
need anymore. 
To achieve the first goal I would go ahead and implement a backup-strategy with 
three pools(one incremental-, one differential-, one full-pool) and rotating 
tapes(rewriting them after a given time).
Should one of your NASs fail on you, you will be able to restore the files by 
yourself fast, keeping offline-time short, and therefore blackout-costs small.

To achieve the second goal I would go ahead and search for companies that are 
specialized in securing data for a long time(I've read about them in a book). 
The first idea I have had, was using a tape like a "special harddrive". Collect 
the files your users don't need anymore somewhere and write them once to a tape 
and label it by hand. If something happens to this tape, the data will be gone. 
I don't like this idea. I wouldn't do that. Probably the best idea would be to 
call a data-securing-company which do that job for you. Either way I wouldn't 
keep the productiveData-tapes and the archiv-tapes at the same spot(that would 
be another pro for the data-securing-company), because it violates the 
3-2-1-backup-rule(everything will be gone when disaster strikes(flood, fire, 
hurricane....)). If you don't know about the 3-2-1-backup-rule please look it 
up on the internet(this rule discusses good backups in more detail).

> I'm not sure I fully understand here : you say "since the volume-use duration 
> is set to short." . But I believe it's exactly the contrary here : my 
> volume-use duration is set to 100 years !? isn't it ?.  
Yes, is exactly the contrary. I'm not sure but that shouldn't be a problem. If 
you want to write to it indefinitely you can specify it as 0(the default) as 
specified in the documentation(chapter: configuring the director).

  
> > In the bacula-dir.conf you specified the director ressource type "Messages" 
> > there is a option called "append"
> > A part of my bacula-dir.conf:
> > # WARNING! the following will create a file that you must cycle from
> > # time to time as it will grow indefinitely. However, it will
> > # also keep all your messages if they scroll off the console.
> > append = "/var/log/bacula/bacula.log" = all, !skipped
> > console = all, !skipped
> > 
> > At the end "all, !skipped" are the types or classes of messages which go 
> > into it. They are described in more detail in the "Messages 
> > Resource"-Chapter:
> > https://www.bacula.org/11.0.x-manuals/en/main/Messages_Resource.html
> > 
> > If I type the "messages"-command in the bconsole the output is in my case 
> > in both cases the same.
> > 
> 
> This is regarding logs, right ? Doesn't seem to apply to me here. I'm dealing 
> with big video files being unnecessarily saved 10, 15 or 20 times on 
> tapes.... 
> Or maybe I missed something here ? 
In your last email, you asked "Specifically, how do you go about identifying 
exactly which volumes / jobid are to be "deactivated" and how do you do that?
You know the day where everything came to a halt. Knowing this, you can look 
through your logs which jobs ran on that day. For every Job there is a longer 
list with one tag named "Volume name(s)".
Under this tag are the volumes listed, that got used in that job.
Sorry for making it not clearer.


Maybe there is someone who has experiences with a similar backup-job or such 
data-securing-companies and can help you better.


Kind regards

Sebastian

-------------------------------------------------------------------------------------------------
FreeMail powered by mail.de - MEHR SICHERHEIT, SERIOSITÄT UND KOMFORT


_______________________________________________
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

Reply via email to