Hi Sebastian,

My answers below.

Thanks once again for your help !

Samuel

Le jeu. 30 sept. 2021 à 02:23, <neumei...@mail.de> a écrit :

> >
> > I mean that my goal is to save any new file, once and permanently.
> > If monday I have file1 on my nas, I want it to be saved on tape.
> > Tuesday I add file2 : I want it to be saved on tape.
> > Wednesday, file1 is deleted from NAS: it's a mistake, and I still want
> to keep file1 forever on tape (and be able to restore it ).
> > Every file that has existed once on my NAS must be saved permanently on
> a tape.
> >
> Okay I understand it like this:
> You have done one full backup at the beginning. After that, you are doing
> incremental backups every night to save every new file. If the tape is full
> it gets packed away as archive and never gets rewritten? Right?
> Your primary goal is to save your current data and archive "deleted" files
> forever?
>

Yes ! Exactly.


> I don't use tapes, but I think if you do incremental backups and you want
> to restore something you need to insert a huge part of the tapes because
> bacula needs to read them.(I'm not sure about that.)
> If bacula should to this, you will have a huge problem if you want to
> restore a file in lets say 10years.
>

Not really. I can do a restore job, searching by finename. If the tape is
not in the library, Bacula asks me to put it in...I've tested this
procedure a few times, it works.


> And being honest I really don't like the idea of doing incremental backups
> endlessly without differential- and full-backups in between(I wrote more
> about that later)
>
>
> > Let me show you my (simplified)  configuration :
> >
> > I mounted ( nfs ) my first NAS on, say, /mnt/NAS1/
> > My file set is :  FileSet {
> > Name = "NAS1"
> > File = /mnt/NAS1
> > }
> >
> > My job is  Job {
> > Name = "BackupNAS1"
> > JobDefs = "DefaultJob"
> > Level = Incremental
> > FileSet="NAS1"
> > #Accurate = yes # Not clear what I should do here. activate to yes
> seemed to add many unwanted files - probably moved/renamed files ?
> > Pool = BACKUP1
> > Storage = ScalarI3-BACKUP1 # this is my tape library
> > Schedule = NAS1Daily #run every day
> >
> > }
> >
> > with
> > JobDefs {
> > Name = "DefaultJob"
> > Type = Backup
> > Level = Incremental
> > Client = lto8-fd
> > FileSet = "Test File Set"
> > Messages = Standard
> > SpoolAttributes = yes
> > Priority = 10
> > Write Bootstrap = "/var/lib/bacula/%c.bsr"
> > }
> >
> > My pool is :
> > Pool {
> > Name = BACKUP1
> > Pool Type = Backup
> > Recycle = no
> > AutoPrune = no
> > Volume Retention = 100 years
> > Job Retention = 100 years
> > Maximum Volume Bytes = 0
> > Maximum Volumes = 1000
> > Storage = ScalarI3-BACKUP1
> > Next Pool = BACKUP1
> > }
>
>
> To your .conf:
> -under JobDefs-DefaultJob :you declare  "FileSet = "Test File Set"" and in
> your jobdef you declare "FileSet="NAS1"" if that's your standard fileset,
> set it like this or try to ommit it in the jobdef. It is a little bit
> confusing.
>
OK

> -you use the "Next Pool"-Ressource in your Pool. Documentation states: it
> belongs under Schedule>Run>Next Pool. Either way it describes a migrating
> job. I think that's not what you want to do?
>
I had tried a "virtual backup", so that all my incremental jobs merge into
one, periodically. I thought it was only virtual, only dealing with the
catalog data, but it seems I can do that only by recreating a whole bunch
of volumes.
I have hundreds of TeraOctets of datas and I don't want to do that ! So I
let the incremental jobs running. Let aside my current problem, it's
convenient for what I need...

>
>
> If I would be in your place, I would do it differently(assumed i got your
> primary goal right that you want to save your current data and archive
> "deleted" files forever?):
> -first I would set nfs user permissions, if nfs or samba doesn't to the
> trick I would straight head to nextcloud(also Open-Source with a pricing
> plan for companies)
> Why?: -> you can set permissions that your users can't delete their files
> and are forced to move it in a archive-folder with a good
> naming-convention, when they want to get rid of it(maybe you can automate
> it and the files go in an archive-folder, when your users hit the delete
> button and not to the bin. Should they do mistakes, it's up to you, to
> figure out the right file(might not be that clean).
> -> having a good naming-convention and some sort of documentation makes it
> 1million times easier to find the right file in the future.
>

We have all this settle, in a way or another. But, I still need to give
some full rights to some users , and the problem is more : what if the NAS
burns or what if 3 HDD crash at the same time etc...
I want a robust and simple backup solution in case of rare event...


>
> I think you have two major goals: 1. keeping the productiveData(the data
> your users are currently using) save and 2. archiving old files, your users
> don't need anymore.

No. Every file could be used any time. Any file is available.

>
> To achieve the first goal I would go ahead and implement a backup-strategy
> with three pools(one incremental-, one differential-, one full-pool) and
> rotating tapes(rewriting them after a given time).
>

I have hundreds of TeraOctets. This would mean doubling or more the space I
need...



> Should one of your NASs fail on you, you will be able to restore the files
> by yourself fast, keeping offline-time short, and therefore blackout-costs
> small.
>
> To achieve the second goal I would go ahead and search for companies that
> are specialized in securing data for a long time(I've read about them in a
> book). The first idea I have had, was using a tape like a "special
> harddrive". Collect the files your users don't need anymore somewhere and
> write them once to a tape and label it by hand. If something happens to
> this tape, the data will be gone. I don't like this idea. I wouldn't do
> that. Probably the best idea would be to call a data-securing-company which
> do that job for you. Either way I wouldn't keep the productiveData-tapes
> and the archiv-tapes at the same spot(that would be another pro for the
> data-securing-company), because it violates the
> 3-2-1-backup-rule(everything will be gone when disaster strikes(flood,
> fire, hurricane....)). If you don't know about the 3-2-1-backup-rule please
> look it up on the internet(this rule discusses good backups in more detail).
>

My idea was : when one volume is full, store it in another place... So it
was OK, I guess.

>
> > I'm not sure I fully understand here : you say "since the volume-use
> duration is set to short." . But I believe it's exactly the contrary here :
> my volume-use duration is set to 100 years !? isn't it ?.
> Yes, is exactly the contrary. I'm not sure but that shouldn't be a
> problem. If you want to write to it indefinitely you can specify it as
> 0(the default) as specified in the documentation(chapter: configuring the
> director).
>
>
> > > In the bacula-dir.conf you specified the director ressource type
> "Messages" there is a option called "append"
> > > A part of my bacula-dir.conf:
> > > # WARNING! the following will create a file that you must cycle from
> > > # time to time as it will grow indefinitely. However, it will
> > > # also keep all your messages if they scroll off the console.
> > > append = "/var/log/bacula/bacula.log" = all, !skipped
> > > console = all, !skipped
> > >
> > > At the end "all, !skipped" are the types or classes of messages which
> go into it. They are described in more detail in the "Messages
> Resource"-Chapter:
> > > https://www.bacula.org/11.0.x-manuals/en/main/Messages_Resource.html
> > >
> > > If I type the "messages"-command in the bconsole the output is in my
> case in both cases the same.
> > >
> >
> > This is regarding logs, right ? Doesn't seem to apply to me here. I'm
> dealing with big video files being unnecessarily saved 10, 15 or 20 times
> on tapes....
> > Or maybe I missed something here ?
> In your last email, you asked "Specifically, how do you go about
> identifying exactly which volumes / jobid are to be "deactivated" and how
> do you do that?
> You know the day where everything came to a halt. Knowing this, you can
> look through your logs which jobs ran on that day. For every Job there is a
> longer list with one tag named "Volume name(s)".
> Under this tag are the volumes listed, that got used in that job.
> Sorry for making it not clearer.
>
I understand very clearly. But this is going to be quite long to
check, because I have also to see what job has got "new" files. I was
hoping there would be a way to "deduplicate" files in jobs, and jobs in
incremental backup ...
Well it seems I have to do this by hand ?



>
>
> Maybe there is someone who has experiences with a similar backup-job or
> such data-securing-companies and can help you better.
>

Anyway, I think my case shows a kind of misconception ( or misconfiguration
? ) : If an incremental job is delayed for some reason, why should it
backup many times the same file !? How to avoid that ?


>
> Kind regards
>

Kind regards and many thanks !! :)

Samuel

>
> Sebastian
>
>
> -------------------------------------------------------------------------------------------------
> FreeMail powered by mail.de - MEHR SICHERHEIT, SERIOSITÄT UND KOMFORT
>
>
> _______________________________________________
> Bacula-users mailing list
> Bacula-users@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/bacula-users
>
_______________________________________________
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

Reply via email to