On 11/5/21 13:43, mark.berg...@pennmedicine.upenn.edu wrote:

We're using the GPFS filesystem, and doing filesystem snapshots every
15 minutes, with a limited set retained for at least 2 months. The
snapshots allow for almost instant restores of recent data and comparison
between different versions of files, without system administrator
intervention.

Because of snapshots, I'm planning to eliminate all nightly incremental
& differential backups to tape. Tape backups would be only for
archival/disaster-recovery purposes and for compliance with grant and
data management requirements.

I'm not sure this is wise. Remember that if you lose an array, you lose all of its snapshots too. Or would you consider that a disaster-recovery scenario?


The new strategy would be to do a full backup every 2 months, kept for
5 months. One backup would be kept for at least 2 years, the others would
be rotated (media reused). For example:

        January 2021            keep until January 2023
        March 2021              keep until August 2021
        May 2021                keep until October 2021
        July 2021               keep until December 2021
        September 2021          re-use March 2021 media, keep until February 
2022
        November 2021           re-use May 2021 media, keep until April 2022
        January 2022            keep until January 2024

This is going to be complex. I think it's DOABLE, but you will need a complex set of Pools and Schedules because of the way you're setting up multiple rentention times for backups of the same jobs at the same levels.

What you might need to do is run all of your Full backups to one Pool that has five months retention, then every six months run a Copy job that archives the most recent set of Full backups to a second Pool with two years retention. This is probably the method that will result in the least tearing out of hair.


All tape backups would be done from a snapshot, so that no files within
the source of the backup change during the process. A "run before job"
script would dump coherent copies of databases, then create a filesystem
snapshot dedicated to the backup. That snapshot would be removed when
the backup is complete.

We've got about 700 top-level directories for user accounts and research
projects. We'll probably run an individual backup job for each group of
directories alphabetically (A*, B*, etc), so that the 400TB will be spread
(unevenly) across about 45 Bacula jobs.


This MOSTLY seems sound, with the proviso that I am not familiar with the details of GPFS. But I've implemented similar schemes on top of Solaris' ZFS.

The scheme of backing up snapshots is sound and a good plan. It entirely sidesteps the problem of files being changed while they are being backed up. Does GPFS offer you a way to create incremental snapshots containing the changes since a stipulated previous snapshot? That might be a way to get viable intermediate incremental or differential backups.



--
  Phil Stracchino
  Babylon Communications
  ph...@caerllewys.net
  p...@co.ordinate.org
  Landline: +1.603.293.8485
  Mobile:   +1.603.998.6958


_______________________________________________
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

Reply via email to