Thanks Rory, the machine has the capacity to pull through pg_dumps but like
u rightly mentioned incremental backups mean that we will need to work with
the wal's.. 18TB is what is the scary part and with compression I dont see
it being less than 2TB a day...

On Fri, 15 May 2020, 17:02 Rory Campbell-Lange, <r...@campbell-lange.net>
wrote:

> On 15/05/20, Suhail Bamzena (suhailsa...@gmail.com) wrote:
> > Hello All,
> > I have very recently inherited an 18 TB DB that is running version 9.2.
> > Apparently this database has never been backed up and I have been tasked
> to
> > set in a periodic backup routine (weekly full & daily incremental) and
> dump
> > it into a NAS. What is the best way to go about this? Did some reading
> and
> > hear that pgbackrest does a good job with such huge sizes. Your expert
> > advise is needed.
>
> Incremental backups suggest the need to backup WAL archives. See
> https://www.postgresql.org/docs/9.2/continuous-archiving.html
>
> pgbackrest looks very cool but we haven't used it.
>
> A very simple solution could be just to dump the database daily with
> pg_dump, if you have the space and machine capacity to do it. Depending
> on what you are storing, you can achieve good compression with this, and
> it is a great way of having a simple file from which to restore a
> database.
>
> Our ~200GB cluster resolves to under 10GB of pg_dump files, although
> 18TB is a whole different order of size.
>
> Rory
>

Reply via email to