Hi All,

Thanks you for all the replies. It was  very helpful for me, but not for the 
management. 
I think they want anything else. Because many of the files have Japanese or 
Chinese characters,
So it's difficult to index the Tar files.

Regards

Bo  

-----Original Message-----
From: ADSM: Dist Stor Manager [mailto:ADSM-L@VM.MARIST.EDU] On Behalf Of Skylar 
Thompson
Sent: 20. januar 2017 15:43
To: ADSM-L@VM.MARIST.EDU
Subject: Re: [ADSM-L] Advice for archiving 80 billion of small files.

Do you need to recover files individually? If so, then image backup (at least 
on its own) won't be a good option. One thing you could do is tar up chunks 
(maybe a million files) and archive/backup those chunks. Keep a catalog 
(hopefully a database with indexes) of which files are in which tar balls, and 
then when you go to restore you only have to recover 1/80000 of your data to 
get one file.

On Fri, Jan 20, 2017 at 02:18:04PM +0000, Bo Nielsen wrote:
> Hi all,
>
> I need advice.
> I must archive 80 billion small files, but that is not possible, as I see it.
> since it will fill in the TSM's Database about 73 Tb.
> The filespace is mounted on a Linux server.
> Is there a way to pack/zip the files, so it's a smaller number of files.
> anybody who has tried this ??
>
> Regards,
>
> Bo Nielsen
>
>
> IT Service
>
>
>
> Technical University of Denmark
>
> IT Service
>
> Frederiksborgvej 399
>
> Building 109
>
> DK - 4000 Roskilde
>
> Denmark
>
> Mobil +45 2337 0271
>
> boa...@dtu.dk<mailto:boa...@dtu.dk>

--
-- Skylar Thompson (skyl...@u.washington.edu)
-- Genome Sciences Department, System Administrator
-- Foege Building S046, (206)-685-7354
-- University of Washington School of Medicine

Reply via email to