ADSM-L

Re: [ADSM-L] Advice for archiving 80 billion of small files.

2017-01-26 08:27:54
Subject: Re: [ADSM-L] Advice for archiving 80 billion of small files.
From: Bo Nielsen <boanie AT DTU DOT DK>
To: ADSM-L AT VM.MARIST DOT EDU
Date: Thu, 26 Jan 2017 13:23:39 +0000
Hi All,

Thanks you for all the replies. It was  very helpful for me, but not for the 
management. 
I think they want anything else. Because many of the files have Japanese or 
Chinese characters,
So it's difficult to index the Tar files.

Regards

Bo  

-----Original Message-----
From: ADSM: Dist Stor Manager [mailto:ADSM-L AT VM.MARIST DOT EDU] On Behalf Of 
Skylar Thompson
Sent: 20. januar 2017 15:43
To: ADSM-L AT VM.MARIST DOT EDU
Subject: Re: [ADSM-L] Advice for archiving 80 billion of small files.

Do you need to recover files individually? If so, then image backup (at least 
on its own) won't be a good option. One thing you could do is tar up chunks 
(maybe a million files) and archive/backup those chunks. Keep a catalog 
(hopefully a database with indexes) of which files are in which tar balls, and 
then when you go to restore you only have to recover 1/80000 of your data to 
get one file.

On Fri, Jan 20, 2017 at 02:18:04PM +0000, Bo Nielsen wrote:
> Hi all,
>
> I need advice.
> I must archive 80 billion small files, but that is not possible, as I see it.
> since it will fill in the TSM's Database about 73 Tb.
> The filespace is mounted on a Linux server.
> Is there a way to pack/zip the files, so it's a smaller number of files.
> anybody who has tried this ??
>
> Regards,
>
> Bo Nielsen
>
>
> IT Service
>
>
>
> Technical University of Denmark
>
> IT Service
>
> Frederiksborgvej 399
>
> Building 109
>
> DK - 4000 Roskilde
>
> Denmark
>
> Mobil +45 2337 0271
>
> boanie AT dtu DOT dk<mailto:boanie AT dtu DOT dk>

--
-- Skylar Thompson (skylar2 AT u.washington DOT edu)
-- Genome Sciences Department, System Administrator
-- Foege Building S046, (206)-685-7354
-- University of Washington School of Medicine