Hi All,
Thanks you for all the replies. It was very helpful for me, but not for the
management.
I think they want anything else. Because many of the files have Japanese or
Chinese characters,
So it's difficult to index the Tar files.
Regards
Bo
-----Original Message-----
From: ADSM: Dist Stor Manager [mailto:ADSM-L AT VM.MARIST DOT EDU] On Behalf Of
Skylar Thompson
Sent: 20. januar 2017 15:43
To: ADSM-L AT VM.MARIST DOT EDU
Subject: Re: [ADSM-L] Advice for archiving 80 billion of small files.
Do you need to recover files individually? If so, then image backup (at least
on its own) won't be a good option. One thing you could do is tar up chunks
(maybe a million files) and archive/backup those chunks. Keep a catalog
(hopefully a database with indexes) of which files are in which tar balls, and
then when you go to restore you only have to recover 1/80000 of your data to
get one file.
On Fri, Jan 20, 2017 at 02:18:04PM +0000, Bo Nielsen wrote:
> Hi all,
>
> I need advice.
> I must archive 80 billion small files, but that is not possible, as I see it.
> since it will fill in the TSM's Database about 73 Tb.
> The filespace is mounted on a Linux server.
> Is there a way to pack/zip the files, so it's a smaller number of files.
> anybody who has tried this ??
>
> Regards,
>
> Bo Nielsen
>
>
> IT Service
>
>
>
> Technical University of Denmark
>
> IT Service
>
> Frederiksborgvej 399
>
> Building 109
>
> DK - 4000 Roskilde
>
> Denmark
>
> Mobil +45 2337 0271
>
> boanie AT dtu DOT dk<mailto:boanie AT dtu DOT dk>
--
-- Skylar Thompson (skylar2 AT u.washington DOT edu)
-- Genome Sciences Department, System Administrator
-- Foege Building S046, (206)-685-7354
-- University of Washington School of Medicine
|