ADSM-L

Re: [ADSM-L] best backup method for millions of small files?

2009-04-30 19:58:44
Subject: Re: [ADSM-L] best backup method for millions of small files?
From: Allan Mills <mill2all AT POLICE.NSW.GOV DOT AU>
To: ADSM-L AT VM.MARIST DOT EDU
Date: Fri, 1 May 2009 10:03:02 +1000
Steve

Had this problem last week and found this Microsoft Technet article.
Fixed the problem after a reboot

http://support.microsoft.com/kb/30401


Regards,
Allan

Allan Mills
Technology Operations
02 8835 8035
0422 208 031



Steven Harris <sjharris AT AU1.IBM DOT COM>
Sent by: "ADSM: Dist Stor Manager" <ADSM-L AT VM.MARIST DOT EDU>
01/05/2009 09:39
Please respond to
"ADSM: Dist Stor Manager" <ADSM-L AT VM.MARIST DOT EDU>


To
ADSM-L AT VM.MARIST DOT EDU
cc

Subject
Re: [ADSM-L] best backup method for millions of small files?






Hi Norman

Your post worries me, as I'm just implementing an email archive solution
that will depend on windows journalling to back up some huge repositories.
The particular product fills up "containers" that  once filled never
change, so the change rate will be low there, but there are also index
files that will change often.

Have you determined whether the memory issue is related to number of files
or number of changes?

Regards

Steve

Steven Harris
TSM Admin, Sydney Australia





             "Gee, Norman"
             <Norman.Gee AT LC DOT CA
             .GOV>                                                      To
             Sent by: "ADSM:           ADSM-L AT VM.MARIST DOT EDU
             Dist Stor                                                  cc
             Manager"
             <[email protected]                                     Subject
             .EDU>                     Re: [ADSM-L] best backup method for
                                       millions of small files?

             01/05/2009 07:12
             AM


             Please respond to
             "ADSM: Dist Stor
                 Manager"
             <[email protected]
                   .EDU>






What options are there when journaling runs out of memory on a 32 bit
Windows server?  I have about 10 million files on one server that the
journal engine runs out of memory. With memory efficient disk cache
method and resource utilization 5, its runs out of memory,  resource
utilization of 4 runs too long.

-----Original Message-----
From: ADSM: Dist Stor Manager [mailto:ADSM-L AT VM.MARIST DOT EDU] On Behalf Of
Huebner,Andy,FORT WORTH,IT
Sent: Thursday, April 30, 2009 8:16 AM
To: ADSM-L AT VM.MARIST DOT EDU
Subject: Re: best backup method for millions of small files?

You have a disk array copy of the data, is that located close or far?
Have you considered a disk array snap shot also?
If you perform a journaled file system backup and an image backup then
you should be able to restore the image and then update the image with
the file system restore.  This might take a long time, I have never
tried it.
What failure are you trying to protect against?  In our case we use the
disk arrays to protect against a data center loss and a corrupt file
system and a TSM file system backup to protect against the loss of a
file.  Our big ones are in the 10 million file range.  Using a 64bit
Windows server we can backup the file system in about 6 - 8 hours
without journaling.  We suspect we could get the time down to around 4
hours if the TSM server was not busy backing up 500 other nodes.

To me the important thing is to figure out what you are protecting
against with each thing you do.  Also be sure and ask what the Recovery
Point Objective (RPO) is.  If it is less than 24 hours then array based
solutions may be the best choice.  Over 24 hours then TSM may be the
best choice.

Andy Huebner

-----Original Message-----
From: ADSM: Dist Stor Manager [mailto:ADSM-L AT VM.MARIST DOT EDU] On Behalf Of
Mehdi Salehi
Sent: Thursday, April 30, 2009 9:39 AM
To: ADSM-L AT VM.MARIST DOT EDU
Subject: Re: [ADSM-L] best backup method for millions of small files?

Hi,
None the two methods that you mean in the user's guide are suitable for
my
case. "Image+normal incremental" that you emphasized in your post means
getting full image backups for example every week. For the incremental
part,
one file-based full backup is needed which is a nightmare for 20
millions.
OK, if I accept the initial incremental backup time (that might take for
days), what happens in restoration?

Naturally, last image backup should be restored first and it will take A
minutes. Provided that image backups are weekly, the progressive
incremental
backups of the week is about 6*20MB=120MB. Now imagine 120MB of 15-20K
files
are to be restored in filesystem with an incredibly big file address
table
and system should create an inode-like entry for each. If this step
takes B
minutes, the total restoration time would be A+B. (A+B/A) ratio is
important
and I will try to measure and share it with the group.

Steven, your solution is excellent for ordinary filesystems with a
limited
number of files. But I think for millions of files, only backup/restore
method that do not care how many files exist in the volume are feasible.
Somehing like pure image backup (like Acronis image incremental backup)
or
the method that FastBack exploites.

Your points are welcomed.

Regards,
Mehdi Salehi


This e-mail (including any attachments) is confidential and may be
legally privileged. If you are not an intended recipient or an
authorized representative of an intended recipient, you are prohibited
from using, copying or distributing the information in this e-mail or
its attachments. If you have received this e-mail in error, please
notify the sender immediately by return e-mail and delete all copies of
this message and any attachments.
Thank you.
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
_ _ _ _ _

All mail is subject to content scanning for possible violation of NSW
Police Force
Electronic Messaging Policy. All NSW Police Force employees are required
to
familiarise themselves with the content of the policy, found under
Policies on the
NSW Police Force Intranet.





_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
_

The information contained in this email is intended for the named recipient(s)
only. It may contain private, confidential, copyright or legally privileged
information.  If you are not the intended recipient or you have received this
email by mistake, please reply to the author and delete this email immediately.
You must not copy, print, forward or distribute this email, nor place reliance
on its contents. This email and any attachment have been virus scanned. However,
you are requested to conduct a virus scan as well.  No liability is accepted
for any loss or damage resulting from a computer virus, or resulting from a 
delay
or defect in transmission of this email or any attached file. This email does 
not
constitute a representation by the NSW Police Force unless the author is legally
entitled to do so.

<Prev in Thread] Current Thread [Next in Thread>