OPERATING SYSTEM = Ubuntu 8.04.3
LTS
BACULA VERSION(upgraded from 2.2.8-5)
bacula-client
2.2.8-5ubuntu7.2 bacula-common
3.0.2-3
bacula-common-mysql 3.0.2-3 bacula-console
3.0.2-3
bacula-director-common 3.0.2-3
bacula-director-mysql
3.0.2-3 bacula-fd
3.0.2-3
- bacula-sd
3.0.2-3
bacula-sd-mysql
3.0.2-3
- bacula-server
2.2.8-5ubuntu7.2
Hello Guys,
I seem to be having some issues with copy and
migration jobs.I have all my backup volumes
sitting on a Storage array, within this storage array i have multiple
pools and multiple volumes, what i am trying to achieve is copy the volumes from
each pool on this storage array to new pools and volumes located on a offsite
storage array.
The problem i am having is when
i run a copy or migration jobs the volumes from the
primary storage's pools are being truncated into
a single volume in the new offsite pools, my volumes are not copying
one for one, which is how i assumed the Copy and Migrate work?
example: i want to copy the below volumes BD0098
and BD0105 from the primary storage to the offsite storage, when i do this i
only end up with one volume on the offsite, not all volumes copied
across.
PRIMARY STORAGE ARRAY
-rw-r----- 1 bacula tape 66M 2009-10-08 09:58
BD0098 -rw-r----- 1 bacula tape 982 2009-10-08 09:59 BD0105
Results in:
OFFSITE STORAGE ARRAY
-rw-r----- 1 bacula tape 67M 2009-10-08 10:02
offsite-0104
Any assistance would be greaty
appriciated
Joe
For example one of my jobs is configured as
such
#######################################################################################
# DATABSE BACKUP JOB
Job { Name =
"icp-qv1-newsfeed1--dkactive-daily"
Type = Backup Level =
Full Client =
icp-qv1-bacula-fd FileSet =
"icp-qv1-newsfeed1--dkactive-daily"
Schedule =
"icp-qv1-newsfeed1--dkactive-daily-schedule"
Storage =
icp-qv1-newsfeed1--dkactive-daily-device
Messages = Standard Pool =
icp-qv1-newsfeed1--dkactive-daily-pool
RunScript
{
RunsWhen =
After
RunsOnFailure =
no
Command = "/etc/bacula/postremove.sh
/mnt/qv1-filer3-backups/icp-qv1-newsfeed1/dkactive/Daily"
} Priority = 10 }
FileSet
{ Name =
"icp-qv1-newsfeed1--dkactive-daily"
Include {
Options {
signature =
MD5
} File =
/mnt/qv1-filer3-backups/icp-qv1-newsfeed1/dkactive/Daily
} #
Exclude {
# File = #
} }
Schedule
{ Name =
"icp-qv1-newsfeed1--dkactive-daily-schedule"
Run = Full mon-sun at 21:00 }
Pool { Name =
icp-qv1-newsfeed1--dkactive-daily-pool
Pool Type = Backup Recycle =
yes
# Bacula can automatically recycle
Volumes AutoPrune =
yes
# Prune expired volumes Use volume
_once_ = yes Volume Retention =
2419200 seconds # Retention of
backup Label Format =
"BD" Next Pool =
icp-qv1-newsfeed1--dkactive-daily-pool-offsite
Storage = icp-qv1-newsfeed1--dkactive-daily-device
Storage { Name =
icp-qv1-newsfeed1--dkactive-daily-device
#Do not use "localhost" here
Address =
icp-qv1-bacula
# N.B. Use a fully qualified name
here SDPort =
9103 Password =
"Cv70F6pf1t6pBopT4vQOnigDrR0v3LT3Cgkiyj"
Device =
icp-qv1-newsfeed1--dkactive-daily-device
Media Type = File Maximum
Concurrent Jobs = 10 }
#############################################################################################################
# OFFSITE BACKUP JOB
Job { Name =
"icp-qv1-newsfeed1--dkactive-daily-offsite"
Type = Migrate Level =
Full Client =
icp-qv1-bacula-fd FileSet =
"icp-qv1-newsfeed1--dkactive-daily"
Messages = Standard Pool =
icp-qv1-newsfeed1--dkactive-daily-pool
Selection Type = Volume Selection
Pattern = "BD" Maximum Concurrent
Jobs = 10 }
Pool { Name =
icp-qv1-newsfeed1--dkactive-daily-pool-offsite
Pool Type = Backup Recycle =
yes
# Bacula can automatically recycle
Volumes AutoPrune =
yes
# Prune expired volumes
VolumeRetention = 30 minutes # Retention of
backup Storage =
icp-qv1-newsfeed1--dkactive-daily-device-offsite
Label Format = "offsite-" #Maximum
Volume Jobs = 1 }
Storage { Name =
icp-qv1-newsfeed1--dkactive-daily-device-offsite
#Do not use "localhost" here Address =
icp-qv1-bacula
# N.B. Use a fully qualified name here
SDPort = 9103 Password =
"Cv70F6pf1t6pBopT4vQOnigDrR0v3LT3Cgkiyj"
Device =
icp-qv1-newsfeed1--dkactive-daily-device-offsite
Media Type = File Maximum Concurrent
Jobs = 10 }
|