On Sat, Sep 24, 2011 at 6:32 PM, Boudewijn Ector
<boudewijn AT boudewijnector DOT nl> wrote:
> Hi Guys,
>
>
> For some time, I've been trying to get concurrent jobs in bacula to work.
> For doing so, I've created a pool for each client, and made sure all
> parts of the setup have got the max concurrent jobs = >1 .
>
> Please allow me to elaborate about my configuration:
>
> This is part of my bacula-dir (well, this is a file for a client 'www',
> and it's being included in bacula-dir, along with some exactly the same
> files except for passwords/hostnames):
>
> JobDefs {
> Name = "www-weekly"
> Type = Backup
> Level = Incremental
> Client = www
> FileSet = "Full Set"
> Schedule = "WeeklyCycle"
> Storage = leiden-filestorage
> Messages = Standard
> Pool = wwwPool
> Priority = 10
> }
>
>
>
> Job {
> Name = "wwwjob"
> JobDefs = "www-weekly"
> Write Bootstrap = "/var/lib/bacula/www.bsr"
> }
>
> Client {
> Name = www
> Address = www.KNIP
> FDPort = 9102
> Catalog = MyCatalog
> Password = "KNIP" # password for FileDaemon
> File Retention = 30 days # 30 days
> Job Retention = 6 months # six months
> AutoPrune = yes # Prune expired Jobs/Files
> }
>
>
> Pool {
> Name = wwwPool
> LabelFormat = "wwwVol"
> Pool Type = Backup
> Recycle = yes # Bacula can automatically
> recycle Volumes
> AutoPrune = yes # Prune expired volumes
> Volume Retention = 365 days # one year
> Volume Use Duration = 23h
> }
>
>
>
> As you can see, I've removed some sensitive information. A clone of this
> config is also used for 'mail', and some more machines. Each has it's
> own pool (because of concurrency).
>
>
> Well the bacula-sd.conf:
>
> Storage { # definition of myself
> Name = leiden-filestorage
> WorkingDirectory = "/var/lib/bacula"
> Pid Directory = "/var/run/bacula"
> Maximum Concurrent Jobs = 50
> SDAddresses = {
> ip = { addr = 192.168.1.44; port = 9103 }
> ip = { addr = 127.0.0.1; port =9103 }
> }
> }
> Director {
> Name = leiden-dir
> Password = "*"
> }
> Director {
> Name = leiden-mon
> Password = "*"
> Monitor = yes
> }
> Device {
> Name = leiden-filestorage
> Media Type = File
> Archive Device = /bacula
> LabelMedia = yes; # lets Bacula label unlabeled media
> Random Access = Yes;
> AutomaticMount = yes; # when device opened, read it
> RemovableMedia = no;
> }
>
> Messages {
> Name = Standard
> director = leiden-dir = all
> }
>
>
>
>
>
> Pretty standard, should I change something in here?
>
>
>
> And my bacula-fd.conf:
>
> Director {
> Name = leiden-dir
> Password = "*"
> }
>
> Director {
> Name = www.*-mon
> Password = "*"
> Monitor = yes
> }
>
> FileDaemon { # this is me
> Name = www.*-fd
> FDport = 9102 # where we listen for the director
> WorkingDirectory = /var/lib/bacula
> Pid Directory = /var/run/bacula
> HeartBeat Interval = 15
> Maximum Concurrent Jobs = 20
> FDAddress = *
> }
>
> Messages {
> Name = Standard
> director = www.*-dir = all, !skipped, !restored
> }
> Also quite boring.
>
>
>
Do you have "Maximum Concurrent Jobs" set in the Director and storage
sections in bacula-dir.conf?
>
> Can someone please explain to me why bacula still is not able to run
> concurrent Jobs? Do I have to create a storage for each client (for
> instance)? And what's the reason for having to do so?
>
Only 1 volume and thus pool can be loaded in a storage device at a
time so if you have several pools that you want to run backups on you
need more than 1 storage device. For disk based backups, I highly
recommend using the bacula virtual autochanger.
http://sourceforge.net/projects/vchanger/
This will greatly simplify the setup of multiple pools, devices and
concurrency. Just send all jobs to the virtual autochanger resource
and let bacula handle the devices.
>
> Furthermore, I've enabled the compression on some clients, but
> nevertheless the system's performance isn't very good. It tends to
> stagger at about 1800kb/s , but both ends of the line are 100mbit... and
> almost not being used at all.
Software compression is a very heavy CPU usage process on the FD and
will certainly slow down your backups.
> The director and sd are on the same machine, attached to a NAS (which
> performs fine by itself), and the machine has a dual-core Atom CPU
> running debian and 2gb of RAM. It also has no other jobs except for
> Nagios (which is not very heavily loaded).
>
>
------------------------------------------------------------------------------
All of the data generated in your IT infrastructure is seriously valuable.
Why? It contains a definitive record of application performance, security
threats, fraudulent activity, and more. Splunk takes this data and makes
sense of it. IT sense. And common sense.
http://p.sf.net/sfu/splunk-d2dcopy2
_______________________________________________
Bacula-users mailing list
Bacula-users AT lists.sourceforge DOT net
https://lists.sourceforge.net/lists/listinfo/bacula-users
|