>>>>> On Tue, 09 Sep 2014 20:25:18 +0200, Kern Sibbald said:
>
> On 09/09/2014 07:46 PM, Martin Simmons wrote:
> > It looks like removing readfifo=yes will not help, because the restore code
> > doesn't look at it.
> >
> > The restore will not work without a process already reading from the fifo.
> The simplest Bacula restore doesn't use a fifo. That is what I am
> recommending that he try.
Yes, but it looks like his backup contains data of type FT_FIFO, so the
restore will always try to write it back into a fifo.
__Martin
> However it seems that the configuration is
> quite complete (at least for me).
> Best regards,
> Kern
> >
> > __Martin
> >
> >
> >>>>>> On Tue, 9 Sep 2014 10:57:59 -0400, Kenny Noe said:
> >> Kern, et al....
> >>
> >> I tried another backup. Here is my cliconfig. I parred it down to just
> >> the restore stuff.
> >> #********************************************************************************
> >> # bluewhale
> >> #********************************************************************************
> >> Client {
> >> Name = bluewhale
> >> Address = bluewhale.backup.bnesystems.com
> >> Catalog = BS01-Catalog
> >> Password = "xxxxxx"
> >> FileRetention = 15 days
> >> JobRetention = 15 days
> >> AutoPrune = yes
> >> MaximumConcurrentJobs = 1
> >> }
> >> Job {
> >> Name = Restore_mail_bluewhale
> >> FileSet = Full_mail_bluewhale
> >> Type = Restore
> >> Pool = Pool_mail_bluewhale
> >> Client = bluewhale
> >> Messages = Standard
> >> }
> >> Pool {
> >> Name = Pool_mail_bluewhale
> >> PoolType = Backup
> >> Storage = Storage_bluewhale
> >> MaximumVolumeJobs = 1
> >> CatalogFiles = yes
> >> AutoPrune = yes
> >> VolumeRetention = 1 week
> >> Recycle = yes
> >> LabelFormat = "mail-"
> >> }
> >> Storage {
> >> Name = Storage_bluewhale
> >> Address = 10.10.10.199
> >> SDPort = 9103
> >> Password = "imadirector"
> >> Device = File_bluewhale
> >> MediaType = NAS_bluewhale
> >> MaximumConcurrentJobs = 1
> >> }
> >> Schedule {
> >> Name = Schedule_mail_bluewhale
> >> Run = Level=Full sun-sat at 01:00
> >> }
> >> FileSet {
> >> Name = Full_mail_bluewhale
> >> Include {
> >> Options {
> >> signature=SHA1
> >> }
> >> File="mail.tar"
> >> }
> >> }
> >>
> >>
> >> During the restore I ran status storage from the console. I get this
> >>
> >> *status sto
> >> The defined Storage resources are:
> >> 1: Storage_asterisk
> >> 2: Storage_besc-4dvapp
> >> 3: Storage_besc-bs01
> >> 4: Storage_besc-unixmgr01
> >> 5: Storage_bluewhale
> >> 6: Storage_demo
> >> 7: Storage_dev
> >> 8: Storage_mako
> >> Select Storage resource (1-8): 5
> >> Connecting to Storage daemon Storage_bluewhale at 10.10.10.199:9103
> >>
> >> BS01-SD1 Version: 5.2.2 (26 November 2011) x86_64-unknown-linux-gnu ubuntu
> >> 11.10
> >> Daemon started 09-Sep-14 09:19. Jobs: run=1, running=0.
> >> Heap: heap=598,016 smbytes=386,922 max_bytes=405,712 bufs=947 max_bufs=949
> >> Sizes: boffset_t=8 size_t=8 int32_t=4 int64_t=8 mode=0,0
> >>
> >> Running Jobs:
> >> Reading: Full Restore job Restore_mail_bluewhale JobId=12922
> >> Volume="mail-0386"
> >> pool="Pool_mail_bluewhale" device="File_bluewhale"
> >> (/nas/bacula/bluewhale)
> >> Files=0 Bytes=0 Bytes/sec=0
> >> FDReadSeqNo=6 in_msg=6 out_msg=2320699 fd=6
> >> ====
> >>
> >> Jobs waiting to reserve a drive:
> >> ====
> >>
> >> Terminated Jobs:
> >> JobId Level Files Bytes Status Finished Name
> >> ===================================================================
> >> 12913 Incr 108 73.85 M OK 08-Sep-14 20:02 Backup_os_dev
> >> 12912 Full 4 61.80 G OK 08-Sep-14 20:24
> >> Backup_app_demo
> >> 12914 Incr 230 57.85 M OK 09-Sep-14 00:00
> >> Backup_os_asterisk
> >> 12916 Incr 31 68.06 M OK 09-Sep-14 00:01
> >> Backup_os_besc-unixmgr01
> >> 12917 Incr 0 0 Cancel 09-Sep-14 00:03
> >> Backup_os_bluewhale
> >> 12918 Full 4 501.3 M OK 09-Sep-14 00:04 Backup_app_dev
> >> 12915 Incr 256 1.099 G OK 09-Sep-14 00:06
> >> Backup_os_besc-bs01
> >> 12919 Full 4 54.41 G OK 09-Sep-14 01:04
> >> Backup_app_mako
> >> 12920 0 0 Cancel 09-Sep-14 09:17
> >> Restore_mail_bluewhale
> >> 12921 0 0 OK 09-Sep-14 09:43
> >> Restore_mail_bluewhale
> >> ====
> >>
> >> Device status:
> >> Device "File_asterisk" (/nas/bacula/asterisk) is not open.
> >> Device "File_besc-4dvapp" (/nas/bacula/besc-4dvapp) is not open.
> >> Device "File_besc-bs01" (/nas/bacula/besc-bs01) is not open.
> >> Device "File_besc-unixmgr01" (/nas/bacula/besc-unixmgr01) is not open.
> >> Device "File_bluewhale" (/nas/bacula/bluewhale) is mounted with:
> >> Volume: mail-0386
> >> Pool: *unknown*
> >> Media type: NAS_bluewhale
> >> Total Bytes Read=17,923,756,032 Blocks Read=277,836 Bytes/block=64,512
> >> Positioned at File=4 Block=743,886,199
> >> Device "File_demo" (/nas/bacula/demo) is not open.
> >> Device "File_dev" (/nas/bacula/dev) is not open.
> >> Device "File_mako" (/nas/bacula/mako) is not open.
> >> Device "File_qa" (/nas/bacula/qa) is not open.
> >> Device "File_qa2" (/nas/bacula/qa2) is not open.
> >> Device "File_smart" (/nas/bacula/smart) is not open.
> >> ====
> >>
> >> Used Volume status:
> >> mail-0386 on device "File_bluewhale" (/nas/bacula/bluewhale)
> >> Reader=1 writers=0 devres=0 volinuse=1
> >> mail-0386 read volume JobId=12922
> >> ====
> >>
> >> ====
> >>
> >>
> >>
> >> Why is Pool "*unknown*?? Device status show Total Bytes Read increasing
> >> each time I complete a status stirage check, BUT under the "Reading"
> >> section it shows "Bytes=0" and Bytes/sec=0"
> >>
> >> Finally the "job" completes after approx 25 minutes and the log captures
> >> this
> >>
> >>
> >> 09-Sep 10:11 BS01-DIR1 JobId 12922: Start Restore Job
> >> Restore_mail_bluewhale.2014-09-09_10.11.16_04
> >> 09-Sep 10:11 BS01-DIR1 JobId 12922: Using Device "File_bluewhale"
> >> 09-Sep 10:11 BS01-SD1 JobId 12922: Ready to read from volume "mail-0386" on
> >> device "File_bluewhale" (/nas/bacula/bluewhale).
> >> 09-Sep 10:11 BS01-SD1 JobId 12922: Forward spacing Volume "mail-0386" to
> >> file:block 0:219.
> >> 09-Sep 10:33 BS01-SD1 JobId 12922: End of Volume at file 28 on device
> >> "File_bluewhale" (/nas/bacula/bluewhale), Volume "mail-0386"
> >> 09-Sep 10:33 BS01-SD1 JobId 12922: End of all volumes.
> >> 09-Sep 10:12 BS01-FD1 JobId 12922: Error: create_file.c:292 Could not open
> >> /tmp/data/backups/mail/fifo/mail.tar: ERR=Interrupted system call
> >> 09-Sep 10:33 BS01-DIR1 JobId 12922: Bacula BS01-DIR1 5.2.2 (26Nov11):
> >> Build OS: x86_64-unknown-linux-gnu ubuntu 11.10
> >> JobId: 12922
> >> Job: Restore_mail_bluewhale.2014-09-09_10.11.16_04
> >> Restore Client: besc-bs01
> >> Start time: 09-Sep-2014 10:11:18
> >> End time: 09-Sep-2014 10:33:31
> >> Files Expected: 1
> >> Files Restored: 0
> >> Bytes Restored: 0
> >> Rate: 0.0 KB/s
> >> FD Errors: 0
> >> FD termination status: OK
> >> SD termination status: OK
> >> Termination: Restore OK -- warning file count mismatch
> >>
> >> 09-Sep 10:33 BS01-DIR1 JobId 12922: Begin pruning Jobs older than 15 days .
> >> 09-Sep 10:33 BS01-DIR1 JobId 12922: No Jobs found to prune.
> >> 09-Sep 10:33 BS01-DIR1 JobId 12922: Begin pruning Files.
> >> 09-Sep 10:33 BS01-DIR1 JobId 12922: No Files found to prune.
> >> 09-Sep 10:33 BS01-DIR1 JobId 12922: End auto prune.
> >>
> >>
> >> Thoughts??
> >>
> >> Thanks ----Kenny
> >>
> >>
> >>
> >>
> >>
> >> On Tue, Sep 9, 2014 at 9:56 AM, Kern Sibbald <kern AT sibbald DOT com>
> >> wrote:
> >>
> >>> Hello,
> >>>
> >>> I would remove the
> >>>
> >>> readfifo=yes
> >>>
> >>> though I am not 100% sure it is used on a restore.
> >>>
> >>> Then simply restore the file "mail.tar" making absolutely sure you have
> >>> not marked any directories for restore. Do the restore to /tmp. Then you
> >>> will have the mail.tar file that you can detar manually to get your files
> >>> back.
> >>>
> >>> Best regards,
> >>> Kern
> >>>
> >>> On 09/09/2014 03:25 PM, Kenny Noe wrote:
> >>>
> >>> Ana, Hi! Thanks for the reply... I get the same error no matter where
> >>> I try and write to. I've tried to a remote NAS and to local /tmp.
> >>>
> >>> Kern, below is my Fileset. Should I remove the "Include" statment?
> >>>
> >>> FileSet {
> >>> Name = Full_mail_bluewhale
> >>> Include {
> >>> Options {
> >>> signature=SHA1
> >>> readfifo=yes
> >>> }
> >>> File="/data/backups/mail/fifo/mail.tar"
> >>> }
> >>> }
> >>>
> >>>
> >>> Thank you all for your comments so far. I'm still trying to complete
> >>> the restore, so any input would be appreciated.
> >>>
> >>> Sincerely,
> >>> --Kenny
> >>>
> >>>
> >>> ...
> >>>
>
------------------------------------------------------------------------------
Want excitement?
Manually upgrade your production database.
When you want reliability, choose Perforce
Perforce version control. Predictably reliable.
http://pubads.g.doubleclick.net/gampad/clk?id=157508191&iu=/4140/ostg.clktrk
_______________________________________________
Bacula-users mailing list
Bacula-users AT lists.sourceforge DOT net
https://lists.sourceforge.net/lists/listinfo/bacula-users
|