Heitor,

Hi!  Thanks for the reply....   I'm utilizing a console and the Bacula
console and executing a "restore" command then walking thru the prompts
given.  What is this "run" command?  Yes, I believe I'm trying to restore
the same file, but it's not working.

Thoughts?

Thanks    --Kenny


On Mon, Sep 8, 2014 at 12:41 PM, Heitor Faria <hei...@bacula.com.br> wrote:

> Kenny,
>
> First of all, are you using the "run" command to submit a prior configured
> Restore Job? I think this is not advisable, since there are several restore
> variables that only the "restore" command can fetch.
> Did you try to restore the same file with the restore command?
>
> Regards,
>
>
> On Mon, Sep 8, 2014 at 1:28 PM, Kenny Noe <knoe...@gmail.com> wrote:
>
>> Dan,
>>
>> Thanks for the reply.   I tried this this morning and still failed the
>> restore.   I see during the Status Storage - Storage_Bluewhale where the
>> "Running Jobs section shows Files=0, Bytes=0 Bytes/sec=0.  However in the
>> Device Status section, Device "File_bluewhale" is mounted and the Total
>> Bytes Read and Block Read go up....   Now with the simplefied config It
>> seems to have lost it's Pool.
>>
>> I'm confused...  I changed the where to restore to
>> /nas/users/admin/backups and changed in the client config to remove the
>> fifo headache...  But it still is trying to use it...
>>
>> Here is the error from the log :
>> 08-Sep 11:51 BS01-DIR1 JobId 12897: Start Restore Job
>> Restore_mail_bluewhale.2014-09-08_11.51.55_03
>> 08-Sep 11:51 BS01-DIR1 JobId 12897: Using Device "File_bluewhale"
>> 08-Sep 11:51 BS01-SD1 JobId 12897: Ready to read from volume "mail-0386"
>> on device "File_bluewhale" (/nas/bacula/bluewhale).
>> 08-Sep 11:51 BS01-SD1 JobId 12897: Forward spacing Volume "mail-0386" to
>> file:block 0:219.
>> 08-Sep 12:14 BS01-SD1 JobId 12897: End of Volume at file 28 on device
>> "File_bluewhale" (/nas/bacula/bluewhale), Volume "mail-0386"
>> 08-Sep 12:14 BS01-SD1 JobId 12897: End of all volumes.
>> 08-Sep 11:52 BS01-FD1 JobId 12897: Error: create_file.c:292 Could not
>> open /nas/users/admin/backups/data/backups/mail/fifo/mail.tar:
>> ERR=Interrupted system call
>> 08-Sep 12:14 BS01-DIR1 JobId 12897: Bacula BS01-DIR1 5.2.2 (26Nov11):
>>   Build OS:               x86_64-unknown-linux-gnu ubuntu 11.10
>>   JobId:                  12897
>>   Job:                    Restore_mail_bluewhale.2014-09-08_11.51.55_03
>>   Restore Client:         besc-bs01
>>   Start time:             08-Sep-2014 11:51:57
>>   End time:               08-Sep-2014 12:14:09
>>   Files Expected:         1
>>   Files Restored:         0
>>   Bytes Restored:         0
>>   Rate:                   0.0 KB/s
>>   FD Errors:              0
>>   FD termination status:  OK
>>   SD termination status:  OK
>>   Termination:            Restore OK -- warning file count mismatch
>>
>> 08-Sep 12:14 BS01-DIR1 JobId 12897: Begin pruning Jobs older than 15 days
>> .
>> 08-Sep 12:14 BS01-DIR1 JobId 12897: No Jobs found to prune.
>> 08-Sep 12:14 BS01-DIR1 JobId 12897: Begin pruning Files.
>> 08-Sep 12:14 BS01-DIR1 JobId 12897: No Files found to prune.
>> 08-Sep 12:14 BS01-DIR1 JobId 12897: End auto prune.
>>
>>
>> What is a file count mismatch?
>>
>> Here is the status during a restore :
>> Connecting to Storage daemon Storage_bluewhale at 10.10.10.199:9103
>>
>> BS01-SD1 Version: 5.2.2 (26 November 2011) x86_64-unknown-linux-gnu
>> ubuntu 11.10
>> Daemon started 08-Sep-14 11:48. Jobs: run=0, running=0.
>>  Heap: heap=598,016 smbytes=386,922 max_bytes=405,712 bufs=947
>> max_bufs=949
>>  Sizes: boffset_t=8 size_t=8 int32_t=4 int64_t=8 mode=0,0
>>
>> Running Jobs:
>> Reading: Full Restore job Restore_mail_bluewhale JobId=12897
>> Volume="mail-0386"
>>     pool="Pool_mail_bluewhale" device="File_bluewhale"
>> (/nas/bacula/bluewhale)
>>     Files=0 Bytes=0 Bytes/sec=0
>>     FDReadSeqNo=6 in_msg=6 out_msg=84529 fd=6
>> ====
>>
>> Jobs waiting to reserve a drive:
>> ====
>>
>> Terminated Jobs:
>>  JobId  Level    Files      Bytes   Status   Finished        Name
>> ===================================================================
>>  12889  Incr         31    67.94 M  OK       08-Sep-14 00:01
>> Backup_os_besc-unixmgr01
>>  12891  Full          4    501.0 M  OK       08-Sep-14 00:05
>> Backup_app_dev
>>  12888  Incr        437    1.158 G  OK       08-Sep-14 00:06
>> Backup_os_besc-bs01
>>  12890  Incr          0         0   Other    08-Sep-14 00:30
>> Backup_os_bluewhale
>>  12893  Full          0         0   Other    08-Sep-14 01:30
>> Backup_mail_bluewhale
>>  12884  Full   2,361,101    154.6 G  OK       08-Sep-14 04:46
>> Backup_os_mako
>>  12892  Full          4    54.40 G  OK       08-Sep-14 05:56
>> Backup_app_mako
>>  12894                0         0   OK       08-Sep-14 08:53
>> Restore_mail_bluewhale
>>  12895                0         0   OK       08-Sep-14 09:28
>> Restore_mail_bluewhale
>>  12896                0         0   OK       08-Sep-14 10:10
>> Restore_mail_bluewhale
>> ====
>>
>> Device status:
>> Device "File_asterisk" (/nas/bacula/asterisk) is not open.
>> Device "File_besc-4dvapp" (/nas/bacula/besc-4dvapp) is not open.
>> Device "File_besc-bs01" (/nas/bacula/besc-bs01) is not open.
>> Device "File_besc-unixmgr01" (/nas/bacula/besc-unixmgr01) is not open.
>> Device "File_bluewhale" (/nas/bacula/bluewhale) is mounted with:
>>     Volume:      mail-0386
>>     Pool:        *unknown*
>>     Media type:  NAS_bluewhale
>>     Total Bytes Read=1,121,412,096 Blocks Read=17,383 Bytes/block=64,512
>>     Positioned at File=0 Block=1,121,412,275
>> Device "File_demo" (/nas/bacula/demo) is not open.
>> Device "File_dev" (/nas/bacula/dev) is not open.
>> Device "File_mako" (/nas/bacula/mako) is not open.
>> Device "File_qa" (/nas/bacula/qa) is not open.
>> Device "File_qa2" (/nas/bacula/qa2) is not open.
>> Device "File_smart" (/nas/bacula/smart) is not open.
>> ====
>>
>> Used Volume status:
>> mail-0386 on device "File_bluewhale" (/nas/bacula/bluewhale)
>>     Reader=1 writers=0 devres=0 volinuse=1
>> mail-0386 read volume JobId=12897
>> ====
>>
>> ====
>>
>>
>>
>>
>> Here is my simplified client config:
>>
>>
>> #********************************************************************************
>> # bluewhale
>>
>> #********************************************************************************
>>    Client {
>>       Name                   = bluewhale
>>       Address                = bluewhale.bnesystems.com
>>       Catalog                = BS01-Catalog
>>       Password               = "xxxxxxxxx"
>>       FileRetention          = 365 days
>>       JobRetention           = 365 days
>>       AutoPrune              = yes
>>       MaximumConcurrentJobs  = 1
>>    }
>>    Job {
>>       Name                   = Restore_mail_bluewhale
>>       FileSet                = Full_mail_bluewhale
>>       Type                   = Restore
>>       Pool                   = Pool_mail_bluewhale
>>       Client                 = bluewhale
>>       Messages               = Standard
>>    }
>>    Pool {
>>       Name                   = Pool_mail_bluewhale
>>       PoolType               = Backup
>>       Storage                = Storage_bluewhale
>>       MaximumVolumeJobs      = 1
>>       CatalogFiles           = yes
>>       AutoPrune              = yes
>>       VolumeRetention        = 365 days
>>       Recycle                = yes
>>       LabelFormat            = "mail-"
>>    }
>>    Storage {
>>       Name                   = Storage_bluewhale
>>       Address                = 10.10.10.199
>>       SDPort                 = 9103
>>       Password               = "xxxxxxx"
>>       Device                 = File_bluewhale
>>       MediaType              = NAS_bluewhale
>>       MaximumConcurrentJobs  = 1
>>    }
>>    FileSet {
>>       Name = Full_mail_bluewhale
>>       Include {
>>          Options {
>>             signature=SHA1
>> #            readfifo=yes
>>          }
>>          File="/mail.tar"
>>       }
>>    }
>>
>>
>>
>> Thanks for the help.   I appreciate all the input.
>>
>> --Kenny
>>
>>
>> On Sun, Sep 7, 2014 at 8:22 AM, Dan Langille <d...@langille.org> wrote:
>>
>>> I suggest removing the before & after scripts.
>>>
>>> --
>>> Dan Langille
>>> http://langille.org/
>>>
>>>
>>> > On Sep 6, 2014, at 8:38 PM, Kenny Noe <knoe...@gmail.com> wrote:
>>> >
>>> > Dan,
>>> >
>>> > Appreciate the reply....   Yes this is exactly what I want to do.
>>> > However when I try to just do a "simple" restore, the job finishes
>>> > with the error previously given.
>>> >
>>> > Any suggestions to do this would be appreciated.
>>> >
>>> > Thanks    --Kenny
>>> >
>>> >> On Sat, Sep 6, 2014 at 5:51 PM, Dan Langille <d...@langille.org>
>>> wrote:
>>> >>
>>> >> On Sep 5, 2014, at 5:48 PM, Kenny Noe <knoe...@gmail.com> wrote:
>>> >>
>>> >> Birre,
>>> >>
>>> >> Thanks for the reply.   I guess this is where I get lost...
>>> >>
>>> >>
>>> >>
>>> >> The fifo is reading a file that was created in the pre-process called
>>> >> mail.tar.  The mail.tar is made from the following directories
>>> /opt/zimbra
>>> >> and /var/mail/zimbra.  This is where the Zimbra files and mailstore
>>> were
>>> >> kept.
>>> >>
>>> >> This pre-process is a script that has this :
>>> >>
>>> >> MailBackup.bash
>>> >> #!/bin/bash
>>> >>
>>> >> exec >/dev/null
>>> >>
>>> >> MKDIR="/bin/mkdir"
>>> >> MKFIFO="/usr/bin/mkfifo"
>>> >> RM="/bin/rm"
>>> >> TAR="/bin/tar"
>>> >>
>>> >> DEFCODE=0
>>> >> DUMPBASE="/data/backups"
>>> >>
>>> >> errCode=${DEFCODE}
>>> >> mailDir="/var/mail/zimbra"
>>> >> zimbraDir="/opt/zimbra"
>>> >>
>>> >> Main()
>>> >>   {
>>> >>   service zimbra stop
>>> >>
>>> >>   RunMailRestore
>>> >>
>>> >>   service zimbra start
>>> >>
>>> >>   ExitScript ${errCode}
>>> >>   }
>>> >>
>>> >> RunMailRestore()
>>> >>   {
>>> >>   EXTENSION=".tar"
>>> >>
>>> >>   dumpDir="${DUMPBASE}/mail"
>>> >>   fifoDir="${dumpDir}/fifo"
>>> >>
>>> >>   RebuildFifoDir
>>> >>
>>> >>   ${MKFIFO} ${fifoDir}/mail${EXTENSION}
>>> >>
>>> >>   ${TAR} -xpf ${fifoDir}/mail${EXTENSION} 2>&1 </dev/null &
>>> >>   }
>>> >>
>>> >> RebuildFifoDir()
>>> >>   {
>>> >>   if [ -d ${fifoDir} ]
>>> >>   then
>>> >>      ${RM} -rf ${fifoDir}
>>> >>   fi
>>> >>
>>> >>   ${MKDIR} -p ${fifoDir}
>>> >>   }
>>> >>
>>> >> ExitScript()
>>> >>   {
>>> >>   exit ${1}
>>> >>   }
>>> >>
>>> >> Main
>>> >>
>>> >> The restore script simply does a tar xpf instead of a tar cpf.
>>> >>
>>> >>
>>> >> Perhaps instead of doing that, just restore the data, and then do the
>>> tar
>>> >> xpf later.
>>>
>>
>>
>>
>> ------------------------------------------------------------------------------
>> Want excitement?
>> Manually upgrade your production database.
>> When you want reliability, choose Perforce
>> Perforce version control. Predictably reliable.
>>
>> http://pubads.g.doubleclick.net/gampad/clk?id=157508191&iu=/4140/ostg.clktrk
>> _______________________________________________
>> Bacula-users mailing list
>> Bacula-users@lists.sourceforge.net
>> https://lists.sourceforge.net/lists/listinfo/bacula-users
>>
>>
>
>
> --
> ============================================
> Heitor Medrado de Faria | Need Bacula training? 10% discount coupon code
> at Udemy: bacula-users
> <https://www.udemy.com/bacula-backup-software/?couponCode=bacula-users>
> +55 61 2021-8260
> +55 61 8268-4220
> Site: www.bacula.com.br
> Facebook: heitor.faria <http://www.facebook.com/heitor.faria>
> Gtalk: heitorfa...@gmail.com
> ============================================
>
------------------------------------------------------------------------------
Want excitement?
Manually upgrade your production database.
When you want reliability, choose Perforce
Perforce version control. Predictably reliable.
http://pubads.g.doubleclick.net/gampad/clk?id=157508191&iu=/4140/ostg.clktrk
_______________________________________________
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

Reply via email to