Sorry,  failed to include the list...

---Kenny

---------- Forwarded message ----------
From: Kenny Noe <knoe...@gmail.com>
Date: Mon, Sep 8, 2014 at 1:25 PM
Subject: Re: [Bacula-users] Restore from dead client
To: Dan Langille <d...@langille.org>


Gents....  Following instructions.

Appreciate your time...

--Kenny

*************************************
Complete screen capture of restore....

*restore

First you select one or more JobIds that contain files
to be restored. You will be presented several methods
of specifying the JobIds. Then you will be allowed to
select which files from those JobIds are to be restored.

To select the JobIds, you have the following choices:
     1: List last 20 Jobs run
     2: List Jobs where a given File is saved
     3: Enter list of comma separated JobIds to select
     4: Enter SQL list command
     5: Select the most recent backup for a client
     6: Select backup for a client before a specified time
     7: Enter a list of files to restore
     8: Enter a list of files to restore before a specified time
     9: Find the JobIds of the most recent backup for a client
    10: Find the JobIds for a backup for a client before a specified time
    11: Enter a list of directories to restore for found JobIds
    12: Select full restore to a specified Job date
    13: Cancel
Select item:  (1-13): 3
Enter JobId(s), comma separated, to restore: 12510
You have selected the following JobId: 12510

Building directory tree for JobId(s) 12510 ...
1 files inserted into the tree.

You are now entering file selection mode where you add (mark) and
remove (unmark) files to be restored. No files are initially added, unless
you used the "all" keyword on the command line.
Enter "done" to leave this mode.

cwd is: /
$ mark *
1 file marked.
$ done
Bootstrap records written to /etc/bacula/BS01-DIR1.restore.4.bsr

The job will require the following
   Volume(s)                 Storage(s)                SD Device(s)
===========================================================================

    mail-0386                 Storage_bluewhale         File_bluewhale

Volumes marked with "*" are online.


1 file selected to be restored.

The defined Restore Job resources are:
     1: Restore_os_asterisk
     2: Restore_os_besc-4dvapp
     3: Restore_db_besc-bs01
     4: Restore_os_besc-bs01
     5: Restore_os_besc-unixmgr01
     6: Restore_mail_bluewhale
     7: Restore_os_demo
     8: Restore_app_demo
     9: Restore_os_dev
    10: Restore_app_dev
    11: Restore_os_mako
    12: Restore_app_mako
Select Restore Job (1-12): 6
Defined Clients:
     1: asterisk
     2: besc-4dvapp
     3: besc-bs01
     4: besc-unixmgr01
     5: bluewhale
     6: demo
     7: dev
     8: mako
     9: qa
    10: qa2
    11: smart
Select the Client (1-11): 5
Run Restore job
JobName:         Restore_mail_bluewhale
Bootstrap:       /etc/bacula/BS01-DIR1.restore.4.bsr
Where:           *None*
Replace:         always
FileSet:         Full_mail_bluewhale
Backup Client:   bluewhale
Restore Client:  bluewhale
Storage:         Storage_bluewhale
When:            2014-09-08 13:22:56
Catalog:         BS01-Catalog
Priority:        10
Plugin Options:  *None*
OK to run? (yes/mod/no): mod
Parameters to modify:
     1: Level
     2: Storage
     3: Job
     4: FileSet
     5: Restore Client
     6: When
     7: Priority
     8: Bootstrap
     9: Where
    10: File Relocation
    11: Replace
    12: JobId
    13: Plugin Options
Select parameter to modify (1-13): 5
The defined Client resources are:
     1: asterisk
     2: besc-4dvapp
     3: besc-bs01
     4: besc-unixmgr01
     5: bluewhale
     6: demo
     7: dev
     8: mako
Select Client (File daemon) resource (1-8): 3
Run Restore job
JobName:         Restore_mail_bluewhale
Bootstrap:       /etc/bacula/BS01-DIR1.restore.4.bsr
Where:           *None*
Replace:         always
FileSet:         Full_mail_bluewhale
Backup Client:   bluewhale
Restore Client:  besc-bs01
Storage:         Storage_bluewhale
When:            2014-09-08 13:22:56
Catalog:         BS01-Catalog
Priority:        10
Plugin Options:  *None*
OK to run? (yes/mod/no): mod
Parameters to modify:
     1: Level
     2: Storage
     3: Job
     4: FileSet
     5: Restore Client
     6: When
     7: Priority
     8: Bootstrap
     9: Where
    10: File Relocation
    11: Replace
    12: JobId
    13: Plugin Options
Select parameter to modify (1-13): 9
Please enter path prefix for restore (/ for none): /nas/users/admin/backups
Run Restore job
JobName:         Restore_mail_bluewhale
Bootstrap:       /etc/bacula/BS01-DIR1.restore.4.bsr
Where:           /nas/users/admin/backups
Replace:         always
FileSet:         Full_mail_bluewhale
Backup Client:   bluewhale
Restore Client:  besc-bs01
Storage:         Storage_bluewhale
When:            2014-09-08 13:22:56
Catalog:         BS01-Catalog
Priority:        10
Plugin Options:  *None*
OK to run? (yes/mod/no): yes
Job queued. JobId=12899
*





On Mon, Sep 8, 2014 at 1:16 PM, Dan Langille <d...@langille.org> wrote:

> Kenny:
>
> Please run the restore job again and then copy/paste the entire session
> from bcsonole.  Thank you.
>
>
> On Sep 8, 2014, at 1:15 PM, Heitor Faria <hei...@bacula.com.br> wrote:
>
> Kenny,
>
> Ok. Sorry for the confusion about the bluewhale client. This is why we
> need your full restore command, to see exactly what's happening..
> You should not change your Backup Client at the restore, even if the
> restore is dead. You should change your Restore Client. It would be nice to
> inform its version too.
>
> Regards,
>
> On Mon, Sep 8, 2014 at 2:11 PM, Kenny Noe <knoe...@gmail.com> wrote:
>
>> Heitor,
>>
>> Pardon my ignorance, but i don't follow your questions...
>>
>> My original client "bluewhale" experienced a terrible HD failure thus the
>> need to recover.   I have been changing the "client" and "where" parameters
>> when walking thru the restore procedure.  When the restore process runs it
>> creates the requested file "mail.tar" but there are zero bytes.
>>
>> If you get me a bit more to go on I'll gladly share input / output for
>> the request.
>>
>> Thanks   --Kenny
>>
>>
>>
>>
>> On Mon, Sep 8, 2014 at 12:57 PM, Heitor Faria <hei...@bacula.com.br>
>> wrote:
>>
>>> Kenny,
>>>
>>> 1. Could you reproduce the input you used at the restore submission?
>>> 2. Can you tell the client bluewhale version?
>>>
>>> Regards,
>>>
>>> On Mon, Sep 8, 2014 at 1:48 PM, Kenny Noe <knoe...@gmail.com> wrote:
>>>
>>>> Heitor,
>>>>
>>>> Hi!  Thanks for the reply....   I'm utilizing a console and the Bacula
>>>> console and executing a "restore" command then walking thru the prompts
>>>> given.  What is this "run" command?  Yes, I believe I'm trying to restore
>>>> the same file, but it's not working.
>>>>
>>>> Thoughts?
>>>>
>>>> Thanks    --Kenny
>>>>
>>>>
>>>> On Mon, Sep 8, 2014 at 12:41 PM, Heitor Faria <hei...@bacula.com.br>
>>>> wrote:
>>>>
>>>>> Kenny,
>>>>>
>>>>> First of all, are you using the "run" command to submit a prior
>>>>> configured Restore Job? I think this is not advisable, since there are
>>>>> several restore variables that only the "restore" command can fetch.
>>>>> Did you try to restore the same file with the restore command?
>>>>>
>>>>> Regards,
>>>>>
>>>>>
>>>>> On Mon, Sep 8, 2014 at 1:28 PM, Kenny Noe <knoe...@gmail.com> wrote:
>>>>>
>>>>>> Dan,
>>>>>>
>>>>>> Thanks for the reply.   I tried this this morning and still failed
>>>>>> the restore.   I see during the Status Storage - Storage_Bluewhale where
>>>>>> the "Running Jobs section shows Files=0, Bytes=0 Bytes/sec=0.  However in
>>>>>> the Device Status section, Device "File_bluewhale" is mounted and the 
>>>>>> Total
>>>>>> Bytes Read and Block Read go up....   Now with the simplefied config It
>>>>>> seems to have lost it's Pool.
>>>>>>
>>>>>> I'm confused...  I changed the where to restore to
>>>>>> /nas/users/admin/backups and changed in the client config to remove
>>>>>> the fifo headache...  But it still is trying to use it...
>>>>>>
>>>>>> Here is the error from the log :
>>>>>> 08-Sep 11:51 BS01-DIR1 JobId 12897: Start Restore Job
>>>>>> Restore_mail_bluewhale.2014-09-08_11.51.55_03
>>>>>> 08-Sep 11:51 BS01-DIR1 JobId 12897: Using Device "File_bluewhale"
>>>>>> 08-Sep 11:51 BS01-SD1 JobId 12897: Ready to read from volume
>>>>>> "mail-0386" on device "File_bluewhale" (/nas/bacula/bluewhale).
>>>>>> 08-Sep 11:51 BS01-SD1 JobId 12897: Forward spacing Volume "mail-0386"
>>>>>> to file:block 0:219.
>>>>>> 08-Sep 12:14 BS01-SD1 JobId 12897: End of Volume at file 28 on device
>>>>>> "File_bluewhale" (/nas/bacula/bluewhale), Volume "mail-0386"
>>>>>> 08-Sep 12:14 BS01-SD1 JobId 12897: End of all volumes.
>>>>>> 08-Sep 11:52 BS01-FD1 JobId 12897: Error: create_file.c:292 Could not
>>>>>> open /nas/users/admin/backups/data/backups/mail/fifo/mail.tar:
>>>>>> ERR=Interrupted system call
>>>>>> 08-Sep 12:14 BS01-DIR1 JobId 12897: Bacula BS01-DIR1 5.2.2 (26Nov11):
>>>>>>   Build OS:               x86_64-unknown-linux-gnu ubuntu 11.10
>>>>>>   JobId:                  12897
>>>>>>   Job:
>>>>>>  Restore_mail_bluewhale.2014-09-08_11.51.55_03
>>>>>>   Restore Client:         besc-bs01
>>>>>>   Start time:             08-Sep-2014 11:51:57
>>>>>>   End time:               08-Sep-2014 12:14:09
>>>>>>   Files Expected:         1
>>>>>>   Files Restored:         0
>>>>>>   Bytes Restored:         0
>>>>>>   Rate:                   0.0 KB/s
>>>>>>   FD Errors:              0
>>>>>>   FD termination status:  OK
>>>>>>   SD termination status:  OK
>>>>>>   Termination:            Restore OK -- warning file count mismatch
>>>>>>
>>>>>> 08-Sep 12:14 BS01-DIR1 JobId 12897: Begin pruning Jobs older than 15
>>>>>> days .
>>>>>> 08-Sep 12:14 BS01-DIR1 JobId 12897: No Jobs found to prune.
>>>>>> 08-Sep 12:14 BS01-DIR1 JobId 12897: Begin pruning Files.
>>>>>> 08-Sep 12:14 BS01-DIR1 JobId 12897: No Files found to prune.
>>>>>> 08-Sep 12:14 BS01-DIR1 JobId 12897: End auto prune.
>>>>>>
>>>>>>
>>>>>> What is a file count mismatch?
>>>>>>
>>>>>> Here is the status during a restore :
>>>>>> Connecting to Storage daemon Storage_bluewhale at 10.10.10.199:9103
>>>>>>
>>>>>> BS01-SD1 Version: 5.2.2 (26 November 2011) x86_64-unknown-linux-gnu
>>>>>> ubuntu 11.10
>>>>>> Daemon started 08-Sep-14 11:48. Jobs: run=0, running=0.
>>>>>>  Heap: heap=598,016 smbytes=386,922 max_bytes=405,712 bufs=947
>>>>>> max_bufs=949
>>>>>>  Sizes: boffset_t=8 size_t=8 int32_t=4 int64_t=8 mode=0,0
>>>>>>
>>>>>> Running Jobs:
>>>>>> Reading: Full Restore job Restore_mail_bluewhale JobId=12897
>>>>>> Volume="mail-0386"
>>>>>>     pool="Pool_mail_bluewhale" device="File_bluewhale"
>>>>>> (/nas/bacula/bluewhale)
>>>>>>     Files=0 Bytes=0 Bytes/sec=0
>>>>>>     FDReadSeqNo=6 in_msg=6 out_msg=84529 fd=6
>>>>>> ====
>>>>>>
>>>>>> Jobs waiting to reserve a drive:
>>>>>> ====
>>>>>>
>>>>>> Terminated Jobs:
>>>>>>  JobId  Level    Files      Bytes   Status   Finished        Name
>>>>>> ===================================================================
>>>>>>  12889  Incr         31    67.94 M  OK       08-Sep-14 00:01
>>>>>> Backup_os_besc-unixmgr01
>>>>>>  12891  Full          4    501.0 M  OK       08-Sep-14 00:05
>>>>>> Backup_app_dev
>>>>>>  12888  Incr        437    1.158 G  OK       08-Sep-14 00:06
>>>>>> Backup_os_besc-bs01
>>>>>>  12890  Incr          0         0   Other    08-Sep-14 00:30
>>>>>> Backup_os_bluewhale
>>>>>>  12893  Full          0         0   Other    08-Sep-14 01:30
>>>>>> Backup_mail_bluewhale
>>>>>>  12884  Full   2,361,101    154.6 G  OK       08-Sep-14 04:46
>>>>>> Backup_os_mako
>>>>>>  12892  Full          4    54.40 G  OK       08-Sep-14 05:56
>>>>>> Backup_app_mako
>>>>>>  12894                0         0   OK       08-Sep-14 08:53
>>>>>> Restore_mail_bluewhale
>>>>>>  12895                0         0   OK       08-Sep-14 09:28
>>>>>> Restore_mail_bluewhale
>>>>>>  12896                0         0   OK       08-Sep-14 10:10
>>>>>> Restore_mail_bluewhale
>>>>>> ====
>>>>>>
>>>>>> Device status:
>>>>>> Device "File_asterisk" (/nas/bacula/asterisk) is not open.
>>>>>> Device "File_besc-4dvapp" (/nas/bacula/besc-4dvapp) is not open.
>>>>>> Device "File_besc-bs01" (/nas/bacula/besc-bs01) is not open.
>>>>>> Device "File_besc-unixmgr01" (/nas/bacula/besc-unixmgr01) is not open.
>>>>>> Device "File_bluewhale" (/nas/bacula/bluewhale) is mounted with:
>>>>>>     Volume:      mail-0386
>>>>>>     Pool:        *unknown*
>>>>>>     Media type:  NAS_bluewhale
>>>>>>     Total Bytes Read=1,121,412,096 Blocks Read=17,383
>>>>>> Bytes/block=64,512
>>>>>>     Positioned at File=0 Block=1,121,412,275
>>>>>> Device "File_demo" (/nas/bacula/demo) is not open.
>>>>>> Device "File_dev" (/nas/bacula/dev) is not open.
>>>>>> Device "File_mako" (/nas/bacula/mako) is not open.
>>>>>> Device "File_qa" (/nas/bacula/qa) is not open.
>>>>>> Device "File_qa2" (/nas/bacula/qa2) is not open.
>>>>>> Device "File_smart" (/nas/bacula/smart) is not open.
>>>>>> ====
>>>>>>
>>>>>> Used Volume status:
>>>>>> mail-0386 on device "File_bluewhale" (/nas/bacula/bluewhale)
>>>>>>     Reader=1 writers=0 devres=0 volinuse=1
>>>>>> mail-0386 read volume JobId=12897
>>>>>> ====
>>>>>>
>>>>>> ====
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Here is my simplified client config:
>>>>>>
>>>>>>
>>>>>> #********************************************************************************
>>>>>> # bluewhale
>>>>>>
>>>>>> #********************************************************************************
>>>>>>    Client {
>>>>>>       Name                   = bluewhale
>>>>>>       Address                = bluewhale.bnesystems.com
>>>>>>       Catalog                = BS01-Catalog
>>>>>>       Password               = "xxxxxxxxx"
>>>>>>       FileRetention          = 365 days
>>>>>>       JobRetention           = 365 days
>>>>>>       AutoPrune              = yes
>>>>>>       MaximumConcurrentJobs  = 1
>>>>>>    }
>>>>>>    Job {
>>>>>>       Name                   = Restore_mail_bluewhale
>>>>>>       FileSet                = Full_mail_bluewhale
>>>>>>       Type                   = Restore
>>>>>>       Pool                   = Pool_mail_bluewhale
>>>>>>       Client                 = bluewhale
>>>>>>       Messages               = Standard
>>>>>>    }
>>>>>>    Pool {
>>>>>>       Name                   = Pool_mail_bluewhale
>>>>>>       PoolType               = Backup
>>>>>>       Storage                = Storage_bluewhale
>>>>>>       MaximumVolumeJobs      = 1
>>>>>>       CatalogFiles           = yes
>>>>>>       AutoPrune              = yes
>>>>>>       VolumeRetention        = 365 days
>>>>>>       Recycle                = yes
>>>>>>       LabelFormat            = "mail-"
>>>>>>    }
>>>>>>    Storage {
>>>>>>       Name                   = Storage_bluewhale
>>>>>>       Address                = 10.10.10.199
>>>>>>       SDPort                 = 9103
>>>>>>       Password               = "xxxxxxx"
>>>>>>       Device                 = File_bluewhale
>>>>>>       MediaType              = NAS_bluewhale
>>>>>>       MaximumConcurrentJobs  = 1
>>>>>>    }
>>>>>>    FileSet {
>>>>>>       Name = Full_mail_bluewhale
>>>>>>       Include {
>>>>>>          Options {
>>>>>>             signature=SHA1
>>>>>> #            readfifo=yes
>>>>>>          }
>>>>>>          File="/mail.tar"
>>>>>>       }
>>>>>>    }
>>>>>>
>>>>>>
>>>>>>
>>>>>> Thanks for the help.   I appreciate all the input.
>>>>>>
>>>>>> --Kenny
>>>>>>
>>>>>>
>>>>>> On Sun, Sep 7, 2014 at 8:22 AM, Dan Langille <d...@langille.org>
>>>>>> wrote:
>>>>>>
>>>>>>> I suggest removing the before & after scripts.
>>>>>>>
>>>>>>> --
>>>>>>> Dan Langille
>>>>>>> http://langille.org/
>>>>>>>
>>>>>>>
>>>>>>> > On Sep 6, 2014, at 8:38 PM, Kenny Noe <knoe...@gmail.com> wrote:
>>>>>>> >
>>>>>>> > Dan,
>>>>>>> >
>>>>>>> > Appreciate the reply....   Yes this is exactly what I want to do.
>>>>>>> > However when I try to just do a "simple" restore, the job finishes
>>>>>>> > with the error previously given.
>>>>>>> >
>>>>>>> > Any suggestions to do this would be appreciated.
>>>>>>> >
>>>>>>> > Thanks    --Kenny
>>>>>>> >
>>>>>>> >> On Sat, Sep 6, 2014 at 5:51 PM, Dan Langille <d...@langille.org>
>>>>>>> wrote:
>>>>>>> >>
>>>>>>> >> On Sep 5, 2014, at 5:48 PM, Kenny Noe <knoe...@gmail.com> wrote:
>>>>>>> >>
>>>>>>> >> Birre,
>>>>>>> >>
>>>>>>> >> Thanks for the reply.   I guess this is where I get lost...
>>>>>>> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> The fifo is reading a file that was created in the pre-process
>>>>>>> called
>>>>>>> >> mail.tar.  The mail.tar is made from the following directories
>>>>>>> /opt/zimbra
>>>>>>> >> and /var/mail/zimbra.  This is where the Zimbra files and
>>>>>>> mailstore were
>>>>>>> >> kept.
>>>>>>> >>
>>>>>>> >> This pre-process is a script that has this :
>>>>>>> >>
>>>>>>> >> MailBackup.bash
>>>>>>> >> #!/bin/bash
>>>>>>> >>
>>>>>>> >> exec >/dev/null
>>>>>>> >>
>>>>>>> >> MKDIR="/bin/mkdir"
>>>>>>> >> MKFIFO="/usr/bin/mkfifo"
>>>>>>> >> RM="/bin/rm"
>>>>>>> >> TAR="/bin/tar"
>>>>>>> >>
>>>>>>> >> DEFCODE=0
>>>>>>> >> DUMPBASE="/data/backups"
>>>>>>> >>
>>>>>>> >> errCode=${DEFCODE}
>>>>>>> >> mailDir="/var/mail/zimbra"
>>>>>>> >> zimbraDir="/opt/zimbra"
>>>>>>> >>
>>>>>>> >> Main()
>>>>>>> >>   {
>>>>>>> >>   service zimbra stop
>>>>>>> >>
>>>>>>> >>   RunMailRestore
>>>>>>> >>
>>>>>>> >>   service zimbra start
>>>>>>> >>
>>>>>>> >>   ExitScript ${errCode}
>>>>>>> >>   }
>>>>>>> >>
>>>>>>> >> RunMailRestore()
>>>>>>> >>   {
>>>>>>> >>   EXTENSION=".tar"
>>>>>>> >>
>>>>>>> >>   dumpDir="${DUMPBASE}/mail"
>>>>>>> >>   fifoDir="${dumpDir}/fifo"
>>>>>>> >>
>>>>>>> >>   RebuildFifoDir
>>>>>>> >>
>>>>>>> >>   ${MKFIFO} ${fifoDir}/mail${EXTENSION}
>>>>>>> >>
>>>>>>> >>   ${TAR} -xpf ${fifoDir}/mail${EXTENSION} 2>&1 </dev/null &
>>>>>>> >>   }
>>>>>>> >>
>>>>>>> >> RebuildFifoDir()
>>>>>>> >>   {
>>>>>>> >>   if [ -d ${fifoDir} ]
>>>>>>> >>   then
>>>>>>> >>      ${RM} -rf ${fifoDir}
>>>>>>> >>   fi
>>>>>>> >>
>>>>>>> >>   ${MKDIR} -p ${fifoDir}
>>>>>>> >>   }
>>>>>>> >>
>>>>>>> >> ExitScript()
>>>>>>> >>   {
>>>>>>> >>   exit ${1}
>>>>>>> >>   }
>>>>>>> >>
>>>>>>> >> Main
>>>>>>> >>
>>>>>>> >> The restore script simply does a tar xpf instead of a tar cpf.
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> Perhaps instead of doing that, just restore the data, and then do
>>>>>>> the tar
>>>>>>> >> xpf later.
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> ------------------------------------------------------------------------------
>>>>>> Want excitement?
>>>>>> Manually upgrade your production database.
>>>>>> When you want reliability, choose Perforce
>>>>>> Perforce version control. Predictably reliable.
>>>>>>
>>>>>> http://pubads.g.doubleclick.net/gampad/clk?id=157508191&iu=/4140/ostg.clktrk
>>>>>> _______________________________________________
>>>>>> Bacula-users mailing list
>>>>>> Bacula-users@lists.sourceforge.net
>>>>>> https://lists.sourceforge.net/lists/listinfo/bacula-users
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> ============================================
>>>>> Heitor Medrado de Faria | Need Bacula training? 10% discount coupon
>>>>> code at Udemy: bacula-users
>>>>> <https://www.udemy.com/bacula-backup-software/?couponCode=bacula-users>
>>>>> +55 61 2021-8260
>>>>> +55 61 8268-4220
>>>>> Site: www.bacula.com.br
>>>>> Facebook: heitor.faria <http://www.facebook.com/heitor.faria>
>>>>> Gtalk: heitorfa...@gmail.com
>>>>> ============================================
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> ============================================
>>> Heitor Medrado de Faria | Need Bacula training? 10% discount coupon
>>> code at Udemy: bacula-users
>>> <https://www.udemy.com/bacula-backup-software/?couponCode=bacula-users>
>>> +55 61 2021-8260
>>> +55 61 8268-4220
>>> Site: www.bacula.com.br
>>> Facebook: heitor.faria <http://www.facebook.com/heitor.faria>
>>> Gtalk: heitorfa...@gmail.com
>>> ============================================
>>>
>>
>>
>
>
> --
> ============================================
> Heitor Medrado de Faria | Need Bacula training? 10% discount coupon code
> at Udemy: bacula-users
> <https://www.udemy.com/bacula-backup-software/?couponCode=bacula-users>
> +55 61 2021-8260
> +55 61 8268-4220
> Site: www.bacula.com.br
> Facebook: heitor.faria <http://www.facebook.com/heitor.faria>
> Gtalk: heitorfa...@gmail.com
> ============================================
>
>
>
------------------------------------------------------------------------------
Want excitement?
Manually upgrade your production database.
When you want reliability, choose Perforce
Perforce version control. Predictably reliable.
http://pubads.g.doubleclick.net/gampad/clk?id=157508191&iu=/4140/ostg.clktrk
_______________________________________________
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

Reply via email to