Hi Alexander,

it looks like we ran into the same problem (and wrote the post at the exact 
same time. My problem is described here: 
https://groups.google.com/g/bareos-users/c/6YBoLojr2ac

Alexander P schrieb am Donnerstag, 28. März 2024 um 09:28:33 UTC+1:

> Hello everyone,
> Yesterday I updated from Bareos 22.1.2 to 23.0.3.
> I also updated the database as described here
> https://docs.benbenos.org/introductionandtutorial/updatingbareos.html
>
> Now I have the problem that copy jobs from a storage to another. After 
> exactly one minute I get the email that the job failed.
> But when I look at the two storages then I can see that they still copied.
>
> There is nowhere else a message or error (Syslog/Postgress/....). Only 
> this is what makes it difficult to find the error.
>
>
>
> Logfile Bareos Director
> ##############
> 28-Mar 09:13 bareos-dir.mgm.domain.tld JobId 418788: Bareos 
> bareos-dir.mgm.domain.tld 23.0.3~pre47.36e516c0b (19Mar24):
>   Build OS:               Ubuntu 20.04.5 LTS
>   Current JobId:          418788
>   Current Job:            Copy_OVH_Full_To_Muc.2024-03-28_09.13.55_57
>   Catalog:                "MyCatalog" (From Default catalog)
>   Start time:             28-Mar-2024 09:13:57
>   End time:               28-Mar-2024 09:13:58
>   Elapsed time:           1 sec
>   Priority:               100
>   Bareos binary info:     Bareos community build (UNSUPPORTED): Get 
> professional support from https://www.bareos.com
>   Job triggered by:       User
>   Termination:            Copying OK
>
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418789: Start Copying JobId 
> 418789, Job=Copy_OVH_Full_To_Muc.2024-03-28_09.13.57_01
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418789: Connected Storage 
> daemon at bareos-sd01.mgm.domain.tld:9103, encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418789:  Encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: Connected Storage 
> daemon at bareos-sd01.mgm.muc01.fti.int:9103, encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790:  Encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418789: Using Device 
> "Disk_Drive000" to read.
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: There are no more 
> Jobs associated with Volume "MUC_Disk-Tape-3779". Marking it purged.
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: All records pruned 
> from Volume "MUC_Disk-Tape-3779"; marking it "Purged"
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: Recycled volume 
> "MUC_Disk-Tape-3779"
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: Using Device 
> "Disk_Drive000" to write.
> 28-Mar 09:14 bareos-sd01.mgm.domain.tld JobId 418789: Connected Storage 
> daemon at bareos-sd01.mgm.muc01.fti.int:9103, encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-sd01.mgm.domain.tld JobId 418789: Ready to read from 
> volume "OVH_Disk-Tape-3110" on device "Disk_Drive000" 
> (/var/lib/bareos/storage_ovh/Disk_Drive000).
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: 3307 Issuing 
> autochanger "unload slot 5106, drive 0" command.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: 3304 Issuing 
> autochanger "load slot 3779, drive 0" command.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: 3305 Autochanger 
> "load slot 3779, drive 0", status is OK.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: Recycled volume 
> "MUC_Disk-Tape-3779" on device "Disk_Drive000" 
> (/var/lib/bareos/storage_muc/Disk_Drive000), all previous data lost.
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: Max Volume jobs=1 
> exceeded. Marking Volume "MUC_Disk-Tape-3779" as Used.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: autoxflate-sd: 
> MUC_File-Autochanger OUT:[SD->inflate=yes->deflate=no->DEV] 
> IN:[DEV->inflate=yes->deflate=no->SD]
> 28-Mar 09:14 bareos-sd01.mgm.domain.tld JobId 418789: Forward spacing 
> Volume "OVH_Disk-Tape-3110" to file:block 0:257.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: Spooling data ...
> 28-Mar 09:15 bareos-dir.mgm.domain.tld JobId 418789: Error: Bareos 
> bareos-dir.mgm.domain.tld 23.0.3~pre47.36e516c0b (19Mar24):
>   Build OS:               Ubuntu 20.04.5 LTS
>   Prev Backup JobId:      418773
>   Prev Backup Job:       
>  fai-salt.mgm.domain.tld_LinuxAll_Job.2024-03-28_07.58.55_32
>   New Backup JobId:       418790
>   Current JobId:          418789
>   Current Job:            Copy_OVH_Full_To_Muc.2024-03-28_09.13.57_01
>   Backup Level:           Incremental
>   Client:                 fai-salt.mgm.domain.tld
>   FileSet:                "LinuxAll"
>   Read Pool:              "OVH_Full" (From Job resource)
>   Read Storage:           "OVH_File-Autochanger" (From Pool resource)
>   Write Pool:             "MUC_Copy_Full" (From Job Pool's NextPool 
> resource)
>   Write Storage:          "MUC_File-Autochanger" (From Storage from Pool's 
> NextPool resource)
>   Next Pool:              "MUC_Copy_Full" (From Job Pool's NextPool 
> resource)
>   Catalog:                "MyCatalog" (From Default catalog)
>   Start time:             28-Mar-2024 09:14:00
>   End time:               28-Mar-2024 09:15:00
>   Elapsed time:           1 min
>   Priority:               100
>   SD Files Written:       0
>   SD Bytes Written:       0 (0 B)
>   Rate:                   0.0 KB/s
>   Volume name(s):         MUC_Disk-Tape-3779
>   Volume Session Id:      73
>   Volume Session Time:    1711600421
>   Last Volume Bytes:      254 (254 B)
>   SD Errors:              0
>   SD termination status:  Running
>   Bareos binary info:     Bareos community build (UNSUPPORTED): Get 
> professional support from https://www.bareos.com
>   Job triggered by:       User
>   Termination:            *** Copying Error ***
>
>
>
> Email Message
> ##############
> 28-Mar 09:13 bareos-dir.mgm.domain.tld JobId 418789: Copying using 
> JobId=418773 Job=fai-salt.mgm.domain.tld_LinuxAll_Job.2024-03-28_07.58.55_32
> 28-Mar 09:13 bareos-dir.mgm.domain.tld JobId 418789: Bootstrap records 
> written to /var/lib/bareos/bareos-dir.mgm.domain.tld.restore.2.bsr
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418789: Start Copying JobId 
> 418789, Job=Copy_OVH_Full_To_Muc.2024-03-28_09.13.57_01
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418789: Connected Storage 
> daemon at bareos-sd01.mgm.domain.tld:9103, encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418789:  Encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: Connected Storage 
> daemon at bareos-sd01.mgm.muc01.fti.int:9103, encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790:  Encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418789: Using Device 
> "Disk_Drive000" to read.
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: There are no more 
> Jobs associated with Volume "MUC_Disk-Tape-3779". Marking it purged.
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: All records pruned 
> from Volume "MUC_Disk-Tape-3779"; marking it "Purged"
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: Recycled volume 
> "MUC_Disk-Tape-3779"
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: Using Device 
> "Disk_Drive000" to write.
> 28-Mar 09:14 bareos-sd01.mgm.domain.tld JobId 418789: Connected Storage 
> daemon at bareos-sd01.mgm.muc01.fti.int:9103, encryption: 
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 28-Mar 09:14 bareos-sd01.mgm.domain.tld JobId 418789: Ready to read from 
> volume "OVH_Disk-Tape-3110" on device "Disk_Drive000" 
> (/var/lib/bareos/storage_ovh/Disk_Drive000).
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: 3307 Issuing 
> autochanger "unload slot 5106, drive 0" command.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: 3304 Issuing 
> autochanger "load slot 3779, drive 0" command.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: 3305 Autochanger 
> "load slot 3779, drive 0", status is OK.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: Recycled volume 
> "MUC_Disk-Tape-3779" on device "Disk_Drive000" 
> (/var/lib/bareos/storage_muc/Disk_Drive000), all previous data lost.
> 28-Mar 09:14 bareos-dir.mgm.domain.tld JobId 418790: Max Volume jobs=1 
> exceeded. Marking Volume "MUC_Disk-Tape-3779" as Used.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: autoxflate-sd: 
> MUC_File-Autochanger OUT:[SD->inflate=yes->deflate=no->DEV] 
> IN:[DEV->inflate=yes->deflate=no->SD]
> 28-Mar 09:14 bareos-sd01.mgm.domain.tld JobId 418789: Forward spacing 
> Volume "OVH_Disk-Tape-3110" to file:block 0:257.
> 28-Mar 09:14 bareos-sd01.mgm.muc01.fti.int JobId 418790: Spooling data ...
> 28-Mar 09:15 bareos-dir.mgm.domain.tld JobId 418789: Error: Bareos 
> bareos-dir.mgm.domain.tld 23.0.3~pre47.36e516c0b (19Mar24):
>   Build OS:               Ubuntu 20.04.5 LTS
>   Prev Backup JobId:      418773
>   Prev Backup Job:       
>  fai-salt.mgm.domain.tld_LinuxAll_Job.2024-03-28_07.58.55_32
>   New Backup JobId:       418790
>   Current JobId:          418789
>   Current Job:            Copy_OVH_Full_To_Muc.2024-03-28_09.13.57_01
>   Backup Level:           Incremental
>   Client:                 fai-salt.mgm.domain.tld
>   FileSet:                "LinuxAll"
>   Read Pool:              "OVH_Full" (From Job resource)
>   Read Storage:           "OVH_File-Autochanger" (From Pool resource)
>   Write Pool:             "MUC_Copy_Full" (From Job Pool's NextPool 
> resource)
>   Write Storage:          "MUC_File-Autochanger" (From Storage from Pool's 
> NextPool resource)
>   Next Pool:              "MUC_Copy_Full" (From Job Pool's NextPool 
> resource)
>   Catalog:                "MyCatalog" (From Default catalog)
>   Start time:             28-Mar-2024 09:14:00
>   End time:               28-Mar-2024 09:15:00
>   Elapsed time:           1 min 
>   Priority:               100
>   SD Files Written:       0
>   SD Bytes Written:       0 (0 B)
>   Rate:                   0.0 KB/s
>   Volume name(s):         MUC_Disk-Tape-3779
>   Volume Session Id:      73
>   Volume Session Time:    1711600421
>   Last Volume Bytes:      254 (254 B)
>   SD Errors:              0
>   SD termination status:  Running
>   Bareos binary info:     Bareos community build (UNSUPPORTED): Get 
> professional support from https://www.bareos.com
>   Job triggered by:       User
>   Termination:            *** Copying Error ***
>
>
> The Jobs Are running
>
> Source Storage
> ##############
> JobId=418789 Level=Incremental Type=Copy Name=Copy_OVH_Full_To_Muc 
> Status=Running
> Reading: Volume="OVH_Disk-Tape-3110"
>     pool="OVH_Full" device="Disk_Drive000" 
> (/var/lib/bareos/storage_ovh/Disk_Drive000)
> Writing: Volume="OVH_Disk-Tape-3110"
>     pool="OVH_Full" device="Disk_Drive000" 
> (/var/lib/bareos/storage_ovh/Disk_Drive000)
>     spooling=0 despooling=0 despool_wait=0
>     Files=173,960 Bytes=7,699,083,169 AveBytes/sec=4 
> LastBytes/sec=23,363,584
>     FDSocket closed
>
>
> Destination Storage 
> ##############
> JobId=418790 Level=Full Type=Backup 
> Name=fai-salt.mgm.domain.tld_LinuxAll_Job Status=Running
> Writing: Volume="MUC_Disk-Tape-3779"
>     pool="MUC_Copy_Full" device="Disk_Drive000" 
> (/var/lib/bareos/storage_muc/Disk_Drive000)
>     spooling=1 despooling=0 despool_wait=0
>     Files=174,615 Bytes=8,043,094,315 AveBytes/sec=17,447,059 
> LastBytes/sec=16,173,213
>     FDSocket closed
>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to bareos-users+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/bareos-users/cede4014-e680-4324-87e9-85dcd4a773ffn%40googlegroups.com.

Reply via email to