View Issue Details

IDProjectCategoryView StatusLast Update
0001427bareos-coredirectorpublic2022-02-10 19:57
Reporterjcottin Assigned Toarogge  
PriorityhighSeverityblockReproducibilityalways
Status closedResolutionno change required 
PlatformLinuxOSDebianOS Version10
Product Version21.0.0 
Summary0001427: Consolidate Job block everything.
DescriptionHi,

I tried to follow the documentation regarding the always incremental scheme.
But now the consolidate job is "failling".

https://docs.bareos.org/TasksAndConcepts/AlwaysIncrementalBackupScheme.html?highlight=consolidate#consolidate-job
Job {
    Name = "Consolidate"
    Type = "Consolidate"
    Accurate = "yes"
    JobDefs = "DefaultJob"
    Schedule = "WeeklyCycleConsolidate"
}

After the job end, it create other Jobs with level = VirtualFull
But these jobs seems to get stuck.
After waiting 5 days, nothing happens.
During this period, all the other backup task got the status "WAITING".

5 2022-02-07 22:28:56 bareos-dir JobId 363: Using Device "FileStorage-AI" to read.
4 2022-02-07 22:28:55 bareos-dir JobId 363: Connected Storage daemon at 10.0.9.100:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
3 2022-02-07 22:28:55 bareos-dir JobId 363: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.1.bsr
2 2022-02-07 22:28:55 bareos-dir JobId 363: Consolidating JobIds 14,22,29,36,44,52,60,68,76
1 2022-02-07 22:28:55 bareos-dir JobId 363: Start Virtual Backup JobId 363, Job=vm-prevor-web-new-job.2022-02-07_22.28.55_09
Additional Informationexample job configuration

Job {
  Name = vm1-job
  Client = vm1-fd

  Accurate = yes
  Always Incremental = yes
  Always Incremental Job Retention = 30 days
  Always Incremental Keep Number = 15
  Always Incremental Max Full Age = 60 days

  Level = Incremental
  Type = Backup
...
TagsNo tags attached.

Activities

jcottin

jcottin

2022-02-07 15:59

reporter  

consolidate-bareos.png (70,810 bytes)   
consolidate-bareos.png (70,810 bytes)   
jcottin

jcottin

2022-02-07 21:05

reporter   ~0004501

I found the solution there : https://bugs.bareos.org/view.php?id=854

I have two storage.
I duplicated the Device 16 times, and now the consolidation job is OK

Storage {
  Name = FileStorage-AI-Incremental
  Address = 10.0.1.100 # N.B. Use a fully qualified name here (do not use "localhost" here).
  Password = "OqHUkxXFKs1CCg2HkMoTWnhG7D+EbmOzVawzi2bZUREU"
  Device = FileStorage-AI1
  Device = FileStorage-AI2
  Device = FileStorage-AI3
  Device = FileStorage-AI4
  Device = FileStorage-AI5
  Device = FileStorage-AI6
  Device = FileStorage-AI7
  Device = FileStorage-AI8
  Media Type = File
}

Storage {
  Name = FileStorage-AI-Consolidated-10.0.2
  Address = 10.0.2.100 # N.B. Use a fully qualified name here (do not use "localhost" here).
  Password = "OqHUkxXFKs1CCg2HkMoTWnhG7D+EbmOzVawzi2bZUREU"
  Device = FileStorage-AI1-10.0.2
  Device = FileStorage-AI2-10.0.2
  Device = FileStorage-AI3-10.0.2
  Device = FileStorage-AI4-10.0.2
  Device = FileStorage-AI5-10.0.2
  Device = FileStorage-AI6-10.0.2
  Device = FileStorage-AI7-10.0.2
  Device = FileStorage-AI8-10.0.2
  Media Type = File
}
jcottin

jcottin

2022-02-09 09:26

reporter   ~0004502

I don't understand why but the next day it doesn't work anymore.

09-Feb 10:00 bareos-dir JobId 395: Job queued. JobId=397
09-Feb 10:00 bareos-dir JobId 395: Consolidating JobId 397 started.
09-Feb 10:00 bareos-dir JobId 397: Start Virtual Backup JobId 397, Job=vm-e2c-git-job.2022-02-09_10.00.02_33
09-Feb 10:00 bareos-dir JobId 397: Consolidating JobIds 382,84
09-Feb 10:00 bareos-dir JobId 397: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.7.bsr
09-Feb 10:00 bareos-dir JobId 397: Connected Storage daemon at 10.0.6.201:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
09-Feb 10:00 bareos-dir JobId 397: Using Device "FileStorage-AI1-10.0.6" to read.
09-Feb 10:00 bareos-dir JobId 397: Using Device "FileStorage-AI2-10.0.6" to write.
09-Feb 10:00 bareos-sd JobId 397: Volume "AI-Consolidated-vm-e2c-git-0005" previously written, moving to end of data.
09-Feb 10:00 bareos-sd JobId 397: Ready to append to end of Volume "AI-Consolidated-vm-e2c-git-0005" size=8776865034
09-Feb 10:00 bareos-sd JobId 397: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-e2c-git-0005 from dev="FileStorage-AI2-10.0.6" (/var/lib/bareos/storage-AI) to "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI)
09-Feb 10:00 bareos-sd JobId 397: Warning: stored/acquire.cc:336 Read acquire: stored/label.cc:255 Could not reserve volume AI-Consolidated-vm-e2c-git-0005 on "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI)
09-Feb 10:00 bareos-sd JobId 397: Please mount read Volume "AI-Consolidated-vm-e2c-git-0005" for:
09-Feb 10:05 bareos-sd JobId 397: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-e2c-git-0005 from dev="FileStorage-AI2-10.0.6" (/var/lib/bareos/storage-AI) to "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI)
jcottin

jcottin

2022-02-09 09:28

reporter   ~0004503

I should have enough storage and devices.

Storage {
  Name = FileStorage-AI-Consolidated-10.0.6
  Address = 10.0.6.201 # N.B. Use a fully qualified name here (do not use "localhost" here).
  Password = "OqHUkxXFKs1CCg2HkMoTWnhG7D+EbmOzVawzi2bZUREU"
  Device = FileStorage-AI1-10.0.6
  Device = FileStorage-AI2-10.0.6
  Device = FileStorage-AI3-10.0.6
  Device = FileStorage-AI4-10.0.6
  Device = FileStorage-AI5-10.0.6
  Device = FileStorage-AI6-10.0.6
  Device = FileStorage-AI7-10.0.6
  Device = FileStorage-AI8-10.0.6
  Media Type = File
}

Storage {
  Name = FileStorage-AI-Incremental-10.0.6
  Address = 10.0.6.201 # N.B. Use a fully qualified name here (do not use "localhost" here).
  Password = "OqHUkxXFKs1CCg2HkMoTWnhG7D+EbmOzVawzi2bZUREU"
  Device = FileStorage-AI1-10.0.6
  Device = FileStorage-AI2-10.0.6
  Device = FileStorage-AI3-10.0.6
  Device = FileStorage-AI4-10.0.6
  Device = FileStorage-AI5-10.0.6
  Device = FileStorage-AI6-10.0.6
  Device = FileStorage-AI7-10.0.6
  Device = FileStorage-AI8-10.0.6
  Media Type = File
}
jcottin

jcottin

2022-02-09 09:51

reporter   ~0004504

I added two directive for each storage and devices.

storage {
...
Maximum Concurrent Jobs = 8
}

device {
...
Maximum Concurrent Jobs = 1
}

Still have the problem though.
It's random.

09-Feb 16:47 bareos-dir JobId 401: Job queued. JobId=405
09-Feb 16:47 bareos-dir JobId 401: Consolidating JobId 405 started.
09-Feb 16:47 bareos-dir JobId 405: Start Virtual Backup JobId 405, Job=vm-aiqiit-web-job.2022-02-09_16.47.44_20
09-Feb 16:47 bareos-dir JobId 405: Consolidating JobIds 385,80
09-Feb 16:47 bareos-dir JobId 405: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.4.bsr
09-Feb 16:47 bareos-dir JobId 405: Connected Storage daemon at 10.0.9.100:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
09-Feb 16:47 bareos-dir JobId 405: Using Device "FileStorage-AI3" to read.
09-Feb 16:47 bareos-dir JobId 405: Using Device "FileStorage-AI4" to write.
09-Feb 16:47 bareos-sd JobId 405: Volume "AI-Consolidated-vm-aiqiit-web-0004" previously written, moving to end of data.
09-Feb 16:47 bareos-sd JobId 405: Ready to append to end of Volume "AI-Consolidated-vm-aiqiit-web-0004" size=4492689890
09-Feb 16:47 bareos-sd JobId 405: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-aiqiit-web-0004 from dev="FileStorage-AI4" (/var/lib/bareos/storage-AI) to "FileStorage-AI3" (/var/lib/bareos/storage-AI)
09-Feb 16:47 bareos-sd JobId 405: Warning: stored/acquire.cc:336 Read acquire: stored/label.cc:255 Could not reserve volume AI-Consolidated-vm-aiqiit-web-0004 on "FileStorage-AI3" (/var/lib/bareos/storage-AI)
09-Feb 16:47 bareos-sd JobId 405: Please mount read Volume "AI-Consolidated-vm-aiqiit-web-0004" for:
Job: vm-aiqiit-web-job.2022-02-09_16.47.44_20
    Storage: "FileStorage-AI3" (/var/lib/bareos/storage-AI)
    Pool: AI-Incremental-vm-aiqiit-web
    Media type: File
jcottin

jcottin

2022-02-09 10:09

reporter   ~0004505

jcottin

jcottin

2022-02-09 10:12

reporter   ~0004506

At some point it worked.
not sure why.

09-Feb 17:10 bareos-dir JobId 413: Start Virtual Backup JobId 413, Job=vm-aiqiit-web-job.2022-02-09_17.10.30_07
09-Feb 17:10 bareos-dir JobId 413: Consolidating JobIds 385,80
09-Feb 17:10 bareos-dir JobId 413: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.1.bsr
09-Feb 17:10 bareos-dir JobId 413: Connected Storage daemon at 10.0.9.100:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
09-Feb 17:10 bareos-dir JobId 413: Using Device "FileStorage-AI1" to read.
09-Feb 17:10 bareos-dir JobId 413: Using Device "FileStorage-AI2" to write.
09-Feb 17:10 bareos-sd JobId 413: Volume "AI-Consolidated-vm-aiqiit-web-0178" previously written, moving to end of data.
09-Feb 17:10 bareos-sd JobId 413: Ready to append to end of Volume "AI-Consolidated-vm-aiqiit-web-0178" size=253
09-Feb 17:10 bareos-sd JobId 413: Ready to read from volume "AI-Consolidated-vm-aiqiit-web-0004" on device "FileStorage-AI1" (/var/lib/bareos/storage-AI).
09-Feb 17:10 bareos-sd JobId 413: Forward spacing Volume "AI-Consolidated-vm-aiqiit-web-0004" to file:block 1:152222568.
09-Feb 17:10 bareos-sd JobId 413: End of Volume at file 1 on device "FileStorage-AI1" (/var/lib/bareos/storage-AI), Volume "AI-Consolidated-vm-aiqiit-web-0004"
09-Feb 17:10 bareos-sd JobId 413: Ready to read from volume "AI-Incremental-vm-aiqiit-web0054" on device "FileStorage-AI1" (/var/lib/bareos/storage-AI).
09-Feb 17:10 bareos-sd JobId 413: Forward spacing Volume "AI-Incremental-vm-aiqiit-web0054" to file:block 0:250.
09-Feb 17:10 bareos-sd JobId 413: End of Volume at file 0 on device "FileStorage-AI1" (/var/lib/bareos/storage-AI), Volume "AI-Incremental-vm-aiqiit-web0054"
09-Feb 17:10 bareos-sd JobId 413: End of all volumes.
09-Feb 17:10 bareos-sd JobId 413: Elapsed time=00:00:01, Transfer rate=45.81 M Bytes/second
09-Feb 17:10 bareos-sd JobId 413: Releasing device "FileStorage-AI2" (/var/lib/bareos/storage-AI).
09-Feb 17:10 bareos-sd JobId 413: Releasing device "FileStorage-AI1" (/var/lib/bareos/storage-AI).
09-Feb 17:10 bareos-dir JobId 413: Insert of attributes batch table with 49 entries start
09-Feb 17:10 bareos-dir JobId 413: Insert of attributes batch table done
09-Feb 17:10 bareos-dir JobId 413: Joblevel was set to joblevel of first consolidated job: Incremental
09-Feb 17:10 bareos-dir JobId 413: Bareos bareos-dir 21.0.0 (21Dec21):
  Build OS: Debian GNU/Linux 10 (buster)
  JobId: 413
  Job: vm-aiqiit-web-job.2022-02-09_17.10.30_07
  Backup Level: Virtual Full
  Client: "vm-aiqiit-web-fd" 20.0.1 (02Mar21) Debian GNU/Linux 9.13 (stretch),debian
  FileSet: "LinuxAll-vm-aiqiit-web" 2021-12-31 10:52:36
  Pool: "AI-Consolidated-vm-aiqiit-web" (From Job Pool's NextPool resource)
  Catalog: "MyCatalog" (From Client resource)
  Storage: "FileStorage-AI-Consolidated" (From Storage from Pool's NextPool resource)
  Scheduled time: 09-Feb-2022 17:10:30
  Start time: 09-Jan-2022 20:00:00
  End time: 09-Jan-2022 20:00:05
  Elapsed time: 5 secs
  Priority: 10
  SD Files Written: 49
  SD Bytes Written: 45,816,631 (45.81 MB)
  Rate: 9163.3 KB/s
  Volume name(s): AI-Consolidated-vm-aiqiit-web-0178
  Volume Session Id: 1
  Volume Session Time: 1644397818
  Last Volume Bytes: 45,852,636 (45.85 MB)
  SD Errors: 0
  SD termination status: OK
  Accurate: yes
  Bareos binary info: bareos.org build: Get official binaries and vendor support on bareos.com
  Job triggered by: User
  Termination: Backup OK

09-Feb 17:10 bareos-dir JobId 413: purged JobIds 385,80 as they were consolidated into Job 413
09-Feb 17:10 bareos-dir JobId 413: console command: run AfterJob ".bvfs_update jobid=413"
jcottin

jcottin

2022-02-09 13:47

reporter   ~0004507

I don't understand why, no it's not working again.
Is it a bug or a configuration problem ?
jcottin

jcottin

2022-02-09 13:57

reporter   ~0004508

After cancelling the job multiple times to restart the consolidating job.
At some point, the process succeed for one Job.

09-Feb 20:50 bareos-dir JobId 435: Consolidating JobId 436 started.
09-Feb 20:50 bareos-dir JobId 436: Start Virtual Backup JobId 436, Job=vm-e2c-git-job.2022-02-09_20.50.13_07
09-Feb 20:50 bareos-dir JobId 436: Consolidating JobIds 402,91
09-Feb 20:50 bareos-dir JobId 436: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.1.bsr
09-Feb 20:50 bareos-dir JobId 436: Connected Storage daemon at 10.0.6.201:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
09-Feb 20:50 bareos-dir JobId 436: Using Device "FileStorage-AI1-10.0.6" to read.
09-Feb 20:50 bareos-dir JobId 436: Using Device "FileStorage-AI2-10.0.6" to write.
09-Feb 20:50 bareos-sd JobId 436: Volume "AI-Consolidated-vm-e2c-git-0005" previously written, moving to end of data.
09-Feb 20:50 bareos-sd JobId 436: Ready to append to end of Volume "AI-Consolidated-vm-e2c-git-0005" size=8776865034
09-Feb 20:50 bareos-sd JobId 436: Ready to read from volume "AI-Consolidated-vm-e2c-git-0175" on device "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI).
09-Feb 20:50 bareos-sd JobId 436: Forward spacing Volume "AI-Consolidated-vm-e2c-git-0175" to file:block 0:247.
09-Feb 20:50 bareos-sd JobId 436: End of Volume at file 0 on device "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI), Volume "AI-Consolidated-vm-e2c-git-0175"
09-Feb 20:50 bareos-sd JobId 436: Ready to read from volume "AI-Incremental-vm-e2c-git0060" on device "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI).
09-Feb 20:50 bareos-sd JobId 436: Forward spacing Volume "AI-Incremental-vm-e2c-git0060" to file:block 0:244.
09-Feb 20:50 bareos-sd JobId 436: Elapsed time=00:00:04, Transfer rate=257.1 M Bytes/second
09-Feb 20:50 bareos-sd JobId 436: Releasing device "FileStorage-AI2-10.0.6" (/var/lib/bareos/storage-AI).
09-Feb 20:50 bareos-sd JobId 436: Releasing device "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI).
09-Feb 20:50 bareos-dir JobId 436: Insert of attributes batch table with 728 entries start
09-Feb 20:50 bareos-dir JobId 436: Insert of attributes batch table done
09-Feb 20:50 bareos-dir JobId 436: Joblevel was set to joblevel of first consolidated job: Incremental
09-Feb 20:50 bareos-dir JobId 436: Bareos bareos-dir 21.0.0 (21Dec21):
09-Feb 20:50 bareos-dir JobId 436: purged JobIds 402,91 as they were consolidated into Job 436
09-Feb 20:50 bareos-dir JobId 436: console command: run AfterJob ".bvfs_update jobid=436"

While other job still have problems.

09-Feb 20:50 bareos-dir JobId 435: Consolidating JobId 437 started.
09-Feb 20:50 bareos-dir JobId 437: Start Virtual Backup JobId 437, Job=vm-aiqiit-windows-test-backup-job.2022-02-09_20.50.13_08
09-Feb 20:50 bareos-dir JobId 437: Consolidating JobIds 404,88
09-Feb 20:50 bareos-dir JobId 437: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.2.bsr
09-Feb 20:50 bareos-dir JobId 437: Connected Storage daemon at 10.0.9.100:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
09-Feb 20:50 bareos-dir JobId 437: Using Device "FileStorage-AI1" to read.
09-Feb 20:50 bareos-dir JobId 437: Using Device "FileStorage-AI2" to write.
09-Feb 20:50 bareos-sd JobId 437: Volume "AI-Consolidated-vm-aiqiit-windows-test-backup-0176" previously written, moving to end of data.
09-Feb 20:50 bareos-sd JobId 437: Ready to append to end of Volume "AI-Consolidated-vm-aiqiit-windows-test-backup-0176" size=1160038883
09-Feb 20:50 bareos-sd JobId 437: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-aiqiit-windows-test-backup-0176 from dev="FileStorage-AI2" (/var/lib/bareos/storage-AI) to "FileStorage-AI1" (/var/lib/bareos/storage-AI)
09-Feb 20:50 bareos-sd JobId 437: Warning: stored/acquire.cc:336 Read acquire: stored/label.cc:255 Could not reserve volume AI-Consolidated-vm-aiqiit-windows-test-backup-0176 on "FileStorage-AI1" (/var/lib/bareos/storage-AI)
09-Feb 20:50 bareos-sd JobId 437: Please mount read Volume "AI-Consolidated-vm-aiqiit-windows-test-backup-0176" for:
09-Feb 20:55 bareos-sd JobId 437: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-aiqiit-windows-test-backup-0176 from dev="FileStorage-AI2" (/var/lib/bareos/storage-AI) to "FileStorage-AI1" (/var/lib/bareos/storage-AI)
arogge

arogge

2022-02-10 14:13

manager   ~0004510

You may be hitting a known issue - or you're having a configuration problem.

Could you take a look at status storage when the system is idle (i.e. has no jobs running). You can either post the output or just check if there are eny entries at the bottom in the used volume list.
If you see entries there, that's a bug we're currently working on (I'm not sure it is in the bugtracker yet).
If you don't see entries there, you're just having a configuration problem. Point is: you cannot read from and write to the same volume. That will just fail exactly the way you're describing. Basically you need to make sure that if Consolidation happens on "day 2" it doesn't try to write to the same volume as before, because it will also need to read from that one.
jcottin

jcottin

2022-02-10 19:54

reporter   ~0004517

*status storage
The defined Storage resources are:
1: FileStorage-AI-Consolidated
2: FileStorage-AI-Consolidated-10.0.6
3: FileStorage-AI-Incremental
4: FileStorage-AI-Incremental-10.0.6
5: FileStorageBareOS-Catalog
Select Storage resource (1-5):

No line for each storage.
Used Volume status:
====
====

Thanks for your help.
It seems it's not a bug so.

I don't want to take to much of your time but, can you please help me with the following ?

Yesterday I used the other bug report has a base and changed the pools configuration.

Added this line for every pool (Consolidated and Incremental):
@/etc/bareos/pool-defaults.conf
=======================
egrep -v '^#' /etc/bareos/pool-defaults.conf
        Pool Type = Backup
        Recycle = yes # Bareos can automatically recycle Volumes
        Auto Prune = yes # Prune expired volumes
        Volume Use Duration = 23h
        Action On Purge = Truncate
=======================

Will this do the trick ?
Or is it better to also specify Maximum Volume Jobs = 1 ?
arogge

arogge

2022-02-10 19:57

manager   ~0004518

Could you please contact the mailing-list for this.
I'm closing this, as it is not a bug.

Issue History

Date Modified Username Field Change
2022-02-07 15:59 jcottin New Issue
2022-02-07 15:59 jcottin File Added: consolidate-bareos.png
2022-02-07 21:05 jcottin Note Added: 0004501
2022-02-09 09:26 jcottin Note Added: 0004502
2022-02-09 09:28 jcottin Note Added: 0004503
2022-02-09 09:51 jcottin Note Added: 0004504
2022-02-09 10:09 jcottin Note Added: 0004505
2022-02-09 10:09 jcottin File Added: bareos-always-incremental-consolidation.PNG
2022-02-09 10:12 jcottin Note Added: 0004506
2022-02-09 13:47 jcottin Note Added: 0004507
2022-02-09 13:47 jcottin File Added: bareos-always-incremental-consolidation-problem.PNG
2022-02-09 13:57 jcottin Note Added: 0004508
2022-02-10 14:13 arogge Assigned To => arogge
2022-02-10 14:13 arogge Status new => feedback
2022-02-10 14:13 arogge Note Added: 0004510
2022-02-10 19:54 jcottin Note Added: 0004517
2022-02-10 19:54 jcottin Status feedback => assigned
2022-02-10 19:57 arogge Status assigned => resolved
2022-02-10 19:57 arogge Resolution open => no change required
2022-02-10 19:57 arogge Note Added: 0004518
2022-02-10 19:57 arogge Status resolved => closed