View Issue Details
ID | Project | Category | View Status | Date Submitted | Last Update |
---|---|---|---|---|---|
0001427 | bareos-core | director | public | 2022-02-07 15:59 | 2022-02-10 19:57 |
Reporter | jcottin | Assigned To | arogge | ||
Priority | high | Severity | block | Reproducibility | always |
Status | closed | Resolution | no change required | ||
Platform | Linux | OS | Debian | OS Version | 10 |
Product Version | 21.0.0 | ||||
Summary | 0001427: Consolidate Job block everything. | ||||
Description | Hi, I tried to follow the documentation regarding the always incremental scheme. But now the consolidate job is "failling". https://docs.bareos.org/TasksAndConcepts/AlwaysIncrementalBackupScheme.html?highlight=consolidate#consolidate-job Job { Name = "Consolidate" Type = "Consolidate" Accurate = "yes" JobDefs = "DefaultJob" Schedule = "WeeklyCycleConsolidate" } After the job end, it create other Jobs with level = VirtualFull But these jobs seems to get stuck. After waiting 5 days, nothing happens. During this period, all the other backup task got the status "WAITING". 5 2022-02-07 22:28:56 bareos-dir JobId 363: Using Device "FileStorage-AI" to read. 4 2022-02-07 22:28:55 bareos-dir JobId 363: Connected Storage daemon at 10.0.9.100:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 3 2022-02-07 22:28:55 bareos-dir JobId 363: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.1.bsr 2 2022-02-07 22:28:55 bareos-dir JobId 363: Consolidating JobIds 14,22,29,36,44,52,60,68,76 1 2022-02-07 22:28:55 bareos-dir JobId 363: Start Virtual Backup JobId 363, Job=vm-prevor-web-new-job.2022-02-07_22.28.55_09 | ||||
Additional Information | example job configuration Job { Name = vm1-job Client = vm1-fd Accurate = yes Always Incremental = yes Always Incremental Job Retention = 30 days Always Incremental Keep Number = 15 Always Incremental Max Full Age = 60 days Level = Incremental Type = Backup ... | ||||
Tags | No tags attached. | ||||
|
|
I found the solution there : https://bugs.bareos.org/view.php?id=854 I have two storage. I duplicated the Device 16 times, and now the consolidation job is OK Storage { Name = FileStorage-AI-Incremental Address = 10.0.1.100 # N.B. Use a fully qualified name here (do not use "localhost" here). Password = "OqHUkxXFKs1CCg2HkMoTWnhG7D+EbmOzVawzi2bZUREU" Device = FileStorage-AI1 Device = FileStorage-AI2 Device = FileStorage-AI3 Device = FileStorage-AI4 Device = FileStorage-AI5 Device = FileStorage-AI6 Device = FileStorage-AI7 Device = FileStorage-AI8 Media Type = File } Storage { Name = FileStorage-AI-Consolidated-10.0.2 Address = 10.0.2.100 # N.B. Use a fully qualified name here (do not use "localhost" here). Password = "OqHUkxXFKs1CCg2HkMoTWnhG7D+EbmOzVawzi2bZUREU" Device = FileStorage-AI1-10.0.2 Device = FileStorage-AI2-10.0.2 Device = FileStorage-AI3-10.0.2 Device = FileStorage-AI4-10.0.2 Device = FileStorage-AI5-10.0.2 Device = FileStorage-AI6-10.0.2 Device = FileStorage-AI7-10.0.2 Device = FileStorage-AI8-10.0.2 Media Type = File } |
|
I don't understand why but the next day it doesn't work anymore. 09-Feb 10:00 bareos-dir JobId 395: Job queued. JobId=397 09-Feb 10:00 bareos-dir JobId 395: Consolidating JobId 397 started. 09-Feb 10:00 bareos-dir JobId 397: Start Virtual Backup JobId 397, Job=vm-e2c-git-job.2022-02-09_10.00.02_33 09-Feb 10:00 bareos-dir JobId 397: Consolidating JobIds 382,84 09-Feb 10:00 bareos-dir JobId 397: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.7.bsr 09-Feb 10:00 bareos-dir JobId 397: Connected Storage daemon at 10.0.6.201:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 09-Feb 10:00 bareos-dir JobId 397: Using Device "FileStorage-AI1-10.0.6" to read. 09-Feb 10:00 bareos-dir JobId 397: Using Device "FileStorage-AI2-10.0.6" to write. 09-Feb 10:00 bareos-sd JobId 397: Volume "AI-Consolidated-vm-e2c-git-0005" previously written, moving to end of data. 09-Feb 10:00 bareos-sd JobId 397: Ready to append to end of Volume "AI-Consolidated-vm-e2c-git-0005" size=8776865034 09-Feb 10:00 bareos-sd JobId 397: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-e2c-git-0005 from dev="FileStorage-AI2-10.0.6" (/var/lib/bareos/storage-AI) to "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI) 09-Feb 10:00 bareos-sd JobId 397: Warning: stored/acquire.cc:336 Read acquire: stored/label.cc:255 Could not reserve volume AI-Consolidated-vm-e2c-git-0005 on "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI) 09-Feb 10:00 bareos-sd JobId 397: Please mount read Volume "AI-Consolidated-vm-e2c-git-0005" for: 09-Feb 10:05 bareos-sd JobId 397: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-e2c-git-0005 from dev="FileStorage-AI2-10.0.6" (/var/lib/bareos/storage-AI) to "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI) |
|
I should have enough storage and devices. Storage { Name = FileStorage-AI-Consolidated-10.0.6 Address = 10.0.6.201 # N.B. Use a fully qualified name here (do not use "localhost" here). Password = "OqHUkxXFKs1CCg2HkMoTWnhG7D+EbmOzVawzi2bZUREU" Device = FileStorage-AI1-10.0.6 Device = FileStorage-AI2-10.0.6 Device = FileStorage-AI3-10.0.6 Device = FileStorage-AI4-10.0.6 Device = FileStorage-AI5-10.0.6 Device = FileStorage-AI6-10.0.6 Device = FileStorage-AI7-10.0.6 Device = FileStorage-AI8-10.0.6 Media Type = File } Storage { Name = FileStorage-AI-Incremental-10.0.6 Address = 10.0.6.201 # N.B. Use a fully qualified name here (do not use "localhost" here). Password = "OqHUkxXFKs1CCg2HkMoTWnhG7D+EbmOzVawzi2bZUREU" Device = FileStorage-AI1-10.0.6 Device = FileStorage-AI2-10.0.6 Device = FileStorage-AI3-10.0.6 Device = FileStorage-AI4-10.0.6 Device = FileStorage-AI5-10.0.6 Device = FileStorage-AI6-10.0.6 Device = FileStorage-AI7-10.0.6 Device = FileStorage-AI8-10.0.6 Media Type = File } |
|
I added two directive for each storage and devices. storage { ... Maximum Concurrent Jobs = 8 } device { ... Maximum Concurrent Jobs = 1 } Still have the problem though. It's random. 09-Feb 16:47 bareos-dir JobId 401: Job queued. JobId=405 09-Feb 16:47 bareos-dir JobId 401: Consolidating JobId 405 started. 09-Feb 16:47 bareos-dir JobId 405: Start Virtual Backup JobId 405, Job=vm-aiqiit-web-job.2022-02-09_16.47.44_20 09-Feb 16:47 bareos-dir JobId 405: Consolidating JobIds 385,80 09-Feb 16:47 bareos-dir JobId 405: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.4.bsr 09-Feb 16:47 bareos-dir JobId 405: Connected Storage daemon at 10.0.9.100:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 09-Feb 16:47 bareos-dir JobId 405: Using Device "FileStorage-AI3" to read. 09-Feb 16:47 bareos-dir JobId 405: Using Device "FileStorage-AI4" to write. 09-Feb 16:47 bareos-sd JobId 405: Volume "AI-Consolidated-vm-aiqiit-web-0004" previously written, moving to end of data. 09-Feb 16:47 bareos-sd JobId 405: Ready to append to end of Volume "AI-Consolidated-vm-aiqiit-web-0004" size=4492689890 09-Feb 16:47 bareos-sd JobId 405: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-aiqiit-web-0004 from dev="FileStorage-AI4" (/var/lib/bareos/storage-AI) to "FileStorage-AI3" (/var/lib/bareos/storage-AI) 09-Feb 16:47 bareos-sd JobId 405: Warning: stored/acquire.cc:336 Read acquire: stored/label.cc:255 Could not reserve volume AI-Consolidated-vm-aiqiit-web-0004 on "FileStorage-AI3" (/var/lib/bareos/storage-AI) 09-Feb 16:47 bareos-sd JobId 405: Please mount read Volume "AI-Consolidated-vm-aiqiit-web-0004" for: Job: vm-aiqiit-web-job.2022-02-09_16.47.44_20 Storage: "FileStorage-AI3" (/var/lib/bareos/storage-AI) Pool: AI-Incremental-vm-aiqiit-web Media type: File |
|
|
|
At some point it worked. not sure why. 09-Feb 17:10 bareos-dir JobId 413: Start Virtual Backup JobId 413, Job=vm-aiqiit-web-job.2022-02-09_17.10.30_07 09-Feb 17:10 bareos-dir JobId 413: Consolidating JobIds 385,80 09-Feb 17:10 bareos-dir JobId 413: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.1.bsr 09-Feb 17:10 bareos-dir JobId 413: Connected Storage daemon at 10.0.9.100:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 09-Feb 17:10 bareos-dir JobId 413: Using Device "FileStorage-AI1" to read. 09-Feb 17:10 bareos-dir JobId 413: Using Device "FileStorage-AI2" to write. 09-Feb 17:10 bareos-sd JobId 413: Volume "AI-Consolidated-vm-aiqiit-web-0178" previously written, moving to end of data. 09-Feb 17:10 bareos-sd JobId 413: Ready to append to end of Volume "AI-Consolidated-vm-aiqiit-web-0178" size=253 09-Feb 17:10 bareos-sd JobId 413: Ready to read from volume "AI-Consolidated-vm-aiqiit-web-0004" on device "FileStorage-AI1" (/var/lib/bareos/storage-AI). 09-Feb 17:10 bareos-sd JobId 413: Forward spacing Volume "AI-Consolidated-vm-aiqiit-web-0004" to file:block 1:152222568. 09-Feb 17:10 bareos-sd JobId 413: End of Volume at file 1 on device "FileStorage-AI1" (/var/lib/bareos/storage-AI), Volume "AI-Consolidated-vm-aiqiit-web-0004" 09-Feb 17:10 bareos-sd JobId 413: Ready to read from volume "AI-Incremental-vm-aiqiit-web0054" on device "FileStorage-AI1" (/var/lib/bareos/storage-AI). 09-Feb 17:10 bareos-sd JobId 413: Forward spacing Volume "AI-Incremental-vm-aiqiit-web0054" to file:block 0:250. 09-Feb 17:10 bareos-sd JobId 413: End of Volume at file 0 on device "FileStorage-AI1" (/var/lib/bareos/storage-AI), Volume "AI-Incremental-vm-aiqiit-web0054" 09-Feb 17:10 bareos-sd JobId 413: End of all volumes. 09-Feb 17:10 bareos-sd JobId 413: Elapsed time=00:00:01, Transfer rate=45.81 M Bytes/second 09-Feb 17:10 bareos-sd JobId 413: Releasing device "FileStorage-AI2" (/var/lib/bareos/storage-AI). 09-Feb 17:10 bareos-sd JobId 413: Releasing device "FileStorage-AI1" (/var/lib/bareos/storage-AI). 09-Feb 17:10 bareos-dir JobId 413: Insert of attributes batch table with 49 entries start 09-Feb 17:10 bareos-dir JobId 413: Insert of attributes batch table done 09-Feb 17:10 bareos-dir JobId 413: Joblevel was set to joblevel of first consolidated job: Incremental 09-Feb 17:10 bareos-dir JobId 413: Bareos bareos-dir 21.0.0 (21Dec21): Build OS: Debian GNU/Linux 10 (buster) JobId: 413 Job: vm-aiqiit-web-job.2022-02-09_17.10.30_07 Backup Level: Virtual Full Client: "vm-aiqiit-web-fd" 20.0.1 (02Mar21) Debian GNU/Linux 9.13 (stretch),debian FileSet: "LinuxAll-vm-aiqiit-web" 2021-12-31 10:52:36 Pool: "AI-Consolidated-vm-aiqiit-web" (From Job Pool's NextPool resource) Catalog: "MyCatalog" (From Client resource) Storage: "FileStorage-AI-Consolidated" (From Storage from Pool's NextPool resource) Scheduled time: 09-Feb-2022 17:10:30 Start time: 09-Jan-2022 20:00:00 End time: 09-Jan-2022 20:00:05 Elapsed time: 5 secs Priority: 10 SD Files Written: 49 SD Bytes Written: 45,816,631 (45.81 MB) Rate: 9163.3 KB/s Volume name(s): AI-Consolidated-vm-aiqiit-web-0178 Volume Session Id: 1 Volume Session Time: 1644397818 Last Volume Bytes: 45,852,636 (45.85 MB) SD Errors: 0 SD termination status: OK Accurate: yes Bareos binary info: bareos.org build: Get official binaries and vendor support on bareos.com Job triggered by: User Termination: Backup OK 09-Feb 17:10 bareos-dir JobId 413: purged JobIds 385,80 as they were consolidated into Job 413 09-Feb 17:10 bareos-dir JobId 413: console command: run AfterJob ".bvfs_update jobid=413" |
|
I don't understand why, no it's not working again. Is it a bug or a configuration problem ? |
|
After cancelling the job multiple times to restart the consolidating job. At some point, the process succeed for one Job. 09-Feb 20:50 bareos-dir JobId 435: Consolidating JobId 436 started. 09-Feb 20:50 bareos-dir JobId 436: Start Virtual Backup JobId 436, Job=vm-e2c-git-job.2022-02-09_20.50.13_07 09-Feb 20:50 bareos-dir JobId 436: Consolidating JobIds 402,91 09-Feb 20:50 bareos-dir JobId 436: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.1.bsr 09-Feb 20:50 bareos-dir JobId 436: Connected Storage daemon at 10.0.6.201:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 09-Feb 20:50 bareos-dir JobId 436: Using Device "FileStorage-AI1-10.0.6" to read. 09-Feb 20:50 bareos-dir JobId 436: Using Device "FileStorage-AI2-10.0.6" to write. 09-Feb 20:50 bareos-sd JobId 436: Volume "AI-Consolidated-vm-e2c-git-0005" previously written, moving to end of data. 09-Feb 20:50 bareos-sd JobId 436: Ready to append to end of Volume "AI-Consolidated-vm-e2c-git-0005" size=8776865034 09-Feb 20:50 bareos-sd JobId 436: Ready to read from volume "AI-Consolidated-vm-e2c-git-0175" on device "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI). 09-Feb 20:50 bareos-sd JobId 436: Forward spacing Volume "AI-Consolidated-vm-e2c-git-0175" to file:block 0:247. 09-Feb 20:50 bareos-sd JobId 436: End of Volume at file 0 on device "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI), Volume "AI-Consolidated-vm-e2c-git-0175" 09-Feb 20:50 bareos-sd JobId 436: Ready to read from volume "AI-Incremental-vm-e2c-git0060" on device "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI). 09-Feb 20:50 bareos-sd JobId 436: Forward spacing Volume "AI-Incremental-vm-e2c-git0060" to file:block 0:244. 09-Feb 20:50 bareos-sd JobId 436: Elapsed time=00:00:04, Transfer rate=257.1 M Bytes/second 09-Feb 20:50 bareos-sd JobId 436: Releasing device "FileStorage-AI2-10.0.6" (/var/lib/bareos/storage-AI). 09-Feb 20:50 bareos-sd JobId 436: Releasing device "FileStorage-AI1-10.0.6" (/var/lib/bareos/storage-AI). 09-Feb 20:50 bareos-dir JobId 436: Insert of attributes batch table with 728 entries start 09-Feb 20:50 bareos-dir JobId 436: Insert of attributes batch table done 09-Feb 20:50 bareos-dir JobId 436: Joblevel was set to joblevel of first consolidated job: Incremental 09-Feb 20:50 bareos-dir JobId 436: Bareos bareos-dir 21.0.0 (21Dec21): 09-Feb 20:50 bareos-dir JobId 436: purged JobIds 402,91 as they were consolidated into Job 436 09-Feb 20:50 bareos-dir JobId 436: console command: run AfterJob ".bvfs_update jobid=436" While other job still have problems. 09-Feb 20:50 bareos-dir JobId 435: Consolidating JobId 437 started. 09-Feb 20:50 bareos-dir JobId 437: Start Virtual Backup JobId 437, Job=vm-aiqiit-windows-test-backup-job.2022-02-09_20.50.13_08 09-Feb 20:50 bareos-dir JobId 437: Consolidating JobIds 404,88 09-Feb 20:50 bareos-dir JobId 437: Bootstrap records written to /var/lib/bareos/bareos-dir.restore.2.bsr 09-Feb 20:50 bareos-dir JobId 437: Connected Storage daemon at 10.0.9.100:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 09-Feb 20:50 bareos-dir JobId 437: Using Device "FileStorage-AI1" to read. 09-Feb 20:50 bareos-dir JobId 437: Using Device "FileStorage-AI2" to write. 09-Feb 20:50 bareos-sd JobId 437: Volume "AI-Consolidated-vm-aiqiit-windows-test-backup-0176" previously written, moving to end of data. 09-Feb 20:50 bareos-sd JobId 437: Ready to append to end of Volume "AI-Consolidated-vm-aiqiit-windows-test-backup-0176" size=1160038883 09-Feb 20:50 bareos-sd JobId 437: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-aiqiit-windows-test-backup-0176 from dev="FileStorage-AI2" (/var/lib/bareos/storage-AI) to "FileStorage-AI1" (/var/lib/bareos/storage-AI) 09-Feb 20:50 bareos-sd JobId 437: Warning: stored/acquire.cc:336 Read acquire: stored/label.cc:255 Could not reserve volume AI-Consolidated-vm-aiqiit-windows-test-backup-0176 on "FileStorage-AI1" (/var/lib/bareos/storage-AI) 09-Feb 20:50 bareos-sd JobId 437: Please mount read Volume "AI-Consolidated-vm-aiqiit-windows-test-backup-0176" for: 09-Feb 20:55 bareos-sd JobId 437: Warning: stored/vol_mgr.cc:520 Need volume from other drive, but swap not possible. Status: read=0 num_writers=1 num_reserve=0 swap=0 vol=AI-Consolidated-vm-aiqiit-windows-test-backup-0176 from dev="FileStorage-AI2" (/var/lib/bareos/storage-AI) to "FileStorage-AI1" (/var/lib/bareos/storage-AI) |
|
You may be hitting a known issue - or you're having a configuration problem. Could you take a look at status storage when the system is idle (i.e. has no jobs running). You can either post the output or just check if there are eny entries at the bottom in the used volume list. If you see entries there, that's a bug we're currently working on (I'm not sure it is in the bugtracker yet). If you don't see entries there, you're just having a configuration problem. Point is: you cannot read from and write to the same volume. That will just fail exactly the way you're describing. Basically you need to make sure that if Consolidation happens on "day 2" it doesn't try to write to the same volume as before, because it will also need to read from that one. |
|
*status storage The defined Storage resources are: 1: FileStorage-AI-Consolidated 2: FileStorage-AI-Consolidated-10.0.6 3: FileStorage-AI-Incremental 4: FileStorage-AI-Incremental-10.0.6 5: FileStorageBareOS-Catalog Select Storage resource (1-5): No line for each storage. Used Volume status: ==== ==== Thanks for your help. It seems it's not a bug so. I don't want to take to much of your time but, can you please help me with the following ? Yesterday I used the other bug report has a base and changed the pools configuration. Added this line for every pool (Consolidated and Incremental): @/etc/bareos/pool-defaults.conf ======================= egrep -v '^#' /etc/bareos/pool-defaults.conf Pool Type = Backup Recycle = yes # Bareos can automatically recycle Volumes Auto Prune = yes # Prune expired volumes Volume Use Duration = 23h Action On Purge = Truncate ======================= Will this do the trick ? Or is it better to also specify Maximum Volume Jobs = 1 ? |
|
Could you please contact the mailing-list for this. I'm closing this, as it is not a bug. |
|
Date Modified | Username | Field | Change |
---|---|---|---|
2022-02-07 15:59 | jcottin | New Issue | |
2022-02-07 15:59 | jcottin | File Added: consolidate-bareos.png | |
2022-02-07 21:05 | jcottin | Note Added: 0004501 | |
2022-02-09 09:26 | jcottin | Note Added: 0004502 | |
2022-02-09 09:28 | jcottin | Note Added: 0004503 | |
2022-02-09 09:51 | jcottin | Note Added: 0004504 | |
2022-02-09 10:09 | jcottin | Note Added: 0004505 | |
2022-02-09 10:09 | jcottin | File Added: bareos-always-incremental-consolidation.PNG | |
2022-02-09 10:12 | jcottin | Note Added: 0004506 | |
2022-02-09 13:47 | jcottin | Note Added: 0004507 | |
2022-02-09 13:47 | jcottin | File Added: bareos-always-incremental-consolidation-problem.PNG | |
2022-02-09 13:57 | jcottin | Note Added: 0004508 | |
2022-02-10 14:13 | arogge | Assigned To | => arogge |
2022-02-10 14:13 | arogge | Status | new => feedback |
2022-02-10 14:13 | arogge | Note Added: 0004510 | |
2022-02-10 19:54 | jcottin | Note Added: 0004517 | |
2022-02-10 19:54 | jcottin | Status | feedback => assigned |
2022-02-10 19:57 | arogge | Status | assigned => resolved |
2022-02-10 19:57 | arogge | Resolution | open => no change required |
2022-02-10 19:57 | arogge | Note Added: 0004518 | |
2022-02-10 19:57 | arogge | Status | resolved => closed |