-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Differential Ceph backups scan all allocated space (Ceph bug) #114
Comments
Is this possibly the cause? The sample diff output shows 'exists' being true or false whereas the manual indicates something else: Not aware of RBD parameters being positional (ie sensitive to order)... From the script:
|
My tests appear to indicate a bug with Ceph Octopus 15.2.11 when the '--whole-object' switch is used:
This is 100% correct, if I however add the '--whole-object' switch it lists every allocated object from the entire image:
|
Opened a bug with Ceph here: |
Pretty sure this change in behaviour is a regression due to the patch introduced in January 2020 for Ceph issue https://tracker.ceph.com/issues/42248 Have logged a bug with Ceph directly: |
@bbs2web just saw your ticket. Thanks for all the information. Hopefully this will be fixed soon. |
Herewith confirmation that the fix for this has been merged in to the main development branch and also backported to Ceph Pacific 16.2.7 and pending for back porting to Ceph Octopus. Just tested on a test cluster that we upgraded to 16.2.7 today and it's working as expected again. As per the steps to reproduce the problem whilst running 16.2.6 (https://tracker.ceph.com/issues/50787) it now works as expected:
|
I'm a benji backup convert but since upgrading to 0.14.1 differential backups appear to select all allocated space which each run.
We regularly schedule trim/discard operations on guests for space reclamation reasons. Two consecutive backups without running Defragment and Optimise Drive in Windows still results in an incremental backup of 195983 blocks (at 4 MiB a block that's 765 GiB out of the 1 TiB volume).
Backups themselves aren't requiring any more space or data transfer (vlan1) but reads from source VMs being backed up (vlan100) shows the increased usage since updating:
We are using https://github.com/elemental-lf/benji/blob/master/scripts/ceph.sh as the '/usr/local/bin/benji-backup.sh' function script with "${BENJI_LOG_LEVEL:=INFO}"
I've stepped through the scripts and have a sample Ceph RBD diff which contains information such as:
Can share the entire file but presume it not to be relevant...
The text was updated successfully, but these errors were encountered: