-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
velero is unable to prune PartiallyFailed backup from object storage #5940
Comments
@jhuisss Please also share the podVolumeBackup CRs included in the partially failed backup |
Hi, now attached logbundle for the PartiallyFailed backup and podVolumeBackup CR yaml file. |
This looks like a bug of handling Restic snapshot erroneously:
However, on Velero side:
|
Reproduce steps:
|
This problem also happens with Kopia uploader |
After double confirm from the code and testing, Kopia path doesn't have this problem because the snapshot for Kopia path is saved by Velero, if Kopia uploader reports any error, Velero will abort saving the snapshot, so the backed up data will be removed by Kopia GC some time later. Since in v1.12 and later, we will suppress Restic path, this issue is not with high priority. |
@Lyndon-Li note that restic will still be available as a non-default option in 1.12 and (assuming we deprecate this in 1.12), until 1.14, assuming the proposed deprecation policy is the one we end up with -- i.e. earliest time we can remove a feature is 2 releases post-deprecation. This assumes that we deprecate restic in the same release we make it non-default. It we deprecate later, then those numbers shift as well. |
@sseago |
also experience this volume backups using CSI snapshots with datamover, after the BackupDeleteRequest is completed the objects for the volume backup remain in the S3 bucket. |
What steps did you take and what happened:
I'm using velero 1.9.5, there is a PartiallyFailed backup (restic backup with nfs pv), which results in 4.1G disk usage on the minio server. When I use
velero backup delete <partiallyfailed backup>
cmd to delete the backup, it did delete the backup CR, but it won't release disk space in bucket in the object storage, even though I have setdefault-restic-prune-frequency=0h30m0s
and make sure in the resticrepository, themaintenanceFrequency
equals to 30m0s.What did you expect to happen:
When running cmd
velero backup delete <partiallyfailed backup>
, after half an hour, the disk space used by this backup should be released.The following information will help us better understand what's going on:
One wired thing is that deleting
Completed
backup will release the disk used in objectstorage, but toPartiallyFailed
backup, it won’t release disk space.BTW, Only 1 backup in this namespace and this object storage bucket, no incremental backup, so delete the backup should release all disk usage in this bucket.
Anything else you would like to add:
velero logs during running
velero backup delete <partiallyfailed backup>
:Environment:
velero version
): 1.9.5velero client config get features
):kubectl version
):/etc/os-release
):Vote on this issue!
This is an invitation to the Velero community to vote on issues, you can see the project's top voted issues listed here.
Use the "reaction smiley face" up to the right of this comment to vote.
The text was updated successfully, but these errors were encountered: