-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pod Volume Backup Failed while backing up volumes. #5188
Comments
Do you configure a customized CA in the BackupStorageLocation? |
Yes, it's configured and BackupStorageLocation is available.
|
I have also found out similar issue created for PodVolumeBackup failed I have followed issue and added changes specified in cherrypick 5145 and tested again. Now i can see PodVolumeBackups are getting completed but i am having another error. due to which my backups still getting partially failed. attaching the logs of PodVolumeBackups
also attaching the velero logs for reference
|
Are you using a dev build of Velero with your own patch? Could you try the v1.9.1-rc2? BTW, the status of the PodVolumeBackup is weired, there is a error message but the phase is completed, did you patch the status manually?:
|
Yes, i have manually modified the files. Let me try to use the v1.9.1-rc2 and test. |
Hi @ywk253100, I have used the v1.9.1-rc2 patch and tested the backup again.
|
Is this a fresh installation or an upgrade one? Did you update the CRD as well if it is upgraded from the previous version? Could you try the fresh installation? If the error is still there, please run velero debug to collect all the necessary information |
Yes, we did the fresh installation with upgrading all crds with latest version as you mentioned. Please find the attached velero logs in debug mode. |
any update @ywk253100 ? |
any update @ywk253100 |
@nwakalka Sorry for the late reply. I didn't find any useful information in the log file. I'm not sure whether this issue is related to your environment/configuration or it's a bug of Velero, but I'm afraid have no time to do more investigation. Could you debug it further in your local environment and appreciate it if you can find out the root cause? |
Hi @ywk253100 , After going through logs again, we have observed incorrect order of restic snapshots and initialization of restic repository.
Execution it's not in correct order, restic snapshots are getting executed before initializing restic repository. |
please the attached logs of restic, as i don't see any error here.
|
Error string |
Hi @ywk253100 , @sseago , I checked with release 1.7 version and it's working as expected. but for v1.8 i am facing below error
|
Hi @ywk253100 , @sseago , Currently our issue is resolved with velero 1.9, as we have two namespaces and both namespaces having velero and restic pods. once removed velero and restic from one of namespace our backups and restore are success. till velero 1.7 both namespaces have velero and restic and are working independently. do we have any other provision now ? |
@nwakalka Were both velero installations the same version? Two independent velero installations should still work as of 1.9. Nothing has been changed that was intended to break this, although it's possible that some of the refactoring that was done introduced a regression. However, if you have two different velero versions installed, then one of them will have the wrong CRDs, since CRDs are cluster-scoped -- in that case the installation with the wrong CRDs will probably not work properly. |
Hi @sseago , Thank you for replying. yes, we are deploying same versions of velero and yes correct CRDs of same versions. |
Hi all, I had some time to dig into this issue (disclaimer: I am a colleage of @nwakalka). What is missed here is the fact that we run two complete stacks of velero in two separate namespaces, including restic. There is one distinct daemonset for restic in each namespace. With up until (and including) v1.7.2 velero & restic have acted only on podvolumebackups which were created in their respective namespace. Starting with v1.8, both restic instances will reconcile a podvolumebackup, ignoring the actual namespace. v1.7.2 behavior: v1.8 (and above) behavior: This new behavior results in the described status above. Now my questions:
|
I think "v1.8 behavior" if true is a bug. |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
Closing the stale issue. |
What steps did you take and what happened:
We were trying to run velero backups with pod volume backups. where pod volumes that need to be backed up using Restic.
here you can see our velero backups are partiallyFailing. attaching the velero backup logs for more info
Later checked for pod volume backups which was failed
Describing the pod volumebackup
What did you expect to happen:
Pod volume backup need to be completed and hence velero backup need to completed.
Anything else you would like to add:
Environment:
velero version
): 1.9.0-1velero client config get features
): features:kubectl version
): v1.23.9/etc/os-release
): Ubuntu 20.04.4 LTSThe text was updated successfully, but these errors were encountered: