-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Restoring a file system backup to a different cluster failed due to Kopia snapshot not found #8019
Comments
This means that Kopia uploader could not find the snapshot in the object store location specified in the BSL. So please double check objects in the object store where Kopia repository data is stored as indicated by the BSL, and make sure the BSLs in the source cluster and dest cluster points to the same object store location. |
If the error is |
@Lyndon-Li Thank you for your response, here is the bundle you requested: Another update: we checked with Kopia CLI and we can't find the snapshot either, but the cluster is connected to the right backup bucket, the BackupStorageLocation is listed as Available. |
Since you have connected to the kopia repo, could you run |
@Lyndon-Li sure.
|
From the above output, the repo is empty. |
@Lyndon-Li I retried with a new backup, made sure to connect the right bucket to the cluster where I restore, I verified the BackupStorageLocation, its the same as the other cluster and it says its available. Even when I run And another thing, when I connect to the bucket and list Kopia snapshots, I still don't find anything, its empty. |
What do you see in this bucket? Do you see a |
@Lyndon-Li Deleting all the contents of the backup bucket solved the issue, but that is not a good solution, just a temporary fix to keep implementing. We cannot do this in a production environment. There is another error complaining about sync, similar to this one in this issue: kopia/kopia#1938 Any suggestions to what might have happened or how to actually fix the issue from your side? |
I don't think it is related to kopia issue 1938, because there is no error in the log you shared. Therefore, I do need some more info to get what was happening, e.g., the questions I asked in #8019 (comment) |
@Lyndon-Li sure, I can see a Kopia folder inside the bucket, inside it there was only some files that start with _log_* , it seems to be log files. |
We encountered the Maybe a short word about what we're doing: we're heavily switching between clusters for our backup / restore process development. Means we have a source-cluster where velero is running and creating backups and a target cluster where we do restores via velero. The S3 bucket is only accessible by one velero installation at a time (we can guarantee this because we use aws s3api put-bucket-policy with only one unique Principal). Maybe we trigger some weird caching effects during this switching back-and-forth. We'll be still be attentive if it occurs another time. What we've learnt during the debugging:
|
Answer for all the related problems:
|
This issue is stale because it has been open 60 days with no activity. Remove stale label or comment or this will be closed in 14 days. If a Velero team member has requested log or more information, please provide the output of the shared commands. |
This issue was closed because it has been stalled for 14 days with no activity. |
What steps did you take and what happened:
I am creating a file system backup from a particular namespace in a K8s cluster and restoring it to another cluster. But the Restore is stuck in "In Progress" and it fails after timeout (I am also backing up and restoring the Pod to which the volume is mounted, along with some Secrets and configMaps).
The backup is stored in an S3 bucket and I made sure that the same bucket is linked to the new cluster.
After investigating, I can see that for some reason, the PodVolumeRestore failed with the error:
data path restore failed: Failed to run Kopia restore: Unable to load snapshot 2e97d1c5b03468f979e3143149d46239: snapshot not found
What did you expect to happen:
Restore to complete without an issue.
The following information will help us better understand what's going on:
Anything else you would like to add:
Restoring the backup to the same cluster it was taken from works with no issues, this only happens when I restore to a different cluster.
Environment:
velero version
):Velero features (use
velero client config get features
):features: <NOT SET>
Kubernetes version (use
kubectl version
):The text was updated successfully, but these errors were encountered: