-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Take statefulsets backup on ACTIVE Cluster and restore it to the STAND BY cluster B using velero giving error #7737
Comments
What Velero version are you using? Can you help provide us with more debug info by using the command from this doc. |
@allenxu404 Please let me know if there is any additional information requires |
Log given above looks normal. PV was successfully restored from snapshot as below log message shows:
It seems that the VolumeId was not available for cluster B for some reason. I think you can further troubleshoot it by restore PV on ACTIVE cluster instead of STAND BY cluster B. I assume the restore will work in that case. |
HI @allenxu404 |
It seems the CSI snapshot related CRDs are missed from the cluster. |
HI @blackpiglet I have already installed volume snapshot crd's
Thanks |
@kish5430 Can you help verify the status of the associated PV and PVC to confirm their functionality? Additionally, Can you access the AWS console to validate the volume's creation and ensure its proper configuration in the backend? |
What steps did you take and what happened:
While working on an active EKS cluster, I deployed an application with three etcd pods. I took a backup of these etcd pods using Velero. Later, I switched to a standby cluster and attempted to restore the backup. Although the restore process was successful and the pods were deployed and not running, there was a failure in attaching volumes to the etcd pods.
Command: velero backup create milvus-stg-east1-etcd-backup --selector 'app.kubernetes.io/name=etcd'
What did you expect to happen:
Volume attachment should happen and etcd pods run without any issue.
Etcd Pod logs:
Warning FailedAttachVolume 101s (x11 over 34m) attachdetach-controller (combined from similar events): AttachVolume.Attach failed for volume "pvc-ed7a6088-9f9e-46fc-88ab-bbe8364a28f7" : rpc error: code = Internal desc = Could not attach volume "vol-00c1e0e23881130c9" to node "i-03a2b2d33c76ccef2": could not attach volume "vol-00c1e0e23881130c9" to node "i-03a2b2d33c76ccef2": InvalidVolume.NotFound: The volume 'vol-00c1e0e23881130c9' does not exist.
status code: 400, request id: 4160e339-013b-4b3b-8f39-c3990cf66c2e
Here volume 'vol-00c1e0e23881130c9'' is not exist in volumes in aws
Please find the attached velero restore logs.
velero_restore.txt
The text was updated successfully, but these errors were encountered: