-
Notifications
You must be signed in to change notification settings - Fork 748
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pods stuck in ContainerCreating or aws-node restarting after Service Account role for managed Addon is added/updated #1338
Comments
I experienced the same issue and able to see the IPAMD got the ipamd
kubelet
|
Hello! has anyone tried the amazon-k8s-cni update? logs:
describe:
|
@cp38510 Exact same problem here. Unmanaged EKS 1.17 with amazon-k8s-cni:v1.7.9 |
Same problem here. We are using EKS in v1.21 and cni 1.7.5. We need to specify some configurations into cni daemonset of the cluster and we deploy the aws-k8s-cni.yaml of v1.7.5. The problem are the aws-node pods stuck and restarting again and again and some time later, some cluster nodes entry into not-ready status. Some idea? Node description:
|
This issue is stale because it has been open 60 days with no activity. Remove stale label or comment or this will be closed in 14 days |
Issue closed due to inactivity. |
What happened:
On adding or updating the aws-node managed add on's service account role from the console the service account was updated with the role ARN but a rolling restart on the aws-node was not triggered which lead to it still using the previous identity to make all the EC2 API Calls. If the previous identity no longer has the
AmazonEKS_CNI_Policy
then the EC2 API call will start to fail withUnauthorizedOperation
and it's possible that the aws-node may not come up or new pods are stuck in ContainerCreating.It can be verified by checking the ipamd logs for Unauthorized errors on calling the EC2 API.
Current Workaround
After updating the role associated with the SA from console manually triggering a rolling restart should recreate the pod with the right identity.
To verify the new pods are using the right role ARN
Attach logs
In case the aws-node keeps on restarting
Pod Events
Logs
In case aws-node starts but unable to invoke EC2 API
What you expected to happen:
Managed Add on should provide an option to do rolling restart when the Service account role is updated or provide instructions to do the same manually when updating the service account role.
How to reproduce it (as minimally and precisely as possible):
AmazonEKS_CNI_Policy
form the node instance role.Environment:
kubectl version
): v1.18The text was updated successfully, but these errors were encountered: