-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
AWS Load Balancer Controller v2.1.3 + EKSCTL (Multiple tagged SGs) #3459
Comments
Thanks for opening the issue @marciogmorales ! https://github.com/kubernetes/kubernetes/blob/68b4e26caf6ede7af577db4af62fb405b4dd47e6/staging/src/k8s.io/legacy-cloud-providers/aws/aws.go#L4111-L4138 the error is coming back from the AWS cloud provider. It looks for the The error your getting would indicate that the Which nodegroup type is instance |
Thanks @marciogmorales. I don't see the |
closing due to inactivity |
Hi @aclevername, I've run into this specific error as well and hope to revive this debugging effort ( UPDATE: RESOLVEDThe issue was caused by having a tag called Info: background
Info:
|
core-a (a unmanaged nodepool, the first in my eksctl config list) | ClusterSharedNodeSecurityGroup |
---|---|
![]() |
![]() |
Info: all security groups
All security groups related to the VPC at least, I omitted a few related to another VPC.
Info: about associated instance
instance info: tags section
instance info: security section
Info: eksctl-cluster-config.yaml
apiVersion: eksctl.io/v1alpha5
kind: ClusterConfig
metadata:
name: jmte
region: us-west-2
version: "1.19"
tags:
2i2c.org/project: jmte
availabilityZones: [us-west-2d, us-west-2b, us-west-2a]
iam:
withOIDC: true
nodeGroups:
- name: core-a
availabilityZones: [us-west-2d]
instanceType: m5.large
minSize: 0
maxSize: 2
desiredCapacity: 1
volumeSize: 80
labels:
hub.jupyter.org/node-purpose: core
tags:
k8s.io/cluster-autoscaler/node-template/label/hub.jupyter.org/node-purpose: core
iam:
withAddonPolicies:
autoScaler: true
efs: true
- name: user-a
availabilityZones: [us-west-2d]
instanceType: m5.xlarge # 57 pods, 4 cpu, 16 GB
minSize: 0
maxSize: 20
desiredCapacity: 0
volumeSize: 80
labels:
hub.jupyter.org/node-purpose: user
tags:
k8s.io/cluster-autoscaler/node-template/label/hub.jupyter.org/node-purpose: user
iam:
withAddonPolicies:
autoScaler: true
efs: true
- name: worker-xlarge
availabilityZones: &availabilityZones [us-west-2d, us-west-2b, us-west-2a]
minSize: &minSize 0
maxSize: &maxSize 8
desiredCapacity: &desiredCapacity 0
volumeSize: &volumeSize 80
labels: &labels
k8s.dask.org/node-purpose: worker
taints: &taints
k8s.dask.org_dedicated: worker:NoSchedule
tags: &tags
k8s.io/cluster-autoscaler/node-template/label/k8s.dask.org/node-purpose: worker
k8s.io/cluster-autoscaler/node-template/taint/k8s.dask.org_dedicated: worker:NoSchedule
iam: &iam
withAddonPolicies:
autoScaler: true
efs: true
instancesDistribution:
instanceTypes:
- m5a.xlarge # 57 pods, 4 cpu, 16 GB (AMD, 10 GBits network, 100% cost)
- m5.xlarge # 57 pods, 4 cpu, 16 GB (Intel, 10 GBits network, ~112% cost)
# - m5n.xlarge # 57 pods, 4 cpu, 16 GB (Intel, 25 GBits network, ~139% cost)
onDemandBaseCapacity: &onDemandBaseCapacity 0
onDemandPercentageAboveBaseCapacity: &onDemandPercentageAboveBaseCapacity 0
spotAllocationStrategy: &spotAllocationStrategy capacity-optimized
- name: worker-2xlarge
availabilityZones: *availabilityZones
minSize: *minSize
maxSize: *maxSize
desiredCapacity: *desiredCapacity
volumeSize: *volumeSize
labels: *labels
taints: *taints
tags: *tags
iam: *iam
instancesDistribution:
instanceTypes:
- m5a.2xlarge # 57 pods, 8 cpu, 32 GB (AMD, 10 GBits network, 100% cost)
- m5.2xlarge # 57 pods, 8 cpu, 32 GB (Intel, 10 GBits network, ~112% cost)
# - m5n.2xlarge # 57 pods, 8 cpu, 32 GB (Intel, 25 GBits network, ~139% cost)
onDemandBaseCapacity: *onDemandBaseCapacity
onDemandPercentageAboveBaseCapacity: *onDemandPercentageAboveBaseCapacity
spotAllocationStrategy: *spotAllocationStrategy
# ... repeating entries for worker-4xlarge, worker-8xlarge, and worker-16xlarge omitted
I updated my comment with the resolution for me to the issue, see #3459 (comment). |
Hello,
I'm using eksctl to launch an Amazon EKS cluster consisting of:
1 Managed Linux node
2 Windows Nodes
The Windows nodes by default has two Security Groups created and attached:
1 - ClusterSharedNodeSecurityGroup
2 - SG for communication between the control plane and worker nodes in group windows-ng-ltsc
However,
When trying to launch a Type: Loadbalancer (NLB) I receive the following error on the Service:
Warning SyncLoadBalancerFailed 23s (x5 over 101s) service-controller Error syncing load balancer: failed to ensure load balancer: Multiple tagged security groups found for instance i-0f7XXXX; ensure only the k8s security group is tagged; the tagged groups were sg-0cf0274a15cb31e9a(eksctl-eks-windows-nodegroup-windows-ng-ltsc-SG-7ZXXXXXZ) sg-04be95eb6f37ad123(eks-cluster-sg-eks-windows-1XXXX7)
I found some people complain about the same some years ago, but sounds like the issue still persists. Any clue on how to solve it?
The text was updated successfully, but these errors were encountered: