Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Failure during planning phase when node pool resources are exhausted #4116

Closed
coryodaniel opened this issue Jul 26, 2019 · 3 comments
Closed
Assignees
Labels
bug forward/review In review; remove label to forward service/container

Comments

@coryodaniel
Copy link

When a machine type is exhausted in a zone, terraform plan fails making it impossible to destroy, taint, or change the nodepool.

To continue you have to delete the node pool in the GCP UI or CLI and then remove it from the terraform state.

Community Note

  • Please vote on this issue by adding a 👍 reaction to the original issue to help the community and maintainers prioritize this request
  • Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
  • If you are interested in working on this issue or have submitted a pull request, please leave a comment
  • If an issue is assigned to the "modular-magician" user, it is either in the process of being autogenerated, or is planned to be autogenerated soon. If an issue is assigned to a user, that user is claiming responsibility for the issue. If an issue is assigned to "hashibot", a community member has claimed the issue already.

Terraform Version

Affected Resource(s)

  • google_container_node_pool

Terraform Configuration Files

Applies to any GCP machine type that is exhausted.

Debug Output

Error: Error reading NodePool "pool-worker-preemptible" from cluster "CLUSTER_NAME": Nodepool "pool-worker-preemptible" has status "RUNNING_WITH_ERROR" with message "asia-northeast1-a: Deploy error: Not all instances running in IGM after 4m4.752299662s. Expect 1. Current errors: [ZONE_RESOURCE_POOL_EXHAUSTED_WITH_DETAILS]: Instance 'INSTANCE_NAME' creation failed: The zone 'projects/ox-delivery-prod/zones/asia-northeast1-a' does not have enough resources available to fulfill the request. '(resource type:compute)'."

Expected Behavior

The plan reads the current state and does not fail.

Actual Behavior

The plan reads and encounters an error on the resource and crashes.

Steps to Reproduce

  1. terraform apply
  2. Wait for the machine type to be exhausted in a region
  3. terraform plan # this will crash

Important Factoids

Machine type is exhausted. Currently requesting n1-standard-96 pvms. These are easy to exhaust (and I know we shouldn't be using them), but the bug exists for other machine types as well.

References

@ghost ghost added the bug label Jul 26, 2019
@paddycarver paddycarver self-assigned this Aug 1, 2019
@paddycarver
Copy link
Contributor

I think this is definitely related to #3304. I think the crux of the problem is that we need to detect an error for when resources are exhausted, but we need to distinguish that from other errors, and I think that's trickier than it would appear at first blush.

@rileykarson
Copy link
Collaborator

Closing as dupe of #3304

@ghost
Copy link

ghost commented Mar 28, 2020

I'm going to lock this issue because it has been closed for 30 days ⏳. This helps our maintainers find and focus on the active issues.

If you feel this issue should be reopened, we encourage creating a new issue linking back to this one for added context. If you feel I made an error 🤖 🙉 , please reach out to my human friends 👉 [email protected]. Thanks!

@ghost ghost locked and limited conversation to collaborators Mar 28, 2020
@github-actions github-actions bot added service/container forward/review In review; remove label to forward labels Jan 15, 2025
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
bug forward/review In review; remove label to forward service/container
Projects
None yet
Development

No branches or pull requests

3 participants