-
-
Notifications
You must be signed in to change notification settings - Fork 4.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
EKS 1.28 creation failure using 19.19.1 -- NodeCreationFailure: Instances failed to join the Kubernetes cluster #2819
Comments
We currently run in to this exact same issue. Any updates on this? |
I need a reproduction in order to help troubleshoot |
This comment has been minimized.
This comment has been minimized.
@bryantbiggs I have re-added the terraform code for reproduction. You may need to update the VPC configuration for vpc module. We do not use the terraform-aws-module/vpc internally so couldn't come up with a complete private VPC creation here. Just to be clearer same VPC settings have been working in the past with older EKS versions. It has all the right tags including |
we only need a minimal example that reproduces the issue - are all of those configurations required? Try removing configs, you might find the issue on your own |
Within the eks module, yes all configuration is required. The only additional configuration above is the use of specific names and tagging of launch template, and provisioned resources. I have already gone through multiple deployments removing a lot of configuration which are not required with respect to the working of the cluster. Now, the only option to test is to test it without using terraform-as-module/eks. |
In our case the problem was an S3 gateway endpoint that received Bottlerockets ECR requests through an prefix list route in another VPC (centralized egress via TGW). For some weird reason S3 endpoints only forward traffic for the VPC they are deployed in though so this basically created a blackhole route for our S3 destined traffic. I found this issue by checking the system logs of the EC2 instances. @rajali maybe it’s worth checking in your case too? So all in all not a bug of this module but still rather interesting (at least as far as I am concerned). |
thanks @trc-ikeskin, the cluster is in the same vpc as the s3 gateway endpoint. |
closing this issue for now unless there is a minimal, reproducible example supplied that can demonstrate the issue |
I'm going to lock this issue because it has been closed for 30 days ⏳. This helps our maintainers find and focus on the active issues. If you have found a problem that seems similar to this, please open a new issue and complete the issue template so we can capture all the details necessary to investigate further. |
Description
I am getting this error on a fresh cluster creation
Error: waiting for EKS Node Group (###) create: unexpected state 'CREATE_FAILED', wanted target 'ACTIVE'. last error: ####: NodeCreationFailure: Instances failed to join the kubernetes cluster
Versions
Module version [Required]: 19.19.1
Terraform version:
Terraform v1.6.4
on linux_amd64
Provider version(s):
Reproduction Code [Required]
Steps to reproduce the behavior:
Yes YesRun terraform apply
Expected behavior
The cluster should get created with nodes showing status as "active" on the EKS console
Actual behavior
Error: waiting for EKS Node Group (###) create: unexpected state 'CREATE_FAILED', wanted target 'ACTIVE'. last error: ####: NodeCreationFailure: Instances failed to join the kubernetes cluster
The text was updated successfully, but these errors were encountered: