-
-
Notifications
You must be signed in to change notification settings - Fork 4.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to update Node Groups in place with cluster placement group strategy without EFA #3044
Comments
cc @james-masson ref #2959 |
Consider an EKS cluster deployed across 3 availability zones. A nodegroup that makes use of a "cluster" placement group will only be able to deploy into a single one of these AZs. That's the point of the "cluster" placement group, to put the instances into the same physical rack. It turns out that on initial deployment, this configuration is not a problem. The nodegroup with cluster placement successfully gets deployed into a single AZ of the 3 subnets configured with no errors. Arguably this itself is a bug. However, when it comes time to replace or upgrade the nodegroup, you get the error listed by my colleague @Josephuss This PR tries to fix this issue by adding the concept of a AZ filter into the nodegroups, to allow nodegroups to be deployed into a subset of the subnets that the rest of the cluster is configured with. We're aware that this can be currently worked around by overriding subnet ids. https://github.com/terraform-aws-modules/terraform-aws-eks/blob/master/node_groups.tf#L308 eg.
But this is quite fragile, because:
Hence seeing the general value in an interface like this - it's more practical, understandable and portable, and it has value outside of placement groups too.
|
This issue has been automatically marked as stale because it has been open 30 days |
This issue was automatically closed because of stale in 10 days |
I'm going to lock this issue because it has been closed for 30 days ⏳. This helps our maintainers find and focus on the active issues. If you have found a problem that seems similar to this, please open a new issue and complete the issue template so we can capture all the details necessary to investigate further. |
Description
When using a node group without EFA enabled and a placement group with cluster strategy, updates of the node group fail because the auto scaling group does not restrict the list of availability zones.
The node group with cluster placement successfully gets deployed into a single AZ of the 3 subnets configured with no errors, However, replacing or upgrading the node group fails as listed below due to the availability zone is not filtered and the update does not take place unless the subnet id is overridden in the configs.
Versions
Module version [Required]: 20.8.5
Terraform version: 1.5.7
Provider version(s): 5.40.0
Reproduction Code
This is a copy of the managed node group example with a placement group created. The default placement group strategy is cluster.
Steps to reproduce the behavior:
Expected behavior
Node group is increased as requested.
Actual behavior
Node group update may fail with:
Additional context
The text was updated successfully, but these errors were encountered: