-
Notifications
You must be signed in to change notification settings - Fork 4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
cluster autoscaler deleted non-empty node as ScaleDownEmpty #5790
Comments
Seems like CA's knowledge of the node is stale and since the ToBeDeleted taint is added after the pod is scheduled it's possible that the pod was scheduled after CA obtained node info. Something like this:
|
yeah, this is what it looks like, is there a way to enforce CA to drain node in this scenario or any way to improve on this? |
is there any flag or workaround to tackle this issue? |
@infa-ddeore sorry I don't have enough bandwidth to look at this. If this is important, please bring it up in the sig-autoscaling meeting so that someone else can take a look. 🙏 |
I think this issue is not just related ScaleDownEmpty case, but also related to DrainNode case. From I think the nodeinfo might be stale, and ca will not know some new pod is scheduled to the node. |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs. This bot triages issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /close not-planned |
@k8s-triage-robot: Closing this issue, marking it as "Not Planned". In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Which component are you using?:
cluster-autoscaler
What version of the component are you using?:
1.23
Component version:
What k8s version are you using (
kubectl version
)?: 1.23kubectl version
OutputWhat environment is this in?:
EKS
What did you expect to happen?:
cluster autoscaler should re-check node is empty or not before deleting
What happened instead?:
cluster autoscaler deleted non-empty node
How to reproduce it (as minimally and precisely as possible):
It is difficult to reproduce the issue, this behaviour is shows from the logs,
ToBeDeletedTaint
taint to the nodeAnything else we need to know?:
The text was updated successfully, but these errors were encountered: