-
-
Notifications
You must be signed in to change notification settings - Fork 322
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Google Cloud and Fedora CoreOS Custom worker nodes doesn't join the cluster #942
Comments
I wasn't able to reproduce this, clusters with a worker pool of Fedora CoreOS nodes work as expected. You'll probably need to clarify doesn't join. Do you never see them appear or are they there but NotReady? Check the Kubelet logs, be sure you're using the correct kubeconfig and other settings. If a node is reporting its hostname as localhost, check its hostname, is it still localhost? Did other units fail before the hostname was set maybe? Otherwise, restart Kubelet, a kubelet cannot register itself as localhost. |
Thank you, they don't show in get nodes, and after a while 1 more node only shows "aside from the controllers" labeled with localhost, flatcar works fine - I have uploaded the bucket and created compute image and all is good. |
Have you discovered anything different about your Fedora CoreOS nodes setup from the suggestions or the example? Or you're just switching to Flatcar Linux anyway and I'll close this out. |
Thanks alot @dghubble - I will apply the example shortly and update the issue with the result, any particular info you would want to see to help debug this issue? |
You'd need to provide clear details of why your node didn't join and show an example that is independently reproducible. Otherwise, I've already checked a GCP worker pool setup and found it to work fine. I don't provide debug/support. |
A kubelet cannot register itself as localhost, nobody can talk to a remote node via localhost. If a node is reporting its hostname as localhost, check its hostname, is it still localhost? Check dmesg. Did you have other systemd units fail that prevented the hostname being set? Restart Kubelet to pickup the current hostname. |
Thank you @dghubble I will try to rebuild the clusters again and restart kubelet and check. |
Description
But Custom worker nodes doesn't join the cluster, and some workers node if it joins it shows localhost as hostname in
kubectl get nodes
In cluster on Google Cloud - with Fedora CoreOS Stable.
I have added the custom snippet for workers and controllers in #915 and ssh setup works fine.
I have tried the same setup with Flatcar linux and it works as expected
Steps to Reproduce
Normal as per the documentation google cloud cluster and as per custom worker pool from the docs
Expected behavior
Worker should join the cluster and they should have the correct host name.
Environment
v0.14.7
Possible Solution
Link to a PR or description.
The text was updated successfully, but these errors were encountered: