-
Notifications
You must be signed in to change notification settings - Fork 716
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add retries for kubeadm join / MarkControlPlane #2093
Comments
/assign |
@RA489 i'm going to take this ticket as i have some time later today and tomorrow. |
if 2 minutes is not sufficient what value would you recommend? |
@neolit123 thanks for pointing this out. |
@fabriziopandini: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Is this a BUG REPORT or FEATURE REQUEST?
BUG REPORT
Versions
kubeadm version: v1.17.*
What happened?
While executing Cluster API tests, in some cases it was observed kubeadm join failures when adding the master label to the joining node.
xref kubernetes-sigs/cluster-api#2769
What you expected to happen?
To make mark control plane more resilient by adding a retry loop to this operation
How to reproduce it (as minimally and precisely as possible)?
This error happens only sometimes, most probably when there is a temporary blackout of the load balancer that sits in front of the API servers (HA proxy reloading his configuration).
Also, the error might happen when the new API server enters the load balancing pool but the underlying etcd member is not yet available due to slow network/slow I/O causing delays in etcd getting online or in some cases, also change fo the etcd leader.
Anything else we need to know?
Important: if possible the change should be kept as small and possible and backported
The text was updated successfully, but these errors were encountered: