-
Notifications
You must be signed in to change notification settings - Fork 430
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add health resilience for managed cluster #2169
Comments
@dtzar thanks for raising this issue. For 1), this seems like a clusterctl / Cluster API feature request to me, not one specific to Azure. Would you mind creating an issue in the CAPI repo so we can brainstorm solutions with the other providers? For 2) specifically, this seems like something that should have gotten caught in a webhook instead. I'm actually refactoring the managed clusters surface area (#2168) quite a bit and noticed this too -- there are a lot of validations being done in the managed cluster reconciler that should be done in webhooks instead. That would prevent the user from even creating the resource if it's invalid. I'll open a separate issue for this. |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs. This bot triages issues and PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs. This bot triages issues and PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
/remove-lifecycle rotten |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs. This bot triages issues and PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
I think we can close this. @dtzar @CecileRobertMichon please re-open if I'm wrong |
/kind feature
Describe the challenges
After "successfully" completing
clusterctl init --infrastructure azure
I'm personally finding there are various scenarios where the actual management cluster is in an unhealthy state. When you then runkubectl -f myCapzWorkloadCluster.yaml
the cluster is permanently stuck in the provisioning state.kubectl describe clusters
at the end shows it can't provision because control plane is not ready:The two specific scenarios I've hit why the managed cluster isn't healthy:
Describe potential solutions you'd like to see
Any number of ways to ensure that the management cluster is healthy before attempting to deploy a workload cluster. This could include things like (in no particular order):
Environment:
kubectl version
):1.23.1 client, 1.21.5 server/etc/os-release
): WSL2 on Ubuntu Focal from Windows 11The text was updated successfully, but these errors were encountered: