-
Notifications
You must be signed in to change notification settings - Fork 431
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
VMSS Flex support for MachinePools #2813
Conversation
d873a5c
to
1defca4
Compare
7c21d95
to
97e818f
Compare
13c4dbe
to
f1b9a94
Compare
/test pull-cluster-api-provider-azure-e2e e2e failed to create an HA cluster and e2e-optional failed on the clusterclass spec. Neither appears to be related to these changes, but that's a worrisome number of flakes. ❄️ |
/test pull-cluster-api-provider-azure-e2e-optional |
I updated the |
/test pull-cluster-api-provider-azure-e2e-optional |
/lgtm |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
/assign @jackfrancis
/test pull-cluster-api-provider-azure-e2e Failed to provision HA cluster. |
This has passed the |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
/lgtm
/approve |
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: jackfrancis The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
seems like a race unrelated to this change, will investigate /retest |
/retest I haven't seen this before, yay. I don't think this PR caused it, so retesting. INFO: "With ipv6 worker node" started at Tue, 10 Jan 2023 20:36:35 UTC on Ginkgo node 6 of 10 and junit test report to file /logs/artifacts/test_e2e_junit.e2e_suite.1.xml
<< Timeline
[FAILED] Timed out after 1800.000s.
Expected success, but got an error:
<*errors.fundamental | 0xc003fa89c0>: {
msg: "cannot re-use a name that is still in use",
stack: [0x34dfe66, 0x34dda45, 0x352c418, 0x14ef085, 0x14ee57c, 0x190a67a, 0x190b582, 0x1908d2d, 0x352c09b, 0x351e558, 0x3522130, 0x2f9caf0, 0x35342c8, 0x18e639b, 0x18f9e98, 0x147c741],
}
cannot re-use a name that is still in use
In [It] at: /home/prow/go/src/sigs.k8s.io/cluster-api-provider-azure/test/e2e/helpers.go:879 @ 01/10/23 20:30:17.977
Full Stack Trace
sigs.k8s.io/cluster-api-provider-azure/test/e2e.InstallHelmChart({_, _}, {{0x4223910, 0xc000428190}, {{0xc000d03e30, 0x22}, {0xc000531510, 0x31}, {0xc000533cf3, 0x17}, ...}, ...}, ...)
/home/prow/go/src/sigs.k8s.io/cluster-api-provider-azure/test/e2e/helpers.go:879 +0x5db
sigs.k8s.io/cluster-api-provider-azure/test/e2e.InstallAzureDiskCSIDriverHelmChart({_, _}, {{0x4223910, 0xc000428190}, {{0xc000d03e30, 0x22}, {0xc000531510, 0x31}, {0xc000533cf3, 0x17}, ...}, ...}, ...)
/home/prow/go/src/sigs.k8s.io/cluster-api-provider-azure/test/e2e/cloud-provider-azure.go:79 +0x1f8
sigs.k8s.io/cluster-api-provider-azure/test/e2e.EnsureControlPlaneInitialized({_, _}, {{0x4223910, 0xc000428190}, {{0xc000d03e30, 0x22}, {0xc000531510, 0x31}, {0xc000533cf3, 0x17}, ...}, ...}, ...)
/home/prow/go/src/sigs.k8s.io/cluster-api-provider-azure/test/e2e/common.go:268 +0xb90
sigs.k8s.io/cluster-api/test/framework/clusterctl.ApplyClusterTemplateAndWait({_, _}, {{0x4223910, 0xc000428190}, {{0xc000d03e30, 0x22}, {0xc000531510, 0x31}, {0xc000533cf3, 0x17}, ...}, ...}, ...)
/home/prow/go/pkg/mod/sigs.k8s.io/cluster-api/test@v1.3.1/framework/clusterctl/clusterctl_helpers.go:334 +0xd30
sigs.k8s.io/cluster-api-provider-azure/test/e2e.glob..func1.5.1()
/home/prow/go/src/sigs.k8s.io/cluster-api-provider-azure/test/e2e/azure_test.go:303 +0x488
------------------------------
[SynchronizedAfterSuite] PASSED [0.000 seconds]
[SynchronizedAfterSuite]
/home/prow/go/src/sigs.k8s.io/cluster-api-provider-azure/test/e2e/e2e_suite_test.go:116 |
@mboersma fwiw I saw that last week (on another PR) |
that shouldn't be happening anymore since I made the helm install idempotent in #2915, I'll look into it |
/retest |
👁️ 🩸 |
What type of PR is this?
/kind feature
What this PR does / why we need it:
Implements VMSS Flex mode support for MachinePools.
There is still work to do here, I'm just getting a WIP branch out to help with collaboration. For example:Flex mode still on by defaultFixedMP ReadyReplicas not always syncing with actual VM countFixedfaultDomainCount still hard-coded at 3 because cache code isn't workingFixedWhich issue(s) this PR fixes:
Fixes #999
Fixes #2987
Special notes for your reviewer:
This work was started and mostly completed by @devigned and @jackfrancis. Thanks team!
TODOs:
Release note: