Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Azure cncf-upstream:capi missing PlatformImage (Quick Start issue) #2375

Closed
lmcdasm opened this issue Jun 10, 2022 · 8 comments
Closed

Azure cncf-upstream:capi missing PlatformImage (Quick Start issue) #2375

lmcdasm opened this issue Jun 10, 2022 · 8 comments
Assignees
Labels
kind/bug Categorizes issue or PR as related to a bug.

Comments

@lmcdasm
Copy link

lmcdasm commented Jun 10, 2022

/kind bug

[Before submitting an issue, have you checked the Troubleshooting Guide?]

What steps did you take and what happened:
Following the Quick Start Guide - Using Azure Provider steps and documentation.

After generating the workload cluster configuration (workload-cluster-3 in this case), i was able to do the "kubectl apply -f workload-cluster-3.yaml" the the "kubectl get cluster" comes back as "provisioned".

I can see that the Azure LB, VNET, NSG and other objects are being created, however the VM's fail to create. Complaining of a missing image

ERROR MESSAGE IN AZURE ACTIVITY LOGS/RES MANAGER:
image

I have tried this with v1.24.0 (what my AKS cluster that is hosting the management cluster is running), and tried 1.23.7 (Last 1.23 release from AKS) and 1.23.5.

I would imagine that im missing something simple - reading the documentation it seems to indicate that Azure Marketplace would have the images, however i can only find the Windows one there (from CNCF-UPSTREAM) and a CAPI TKG Vmware one, not Ubuntu one.

What did you expect to happen:

I would have expected the cluster deployment to start up the VM using an image that is available from within Azure without having to add additional configuration. Following the Quick Start guide, we cannot connect using the workload-cluster specific KUBECONFIG / Apply Calico workaround until the VM is up (and the workload-cluster control API is able to respond behind the LB created).

image

Anything else you would like to add:
its very possible that I missed something here / in the Quick start, so apologies. Do i need to pull the image from source and drop to local repo, etc? I couldnt find the image locations in the Docs (again, sorry if this is silly/newbie question) :)

As a side question, are the workload clusters simply "VMs with ubuntu and K8s installed after the fact" ? I was thinking/wondering the following:

  • if im using an AKS as the management cluster host, should we not see either:
  1. A new node pool within the existing AKS that is used as a "User Pool" (from the context of the management cluster) and then runs the workload-cluster "there". AKS Supports having Node pools on different VNETs now (well subnets anyway).

  2. if not within the reach of the Management (Host AKS) cluster, would it not be better to have the Azure provider spin AKS instances with VMSS (and have autoscaling and all that jazz with it)?

Sorry, maybe the above is counter to the cloud Native idea, so you dont want to go too much Azure AKS land?

Environment:
image
image

  • cluster-api-provider-azure version:
    PROVIDER= "infrastructure-azure" - Version = v1.3.1

  • Kubernetes version: (use kubectl version):
    cluster Version = 1.24.1
    server Version = 1.24.0

  • OS (e.g. from /etc/os-release):
    CentOS 7.9.2009

@k8s-ci-robot k8s-ci-robot added the kind/bug Categorizes issue or PR as related to a bug. label Jun 10, 2022
@CecileRobertMichon
Copy link
Contributor

/assign @mboersma

Hi @lmcdasm, the latest batch of images were named differently in the cncf-upstream offer which required #2302 to be able to use the latest versions released (v1.24.1, v1.24.0, v1.23.7, v1.22.10, v1.21.13).

In the meantime, you can use an older Kubernetes version as a workaround (such as v1.23.6).

v1.3.2 will be released today which will fix the issue and make the quickstart work again out of the box again. Apologies for the inconvenience.

@CecileRobertMichon
Copy link
Contributor

Regarding your other questions:

As a side question, are the workload clusters simply "VMs with ubuntu and K8s installed after the fact" ? I was thinking/wondering the following:

if im using an AKS as the management cluster host, should we not see either:
A new node pool within the existing AKS that is used as a "User Pool" (from the context of the management cluster) and then runs the workload-cluster "there". AKS Supports having Node pools on different VNETs now (well subnets anyway).

if not within the reach of the Management (Host AKS) cluster, would it not be better to have the Azure provider spin AKS instances with VMSS (and have autoscaling and all that jazz with it)?

A workload cluster != a node pool. It is a full Kubernetes cluster, including a control-plane. AKS is a managed Kubernetes service offered by Azure. Cluster API on the other hand is a tool for managing the lifecycle of Kubernetes clusters declaratively. When you build a workload cluster with CAPZ with the "default" flavor, it builds a full Kubernetes clusters that you as a user can manage yourself. There are very good reasons for using AKS and we actually recommend that most users do it when it fits their needs as it takes away a lot of the hassle of managing your own k8s clusters. CAPZ is an alternative for users who have requirements for managing their own k8s clusters.

If you are interested in using AKS but still want the declarative lifecycle management that Cluster API offers, check out https://capz.sigs.k8s.io/topics/managedcluster.html, which is an experimental feature of CAPZ to create/manage AKS clusters.

@lmcdasm
Copy link
Author

lmcdasm commented Jun 14, 2022 via email

@lmcdasm
Copy link
Author

lmcdasm commented Jun 14, 2022 via email

@CecileRobertMichon
Copy link
Contributor

For sure! Your original issue should now be fixed as v1.3.2 has been released. Please let us know if you encounter other issues. We're active on k8s slack in the #cluster-api-azure channel as well.

@CecileRobertMichon
Copy link
Contributor

/close

@k8s-ci-robot
Copy link
Contributor

@CecileRobertMichon: Closing this issue.

In response to this:

/close

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@lmcdasm
Copy link
Author

lmcdasm commented Jun 15, 2022

@CecileRobertMichon

Hey there.

Confirmed that i was able to roll out 1.24.0 Azure IaaS based Clusters and was able to finish my deployment.

Thanks again!
Daniel

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/bug Categorizes issue or PR as related to a bug.
Projects
None yet
Development

No branches or pull requests

4 participants