Skip to content

Latest commit

 

History

History
179 lines (137 loc) · 6.49 KB

build.md

File metadata and controls

179 lines (137 loc) · 6.49 KB

Building and testing MicroK8s

Building the snap from source

To build the MicroK8s snap, you need to install Snapcraft.

sudo snap install snapcraft --classic

Building a snap is done by running snapcraft in the root of the project. Snapcraft spawn a VM managed by Multipass and builds MicroK8s inside of it. If you don’t have Multipass installed, snapcraft will first prompt for its automatic installation.

After snapcraft finishes, you can install the newly compiled snap:

sudo snap install microk8s_*_amd64.snap --classic --dangerous

Building the snap using LXD

Alternatively, you can build the snap in an LXC container. This is useful if you are working in a virtual machine without nested virtualization. Snapcraft and LXD are needed in this case:

sudo snap install lxd
sudo apt-get remove lxd* -y
sudo apt-get remove lxc* -y
sudo lxd init
sudo usermod -a -G lxd ${USER}

Build the snap with:

git clone http://github.com/canonical/microk8s
cd ./microk8s/
snapcraft --use-lxd

Finally, install it with:

sudo snap install microk8s_*_amd64.snap --classic --dangerous

Building a custom MicroK8s package

To produce a custom build with specific component versions you cannot use the snapcraft build process on the host OS. You need to prepare an LXC container with Ubuntu 16:04 and snapcraft:

lxc launch ubuntu:16.04 --ephemeral test-build
lxc exec test-build -- snap install snapcraft --classic
lxc exec test-build -- apt update
lxc exec test-build -- git clone https://github.com/canonical/microk8s

You can then set the following environment variables prior to building:

  • KUBE_VERSION: Kubernetes release to package. Defaults to latest stable.
  • ETCD_VERSION: version of etcd.
  • CNI_VERSION: version of CNI tools.
  • KUBE_TRACK: Kubernetes release series (e.g., 1.10) to package. Defaults to latest stable.
  • ISTIO_VERSION: istio release.
  • KNATIVE_SERVING_VERSION: Knative Serving release.
  • KNATIVE_EVENTING_VERSION: Knative Eventing release.
  • RUNC_COMMIT: the commit hash from which to build runc.
  • CONTAINERD_COMMIT: the commit hash from which to build containerd
  • KUBERNETES_REPOSITORY: build the Kubernetes binaries from this repository instead of getting them from upstream
  • KUBERNETES_COMMIT: commit to be used from KUBERNETES_REPOSITORY for building the Kubernetes binaries

For building you prepend the variables you need as well as SNAPCRAFT_BUILD_ENVIRONMENT=host so the current LXC container is used. For example to build the MicroK8s snap for Kubernetes v1.9.6, run:

lxc exec test-build -- sh -c "cd microk8s && SNAPCRAFT_BUILD_ENVIRONMENT=host KUBE_VERSION=v1.9.6 snapcraft"

The produced snap is inside the ephemeral LXC container, you need to copy it to the host:

lxc file pull test-build/root/microk8s/microk8s_v1.9.6_amd64.snap .

After copying it, you can install it with:

snap install microk8s_*_amd64.snap --classic --dangerous

Assembling the Calico CNI manifest

The calico CNI manifest can be found under upgrade-scripts/000-switch-to-calico/resources/calico.yaml. Building the manifest is subject to the upstream calico project k8s installation process. The calico.yaml manifest is a slightly modified version of: https://projectcalico.docs.tigera.io/archive/v3.25/manifests/calico.yaml:

  • Set the calico backend from "bird" to "vxlan": calico_backend: "vxlan"

  • CALICO_IPV4POOL_CIDR was set to "10.1.0.0/16"

  • CNI_NET_DIR had to be added and set to "/var/snap/microk8s/current/args/cni-network"

  • We set the following mount paths:

    1. var-run-calico to /var/snap/microk8s/current/var/run/calico
    2. var-lib-calico to /var/snap/microk8s/current/var/lib/calico
    3. cni-bin-dir to /var/snap/microk8s/current/opt/cni/bin
    4. cni-net-dir to /var/snap/microk8s/current/args/cni-network
    5. cni-log-dir to /var/snap/microk8s/common/var/log/calico/cni
    6. host-local-net-dir to /var/snap/microk8s/current/var/lib/cni/networks
    7. policysync to /var/snap/microk8s/current/var/run/nodeagent
  • We enabled vxlan following the instructions in the official docs.

  • The liveness and readiness probes of bird-live was commented out

  • We set the IP autodetection method to

                - name: IP_AUTODETECTION_METHOD
                value: "first-found"
    
  • In the cni_network_config in the calico manifest we also set:

            "log_file_path": "/var/snap/microk8s/common/var/log/calico/cni/cni.log",
            "nodename_file": "/var/snap/microk8s/current/var/lib/calico/nodename",
    
  • The mount-bpffs pod is commented out. This disables eBPF support but allows the CNI to deploy inside an LXC container.

  • The bpffs mount is commented out:

              - name: bpffs
                mountPath: /sys/fs/bpf
    

    end

          - name: bpffs
            hostPath:
              path: /sys/fs/bpf
              type: Directory
    

Running the tests locally

To successfully run the tests you need to install:

  1. From your distribution's repository:
    • python3
    • pytest
    • pip3
    • docker.io
    • tox
  2. From pip3:
    • requests
    • pyyaml
    • sh

On Ubuntu 20.04, for example, run the following commands to get these dependencies.

sudo apt install python3-pip docker.io tox -y
pip3 install -U pytest requests pyyaml sh

First, run the static analysis using Tox. Run the following command in the root of the project.

tox -e lint

Then, use the "Building the snap from source" instructions to build the snap and install it locally. The tests check this locally installed MicroK8s instance.

Finally, run the tests themselves. The test-simple.py and test-upgrade.py files under the tests directory are the two main files of our test suite. Running the tests is done with pytest:

pytest -s tests/test-simple.py
pytest -s tests/test-upgrade.py

Note: the ingress and dashboard-ingress tests make use of nip.io for wildcard ingress domains on localhost. DNS rebinding protection may prevent the resolution of the domains used in the tests.

A workaround is adding these entries to /etc/hosts:

127.0.0.1 kubernetes-dashboard.127.0.0.1.nip.io
127.0.0.1 microbot.127.0.0.1.nip.io