Cluster API provider for metal-stack

Grigoriy Mikhalkin Thu Mar 25 2021

One of the main goals of the metal-stack project is to provide a fast and convenient way to provision Kubernetes clusters on bare-metal provided by metal-stack’s infrastructure layer.

Until now Gardener was the only way to manage Kubernetes clusters on metal-stack with our provider implementation for Gardener.

Today we’re happy to provide you with a second option: a provider implementation for the Cluster API spec!

What is Cluster API

Cluster API was started by the Kubernetes Special Interest Group (SIG) Cluster Lifecycle and its goal is to simplify provisioning, upgrading, and operating multiple Kubernetes clusters. Thereby it’s using Kubernetes-style APIs to automate cluster lifecycle management. Clusters and nodes can be managed similar to how applications can be managed with Pods and Deployments:

  • to manage clusters, there is a Cluster resource
  • for managing machines, there is a Machine resource (comparable to the Pod resource for apps)
  • and a MachineDeployment resource (comparable to the Deployment resource for apps)

Cluster API

Cluster API components usually run in a separate management cluster. It includes core controllers, like Cluster controller and Machine controller which implement high-level logic for managing the cluster lifecycle. Infrastructure providers contain provider specific logic, like communication with bare-metal machines and setting additional services and functionality (cluster network, firewalls, etc).

For more details, you may check the Cluster API docs.

Architecture

The provider can be seen as an intermediate layer between the metal-stack infrastructure layer, which manages bare-metal machines and the Cluster API, which manages Kubernetes clusters. It communicates with the metal-stack API using our golang client metal-go.

provider architecture

The provider consists of 3 controllers:

For more details about provider components (controllers, resources), take a look at our provider documentation.

Try it out on your machine

Prerequisites

To start our Cluster API provider locally, you need to clone the provider repository and the branch 3-machines of the mini-lab:

git clone https://github.com/metal-stack/cluster-api-provider-metalstack.git
git clone https://github.com/metal-stack/mini-lab.git
cd mini-lab && git checkout 3-machines && cd -

After that, follow the instructions in the README file.

Besides that, you’ll need:

Install Cluster API components

Before initializing the Cluster API management cluster run:

make manager-test

That command generates a manifest file that bundles the provider deployment and all required CRDs and places it in out/ directory. To install Cluster API components (including the metal-stack provider) in the kind cluster deployed by mini-lab, run clusterctl init with generated manifest file passed as --config flag:

clusterctl init --config=out/test/infrastructure-metalstack/clusterctl-{latest_tag}-dirty.yaml --infrastructure=metalstack -v3

Create a cluster

After everything is ready (you can check readiness of pods by running kubectl get pods -A), you can generate a manifest file from template. Some important resources are defined there:

  • KubeadmControlPlane – used for configuring Control Plane nodes of cluster.
  • MetalStackCluster – specifies things like private network, partition and project in which cluster will be provisioned.
  • MachineDeployment – similar to Deployment resource, but for machines. Specifies machine config and number of machines to provision. Used for configuring and replicating worker nodes.

To generate the manifest file, just run make cluster which places it in out/ directory. After it’s ready, run kubectl apply -f out/cluster.yaml which creates the corresponding resources that define the control plane and worker machines. After some time, the control plane and worker nodes become ready and you can fetch the Kubeconfig of the just created cluster and play with it!

Check it out and have fun!