Skip to content

piraeusdatastore/linstor-csi

Repository files navigation

Linstor CSI Plugin

build tests latest version

This CSI plugin allows for the use of LINSTOR volumes on Container Orchestrators that implement CSI, such as Kubernetes.

Building

If you wish to create a docker image for a local registry run make upload REGISTRY=local.example.com.

Deployment

If you are looking to deploy a full LINSTOR setup with LINSTOR controller and satellites, take a look at our operator.

This project ONLY deploys the CSI components, a working LINSTOR cluster is required. Use our example deployment in examples/k8s/deploy as base to deploy only the LINSTOR CSI components. You will need to update every occurence of LINSTOR_CONTROLLER_URL with the actual URL of your LINSTOR Controller, for example like this:

$ LINSTOR_CONTROLLER_URL=http://linstor-controller.example.com:3370
$ kubectl kustomize http://github.com/piraeusdatastore/linstor-csi/examples/k8s/deploy \
  | sed "s#LINSTOR_CONTROLLER_URL#$LINSTOR_CONTROLLER_URL#" \
  | kubectl apply --server-side -f -

Usage

This project must be used in conjunction with a working LINSTOR cluster, version 0.9.11 or better. LINSTOR's documentation is the foremost guide on setting up and administering LINSTOR.

⚠️️ Known issues

  • Due to the way ZFS snapshots work, provisioning new Volumes from existing Volumes does not work using ZFS storage pools. The internal temporary snapshot cannot be deleted after the new volume is created.

    As a workaround, first create a VolumeSnapshot of the existing volume and restore from that snapshot. Also read the issue below!

  • Deletion of VolumeSnapshots will fail for ZFS based volumes if a volume restored from the snapshot exists in the cluster.

Kubernetes

After the plugin has been deployed, you're free to create storage classes that point to the name of the external provisioner associateed with the CSI plugin and have your users start provisioning volumes from them. A basic storage class could look like this:

apiVersion: storage.k8s.io/v1
kind: StorageClass
metadata:
  name: linstor-basic-storage
provisioner: linstor.csi.linbit.com
allowVolumeExpansion: true
parameters:
  linstor.csi.linbit.com/placementCount: "2"
  linstor.csi.linbit.com/storagePool: "my-storage-pool"
  linstor.csi.linbit.com/resourceGroup: "linstor-basic-storage"
  csi.storage.k8s.io/fstype: xfs
  # You can override LINSTOR properties by adding the property.linstor.csi.linbit.com prefix:
  property.linstor.csi.linbit.com/DrbdOptions/auto-quorum: suspend-io

A full list of all parameters usable in a storage class is available here.

Ensure that all kubelets that are expected to use LINSTOR volumes have a running LINSTOR satellite that is configured to work with the LINSTOR controller configured in the plugin's deployment files and that the storage pool indicated in the storage class has been properly configured. This pool does not need to be present on the Kubelets themselves for volumes attached over the network.

Most of the documentation for using this project with Kubernetes is located here.

Kubevirt

An example of using the CSI driver in combination with kubevirt (block device mode, live migration) can be found in the examples/kubevirt/ directory.