Skip to content

Commit

Permalink
Added page for how to use GPU
Browse files Browse the repository at this point in the history
  • Loading branch information
Xartos committed Dec 16, 2024
1 parent 9f49480 commit 04884fa
Show file tree
Hide file tree
Showing 2 changed files with 49 additions and 0 deletions.
48 changes: 48 additions & 0 deletions docs/user-guide/gpu.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,48 @@
---
search:
boost: 2
---

# Using GPU Workload in Welkin

!!! elastisys "For Elastisys Managed Services Customers"
You can order a new Environment with GPU support by filing a [service ticket](https://elastisys.atlassian.net/servicedesk/).
Make sure to specify the need for GPU Nodes in "Additional information or comments".
If you are unsure, get in touch with your account manager.

As the demand for AI, machine learning, and data science workloads grows, Kubernetes provides a flexible and scalable platform to manage these applications.
In this guide, we'll focus on how to use GPU in the Welkin platform.

> [!NOTE]
> Not all infrastructure providers have support for GPU.
> Check with the platform administrator to find out if your environment has support for GPU workload.
## Deployment

To use GPU resources in your cluster, you need to create a deployment that is using the resource `nvidia.com/gpu`.
Here's an example of how to configure GPU resources for a Pod:

```yaml
apiVersion: v1
kind: Pod
metadata:
name: cuda-vectoradd
spec:
restartPolicy: OnFailure
containers:
- name: cuda-vectoradd
image: "nvcr.io/nvidia/k8s/cuda-sample:vectoradd-cuda11.7.1-ubuntu20.04"
resources:
limits:
nvidia.com/gpu: 1
```
> [!NOTE]
> If your cluster is using the cluster autoscaling feature and there's currently not enough resources, the autoscaler will create one for you.
> It might take a couple of minutes for the new node to join the cluster and to install all the pre-requisites.
### Further Reading
- [Kubernetes Schedule GPU Documentation](https://kubernetes.io/docs/tasks/manage-gpus/scheduling-gpus/)
- [Kubernetes Cluster Autoscaler Documentation](https://kubernetes.io/docs/concepts/cluster-administration/cluster-autoscaling/)
- [Cluster Autoscaler FAQ](https://github.com/kubernetes/autoscaler/blob/master/cluster-autoscaler/FAQ.md)
1 change: 1 addition & 0 deletions mkdocs.yml
Original file line number Diff line number Diff line change
Expand Up @@ -150,6 +150,7 @@ nav:
- "Prepare Your Identity Provider (IdP)": "user-guide/prepare-idp.md"
- "Container registry": "user-guide/registry.md"
- "Continuous Development": "user-guide/continous-development.md"
- "GPU": "user-guide/gpu.md"
- "Network Model": "user-guide/network-model.md"
- "Kubernetes API": "user-guide/kubernetes-api.md"
- "Kubernetes UI": "user-guide/kubernetes-ui.md"
Expand Down

0 comments on commit 04884fa

Please sign in to comment.