Kubernetes (Helm)#
This guide walks through installing the AMD GPU Operator on a Kubernetes cluster using Helm.
Prerequisites#
System Requirements#
- Kubernetes cluster v1.29.0 or later 
- Helm v3.2.0 or later 
- kubectlcommand-line tool configured with access to the cluster
- Cluster admin privileges 
Cluster Requirements#
- A functioning Kubernetes cluster with: - All system pods running and ready 
- Properly configured Container Network Interface (CNI) 
- Worker nodes with AMD GPUs 
 
Required Access#
- Access to pull images from: - AMD’s container registry or your configured registry 
- Public container registries (Docker Hub, Quay.io) 
 
Pre-Installation Steps#
1. Verify Cluster Status#
Check that your cluster is healthy and running:
kubectl get nodes
kubectl get pods -A
Expected output should show:
- All nodes in - Readystate
- System pods running (kube-system namespace) 
- CNI pods running (e.g., Flannel, Calico) 
Example of a healthy cluster:
NAMESPACE      NAME                                          READY   STATUS    RESTARTS   AGE
kube-flannel   kube-flannel-ds-7krtk                         1/1     Running   0          10d
kube-system    coredns-7db6d8ff4d-644fp                      1/1     Running   0          2d20h
kube-system    kube-apiserver-control-plane                  1/1     Running   0          64d
kube-system    kube-controller-manager-control-plane         1/1     Running   0          64d
kube-system    kube-scheduler-control-plane                  1/1     Running   0          64d
2. Install Cert-Manager#
Note
If cert-manager is already installed in your cluster, you can skip this step.
The AMD GPU Operator requires cert-manager for TLS certificate management.
- Add the - cert-managerrepository:
helm repo add jetstack https://charts.jetstack.io --force-update
- Install - cert-manager:
helm install cert-manager jetstack/cert-manager \
  --namespace cert-manager \
  --create-namespace \
  --version v1.15.1 \
  --set crds.enabled=true
- Verify the installation: 
kubectl get pods -n cert-manager
Expected output:
NAME                                       READY   STATUS    RESTARTS   AGE
cert-manager-84489bc478-qjwmw             1/1     Running   0          2m
cert-manager-cainjector-7477d56b47-v8nq8  1/1     Running   0          2m
cert-manager-webhook-6d5cb854fc-h6vbk     1/1     Running   0          2m
Installing Operator#
1. Add the AMD Helm Repository#
helm repo add rocm https://rocm.github.io/gpu-operator
helm repo update
2. Install the Operator#
Basic installation:
helm install amd-gpu-operator rocm/gpu-operator-charts \
  --namespace kube-amd-gpu \
  --create-namespace \
  --version=v1.0.0
Note
Installation Options
- Skip NFD installation: - --set node-feature-discovery.enabled=false
- Skip KMM installation: - --set kmm.enabled=false
Warning
It is strongly recommended to use AMD-optimized KMM images included in the operator release.
3. Helm Chart Customization Parameters#
Installation with custom options:
- Prepare your custom configuration in a YAML file (e.g. - values.yaml), then use it with- helm installcommand to deploy your helm charts. An example values.yaml file can be found here for you to edit and use: here
helm install amd-gpu-operator rocm/gpu-operator-charts \
  --namespace kube-amd-gpu \
  --create-namespace \
  --version=v1.0.0 \
  -f values.yaml
The following parameters are able to be configued when using the Helm Chart. In order to view all available options, please refer to this section or run the command helm show values rocm/gpu-operator-charts.
| Key | Type | Default | Description | 
|---|---|---|---|
| controllerManager.manager.image.repository | string | 
 | AMD GPU operator controller manager image repository | 
| controllerManager.manager.image.tag | string | 
 | AMD GPU operator controller manager image tag | 
| controllerManager.manager.imagePullPolicy | string | 
 | Image pull policy for AMD GPU operator controller manager pod | 
| controllerManager.manager.imagePullSecrets | string | 
 | Image pull secret name for pulling AMD GPU operator controller manager image if registry needs credential to pull image | 
| controllerManager.nodeAffinity.nodeSelectorTerms | list | 
 | Node affinity selector terms config for the AMD GPU operator controller manager, set it to [] if you want to make affinity config empty | 
| controllerManager.nodeSelector | object | 
 | Node selector for AMD GPU operator controller manager deployment | 
| installdefaultNFDRule | bool | 
 | Set to true to install default NFD rule for detecting AMD GPU hardware based on pci vendor ID and device ID | 
| kmm.controller.manager.env.relatedImageBuild | string | 
 | KMM kaniko builder image for building driver image within cluster | 
| kmm.controller.manager.env.relatedImageBuildPullSecret | string | 
 | Image pull secret name for pulling KMM kaniko builder image if registry needs credential to pull image | 
| kmm.controller.manager.env.relatedImageSign | string | 
 | KMM signer image for signing driver image’s kernel module with given key pairs within cluster | 
| kmm.controller.manager.env.relatedImageSignPullSecret | string | 
 | Image pull secret name for pulling KMM signer image if registry needs credential to pull image | 
| kmm.controller.manager.env.relatedImageWorker | string | 
 | KMM worker image for loading / unloading driver kernel module on worker nodes | 
| kmm.controller.manager.env.relatedImageWorkerPullSecret | string | 
 | Image pull secret name for pulling KMM worker image if registry needs credential to pull image | 
| kmm.controller.manager.image.repository | string | 
 | KMM controller manager image repository | 
| kmm.controller.manager.image.tag | string | 
 | KMM controller manager image tag | 
| kmm.controller.manager.imagePullPolicy | string | 
 | Image pull policy for KMM controller manager pod | 
| kmm.controller.manager.imagePullSecrets | string | 
 | Image pull secret name for pulling KMM controller manager image if registry needs credential to pull image | 
| kmm.controller.nodeAffinity.nodeSelectorTerms | list | 
 | Node affinity selector terms config for the KMM controller manager deployment, set it to [] if you want to make affinity config empty | 
| kmm.controller.nodeSelector | object | 
 | Node selector for the KMM controller manager deployment | 
| kmm.enabled | bool | 
 | Set to true/false to enable/disable the installation of kernel module management (KMM) operator | 
| kmm.webhookServer.nodeAffinity.nodeSelectorTerms | list | 
 | Node affinity selector terms config for the KMM webhook deployment, set it to [] if you want to make affinity config empty | 
| kmm.webhookServer.nodeSelector | object | 
 | KMM webhook’s deployment node selector | 
| kmm.webhookServer.webhookServer.image.repository | string | 
 | KMM webhook image repository | 
| kmm.webhookServer.webhookServer.image.tag | string | 
 | KMM webhook image tag | 
| kmm.webhookServer.webhookServer.imagePullPolicy | string | 
 | Image pull policy for KMM webhook pod | 
| kmm.webhookServer.webhookServer.imagePullSecrets | string | 
 | Image pull secret name for pulling KMM webhook image if registry needs credential to pull image | 
| node-feature-discovery.enabled | bool | 
 | Set to true/false to enable/disable the installation of node feature discovery (NFD) operator | 
4. Verify the Operator Installation#
Check that all operator components are running:
kubectl get pods -n kube-amd-gpu
Expected output:
NAMESPACE      NAME                                                  READY   STATUS    RESTARTS   AGE
gpu-operator   amd-gpu-operator-controller-manager-6954b68958-ljthg  1/1     Running   0          2m
gpu-operator   amd-gpu-kmm-controller-59b85d48c4-f2hn4               1/1     Running   0          2m
gpu-operator   amd-gpu-kmm-webhook-server-685b9db458-t5qp6           1/1     Running   0          2m
gpu-operator   amd-gpu-nfd-gc-98776b45f-j2hvn                        1/1     Running   0          2m
gpu-operator   amd-gpu-nfd-master-9948b7b76-ncvnz                    1/1     Running   0          2m
gpu-operator   amd-gpu-nfd-worker-dhl7q                              1/1     Running   0          2m
Verify that nodes with AMD GPU hardware are properly labeled:
kubectl get nodes -L feature.node.kubernetes.io/amd-gpu
Install Custom Resource#
After the installation of AMD GPU Operator, you need to create the DeviceConfig custom resource in order to trigger the operator start to work. By preparing the DeviceConfig in the YAML file, you can create the resouce by running kubectl apply -f deviceconfigs.yaml. For custom resource definition and more detailed information, please refer to Custom Resource Installation Guide. Here are some examples for common deployment scenarios.
Inbox or Pre-Installed AMD GPU Drivers#
In order to directly use inbox or pre-installed AMD GPU drivers on the worker node, the operator’s driver installation need to be skipped, thus spec.driver.enable=false need to be specified. By deploying the following custom resource, the operator will directly deploy device plugin, node labeller and metrics exporter on all selected AMD GPU worker nodes.
apiVersion: amd.com/v1alpha1
kind: DeviceConfig
metadata:
  name: test-deviceconfig
  # use the namespace where AMD GPU Operator is running
  namespace: kube-amd-gpu
spec:
  driver:
    # disable the installation of our-of-tree amdgpu kernel module
    enable: false
  devicePlugin:
    devicePluginImage: rocm/k8s-device-plugin:latest
    nodeLabellerImage: rocm/k8s-device-plugin:labeller-latest
        
  # Specify the metrics exporter config
  metricsExporter:
     enable: true
     serviceType: "NodePort"
     # Node port for metrics exporter service, metrics endpoint $node-ip:$nodePort
     nodePort: 32500
     image: docker.io/amd/exporter:v1
  # Specifythe node to be managed by this DeviceConfig Custom Resource
  selector:
    feature.node.kubernetes.io/amd-gpu: "true"
Install out-of-tree AMD GPU Drivers with Operator#
If you want to use the operator to install out-of-tree version AMD GPU drivers (e.g. install specific ROCm verison driver), you need to configure custom resource to trigger the operator to install the specific ROCm version AMD GPU driver. By creating the following custom resource with spec.driver.enable=true, the operator will call KMM operator to trigger the driver installation on the selected worker nodes.
Note
In order to install the out-of-tree version AMD GPU drivers, blacklisting the inbox or pre-installed AMD GPU driver is required, AMD GPU operator can help you push the blacklist option to worker nodes. Please set spec.driver.blacklist=true, create the custom resource and reboot the selected worker nodes to apply the new blacklist config. If amdgpu remains loaded after reboot and worker nodes keep using inbox / pre-installed driver, run sudo update-initramfs -u to update the initial ramdisk with the new modprobe configuration.
apiVersion: amd.com/v1alpha1
kind: DeviceConfig
metadata:
  name: test-deviceconfig
  # use the namespace where AMD GPU Operator is running
  namespace: kube-amd-gpu
spec:
  driver:
    # enable operator to install out-of-tree amdgpu kernel module
    enable: true
    # blacklist is required for installing out-of-tree amdgpu kernel module
    blacklist: true
    # Specify your repository to host driver image
    # DO NOT include the image tag as AMD GPU Operator will automatically manage the image tag for you
    image: docker.io/username/repo
    # (Optional) Specify the credential for your private registry if it requires credential to get pull/push access
    # you can create the docker-registry type secret by running command like:
    # kubectl create secret docker-registry mysecret -n kmm-namespace --docker-username=xxx --docker-password=xxx
    # Make sure you created the secret within the namespace that KMM operator is running
    imageRegistrySecret:
      name: mysecret
    # Specify the driver version by using ROCm version
    version: "6.2.1"
  devicePlugin:
    devicePluginImage: rocm/k8s-device-plugin:latest
    nodeLabellerImage: rocm/k8s-device-plugin:labeller-latest
        
  # Specify the metrics exporter config
  metricsExporter:
     enable: true
     serviceType: "NodePort"
     # Node port for metrics exporter service, metrics endpoint $node-ip:$nodePort
     nodePort: 32500
     image: docker.io/amd/exporter:v1
  # Specifythe node to be managed by this DeviceConfig Custom Resource
  selector:
    feature.node.kubernetes.io/amd-gpu: "true"
Post-Installation Verification#
Verify driver installation status:
kubectl get deviceconfigs -n kube-amd-gpu -oyaml
Verify the AMD GPU allocatable resource:
kubectl get nodes -oyaml | grep "amd.com/gpu"
Verify the AMD GPU node label:
kubectl get nodes -oyaml | grep  "amd.com"
Test GPU Workload Deployment#
Create a simple test pod:
apiVersion: v1
kind: Pod
metadata:
 name: amd-smi
spec:
 containers:
 - image: docker.io/rocm/pytorch:latest
   name: amd-smi
   command: ["/bin/bash"]
   args: ["-c","amd-smi version && amd-smi monitor -ptum"]
   resources:
    limits:
      amd.com/gpu: 1
    requests:
      amd.com/gpu: 1
 restartPolicy: Never
- Create the pod: 
kubectl create -f amd-smi.yaml
- Check the logs and verify the output - amd-smireflects the expected ROCm version and GPU presence:
kubectl logs amd-smi
AMDSMI Tool: 24.6.2+2b02a07 | AMDSMI Library version: 24.6.2.0 | ROCm version: 6.2.2
GPU  POWER  GPU_TEMP  MEM_TEMP  GFX_UTIL  GFX_CLOCK  MEM_UTIL  MEM_CLOCK
  0  126 W     40 °C     32 °C       1 %    182 MHz       0 %    900 MHz
- Delete the pod: 
kubectl delete -f amd-smi.yaml
Troubleshooting#
If you encounter issues during installation:
- Check operator logs: 
kubectl logs -n kube-amd-gpu \
  deployment/amd-gpu-operator-controller-manager
- Check KMM status: 
kubectl get modules -n kube-amd-gpu
- Check NFD status: 
kubectl get nodefeatures -n kube-amd-gpu
For more detailed troubleshooting steps, see our Troubleshooting Guide.
Uninstallation#
Please refer to the Uninstallation document for uninstalling related resources.