KIB for GPU
Create a GPU supported OS image using Konvoy Image Builder
Using the Konvoy Image Builder, you can build an image that has support to use NVIDIA GPU hardware to support GPU workloads.
If the NVIDIA runfile installer has not been downloaded, then retrieve and install the download first by running the following command. The first line in the command below downloads and installs the runfile and the second line places it in the artifacts directory.
curl -O https://download.nvidia.com/XFree86/Linux-x86_64/470.82.01/NVIDIA-Linux-x86_64-470.82.01.run
mv NVIDIA-Linux-x86_64-470.82.01.run artifacts
DKP supported NVIDIA driver version is 470.x.
To build an image for use on GPU enabled hardware, perform the following steps.
In your
overrides/nvidia.yamlfile, add the following to enable GPU builds. You can also access and use the overrides repo or in the documentation under NVIDIA GPU Overrides or Offline NVIDIA Override.CODEgpu: type: - nvidiaBuild your image using the following Konvoy Image Builder commands, making sure to include the flag
--instance-typethat specifies an AWS instance that has an available GPU:
AWS Example:CODEkonvoy-image build --region us-west-2 --instance-type=p2.xlarge --source-ami=ami-12345abcdef images/ami/centos-7.yaml --overrides overrides/nvidia.yamlBy default, your image builds in the
us-west-2region. To specify another region, set the--regionflag:CODEkonvoy-image build --region us-east-1 --instance-type=p2.xlarge --overrides override-source-ami.yaml images/ami/<Your OS>.yaml
NOTE: Ensure that an AMI file is available for your OS selection.
When the command is complete the ami id is printed and written to ./manifest.json.
To use the built ami with Konvoy, specify it with the --ami flag when calling cluster create.
dkp create cluster aws --cluster-name=$(whoami)-aws-cluster --region us-west-2 --ami <ami>
For GPU Steps in Pre-provisioned section of the documentation to use the overrides/nvidia.yaml.
Verification
To verify that the NVIDIA driver is working, connect to the node and execute this command:
nvidia-smi
When drivers are successfully installed, the display looks like the following:
Fri Jun 11 09:05:31 2021
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 460.73.01    Driver Version: 460.73.01    CUDA Version: 11.2     |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|                               |                      |               MIG M. |
|===============================+======================+======================|
|   0  Tesla K80           Off  | 00000000:00:1E.0 Off |                    0 |
| N/A   35C    P0    73W / 149W |      0MiB / 11441MiB |     99%      Default |
|                               |                      |                  N/A |
+-------------------------------+----------------------+----------------------+
+-----------------------------------------------------------------------------+
| Processes:                                                                  |
|  GPU   GI   CI        PID   Type   Process name                  GPU Memory |
|        ID   ID                                                   Usage      |
|=============================================================================|
|  No running processes found                                                 |
+-----------------------------------------------------------------------------+
Additional helpful information can be found in the NVIDIA Device Plug-in for Kubernetes instructions and the Installation Guide of Supported Platforms.
See also: NVIDIA documentation