Skip to main content
Skip table of contents

KIB for GPU

Create a GPU supported OS image using Konvoy Image Builder

Using the Konvoy Image Builder, you can build an image that has support to use NVIDIA GPU hardware to support GPU workloads.

If the NVIDIA runfile installer has not been downloaded, then retrieve and install the download first by running the following command. The first line in the command below downloads and installs the runfile and the second line places it in the artifacts directory.

CODE
curl -O https://download.nvidia.com/XFree86/Linux-x86_64/470.82.01/NVIDIA-Linux-x86_64-470.82.01.run
mv NVIDIA-Linux-x86_64-470.82.01.run artifacts

DKP supported NVIDIA driver version is 470.x.

To build an image for use on GPU enabled hardware, perform the following steps.

  1. In your overrides/nvidia.yaml file, add the following to enable GPU builds. You can also access and use the overrides repo or in the documentation under NVIDIA GPU Overrides or Offline NVIDIA Override.

    CODE
    gpu:
      type:
        - nvidia

  2. Build your image using the following Konvoy Image Builder commands, making sure to include the flag --instance-type that specifies an AWS instance that has an available GPU:
    AWS Example:

    CODE
    konvoy-image build --region us-west-2 --instance-type=p2.xlarge --source-ami=ami-12345abcdef images/ami/centos-7.yaml --overrides overrides/nvidia.yaml

    By default, your image builds in the us-west-2 region. To specify another region, set the --region flag:

    CODE
    konvoy-image build --region us-east-1 --instance-type=p2.xlarge --overrides override-source-ami.yaml images/ami/<Your OS>.yaml

     

NOTE: Ensure that an AMI file is available for your OS selection.

When the command is complete the ami id is printed and written to ./manifest.json.

To use the built ami with Konvoy, specify it with the --ami flag when calling cluster create.

CODE
dkp create cluster aws --cluster-name=$(whoami)-aws-cluster --region us-west-2 --ami <ami>

For GPU Steps in Pre-provisioned section of the documentation to use the overrides/nvidia.yaml.

Verification

To verify that the NVIDIA driver is working, connect to the node and execute this command:

CODE
nvidia-smi

When drivers are successfully installed, the display looks like the following:

CODE
Fri Jun 11 09:05:31 2021
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 460.73.01    Driver Version: 460.73.01    CUDA Version: 11.2     |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|                               |                      |               MIG M. |
|===============================+======================+======================|
|   0  Tesla K80           Off  | 00000000:00:1E.0 Off |                    0 |
| N/A   35C    P0    73W / 149W |      0MiB / 11441MiB |     99%      Default |
|                               |                      |                  N/A |
+-------------------------------+----------------------+----------------------+

+-----------------------------------------------------------------------------+
| Processes:                                                                  |
|  GPU   GI   CI        PID   Type   Process name                  GPU Memory |
|        ID   ID                                                   Usage      |
|=============================================================================|
|  No running processes found                                                 |
+-----------------------------------------------------------------------------+

Additional helpful information can be found in the NVIDIA Device Plug-in for Kubernetes instructions and the Installation Guide of Supported Platforms.

See also: NVIDIA documentation

JavaScript errors detected

Please note, these errors can depend on your browser setup.

If this problem persists, please contact our support.