If your business involves scenarios such as deep learning and high-performance computing, you can use TKE to support the GPU feature, which can help you quickly use a GPU container.
There are many ways to create a GPU CVM instance:
For more information, see Adding a Node. When creating a GPU, you should pay special attention to the following parameters:
On the Select Model page, set Model in Node Model to GPU.
After setting the model, you can select the GPU driver version, CUDA version, and cuDNN version as needed.
- If you select Automatically install GPU driver on the backend, it will be installed automatically during system start, taking 15–25 minutes.
- The supported driver versions are determined by both the operating system and the GPU model.
- If you do not select Automatically install GPU driver on the backend, the GPU driver will be installed by default for some operating systems of earlier versions to ensure the normal use. The complete default driver version information is as shown below:
Operating System Default Driver Version Installed CentOS 7.6, Ubuntu 18, Tencent Linux2.4 450 CentOS 7.2 (not recommended) 384.111 Ubuntu 16 (not recommended) 410.79
With multi-instance GPU (MIG) enabled, an A100 GPU will be divided into seven separate GPU instances to help you improve the GPU utilization when multiple jobs are running. For more information, see NVIDIA Multi-Instance GPU User Guide.
To use the MIG feature, make sure the following conditions are met:
The GPU model is GT4. You have selected **Automatically install GPU driver on the backend** in the console and configured the GPU, CUDA, and cuDNN versions.
For detailed directions, see Adding a Node. When adding a node, you should pay attention to the following:
Was this page helpful?