Optimize GPU settings - Amazon Elastic Compute Cloud
Services or capabilities described in Amazon Web Services documentation might vary by Region. To see the differences applicable to the China Regions, see Getting Started with Amazon Web Services in China (PDF).

Optimize GPU settings

There are several GPU setting optimizations that you can perform to achieve the best performance on NVIDIA GPU instances. With some of these instance types, the NVIDIA driver uses an autoboost feature, which varies the GPU clock speeds. By disabling autoboost and setting the GPU clock speeds to their maximum frequency, you can consistently achieve the maximum performance with your GPU instances. The following procedure helps you to configure the GPU settings to be persistent, disable the autoboost feature if needed, and set the GPU clock speeds to their maximum frequency.

The following steps are for optimizing GPU settings on a Linux instance. For Windows instances, see Optimize GPU settings in the Amazon EC2 User Guide for Windows Instances.

To optimize GPU settings
  1. Configure the GPU settings to be persistent. This command can take several minutes to run.

    [ec2-user ~]$ sudo nvidia-persistenced
  2. [G2, G3, and P2 instances only] Disable the autoboost feature for all GPUs on the instance.

    [ec2-user ~]$ sudo nvidia-smi --auto-boost-default=0
  3. Set all GPU clock speeds to their maximum frequency. Use the memory and graphics clock speeds specified in the following commands.

    Some versions of the NVIDIA driver do not support setting the application clock speed, and display the error "Setting applications clocks is not supported for GPU...", which you can ignore.

    • G3 instances:

      [ec2-user ~]$ sudo nvidia-smi -ac 2505,1177
    • G4dn instances:

      [ec2-user ~]$ sudo nvidia-smi -ac 5001,1590
    • G5 instances:

      [ec2-user ~]$ sudo nvidia-smi -ac 6250,1710
    • P2 instances:

      [ec2-user ~]$ sudo nvidia-smi -ac 2505,875
    • P3 and P3dn instances:

      [ec2-user ~]$ sudo nvidia-smi -ac 877,1530
    • P4d instances:

      [ec2-user ~]$ sudo nvidia-smi -ac 1215,1410
    • P4de instances:

      [ec2-user ~]$ sudo nvidia-smi -ac 1593,1410
    • P5 instances:

      [ec2-user ~]$ sudo nvidia-smi -ac 2619,1980