This page contains proposed changes for a future release of ROCm. Read the latest Linux release of ROCm documentation for your production environments.

Installing PyTorch for ROCm#

Applies to Linux

2024-09-30

15 min read time

PyTorch is an open-source tensor library designed for deep learning. PyTorch on ROCm provides mixed-precision and large-scale training using our MIOpen and RCCL libraries.

To install PyTorch for ROCm, you have the following options:

For hardware, software, and third-party framework compatibility between ROCm and PyTorch, refer to:

Using a Docker image with PyTorch pre-installed#

  1. Download the latest public PyTorch Docker image.

    docker pull rocm/pytorch:latest
    

    You can also download a specific and supported configuration with different user-space ROCm versions, PyTorch versions, and operating systems by filtering through the tags.

    Important

    As of ROCm 6.2.1, rocm/pytorch:latest points to a docker image with the latest ROCm tested release version of PyTorch (for example, version 2.3), similar to rocm/pytorch:latest-release tag. Before ROCm 6.2.1, rocm/pytorch:latest pointed to a development version of PyTorch, which didn’t correspond to a specific PyTorch release.

    Description

    6.2.1 and later

    6.2.0 and earlier

    Latest PyTorch tested release

    rocm/pytorch:latest
    rocm/pytorch:latest-release

    rocm/pytorch:latest-release

    Latest PyTorch dev version

    rocm/pytorch:latest-internal

    rocm/pytorch:latest

  2. Start a Docker container using the image.

    docker run -it --cap-add=SYS_PTRACE --security-opt seccomp=unconfined \
    --device=/dev/kfd --device=/dev/dri --group-add video \
    --ipc=host --shm-size 8G rocm/pytorch:latest
    

    Note

    This will automatically download the image if it does not exist on the host. You can also pass the -v argument to mount any data directories from the host onto the container.

Refer to the Docker image support matrix for a list of prebuilt Docker image tags for PyTorch with ROCm.

Using a wheels package#

PyTorch supports the ROCm platform by providing tested wheels packages. To access this feature, go to pytorch.org/get-started/locally/. For the correct wheels command, you must select ‘Linux’, ‘Python’, ‘pip’, and ‘ROCm’ in the matrix.

Note

The available ROCm release varies between the ‘PyTorch Build’ of Stable or Nightly. More recent releases are generally available through the Nightly builds.

  1. Choose one of the following three options:

    Option 1:

    1. Download a base Docker image with the correct user-space ROCm version.

      Base OS

      Docker Image

      Ubuntu 22.04

      rocm/dev-ubuntu-22.04

      Ubuntu 24.04

      rocm/dev-ubuntu-24.04

    2. Pull the selected image.

    docker pull rocm/dev-ubuntu-22.04:latest
    
    1. Start a Docker container using the downloaded image.

    docker run -it --device=/dev/kfd --device=/dev/dri --group-add video rocm/dev-ubuntu-22.04:latest
    

    Option 2:

    Select a base OS Docker image (Check System requirements (Linux))

    Pull selected base OS image (Ubuntu 22.04 for example)

    docker pull ubuntu:22.04
    

    Start a Docker container using the downloaded image

    docker run -it --device=/dev/kfd --device=/dev/dri --group-add video ubuntu:22.04
    

    Install ROCm using the directions in the ROCm installation overview section.

    Option 3:

    Install on bare metal. Check System requirements (Linux) and install ROCm using the directions in the ROCm installation overview section.

  2. Install the required dependencies for the wheels package.

    sudo apt update
    sudo apt install libjpeg-dev python3-dev python3-pip
    pip3 install wheel setuptools
    
  3. Install torch, torchvision, and torchaudio, as specified in the installation matrix.

    Note

    The following command uses the ROCm 6.2 PyTorch wheel. If you want a different version of ROCm, modify the command accordingly.

    pip3 install --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/rocm6.2/
    
  4. (Optional) Use MIOpen kdb files with ROCm PyTorch wheels.

    PyTorch uses MIOpen for machine learning primitives, which are compiled into kernels at runtime. Runtime compilation causes a small warm-up phase when starting PyTorch, and MIOpen kdb files contain precompiled kernels that can speed up application warm-up phases.

    MIOpen kdb files can be used with ROCm PyTorch wheels. However, the kdb files need to be placed in a specific location with respect to the PyTorch installation path. A helper script simplifies this task by taking the ROCm version and GPU architecture as inputs. This works for Ubuntu and CentOS.

    You can download the helper script here: install_kdb_files_for_pytorch_wheels.sh, or use:

    wget https://raw.githubusercontent.com/wiki/ROCm/pytorch/files/install_kdb_files_for_pytorch_wheels.sh
    

    After installing ROCm PyTorch wheels, run the following code:

    #Optional; replace 'gfx90a' with your architecture and 6.1 with your preferred ROCm version
    export GFX_ARCH=gfx90a
    
    #Optional
    export ROCM_VERSION=6.1
    
    ./install_kdb_files_for_pytorch_wheels.sh
    

Using the PyTorch ROCm base Docker image#

The pre-built base Docker image has all dependencies installed, including:

  • ROCm

  • torchvision

  • Conda packages

  • The compiler toolchain

Additionally, a particular environment flag (BUILD_ENVIRONMENT) is set, which is used by the build scripts to determine the configuration of the build environment.

  1. Download the Docker image. This is the base image, which does not contain PyTorch.

    docker pull rocm/pytorch:latest-base
    
  2. Start a Docker container using the downloaded image.

    docker run -it --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --device=/dev/kfd --device=/dev/dri --group-add video --ipc=host --shm-size 8G rocm/pytorch:latest-base
    

    You can also pass the -v argument to mount any data directories from the host onto the container.

Inside the docker container, run the following steps:

  1. Clone the PyTorch repository.

    cd ~
    git clone https://github.com/pytorch/pytorch.git
    cd pytorch
    git submodule update --init --recursive
    
  2. Set ROCm architecture (optional).

    Note

    By default in the rocm/pytorch:latest-base image, PyTorch builds simultaneously for the following architectures:

    • gfx900

    • gfx906

    • gfx908

    • gfx90a

    • gfx1030

    • gfx1100

    • gfx1101

    • gfx940

    • gfx941

    • gfx942

    If you want to compile only for your microarchitecture (uarch), run:

    export PYTORCH_ROCM_ARCH=<uarch>
    

    Where <uarch> is the architecture reported by the rocminfo command.

    To find your uarch, run:

    rocminfo | grep gfx
    
  3. Build PyTorch.

    .ci/pytorch/build.sh
    

    This converts PyTorch sources for HIP compatibility and builds the PyTorch framework.

    To check if your build is successful, run:

    echo $? # should return 0 if success
    

Using the PyTorch upstream Docker file#

If you don’t want to use a prebuilt base Docker image, you can build a custom base Docker image using scripts from the PyTorch repository. This uses a standard Docker image from operating system maintainers and installs all the required dependencies, including:

  • ROCm

  • torchvision

  • Conda packages

  • The compiler toolchain

  1. Clone the PyTorch repository.

    cd ~
    git clone https://github.com/pytorch/pytorch.git
    cd pytorch
    git submodule update --init --recursive
    
  2. Build the PyTorch Docker image.

    cd .ci/docker
    ./build.sh pytorch-linux-<os-version>-rocm<rocm-version>-py<python-version> -t rocm/pytorch:build_from_dockerfile
    

    Where:

    • <os-version> = ubuntu20.04 (or focal), ubuntu22.04 (or jammy), centos7.5, or centos9

    • <rocm-version> = 6.0, 6.1, or 6.2

    • <python-version> = 3.8 - 3.11

    To verify that your image was successfully created, run:

    docker image ls rocm/pytorch:build_from_dockerfile
    

    If successful, the output looks like this:

    REPOSITORY    TAG                       IMAGE ID         CREATED           SIZE
    rocm/pytorch  build_from_dockerfile     17071499be47     2 minutes ago     32.8GB
    
  3. Start a Docker container using the image with the mounted PyTorch folder.

    docker run -it --cap-add=SYS_PTRACE --security-opt seccomp=unconfined \
    --user root --device=/dev/kfd --device=/dev/dri \
    --group-add video --ipc=host --shm-size 8G \
    -v ~/pytorch:/pytorch rocm/pytorch:build_from_dockerfile
    

    You can also pass the -v argument to mount any data directories from the host onto the container.

  4. Go to the PyTorch directory.

    cd /pytorch
    
  5. Set ROCm architecture.

    To determine your AMD architecture, run:

    rocminfo | grep gfx
    

    The result looks like this (for gfx1030 architecture):

    Name:                    gfx1030
    Name:                    amdgcn-amd-amdhsa--gfx1030
    

    Set the PYTORCH_ROCM_ARCH environment variable to specify the architectures you want to build PyTorch for.

    export PYTORCH_ROCM_ARCH=<uarch>
    

    where <uarch> is the architecture reported by the rocminfo command.

  6. Build PyTorch.

    .ci/pytorch/build.sh
    

    This converts PyTorch sources for HIP compatibility and builds the PyTorch framework.

    To check if your build is successful, run:

    echo $? # should return 0 if success
    

Testing the PyTorch installation#

You can use PyTorch unit tests to validate your PyTorch installation. If you used a prebuilt PyTorch Docker image from AMD ROCm Docker Hub or installed an official wheels package, validation tests are not necessary.

If you want to manually run unit tests to validate your PyTorch installation fully, follow these steps:

  1. Import the torch package in Python to test if PyTorch is installed and accessible.

    Note

    Do not run the following command from the PyTorch home directory.

    python3 -c 'import torch' 2> /dev/null && echo 'Success' || echo 'Failure'
    
  2. Check if the GPU is accessible from PyTorch. In the PyTorch framework, torch.cuda is a generic way to access the GPU. This can only access an AMD GPU if one is available.

    python3 -c 'import torch; print(torch.cuda.is_available())'
    
  3. Run unit tests to validate the PyTorch installation fully.

    Note

    You must run the following command from the PyTorch home directory.

    PYTORCH_TEST_WITH_ROCM=1 python3 test/run_test.py --verbose \
    --include test_nn test_torch test_cuda test_ops \
    test_unary_ufuncs test_binary_ufuncs test_autograd
    

    This command ensures that the required environment variable is set to skip certain unit tests for ROCm. This also applies to wheel installs in a non-controlled environment.

    Note

    Make sure your PyTorch source code corresponds to the PyTorch wheel or the installation in the Docker image. Incompatible PyTorch source code can give errors when running unit tests.

    Some tests may be skipped, as appropriate, based on your system configuration. ROCm doesn’t support all PyTorch features; tests that evaluate unsupported features are skipped. Other tests might be skipped, depending on the host or GPU memory and the number of available GPUs.

    If the compilation and installation are correct, all tests will pass.

  4. (Optional) Run individual unit tests.

    PYTORCH_TEST_WITH_ROCM=1 python3 test/test_nn.py --verbose
    

    You can replace test_nn.py with any other test set.

Running a basic PyTorch example#

The PyTorch examples repository provides basic examples that exercise the functionality of your framework.

Two of our favorite testing databases are:

  • MNIST (Modified National Institute of Standards and Technology): A database of handwritten digits that can be used to train a Convolutional Neural Network for handwriting recognition.

  • ImageNet: A database of images that can be used to train a network for visual object recognition.

MNIST PyTorch example#

  1. Clone the PyTorch examples repository.

    git clone https://github.com/pytorch/examples.git
    
  2. Go to the MNIST example folder.

    cd examples/mnist
    
  3. Follow the instructions in the README.md file in this folder to install the requirements. Then run:

    python3 main.py
    

    This generates the following output:

    ...
    Train Epoch: 14 [58240/60000 (97%)]     Loss: 0.010128
    Train Epoch: 14 [58880/60000 (98%)]     Loss: 0.001348
    Train Epoch: 14 [59520/60000 (99%)]     Loss: 0.005261
    
    Test set: Average loss: 0.0252, Accuracy: 9921/10000 (99%)
    

ImageNet PyTorch example#

  1. Clone the PyTorch examples repository (if you didn’t already do this in the preceding MNIST example).

    git clone https://github.com/pytorch/examples.git
    
  2. Go to the ImageNet example folder.

    cd examples/imagenet
    
  3. Follow the instructions in the README.md file in this folder to install the Requirements. Then run:

    python3 main.py
    

Troubleshooting#

  • What to do if you get the following error when trying to run PyTorch:

    hipErrorNoBinaryForGPU: Unable to find code object for all current devices!
    

    The error denotes that the installation of PyTorch and/or other dependencies or libraries do not support the current GPU. To workaround this issue, use the following steps:

    1. Confirm that the hardware supports the ROCm stack. Refer to System requirements (Linux) and System requirements (Windows).

    2. Determine the gfx target.

      rocminfo | grep gfx
      
    3. Check if PyTorch is compiled with the correct gfx target.

      TORCHDIR=$( dirname $( python3 -c 'import torch; print(torch.__file__)' ) )
      roc-obj-ls -v $TORCHDIR/lib/libtorch_hip.so # check for gfx target
      

      Note

      Recompile PyTorch with the right gfx target if compiling from the source if the hardware is not supported.

  • What if you are unable to access Docker or GPU in user accounts?

    Ensure that the user is added to docker, video, and render Linux groups as described in Setting permissions for groups.

  • Can you install PyTorch directly on bare metal?

    Bare-metal installation of PyTorch is supported through wheels. For more information, see Using a wheels package.

  • How do you profile PyTorch workloads?

    Use the PyTorch Profiler as described in PyTorch Profiler to profile GPU kernels on ROCm.