FlashInfer on ROCm installation#
2026-03-03
3 min read time
System requirements#
To use FlashInfer 0.2.5, you need the following prerequisites:
Install FlashInfer#
To install FlashInfer on ROCm, you have the following options:
Use a prebuilt Docker image with FlashInfer pre-installed#
Docker is the recommended method to set up a FlashInfer environment, as it avoids potential installation issues. The tested, prebuilt image includes FlashInfer, PyTorch, ROCm, and other dependencies.
Pull the Docker image.
docker pull rocm/flashinfer:flashinfer-0.2.5.amd2_rocm7.1.1_ubuntu24.04_py3.12_pytorch2.8
Start a Docker container using the image.
docker run -it --rm \ --privileged \ --network=host --device=/dev/kfd \ --device=/dev/dri --group-add video \ --name=my_flashinfer --cap-add=SYS_PTRACE \ --security-opt seccomp=unconfined \ --ipc=host --shm-size 16G \ rocm/flashinfer:flashinfer-0.2.5.amd2_rocm7.1.1_ubuntu24.04_py3.12_pytorch2.8
The above step will create a Docker container with FlashInfer pre-installed. During this process, the Dockerfile will have a pre-installed and configured micromamba environment with FlashInfer available inside. To use FlashInfer, activate the micromamba environment.
micromamba activate base
Install FlashInfer using pip#
Use a base PyTorch Docker image and follow these steps to install FlashInfer using pip.
Pull the base ROCm PyTorch Docker image.
docker pull rocm/pytorch:rocm7.1.1_ubuntu24.04_py3.12_pytorch_release_2.8.0
Start a Docker container using the image.
docker run -it --rm \ --privileged \ --network=host --device=/dev/kfd \ --device=/dev/dri --group-add video \ --cap-add=SYS_PTRACE --security-opt seccomp=unconfined \ --ipc=host --shm-size 128G \ rocm/pytorch:rocm7.1.1_ubuntu24.04_py3.12_pytorch_release_2.8.0
After setting up the container, install FlashInfer from the AMD-hosted PyPI repository.
pip install amd-flashinfer --index-url https://pypi.amd.com/rocm-7.1.1/simple
Build from source#
FlashInfer on ROCm can be run directly by setting up a Docker container from scratch. A Dockerfile is provided in the ROCm/flashinfer repository to help you get started.
Clone the ROCm/flashinfer repository.
git clone https://github.com/ROCm/flashinfer.git
Enter the directory and build the Dockerfile to create a Docker image.
cd flashinfer docker build \ --build-arg USERNAME=$USER \ --build-arg USER_UID=$(id -u) \ --build-arg USER_GID=$(id -g) \ -f .devcontainer/rocm/Dockerfile \ -t rocm-flashinfer-dev .
Start a Docker container using the image.
docker run -it --rm \ --privileged --network=host --device=/dev/kfd \ --device=/dev/dri --group-add video \ --cap-add=SYS_PTRACE \ --security-opt seccomp=unconfined \ --ipc=host --shm-size 16G \ -v $PWD:/workspace \ rocm-flashinfer-dev
Once you are inside the container, the micromamba environment is automatically activated. You can now install FlashInfer inside it.
cd /workspace FLASHINFER_HIP_ARCHITECTURES=gfx942 python -m pip wheel . --wheel-dir=./dist/ --no-deps --no-build-isolation -v cd dist && pip install amd_flashinfer-*.whl
Test the FlashInfer installation#
Once you have installed FlashInfer, run the following command. If it outputs 0.2.5+amd.2, then FlashInfer is installed correctly. You can now use FlashInfer in your projects.
python -c "import flashinfer; print(flashinfer.__version__)"