Pull the ROCm vLLM Docker image.
docker pull rocm/vllm:rocm7.12.0_gfx950-dcgpu_ubuntu24.04_py3.12_pytorch_2.9.1_vllm_0.16.0
Start the Docker container.
docker run -it --rm \
--network=host \
--group-add=video \
--ipc=host \
--cap-add=SYS_PTRACE \
--security-opt seccomp=unconfined \
--device /dev/kfd \
--device /dev/dri \
-v <path/to/your/models>:/app/models \
-e HF_HOME="/app/models" \
rocm/vllm:rocm7.12.0_gfx950-dcgpu_ubuntu24.04_py3.12_pytorch_2.9.1_vllm_0.16.0
Pull the ROCm vLLM Docker image.
docker pull rocm/vllm:rocm7.12.0_gfx94X-dcgpu_ubuntu24.04_py3.12_pytorch_2.9.1_vllm_0.16.0
Start the Docker container.
docker run -it --rm \
--network=host \
--group-add=video \
--ipc=host \
--cap-add=SYS_PTRACE \
--security-opt seccomp=unconfined \
--device /dev/kfd \
--device /dev/dri \
-v <path/to/your/models>:/app/models \
-e HF_HOME="/app/models" \
rocm/vllm:rocm7.12.0_gfx94X-dcgpu_ubuntu24.04_py3.12_pytorch_2.9.1_vllm_0.16.0
Pull the ROCm vLLM Docker image.
docker pull rocm/vllm:rocm7.12.0_gfx120X-all_ubuntu24.04_py3.12_pytorch_2.9.1_vllm_0.16.0
Start the Docker container.
docker run -it --rm \
--network=host \
--group-add=video \
--ipc=host \
--cap-add=SYS_PTRACE \
--security-opt seccomp=unconfined \
--device /dev/kfd \
--device /dev/dri \
-v <path/to/your/models>:/app/models \
-e HF_HOME="/app/models" \
rocm/vllm:rocm7.12.0_gfx120X-all_ubuntu24.04_py3.12_pytorch_2.9.1_vllm_0.16.0
Pull the ROCm vLLM Docker image.
docker pull rocm/vllm:rocm7.12.0_gfx1151_ubuntu24.04_py3.12_pytorch_2.9.1_vllm_0.16.0
Start the Docker container.
docker run -it --rm \
--network=host \
--group-add=video \
--ipc=host \
--cap-add=SYS_PTRACE \
--security-opt seccomp=unconfined \
--device /dev/kfd \
--device /dev/dri \
-v <path/to/your/models>:/app/models \
-e HF_HOME="/app/models" \
rocm/vllm:rocm7.12.0_gfx1151_ubuntu24.04_py3.12_pytorch_2.9.1_vllm_0.16.0
Set up your Python virtual environment. If you already have a successful
ROCm 7.12.0 installation using pip, skip
this step.
For example, run the following command to create a virtual environment:
Activate your Python virtual environment. For example:
source .venv/bin/activate
Install ROCm 7.12.0 and PyTorch using pip. See Install PyTorch for details.
Install the appropriate vLLM 0.16.0 build for your GFX architecture from the ROCm package repository.
python -m pip install \
--extra-index-url https://rocm.frameworks.amd.com/whl/gfx950-dcgpu/ \
"vllm==0.16.1.dev10+g11515110f.d20260324.rocm712"
python -m pip install \
--extra-index-url https://rocm.frameworks.amd.com/whl/gfx94X-dcgpu/ \
"vllm==0.16.1.dev10+g11515110f.d20260324.rocm712"
python -m pip install \
--extra-index-url https://rocm.frameworks.amd.com/whl/gfx120X-all/ \
"vllm==0.16.1.dev10+g11515110f.d20260323.rocm712"
python -m pip install \
--extra-index-url https://rocm.frameworks.amd.com/whl/gfx1151/ \
"vllm==0.16.1.dev10+g11515110f.d20260323.rocm712"
After setting up your environment, follow the vLLM 0.16.0 usage documentation to get started: Using
vLLM.