[Documentation][AMD] Add information about prebuilt ROCm vLLM docker for perf validation purpose (#12281)

Signed-off-by: Hongxia Yang <hongxyan@amd.com>
This commit is contained in:
Hongxia Yang 2025-01-21 18:49:22 -05:00 committed by GitHub
parent 69196a9bc7
commit 09ccc9c8f7
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -13,6 +13,14 @@ vLLM supports AMD GPUs with ROCm 6.2.
Currently, there are no pre-built ROCm wheels.
However, the [AMD Infinity hub for vLLM](https://hub.docker.com/r/rocm/vllm/tags) offers a prebuilt, optimized
docker image designed for validating inference performance on the AMD Instinct™ MI300X accelerator.
```{tip}
Please check [LLM inference performance validation on AMD Instinct MI300X](https://rocm.docs.amd.com/en/latest/how-to/performance-validation/mi300x/vllm-benchmark.html)
for instructions on how to use this prebuilt docker image.
```
### Build wheel from source
0. Install prerequisites (skip if you are already in an environment/docker with the following installed):