vLLM inference performance testing version history

vLLM inference performance testing version history#

2025-08-14

3 min read time

Applies to Linux

This table lists previous versions of the ROCm vLLM inference Docker image for inference performance testing. For detailed information about available models for benchmarking, see the version-specific documentation. You can find tagged previous releases of the ROCm/vllm Docker image on Docker Hub.

Docker image tag

Components

Resources

rocm/vllm:rocm6.4.1_vllm_0.10.0_20250812 (latest)

  • ROCm 6.4.1

  • vLLM 0.10.0

  • PyTorch 2.7.0

rocm/vllm:rocm6.4.1_vllm_0.9.1_20250715

  • ROCm 6.4.1

  • vLLM 0.9.1

  • PyTorch 2.7.0

rocm/vllm:rocm6.4.1_vllm_0.9.1_20250702

  • ROCm 6.4.1

  • vLLM 0.9.1

  • PyTorch 2.7.0

rocm/vllm:rocm6.4.1_vllm_0.9.0.1_20250605

  • ROCm 6.4.1

  • vLLM 0.9.0.1

  • PyTorch 2.7.0

rocm/vllm:rocm6.3.1_vllm_0.8.5_20250521

  • ROCm 6.3.1

  • 0.8.5 vLLM (0.8.6.dev)

  • PyTorch 2.7.0

rocm/vllm:rocm6.3.1_vllm_0.8.5_20250513

  • ROCm 6.3.1

  • vLLM 0.8.5

  • PyTorch 2.7.0

rocm/vllm:rocm6.3.1_instinct_vllm0.8.3_20250415

  • ROCm 6.3.1

  • vLLM 0.8.3

  • PyTorch 2.7.0

rocm/vllm:rocm6.3.1_instinct_vllm0.7.3_20250325

  • ROCm 6.3.1

  • vLLM 0.7.3

  • PyTorch 2.7.0

rocm/vllm:rocm6.3.1_mi300_ubuntu22.04_py3.12_vllm_0.6.6

  • ROCm 6.3.1

  • vLLM 0.6.6

  • PyTorch 2.7.0

rocm/vllm:rocm6.2_mi300_ubuntu20.04_py3.9_vllm_0.6.4

  • ROCm 6.2.1

  • vLLM 0.6.4

  • PyTorch 2.5.0

rocm/vllm:rocm6.2_mi300_ubuntu22.04_py3.9_vllm_7c5fd50

  • ROCm 6.2.0

  • vLLM 0.4.3

  • PyTorch 2.4.0