Skip to content

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
Updated to move to ROCm 6.3 and post the issue with saving Tunable Ops due to PyTorch bug.
  • Loading branch information
t-parry authored Dec 12, 2024
1 parent 22f9066 commit 50dfe93
Showing 1 changed file with 5 additions and 6 deletions.
11 changes: 5 additions & 6 deletions docs/dev-docker/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,11 +10,11 @@ This documentation shows some reference performance numbers and the steps to rep

It includes:

- ROCm™ 6.2.2
- ROCm™ 6.3

- vLLM 0.6.3

- PyTorch 2.5dev (nightly)
- PyTorch 2.6dev (nightly)

## System configuration

Expand All @@ -23,7 +23,7 @@ The performance data below was measured on a server with MI300X accelerators wit
| System | MI300X with 8 GPUs |
|---|---|
| BKC | 24.13 |
| ROCm | version ROCm 6.2.2 |
| ROCm | version ROCm 6.3 |
| amdgpu | build 2009461 |
| OS | Ubuntu 22.04 |
| Linux Kernel | 5.15.0-117-generic |
Expand All @@ -45,9 +45,8 @@ You can pull the image with `docker pull rocm/vllm-dev:main`

### What is New

- MoE optimizations for Mixtral 8x22B, FP16
- Llama 3.2 stability improvements
- Llama 3.3 support
- ROCm 6.3 support
- Potential bug with Tunable Ops not saving due to a PyTorch issue


Gemms are tuned using PyTorch's Tunable Ops feature (https://github.com/pytorch/pytorch/blob/main/aten/src/ATen/cuda/tunable/README.md)
Expand Down

0 comments on commit 50dfe93

Please sign in to comment.