Skip to content

Ingest FP8 attn scales and use them in ROCm FlashAttention #1184

Ingest FP8 attn scales and use them in ROCm FlashAttention

Ingest FP8 attn scales and use them in ROCm FlashAttention #1184

Triggered via pull request December 19, 2024 23:12
Status Failure
Total duration 24s
Artifacts

ruff.yml

on: pull_request
Matrix: ruff
Fit to window
Zoom out
Zoom in

Annotations

2 errors and 1 warning
Ruff (E501): vllm/model_executor/models/llama.py#L236
vllm/model_executor/models/llama.py:236:81: E501 Line too long (81 > 80)
ruff (3.12)
Process completed with exit code 1.
ruff (3.12)
ubuntu-latest pipelines will use ubuntu-24.04 soon. For more details, see https://github.com/actions/runner-images/issues/10636