Skip to content

Commit

Permalink
fix logits_soft_cap constructor
Browse files Browse the repository at this point in the history
  • Loading branch information
hliuca committed Dec 2, 2024
1 parent 9242621 commit 8cdb96f
Showing 1 changed file with 3 additions and 2 deletions.
5 changes: 3 additions & 2 deletions vllm/attention/backends/rocm_flash_attn.py
Original file line number Diff line number Diff line change
Expand Up @@ -462,8 +462,9 @@ def __init__(

if logits_soft_cap is None:
# In flash-attn, setting logits_soft_cap as 0 means no soft cap.
logits_soft_cap = 0
self.logits_soft_cap = logits_soft_cap
self.logits_soft_cap = 0
else:
self.logits_soft_cap = logits_soft_cap

Check failure on line 467 in vllm/attention/backends/rocm_flash_attn.py

View workflow job for this annotation

GitHub Actions / mypy (3.9)

Incompatible types in assignment (expression has type "float", variable has type "int") [assignment]

Check failure on line 467 in vllm/attention/backends/rocm_flash_attn.py

View workflow job for this annotation

GitHub Actions / mypy (3.10)

Incompatible types in assignment (expression has type "float", variable has type "int") [assignment]

Check failure on line 467 in vllm/attention/backends/rocm_flash_attn.py

View workflow job for this annotation

GitHub Actions / mypy (3.11)

Incompatible types in assignment (expression has type "float", variable has type "int") [assignment]

Check failure on line 467 in vllm/attention/backends/rocm_flash_attn.py

View workflow job for this annotation

GitHub Actions / mypy (3.12)

Incompatible types in assignment (expression has type "float", variable has type "int") [assignment]

self.num_heads = num_heads
self.head_size = head_size
Expand Down

0 comments on commit 8cdb96f

Please sign in to comment.