From 566ebdbc9a25376fd611ec00b107e610425508f9 Mon Sep 17 00:00:00 2001 From: Hui Liu <96135754+hliuca@users.noreply.github.com> Date: Tue, 3 Dec 2024 15:36:37 -0800 Subject: [PATCH] use 0.0 instead of 0 --- vllm/attention/backends/rocm_flash_attn.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/attention/backends/rocm_flash_attn.py b/vllm/attention/backends/rocm_flash_attn.py index cdf1c52b92af8..b4f4e5bb1500a 100644 --- a/vllm/attention/backends/rocm_flash_attn.py +++ b/vllm/attention/backends/rocm_flash_attn.py @@ -462,7 +462,7 @@ def __init__( if logits_soft_cap is None: # In flash-attn, setting logits_soft_cap as 0 means no soft cap. - self.logits_soft_cap = 0 + self.logits_soft_cap = 0.0 else: self.logits_soft_cap = logits_soft_cap