Skip to content

Commit

Permalink
Fix usage of engine
Browse files Browse the repository at this point in the history
  • Loading branch information
casper-hansen committed Sep 21, 2023
1 parent fbeea40 commit 133dd7a
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion awq/modules/fused/attn.py
Original file line number Diff line number Diff line change
Expand Up @@ -219,7 +219,7 @@ def forward(
xv = xv.view((bsz,) + self.attention_shapes["single_xv_view"])

past_key_value = (xk, xv) if use_cache else None
attention_weight = awq_inference_engine.single_query_attention(
attention_weight = ft_inference_engine.single_query_attention(
xq, # query
xk, # key
xv, # value
Expand Down

0 comments on commit 133dd7a

Please sign in to comment.