Skip to content

Commit

Permalink
remove last line
Browse files Browse the repository at this point in the history
Signed-off-by: charlifu <[email protected]>
  • Loading branch information
charlifu committed Dec 2, 2024
1 parent b412bd6 commit f441c65
Showing 1 changed file with 0 additions and 3 deletions.
3 changes: 0 additions & 3 deletions vllm/model_executor/layers/quantization/utils/w8a8_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -146,9 +146,6 @@ def apply_fp8_linear(

return output[0:input_2d.shape[0], ...].view(*output_shape)

return torch.narrow(output, 0, 0,
input_2d.shape[0]).view(*output_shape)

else:
# Fallback for channelwise case, where we use unfused DQ
# due to limitations with scaled_mm
Expand Down

0 comments on commit f441c65

Please sign in to comment.