Skip to content

Commit

Permalink
Fix: exclude cache_engine argument for input dump
Browse files Browse the repository at this point in the history
Signed-off-by: Dahai Tang <[email protected]>
  • Loading branch information
Dahai Tang committed Dec 16, 2024
1 parent c3a8ae8 commit 0f5e9e3
Showing 1 changed file with 2 additions and 1 deletion.
3 changes: 2 additions & 1 deletion vllm/worker/model_runner.py
Original file line number Diff line number Diff line change
Expand Up @@ -1610,7 +1610,8 @@ def prepare_model_input(
virtual_engine=virtual_engine)

@torch.inference_mode()
@dump_input_when_exception(exclude_args=[0], exclude_kwargs=["self"])
@dump_input_when_exception(exclude_args=[0],
exclude_kwargs=["self", "cache_engine"])
def execute_model(
self,
model_input: ModelInputForGPUWithSamplingMetadata,
Expand Down

0 comments on commit 0f5e9e3

Please sign in to comment.