Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Torch.compile with LLM entrypoint #1055

Open
Giuseppe5 opened this issue Oct 14, 2024 · 0 comments
Open

Torch.compile with LLM entrypoint #1055

Giuseppe5 opened this issue Oct 14, 2024 · 0 comments
Labels
enhancement New feature or request examples Anything related to examples good first issue Good for newcomers

Comments

@Giuseppe5
Copy link
Collaborator

Is your feature request related to a problem? Please describe.
Torch.compile allows faster inference with quantization, examples are provided in imagenet/ptq entrypoint and in the stable_diffusion entrypoint.

Describe the solution you'd like
Investigate whether this can be extended to LLM, in particular compatibility with accelerate

@Giuseppe5 Giuseppe5 added enhancement New feature or request good first issue Good for newcomers labels Oct 14, 2024
@Giuseppe5 Giuseppe5 changed the title Torch.compile with LLM entrypoing Torch.compile with LLM entrypoint Oct 14, 2024
@Giuseppe5 Giuseppe5 added the examples Anything related to examples label Oct 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request examples Anything related to examples good first issue Good for newcomers
Projects
None yet
Development

No branches or pull requests

1 participant