-
-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Model][LoRA]LoRA support added for idefics3 #10281
Conversation
Signed-off-by: B-201 <[email protected]>
Signed-off-by: B-201 <[email protected]>
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can do one of these:
🚀 |
Signed-off-by: B-201 <[email protected]>
Signed-off-by: B-201 <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good, I assume you have already tested this on your own?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
overall lgtm, @DarkLight1337 could you please look at this PR, thanks~
Yes, I trained a LoRA myself and tested it locally. It runs very well, and the results align with those on transformers. |
Signed-off-by: B-201 <[email protected]>
Signed-off-by: B-201 <[email protected]> Signed-off-by: Sumit Dubey <[email protected]>
Signed-off-by: B-201 <[email protected]>
Signed-off-by: B-201 <[email protected]> Signed-off-by: Maxime Fournioux <[email protected]>
Signed-off-by: B-201 <[email protected]> Signed-off-by: Tyler Michael Smith <[email protected]>
Signed-off-by: B-201 <[email protected]>
dummy_data
to exceed the maximum model length setting.