-
-
Notifications
You must be signed in to change notification settings - Fork 4.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Core] Update to outlines >= 0.1.8 #10576
Conversation
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can do one of these:
🚀 |
|
converted back to a draft because I don't have |
I introduce this PR to outlines: The caching issue (or not being able to pickle) also affects the multiprocessing-based engine in References: vllm/vllm/engine/multiprocessing/client.py Line 604 in 7e4bbda
|
1b6e5f6
to
22ea8e8
Compare
Thank you @joennlae - this works for me Here are some performance measurements with (a) older outlines, (b) newer outlines with your patch but no caching, (c) newer outlines with your patch and caching reinstated:
|
I understand that `pickleable` is not your priority right now. But the `RegexGuide` needs to be pickled for `vllm` production use, which is multiprocessing-based. This PR reintroduces this pickling capability + some tests. I understand that this introduces more effort on your side. References: dottxt-ai/outlines#1274 vllm-project/vllm#10490 vllm-project/vllm#10576 vllm-project/vllm#10489 It would also tackle the current caching issues: huggingface/text-generation-inference#2766 dottxt-ai/outlines#1283 Closes: #95
This pull request has merge conflicts that must be resolved before it can be |
22ea8e8
to
4147eab
Compare
I think this will be ready once outlines 0.1.8 is available on pypi. |
4147eab
to
279ccc9
Compare
279ccc9
to
8bcfde4
Compare
It is ! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM pending green CI!
0.1.x prior to 0.1.8 + outlines-core 0.1.18 had issues with serialization that broke vllm integration. Also change our code slightly to account for an API change in outlines. Signed-off-by: Russell Bryant <[email protected]>
8bcfde4
to
2143457
Compare
Signed-off-by: Russell Bryant <[email protected]> Signed-off-by: Akshat Tripathi <[email protected]>
Signed-off-by: Russell Bryant <[email protected]>
This PR updates to the latest release of
outlines
that works with vllm.It is a draft while we wait for 0.1.8 to be on pypi.
FIX #3794
FIX #10489
279ccc9 [Core] Update to outlines >= 0.1.8
commit 279ccc9
Author: Russell Bryant [email protected]
Date: Thu Nov 21 21:25:22 2024 +0000