-
-
Notifications
You must be signed in to change notification settings - Fork 5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: IBM Granite 3.1 tool parser fails #11402
Comments
+1 |
@K-Mistele Thanks for trying out the new model and looking in to this! The fix for parsing the |
|
Your current environment
The output of `python collect_env.py`
Not relevant - running with docker
vllm/vllm-openai:v0.6.5
Model Input Dumps
No response
🐛 Describe the bug
The
granite
tool parser (--tool-call-parser granite
) does not seem to be working for IBM Granite 3.1.Note that this is not related to existing streaming-related bugs; note that
stream
is set tofalse
;temperature=0
has also been set to maximize reproducabilityvLLM Run configuration via docker-compose:
Note
--enable-auto-tool-choice
and--tool-call-parser granite
. This should work per the docs on IBM granite tool calling.Example request and response:
Output:
Expected output: a valid tool call completion, for example:
Seems possibly related to #11039, #11069, and #11307
cc @maxdebayser @tjohnson31415
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: