Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Please make the ollama context length a config or env setting #334

Closed
matiashegoburu opened this issue Nov 19, 2024 · 2 comments
Closed

Comments

@matiashegoburu
Copy link

Hi all, I noticed the context length for ollama has been hardcoded to 32k, would it be possible to make it a configurable value somehow? Either a env variable, or a config setting somewhere? I have the ability to run longer context windows, but I am not sure how to set it up without modifying the code. Thanks!

@chrismahoney
Copy link
Collaborator

Please see a PR that actually was just merged today: #328

The community discussion is here: https://thinktank.ottomator.ai/t/gpu-is-not-being-fully-utilized-and-ollama-qwen2-5-32b-is-slow/870/4

@chrismahoney
Copy link
Collaborator

Anyone with similar issues, please refer to the links above 🙌

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants