You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
HI @haseeb-heaven, first thanks for your great work for tying up all common LLMs with one string. Really appreciateable.
I was looking for Vision AI models. In the documentation, you have written that currently, it supports Vopensource and Vision AI models and Vision APIs, but is there any possibility to add any UI, may be a web page running on localhost to deal with LLms, like prompt writing , image uploading, or a drop-down to switch to LLMs and entering their API keys in text fields?
The text was updated successfully, but these errors were encountered:
Hi
Currently it only support GPT Vision and Google AI Vision models and others are not added yet and about the interface it has only CLI interface but for vision models you need to modify it to have the UI for image upload or chat.
The interface is simple to use for all users but for additionalal features like image upload and chat interface we can create new interface but it will take lots of re-designing the interpreter.py and interpreter_lib.py class
Thanks for the reply. Any plans to upgrade this with Gemini?
Open-source vision models like LLAVA and GUI support will make CODE-Interpreter a more generic and interesting tool.
HI @haseeb-heaven, first thanks for your great work for tying up all common LLMs with one string. Really appreciateable.
I was looking for Vision AI models. In the documentation, you have written that currently, it supports Vopensource and Vision AI models and Vision APIs, but is there any possibility to add any UI, may be a web page running on localhost to deal with LLms, like prompt writing , image uploading, or a drop-down to switch to LLMs and entering their API keys in text fields?
The text was updated successfully, but these errors were encountered: