Release : Alpha 1.0
OSSPITA is a lightweight, fully local, open-source desktop interface that enables users to interact for free and in real time with various Open-Source LLMs. Leveraging the Ollama API, OSSPITA provides a user-friendly experience for engaging with cutting-edge AI technology.
Video demo :
Click the image above to watch the video on YouTube
- Real-time interactions with most Open-Source AI models.
- Lightweight and fully local operations.
- Beginner-friendly installation process.
- Retrieval Augmented Generation so you can probe your own documents while preserving your privacy.
- Web Search functionality for the integration of more up-to-date information.
- Comprehensive inference stats.
- Memory allocation tracking for context length tuning.
- Prompt library with (versioning coming).
- Easy access to the most common LLM settings.
- Persistent conversations.
- Multimodal operations.
- Responsive design.
- Agent chaining for complex task resolution.
- Improved Web Search algorithm with enhanced options.
- Online domain names ranking.
- Improved RAG algorithm with enhanced options (chunk size selection).
- Online prompt and agent sharing platform.
- Prompt versioning system.
- Dedicated coding agent.
- Code syntax highlighting.
- A Dark mode theme.
- Charts generation.
- Voice mode.
- Context autosizing option.
- In-depth RAG stats & data.
- In-depth Web Search stats & data.
- Node.js (v22.11.0 or higher) : https://nodejs.org/en/
- npm (10.9.0 or higher)
- Ollama (installed and running locally) : https://ollama.com/download
- The OSspita backend running.
- At least one open source model.
- Nvidia CUDA if needed : https://developer.nvidia.com/cuda-downloads
Models Recommandations :
GPU VRAM | Model | Purpose | Link |
---|---|---|---|
4GB | Llama 3.2:3b | Conversational | Llama 3.2 on Ollama |
8GB | Aya Expanse:8b | Conversational | Aya Expanse on Ollama |
12GB | Mistral Nemo:12b | Conversational | Mistal Nemo on Ollama |
4GB | starcoder2:3b | Coding | StarCoder2 on Ollama |
8GB & 12GB | qwen2.5-coder:7b | Coding | Qwen 2.5 Coder on Ollama |
Clone the repository :
git clone https://github.com/ask0ldd/OsspitaUI.git
Navigate to the project directory:
cd OsspitaUI
cd front
Install dependencies:
npm install
Start the development server:
npm run dev
Open your browser and visit http://localhost:5173 so that you can be guided through the rest of the installation process.
Video demo :
- React 18.3.1
- Vite
- Ollama API
- ExpressJS
- LokiJS
- Duck-duck-scrape
- Cheerio
- React-pdftotext
- Marked 14.1.1
- Turndown
- Cors
- BodyParser
If you face any bug or if you want me to add / improve any functionality, open an issue with your request. Thanks a lot for your help.
- Georgi Gerganov and all associated contributors for llama.cpp.
- The Ollama team for providing the API used for inference.
- React and Vite communities for their excellent tools.
- All contributors and supporters of the project.
- Everybody producing GGUF models.
- And finally, all the people that worked on any of the libraries I used.