Skip to content

kezzico/LLM-Chat

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LLM CHAT

LLM Chat is an interface for conversing with generic open source LLMs such as Zephyr7B and others.

Currently it uses EventStreams to deliver output from the model to the React application via Flask.

Alt text

Install

Start by initializing the virtual environment in flask-app.

cd flask-app

python -m venv venv

source ./venv/bin/activate

pip install -r requirements.txt

deactivate

Next, initialize react-app.

cd react-app

npm install

Run

Start the flask app by activating the virtual environment again

cd flask-app

source ./venv/bin/activate

python flaskapp.py

Next, from a separate shell, start the React app.

cd react-app

npm run start

NPM should launch your browser automatically to localhost:3000

The React app has proxy settings to pass requests through to the flask app.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published