-
Notifications
You must be signed in to change notification settings - Fork 51
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multiuser and simultaneous requests #8
Comments
Hi, lollms is a mono-user project. |
Thank you for your fast answer! To make lollms multi-user and server concurrent requests should I use some separate web server like Apache / Nginx or is Flask default web server is enough for this? Also, can different users request for generation with different models or all users need use only one particular model which is selected server-side only? |
Hi and sorry for being this late. I'm very very buzzy and lollms is taking me a huge chunk of my night life. Then you install lollms-webui on one or multiple PCS, each PC has its own local database, the server don't store your requets. in the bindings you select lollms remote nodes and you go to its settings and add http://ipaddress:9600 to the hosts list (you can have multiple hosts by the way). For now, the server should be configured via lollms-settings command that will allow you to select the binding, model and mount as many persnoas as you want. The users of the webui use the service with the selected model. technically you can run multiple services with different models but that may require you to have big ressources in terms of GPUs. you can also do this remotely from another place or share servers with friends as the lollms remote nodes supports multiple servers. You can create a network of generators/clients and there is a queuing mechanism in case more queries than servers. I hope this answers your question. |
Congrats on great project! I started playing with it and have two questions so far:
I also saw a binding which name assumes that I can connect this web ui to other lollm server - does it allow multiple inferences happening on the same server then?
The text was updated successfully, but these errors were encountered: