-
Hi everyone It seems to me that when a person is receiving a chat response on a web page chat, another person's chat is put on hold until the first person receives the complete chat response. I am running Flowise on Windows 10 with only one GPU. Is there any way to make it 'multi-user'? Something like 'forking' of the chat bot to serve more than one user simultaneously? Thank you very much. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Hi everyone I asked AI how this can be made to serve multiple users on a PC with the following configuration: It replied: Thank you and cheers! m3svcs |
Beta Was this translation helpful? Give feedback.
Hi everyone
I asked AI how this can be made to serve multiple users on a PC with the following configuration:
AMD Ryzen 7 1700 8 Core
64Gb RAM
NVIDIA GeForce GTX 1650 4Gb
Windows 10
Ollama for ChatFlow and embedding
It replied:
The following environment variables need to be set:
OLLAMA_CUDA=1
OLLAMA_NUM_PARALLEL=4
OLLAMA_NUM_THREADS=64
Thank you and cheers!
m3svcs