-
Notifications
You must be signed in to change notification settings - Fork 5
Running LLM As Chatbot in your cloud
This repository is a fork of deep-diver/LLM-As-Chatbot.
The only difference from the original repo is the .dstack.yml file, which allows you to run LLM-As-Chatbot in your cloud with a single dstack run command, automatically provisioning cloud resources for you.
git clone https://github.com/dstackai/LLM-As-Chatbot.git
cd LLM-As-Chatbotpip install "dstack[aws,gcp,azure,lambda]" -U
dstack startOnce the server is up, make sure to log in and create a project with your cloud credentials (AWS, GCP, or Azure). After creating the project, copy its dstack config command and run it locally to configure the CLI to use this project.
Create a .dstack/profiles.yml file that points to the created project and describes the resources.
Example:
profiles:
- name: gcp
project: gcp
resources:
memory: 48GB
gpu:
memory: 24GB
default: trueRun the dstack init command:
dstack initUse the dstack run command:
dstack run .This command will build the environment and run LLM-As-Chatbot in your cloud.
dstack will automatically forward the port to your local machine, providing secure and convenient access.
For more details on how dstack works, check out its documentation.