You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: samples/managed-llm-provider/README.md
+8-7Lines changed: 8 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,11 +6,12 @@ This sample application demonstrates using Managed LLMs with a Docker Model Prov
6
6
7
7
> Note: This version uses a [Docker Model Provider](https://docs.docker.com/compose/how-tos/model-runner/#provider-services) for managing LLMs. For the version with Defang's [OpenAI Access Gateway](https://docs.defang.io/docs/concepts/managed-llms/openai-access-gateway), please see our [*Managed LLM Sample*](https://github.com/DefangLabs/samples/tree/main/samples/managed-llm) instead.
8
8
9
-
The Docker Model Provider allows users to use AWS Bedrock or Google Cloud Vertex AI models with their application. It is a service in the `compose.yaml` file.
9
+
The Docker Model Provider allows users to run LLMs locally using `docker compose`. It is a service with `provider:` in the `compose.yaml` file.
10
+
Defang will transparently fixup your project to use AWS Bedrock or Google Cloud Vertex AI models during deployment.
10
11
11
-
You can configure the `MODEL` and `ENDPOINT_URL` for the LLM separately for local development and production environments.
12
-
* The `MODEL` is the LLM Model ID you are using.
13
-
* The `ENDPOINT_URL` is the bridge that provides authenticated access to the LLM model.
12
+
You can configure the `LLM_MODEL` and `LLM_URL` for the LLM separately for local development and production environments.
13
+
* The `LLM_MODEL` is the LLM Model ID you are using.
14
+
* The `LLM_URL` will be set by Docker and during deployment Defang will provide authenticated access to the LLM model in the cloud.
14
15
15
16
Ensure you have enabled model access for the model you intend to use. To do this, you can check your [AWS Bedrock model access](https://docs.aws.amazon.com/bedrock/latest/userguide/model-access-modify.html) or [GCP Vertex AI model access](https://cloud.google.com/vertex-ai/generative-ai/docs/control-model-access).
16
17
@@ -38,14 +39,14 @@ docker compose -f compose.dev.yaml up --build
38
39
39
40
## Configuration
40
41
41
-
For this sample, you will need to provide the following [configuration](https://docs.defang.io/docs/concepts/configuration):
42
+
For this sample, you will need to provide the following [configuration](https://docs.defang.io/docs/concepts/configuration):
42
43
43
44
> Note that if you are using the 1-click deploy option, you can set these values as secrets in your GitHub repository and the action will automatically deploy them for you.
44
45
45
-
### `MODEL`
46
+
### `LLM_MODEL`
46
47
The Model ID of the LLM you are using for your application. For example, `anthropic.claude-3-haiku-20240307-v1:0`.
0 commit comments