Skip to content

Commit 9e18a8a

Browse files
authored
Merge pull request #374 from DefangLabs/linda-use-haiku
Use Haiku on Managed LLM
2 parents e0d02ad + 677ccc3 commit 9e18a8a

File tree

4 files changed

+4
-4
lines changed

4 files changed

+4
-4
lines changed

samples/managed-llm-provider/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ For this sample, you will need to provide the following [configuration](https://
4141
> Note that if you are using the 1-click deploy option, you can set these values as secrets in your GitHub repository and the action will automatically deploy them for you.
4242
4343
### `MODEL`
44-
The Model ID of the LLM you are using for your application. For example, `anthropic.claude-3-5-sonnet-20241022-v2:0`.
44+
The Model ID of the LLM you are using for your application. For example, `anthropic.claude-3-haiku-20240307-v1:0`.
4545
```bash
4646
defang config set MODEL
4747
```

samples/managed-llm-provider/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ services:
88
restart: always
99
environment:
1010
- ENDPOINT_URL=http://llm/api/v1/chat/completions # endpoint to the Provider Service
11-
- MODEL=anthropic.claude-3-5-sonnet-20241022-v2:0 # LLM model ID used in the Provider Service
11+
- MODEL=anthropic.claude-3-haiku-20240307-v1:0 # LLM model ID used in the Provider Service
1212
- OPENAI_API_KEY=FAKE_TOKEN # the actual value will be ignored when using the Provider Service
1313
healthcheck:
1414
test: ["CMD", "python3", "-c", "import sys, urllib.request; urllib.request.urlopen(sys.argv[1]).read()", "http://localhost:8000/"]

samples/managed-llm/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ For this sample, you will need to provide the following [configuration](https://
4242
> Note that if you are using the 1-click deploy option, you can set these values as secrets in your GitHub repository and the action will automatically deploy them for you.
4343
4444
### `MODEL`
45-
The Model ID of the LLM you are using for your application. For example, `anthropic.claude-3-5-sonnet-20241022-v2:0`.
45+
The Model ID of the LLM you are using for your application. For example, `anthropic.claude-3-haiku-20240307-v1:0`.
4646
```bash
4747
defang config set MODEL
4848
```

samples/managed-llm/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ services:
88
restart: always
99
environment:
1010
- ENDPOINT_URL=http://llm/api/v1/chat/completions # endpoint to the gateway service
11-
- MODEL=anthropic.claude-3-5-sonnet-20241022-v2:0 # LLM model ID used for the gateway
11+
- MODEL=anthropic.claude-3-haiku-20240307-v1:0 # LLM model ID used for the gateway
1212
- OPENAI_API_KEY=FAKE_TOKEN # the actual value will be ignored when using the gateway, but it should match the one in the llm service
1313
healthcheck:
1414
test: ["CMD", "python3", "-c", "import sys, urllib.request; urllib.request.urlopen(sys.argv[1]).read()", "http://localhost:8000/"]

0 commit comments

Comments
 (0)