You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-deploy-models-llama.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -204,7 +204,7 @@ To create a deployment:
204
204
205
205
---
206
206
207
-
To learn about billing for Meta Llama models deployed as a serverless API, see [Cost and quota considerations for Meta Llama models deployed as a serverless API](#cost-and-quota-considerations-for-meta-llama-3.1-models-deployed-as-a-serverless-api).
207
+
To learn about billing for Meta Llama models deployed as a serverless API, see [Cost and quota considerations for Meta Llama models deployed as a serverless API](#cost-and-quota-considerations-for-meta-llama-31-models-deployed-as-a-serverless-api).
208
208
209
209
### Consume Meta Llama models as a service
210
210
@@ -220,7 +220,7 @@ Models deployed as a service can be consumed using either the chat or the comple
220
220
- For completions models, such as `Llama-3-8B`, use the [`<target_url>/v1/completions`](#completions-api) API.
221
221
- For chat models, such as `Meta-Llama-3.1-405B-Instruct`, use the [`/chat/completions`](#chat-api) API.
222
222
223
-
For more information on using the APIs, see the [reference](#reference-for-meta-llama-3.1-models-deployed-a-serverless-api) section.
223
+
For more information on using the APIs, see the [reference](#reference-for-meta-llama-31-models-deployed-a-serverless-api) section.
224
224
225
225
# [Meta Llama 2](#tab/llama-two)
226
226
@@ -232,7 +232,7 @@ Models deployed as a service can be consumed using either the chat or the comple
232
232
- For completions models, such as `Meta-Llama-2-7B`, use the [`/v1/completions`](#completions-api) API or the [Azure AI Model Inference API](reference-model-inference-api.md) on the route `/completions`.
233
233
- For chat models, such as `Meta-Llama-2-7B-Chat`, use the [`/v1/chat/completions`](#chat-api) API or the [Azure AI Model Inference API](reference-model-inference-api.md) on the route `/chat/completions`.
234
234
235
-
For more information on using the APIs, see the [reference](#reference-for-meta-llama-3.1-models-deployed-a-serverless-api) section.
235
+
For more information on using the APIs, see the [reference](#reference-for-meta-llama-31-models-deployed-a-serverless-api) section.
0 commit comments