You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-studio/how-to/deploy-models-llama.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -165,7 +165,7 @@ To create a deployment:
165
165
166
166
1. You can always find the endpoint's details, URL, and access keys by navigating to the project page and selecting **Deployments** from the left menu.
167
167
168
-
To learn about billing for Meta Llama models deployed with pay-as-you-go, see [Cost and quota considerations for Llama 3 models deployed as a service](#cost-and-quota-considerations-for-meta-llama-3.1-models-deployed-as-a-service).
168
+
To learn about billing for Meta Llama models deployed with pay-as-you-go, see [Cost and quota considerations for Llama 3 models deployed as a service](#cost-and-quota-considerations-for-meta-llama-31-models-deployed-as-a-service).
169
169
170
170
# [Meta Llama 2](#tab/llama-two)
171
171
@@ -202,7 +202,7 @@ To create a deployment:
202
202
1. You can return to the Deployments page, select the deployment, and note the endpoint's **Target** URL and the Secret **Key**, which you can use to call the deployment and generate completions.
203
203
1. You can always find the endpoint's details, URL, and access keys by navigating to your project and selecting **Deployments** from the left menu.
204
204
205
-
To learn about billing for Llama models deployed with pay-as-you-go, see [Cost and quota considerations for Llama 3 models deployed as a service](#cost-and-quota-considerations-for-meta-llama-3.1-models-deployed-as-a-service).
205
+
To learn about billing for Llama models deployed with pay-as-you-go, see [Cost and quota considerations for Llama 3 models deployed as a service](#cost-and-quota-considerations-for-meta-llama-31-models-deployed-as-a-service).
206
206
207
207
---
208
208
@@ -225,7 +225,7 @@ Models deployed as a service can be consumed using either the chat or the comple
225
225
- For completions models, such as `Meta-Llama-3-8B`, use the [`/completions`](#completions-api) API.
226
226
- For chat models, such as `Meta-Llama-3.1-405B-Instruct`, use the [`/chat/completions`](#chat-api) API.
227
227
228
-
For more information on using the APIs, see the [reference](#reference-for-meta-llama-3.1-models-deployed-as-a-service) section.
228
+
For more information on using the APIs, see the [reference](#reference-for-meta-llama-31-models-deployed-as-a-service) section.
0 commit comments