Skip to content

Commit 477a933

Browse files
committed
add tabs for CLI and more
1 parent 8acd051 commit 477a933

File tree

3 files changed

+78
-21
lines changed

3 files changed

+78
-21
lines changed

articles/ai-foundry/how-to/deploy-models-serverless.md

Lines changed: 77 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ zone_pivot_groups: azure-ai-serverless-deployment
2020

2121
In this article, you learn how to deploy an Azure AI Foundry Model as a serverless API deployment. [Certain models in the model catalog](deploy-models-serverless-availability.md) can be deployed as a serverless API deployment. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need. This deployment option doesn't require quota from your subscription.
2222

23-
Although serverless API deployment is one of the ways to deploy Azure AI Foundry Models, we recommend that you deploy Foundry Models to **Azure AI Foundry resources**.
23+
Although serverless API deployment is one option for deploying Azure AI Foundry Models, we recommend that you deploy Foundry Models to **Azure AI Foundry resources**.
2424

2525
[!INCLUDE [deploy-models-to-foundry-resources](../includes/deploy-models-to-foundry-resources.md)]
2626

@@ -203,16 +203,16 @@ Furthermore, models offered through Azure Marketplace are available for deployme
203203
204204
__subscription.yml__
205205
206-
```yml
207-
name: AI21-Jamba-1-5-Large-suawp
208-
model_id: azureml://registries/azureml-ai21/models/AI21-Jamba-1.5-Large
209-
```
206+
```yml
207+
name: AI21-Jamba-1-5-Large-qwerty
208+
model_id: azureml://registries/azureml-ai21/models/AI21-Jamba-1.5-Large
209+
```
210210
211-
Use the previous file to create the subscription:
211+
Use the previous file to create the subscription:
212212
213-
```azurecli
214-
az ml marketplace-subscription create -f subscription.yml
215-
```
213+
```azurecli
214+
az ml marketplace-subscription create -f subscription.yml
215+
```
216216
217217
1. (Optional) At any point, you can see the model offers to which your project is currently subscribed:
218218
@@ -224,14 +224,52 @@ Furthermore, models offered through Azure Marketplace are available for deployme
224224
225225
## Deploy the model to a serverless API
226226
227-
In this section, you create an endpoint for your model with the name **AI21-Jamba-1-5-Large-suawp**.
227+
# [Models sold directly by Azure](#tab/azure-direct)
228+
229+
In this section, you create an endpoint for your model. Name the endpoint **DeepSeek-R1-qwerty**.
230+
231+
1. Create the serverless endpoint.
232+
233+
__endpoint.yml__
234+
235+
```yml
236+
name: DeepSeek-R1-qwerty
237+
model_id: azureml://registries/azureml-deepseek/models/DeepSeek-R1
238+
```
239+
240+
Use the _endpoint.yml_ file to create the endpoint:
241+
242+
```azurecli
243+
az ml serverless-endpoint create -f endpoint.yml
244+
```
245+
246+
1. At any point, you can see the endpoints deployed to your project:
247+
248+
```azurecli
249+
az ml serverless-endpoint list
250+
```
251+
252+
1. The created endpoint uses key authentication for authorization. Use the following steps to get the keys associated with a given endpoint.
253+
254+
```azurecli
255+
az ml serverless-endpoint get-credentials -n DeepSeek-R1-qwerty
256+
```
257+
258+
1. If you need to consume this deployment from a different project or hub, or you plan to use Prompt flow to build intelligent applications, you need to create a connection to the standard deployment. To learn how to configure an existing standard deployment on a new project or hub, see [Consume deployed standard deployment from a different project or from Prompt flow](deploy-models-serverless-connect.md).
259+
260+
> [!TIP]
261+
> If you're using Prompt flow in the same project or hub where the deployment was deployed, you still need to create the connection.
262+
263+
# [Models from Partners and Community](#tab/partner-models)
264+
265+
In this section, you create an endpoint for your model. Name the endpoint **AI21-Jamba-1-5-Large-qwerty**.
228266
229267
1. Create the serverless endpoint.
230268
231269
__endpoint.yml__
232270
233271
```yml
234-
name: AI21-Jamba-1-5-Large-suawp
272+
name: AI21-Jamba-1-5-Large-qwerty
235273
model_id: azureml://registries/azureml-ai21/models/AI21-Jamba-1.5-Large
236274
```
237275
@@ -250,14 +288,15 @@ In this section, you create an endpoint for your model with the name **AI21-Jamb
250288
1. The created endpoint uses key authentication for authorization. Use the following steps to get the keys associated with a given endpoint.
251289
252290
```azurecli
253-
az ml serverless-endpoint get-credentials -n AI21-Jamba-1-5-Large-suawp
291+
az ml serverless-endpoint get-credentials -n AI21-Jamba-1-5-Large-qwerty
254292
```
255293
256294
1. If you need to consume this deployment from a different project or hub, or you plan to use Prompt flow to build intelligent applications, you need to create a connection to the standard deployment. To learn how to configure an existing standard deployment on a new project or hub, see [Consume deployed standard deployment from a different project or from Prompt flow](deploy-models-serverless-connect.md).
257295
258296
> [!TIP]
259297
> If you're using Prompt flow in the same project or hub where the deployment was deployed, you still need to create the connection.
260298
299+
---
261300
262301
## Use the standard deployment
263302
@@ -270,11 +309,29 @@ Read more about the [capabilities of this API](../../ai-foundry/model-inference/
270309
271310
You can delete model subscriptions and endpoints. Deleting a model subscription makes any associated endpoint become *Unhealthy* and unusable.
272311
312+
# [Models sold directly by Azure](#tab/azure-direct)
313+
314+
To delete a standard deployment:
315+
316+
```azurecli
317+
az ml serverless-endpoint delete \
318+
--name "DeepSeek-R1-qwerty"
319+
```
320+
321+
To delete the associated model subscription:
322+
323+
```azurecli
324+
az ml marketplace-subscription delete \
325+
--name "DeepSeek-R1"
326+
```
327+
328+
# [Models from Partners and Community](#tab/partner-models)
329+
273330
To delete a standard deployment:
274331

275332
```azurecli
276333
az ml serverless-endpoint delete \
277-
--name "AI21-Jamba-1-5-Large-suawp"
334+
--name "AI21-Jamba-1-5-Large-qwerty"
278335
```
279336

280337
To delete the associated model subscription:
@@ -284,6 +341,8 @@ az ml marketplace-subscription delete \
284341
--name "AI21-Jamba-1.5-Large"
285342
```
286343

344+
---
345+
287346
::: zone-end
288347

289348

@@ -369,12 +428,12 @@ Furthermore, models offered through Azure Marketplace are available for deployme
369428

370429
## Deploy the model to a serverless API
371430

372-
In this section, you create an endpoint for your model with the name **AI21-Jamba-1-5-Large-suawp**.
431+
In this section, you create an endpoint for your model with the name **AI21-Jamba-1-5-Large-qwerty**.
373432

374433
1. Create the serverless endpoint.
375434

376435
```python
377-
endpoint_name="AI21-Jamba-1-5-Large-suawp"
436+
endpoint_name="AI21-Jamba-1-5-Large-qwerty"
378437

379438
serverless_endpoint = ServerlessEndpoint(
380439
name=endpoint_name,
@@ -389,7 +448,7 @@ In this section, you create an endpoint for your model with the name **AI21-Jamb
389448
1. At any point, you can see the endpoints deployed to your project:
390449

391450
```python
392-
endpoint_name="AI21-Jamba-1-5-Large-suawp"
451+
endpoint_name="AI21-Jamba-1-5-Large-qwerty"
393452

394453
serverless_endpoint = ServerlessEndpoint(
395454
name=endpoint_name,
@@ -520,7 +579,7 @@ Furthermore, models offered through Azure Marketplace are available for deployme
520579

521580
## Deploy the model to a serverless API
522581

523-
In this section, you create an endpoint for your model with the name **AI21-Jamba-1-5-Large-suawp**.
582+
In this section, you create an endpoint for your model with the name **AI21-Jamba-1-5-Large-qwerty**.
524583

525584
1. Create the serverless endpoint. Use the following template to create an endpoint:
526585

@@ -569,9 +628,7 @@ In this section, you create an endpoint for your model with the name **AI21-Jamb
569628
--query "[?type=='Microsoft.MachineLearningServices/workspaces/serverlessEndpoints']"
570629
```
571630

572-
1. The created endpoint uses key authentication for authorization. Use the following steps to get the keys associated with a given endpoint.
573-
574-
Use REST APIs to query this information.
631+
1. The created endpoint uses key authentication for authorization. Get the keys associated with the given endpoint by using REST APIs to query this information.
575632

576633
1. If you need to consume this deployment from a different project or hub, or you plan to use Prompt flow to build intelligent applications, you need to create a connection to the standard deployment. To learn how to configure an existing standard deployment on a new project or hub, see [Consume deployed standard deployment from a different project or from Prompt flow](deploy-models-serverless-connect.md).
577634

articles/ai-foundry/includes/deploy-models-to-foundry-resources.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,4 +10,4 @@ ms.custom: include
1010
---
1111

1212
> [!NOTE]
13-
> We recommend that you deploy Azure AI Foundry Models to **Azure AI Foundry resources**, as this deployment method allows you to consume your models via a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. The endpoint follows the [Azure AI Model Inference API](/rest/api/aifoundry/modelinference/) which all the models in Foundry Models support. To learn how to deploy a Foundry Model to the Azure AI Foundry resources, see [Add and configure models to Azure AI Foundry Models](../model-inference/how-to/create-model-deployments.md).
13+
> We recommend that you deploy Azure AI Foundry Models to **Azure AI Foundry resources** so that you can consume your deployments in the resource via a single endpoint with the same authentication and schema to generate inference. The endpoint follows the [Azure AI Model Inference API](/rest/api/aifoundry/modelinference/) which all the Foundry Models support. To learn how to deploy a Foundry Model to the Azure AI Foundry resources, see [Add and configure models to Azure AI Foundry Models](../model-inference/how-to/create-model-deployments.md).
2.75 KB
Loading

0 commit comments

Comments
 (0)