You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In this article, you learn how to deploy an Azure AI Foundry Model as a serverless API deployment. [Certain models in the model catalog](deploy-models-serverless-availability.md) can be deployed as a serverless API deployment. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need. This deployment option doesn't require quota from your subscription.
22
22
23
-
Although serverless API deployment is one of the ways to deploy Azure AI Foundry Models, we recommend that you deploy Foundry Models to **Azure AI Foundry resources**.
23
+
Although serverless API deployment is one option for deploying Azure AI Foundry Models, we recommend that you deploy Foundry Models to **Azure AI Foundry resources**.
Use the _endpoint.yml_ file to create the endpoint:
241
+
242
+
```azurecli
243
+
az ml serverless-endpoint create -f endpoint.yml
244
+
```
245
+
246
+
1. At any point, you can see the endpoints deployed to your project:
247
+
248
+
```azurecli
249
+
az ml serverless-endpoint list
250
+
```
251
+
252
+
1. The created endpoint uses key authentication for authorization. Use the following steps to get the keys associated with a given endpoint.
253
+
254
+
```azurecli
255
+
az ml serverless-endpoint get-credentials -n DeepSeek-R1-qwerty
256
+
```
257
+
258
+
1. If you need to consume this deployment from a different project or hub, or you plan to use Prompt flow to build intelligent applications, you need to create a connection to the standard deployment. To learn how to configure an existing standard deployment on a new project or hub, see [Consume deployed standard deployment from a different project or from Prompt flow](deploy-models-serverless-connect.md).
259
+
260
+
> [!TIP]
261
+
> If you're using Prompt flow in the same project or hub where the deployment was deployed, you still need to create the connection.
262
+
263
+
# [Models from Partners and Community](#tab/partner-models)
264
+
265
+
In this section, you create an endpoint for your model. Name the endpoint **AI21-Jamba-1-5-Large-qwerty**.
@@ -250,14 +288,15 @@ In this section, you create an endpoint for your model with the name **AI21-Jamb
250
288
1. The created endpoint uses key authentication for authorization. Use the following steps to get the keys associated with a given endpoint.
251
289
252
290
```azurecli
253
-
az ml serverless-endpoint get-credentials -n AI21-Jamba-1-5-Large-suawp
291
+
az ml serverless-endpoint get-credentials -n AI21-Jamba-1-5-Large-qwerty
254
292
```
255
293
256
294
1. If you need to consume this deployment from a different project or hub, or you plan to use Prompt flow to build intelligent applications, you need to create a connection to the standard deployment. To learn how to configure an existing standard deployment on a new project or hub, see [Consume deployed standard deployment from a different project or from Prompt flow](deploy-models-serverless-connect.md).
257
295
258
296
> [!TIP]
259
297
> If you're using Prompt flow in the same project or hub where the deployment was deployed, you still need to create the connection.
260
298
299
+
---
261
300
262
301
## Use the standard deployment
263
302
@@ -270,11 +309,29 @@ Read more about the [capabilities of this API](../../ai-foundry/model-inference/
270
309
271
310
You can delete model subscriptions and endpoints. Deleting a model subscription makes any associated endpoint become *Unhealthy* and unusable.
272
311
312
+
# [Models sold directly by Azure](#tab/azure-direct)
313
+
314
+
To delete a standard deployment:
315
+
316
+
```azurecli
317
+
az ml serverless-endpoint delete \
318
+
--name "DeepSeek-R1-qwerty"
319
+
```
320
+
321
+
To delete the associated model subscription:
322
+
323
+
```azurecli
324
+
az ml marketplace-subscription delete \
325
+
--name "DeepSeek-R1"
326
+
```
327
+
328
+
# [Models from Partners and Community](#tab/partner-models)
329
+
273
330
To delete a standard deployment:
274
331
275
332
```azurecli
276
333
az ml serverless-endpoint delete \
277
-
--name "AI21-Jamba-1-5-Large-suawp"
334
+
--name "AI21-Jamba-1-5-Large-qwerty"
278
335
```
279
336
280
337
To delete the associated model subscription:
@@ -284,6 +341,8 @@ az ml marketplace-subscription delete \
284
341
--name "AI21-Jamba-1.5-Large"
285
342
```
286
343
344
+
---
345
+
287
346
::: zone-end
288
347
289
348
@@ -369,12 +428,12 @@ Furthermore, models offered through Azure Marketplace are available for deployme
369
428
370
429
## Deploy the model to a serverless API
371
430
372
-
In this section, you create an endpoint for your model with the name **AI21-Jamba-1-5-Large-suawp**.
431
+
In this section, you create an endpoint for your model with the name **AI21-Jamba-1-5-Large-qwerty**.
373
432
374
433
1. Create the serverless endpoint.
375
434
376
435
```python
377
-
endpoint_name="AI21-Jamba-1-5-Large-suawp"
436
+
endpoint_name="AI21-Jamba-1-5-Large-qwerty"
378
437
379
438
serverless_endpoint = ServerlessEndpoint(
380
439
name=endpoint_name,
@@ -389,7 +448,7 @@ In this section, you create an endpoint for your model with the name **AI21-Jamb
389
448
1. At any point, you can see the endpoints deployed to your project:
390
449
391
450
```python
392
-
endpoint_name="AI21-Jamba-1-5-Large-suawp"
451
+
endpoint_name="AI21-Jamba-1-5-Large-qwerty"
393
452
394
453
serverless_endpoint = ServerlessEndpoint(
395
454
name=endpoint_name,
@@ -520,7 +579,7 @@ Furthermore, models offered through Azure Marketplace are available for deployme
520
579
521
580
## Deploy the model to a serverless API
522
581
523
-
In this section, you create an endpoint for your model with the name **AI21-Jamba-1-5-Large-suawp**.
582
+
In this section, you create an endpoint for your model with the name **AI21-Jamba-1-5-Large-qwerty**.
524
583
525
584
1. Create the serverless endpoint. Use the following template to create an endpoint:
526
585
@@ -569,9 +628,7 @@ In this section, you create an endpoint for your model with the name **AI21-Jamb
1. The created endpoint uses key authentication for authorization. Use the following steps to get the keys associated with a given endpoint.
573
-
574
-
Use REST APIs to query this information.
631
+
1. The created endpoint uses key authentication for authorization. Get the keys associated with the given endpoint by using REST APIs to query this information.
575
632
576
633
1. If you need to consume this deployment from a different project or hub, or you plan to use Prompt flow to build intelligent applications, you need to create a connection to the standard deployment. To learn how to configure an existing standard deployment on a new project or hub, see [Consume deployed standard deployment from a different project orfrom Prompt flow](deploy-models-serverless-connect.md).
Copy file name to clipboardExpand all lines: articles/ai-foundry/includes/deploy-models-to-foundry-resources.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -10,4 +10,4 @@ ms.custom: include
10
10
---
11
11
12
12
> [!NOTE]
13
-
> We recommend that you deploy Azure AI Foundry Models to **Azure AI Foundry resources**, as this deployment method allows you to consume your models via a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. The endpoint follows the [Azure AI Model Inference API](/rest/api/aifoundry/modelinference/) which all the models in Foundry Models support. To learn how to deploy a Foundry Model to the Azure AI Foundry resources, see [Add and configure models to Azure AI Foundry Models](../model-inference/how-to/create-model-deployments.md).
13
+
> We recommend that you deploy Azure AI Foundry Models to **Azure AI Foundry resources** so that you can consume your deployments in the resource via a single endpoint with the same authentication and schema to generate inference. The endpoint follows the [Azure AI Model Inference API](/rest/api/aifoundry/modelinference/) which all the Foundry Models support. To learn how to deploy a Foundry Model to the Azure AI Foundry resources, see [Add and configure models to Azure AI Foundry Models](../model-inference/how-to/create-model-deployments.md).
0 commit comments