Skip to content

Commit 80dfef5

Browse files
committed
Foundry Models and Azure Direct Models
1 parent 5f4eab2 commit 80dfef5

File tree

26 files changed

+61
-61
lines changed

26 files changed

+61
-61
lines changed

articles/ai-foundry/concepts/content-filtering.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ author: PatrickFarley
2626

2727
The content filtering system is powered by [Azure AI Content Safety](../../ai-services/content-safety/overview.md), and it works by running both the prompt input and completion output through a set of classification models designed to detect and prevent the output of harmful content. Variations in API configurations and application design might affect completions and thus filtering behavior.
2828

29-
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **serverless APIs** have content filtering enabled by default. To learn more about the default content filter enabled for serverless APIs, see [Guardrails & controls for models curated by Azure AI in the model catalog](model-catalog-content-safety.md).
29+
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **serverless APIs** have content filtering enabled by default. To learn more about the default content filter enabled for serverless APIs, see [Guardrails & controls for Azure Direct Models in the model catalog](model-catalog-content-safety.md).
3030

3131
## Language support
3232

articles/ai-foundry/concepts/deployments-overview.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ Deployment options vary depending on the model offering:
2525

2626
Azure AI Foundry offers four different deployment options:
2727

28-
|Name | Azure OpenAI | Azure AI model inference | Standard deployment | Managed compute |
28+
|Name | Azure OpenAI | Azure AI Foundry Models | Standard deployment | Managed compute |
2929
|-------------------------------|----------------------|-------------------|----------------|-----------------|
3030
| Which models can be deployed? | [Azure OpenAI models](../../ai-services/openai/concepts/models.md) | [Azure OpenAI models and Standard deployment](../../ai-foundry/model-inference/concepts/models.md) | [Standard deployment](../how-to/model-catalog-overview.md) | [Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute) |
3131
| Deployment resource | Azure OpenAI resource | Azure AI services resource | AI project resource | AI project resource |
@@ -37,7 +37,7 @@ Azure AI Foundry offers four different deployment options:
3737
| Key-less authentication | Yes | Yes | No | No |
3838
| Best suited when | You're planning to use only OpenAI models | You're planning to take advantage of the flagship models in Azure AI catalog, including OpenAI. | You're planning to use a single model from a specific provider (excluding OpenAI). | If you plan to use open models and you have enough compute quota available in your subscription. |
3939
| Billing bases | Token usage & [provisioned throughput units](../../ai-services/openai/concepts/provisioned-throughput.md) | Token usage | Token usage<sup>1</sup> | Compute core hours<sup>2</sup> |
40-
| Deployment instructions | [Deploy to Azure OpenAI](../how-to/deploy-models-openai.md) | [Deploy to Azure AI model inference](../model-inference/how-to/create-model-deployments.md) | [Deploy to Standard deployment](../how-to/deploy-models-serverless.md) | [Deploy to Managed compute](../how-to/deploy-models-managed.md) |
40+
| Deployment instructions | [Deploy to Azure OpenAI](../how-to/deploy-models-openai.md) | [Deploy to Foundry Models](../model-inference/how-to/create-model-deployments.md) | [Deploy to Standard deployment](../how-to/deploy-models-serverless.md) | [Deploy to Managed compute](../how-to/deploy-models-managed.md) |
4141

4242
<sup>1</sup> A minimal endpoint infrastructure is billed per minute. You aren't billed for the infrastructure that hosts the model in pay-as-you-go. After you delete the endpoint, no further charges accrue.
4343

@@ -50,7 +50,7 @@ Azure AI Foundry offers four different deployment options:
5050

5151
Azure AI Foundry encourages you to explore various deployment options and choose the one that best suites your business and technical needs. In general, Consider using the following approach to select a deployment option:
5252

53-
* Start with [Azure AI model inference](../../ai-foundry/model-inference/overview.md), which is the option with the largest scope. This option allows you to iterate and prototype faster in your application without having to rebuild your architecture each time you decide to change something. If you're using Azure AI Foundry hubs or projects, enable this option by [turning on the Azure AI model inference feature](../model-inference/how-to/quickstart-ai-project.md#configure-the-project-to-use-azure-ai-model-inference).
53+
* Start with [Foundry Models](../../ai-foundry/model-inference/overview.md), which is the option with the largest scope. This option allows you to iterate and prototype faster in your application without having to rebuild your architecture each time you decide to change something. If you're using Azure AI Foundry hubs or projects, enable this option by [turning on the Foundry Models feature](../model-inference/how-to/quickstart-ai-project.md#configure-the-project-to-use-azure-ai-model-inference).
5454

5555
* When you're looking to use a specific model:
5656

@@ -63,8 +63,8 @@ Azure AI Foundry encourages you to explore various deployment options and choose
6363

6464
## Related content
6565

66-
* [Configure your AI project to use Azure AI model inference](../../ai-foundry/model-inference/how-to/quickstart-ai-project.md)
67-
* [Add and configure models to Azure AI model inference](../model-inference/how-to/create-model-deployments.md)
66+
* [Configure your AI project to use Foundry Models](../../ai-foundry/model-inference/how-to/quickstart-ai-project.md)
67+
* [Add and configure models to Foundry Models](../model-inference/how-to/create-model-deployments.md)
6868
* [Deploy Azure OpenAI models with Azure AI Foundry](../how-to/deploy-models-openai.md)
6969
* [Deploy open models with Azure AI Foundry](../how-to/deploy-models-managed.md)
7070
* [Model catalog and collections in Azure AI Foundry portal](../how-to/model-catalog-overview.md)

articles/ai-foundry/concepts/model-catalog-content-safety.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
2-
title: Guardrails & controls for models curated by Azure AI in the model catalog
2+
title: Guardrails & controls for Azure Direct Models
33
titleSuffix: Azure AI Foundry
4-
description: Learn about Guardrails & controls for models deployed using serverless APIs, using Azure AI Foundry.
4+
description: Learn about Guardrails & controls for models deployed via standard deployment.
55
manager: scottpolly
66
ms.service: azure-ai-foundry
77
ms.topic: conceptual
@@ -13,7 +13,7 @@ reviewer: ositanachi
1313
ms.custom:
1414
---
1515

16-
# Guardrails & controls for models curated by Azure AI in the model catalog
16+
# Guardrails & controls for Azure Direct Models
1717

1818
[!INCLUDE [feature-preview](../includes/feature-preview.md)]
1919

@@ -24,7 +24,7 @@ In this article, learn about Guardrails & controls capabilities for models from
2424

2525
Azure AI uses a default configuration of [Azure AI Content Safety](/azure/ai-services/content-safety/overview) content filters to detect harmful content across four categories including hate and fairness, self-harm, sexual, and violence for models deployed via serverless APIs. To learn more about content filtering, see [Understand harm categories](#understand-harm-categories).
2626

27-
The default content filtering configuration for text models is set to filter at the medium severity threshold, filtering any detected content at this level or higher. For image models, the default content filtering configuration is set at the low configuration threshold, filtering at this level or higher. For models deployed using the [Azure AI model inference service](../../ai-foundry/model-inference/how-to/configure-content-filters.md), you can create configurable filters by selecting the **Content filters** tab within the **Guardrails & controls** page of the Azure AI Foundry portal.
27+
The default content filtering configuration for text models is set to filter at the medium severity threshold, filtering any detected content at this level or higher. For image models, the default content filtering configuration is set at the low configuration threshold, filtering at this level or higher. For models deployed using the [Azure AI Foundry Models](../../ai-foundry/model-inference/how-to/configure-content-filters.md), you can create configurable filters by selecting the **Content filters** tab within the **Guardrails & controls** page of the Azure AI Foundry portal.
2828

2929
> [!TIP]
3030
> Content filtering isn't available for certain model types that are deployed via serverless APIs. These model types include embedding models and time series models.
@@ -34,7 +34,7 @@ Content filtering occurs synchronously as the service processes prompts to gener
3434
- When you first deploy a language model
3535
- Later, by selecting the content filtering toggle on the deployment details page
3636

37-
Suppose you decide to use an API other than the [Azure AI Model Inference API](/azure/ai-studio/reference/reference-model-inference-api) to work with a model that is deployed via a serverless API. In such a situation, content filtering isn't enabled unless you implement it separately by using Azure AI Content Safety. To get started with Azure AI Content Safety, see [Quickstart: Analyze text content](/azure/ai-services/content-safety/quickstart-text). You run a higher risk of exposing users to harmful content if you don't use content filtering when working with models that are deployed via serverless APIs.
37+
Suppose you decide to use an API other than the [Azure AI Foundry Models API](/azure/ai-studio/reference/reference-model-inference-api) to work with a model that is deployed via a serverless API. In such a situation, content filtering isn't enabled unless you implement it separately by using Azure AI Content Safety. To get started with Azure AI Content Safety, see [Quickstart: Analyze text content](/azure/ai-services/content-safety/quickstart-text). You run a higher risk of exposing users to harmful content if you don't use content filtering when working with models that are deployed via serverless APIs.
3838

3939
[!INCLUDE [content-safety-harm-categories](../includes/content-safety-harm-categories.md)]
4040

articles/ai-foundry/concepts/model-lifecycle-retirement.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,7 @@ Models labeled _Retired_ are no longer available for use. You can't create new d
5858

5959
- Models are labeled _Deprecated_ and remain in the deprecated state for at least 90 days before being moved to the retired state. During this notification period, you can migrate any existing deployments to newer or replacement models.
6060

61-
- For each subscription that has a model deployed as a severless API or deployed to the Azure AI model inference, members of the _owner_, _contributor_, _reader_, monitoring contributor_, and _monitoring reader_ roles receive a notification when a model deprecation is announced. The notification contains the dates when the model enters legacy, deprecated, and retired states. The notification might provide information about possible replacement model options, if applicable.
61+
- For each subscription that has a model deployed as a standard deployment or deployed to the Azure AI model inference, members of the _owner_, _contributor_, _reader_, monitoring contributor_, and _monitoring reader_ roles receive a notification when a model deprecation is announced. The notification contains the dates when the model enters legacy, deprecated, and retired states. The notification might provide information about possible replacement model options, if applicable.
6262

6363
The following tables list the timelines for models that are on track for retirement. The specified dates are in UTC time.
6464

articles/ai-foundry/includes/content-safety-serverless-models.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ ms.custom: include file
1313
# Also used in Azure Machine Learning documentation
1414
---
1515

16-
For language models deployed via standard deployment, Azure AI implements a default configuration of [Azure AI Content Safety](../../ai-services/content-safety/overview.md) text moderation filters that detect harmful content such as hate, self-harm, sexual, and violent content. To learn more about content filtering, see [Guardrails & controls for models curated by Azure AI in the model catalog](../concepts/model-catalog-content-safety.md).
16+
For language models deployed via standard deployment, Azure AI implements a default configuration of [Azure AI Content Safety](../../ai-services/content-safety/overview.md) text moderation filters that detect harmful content such as hate, self-harm, sexual, and violent content. To learn more about content filtering, see [Guardrails & controls for Azure Direct Models in the model catalog](../concepts/model-catalog-content-safety.md).
1717

1818
> [!TIP]
1919
> Content filtering is not available for certain model types that are deployed via serverless APIs. These model types include embedding models and time series models.

articles/ai-foundry/model-inference/includes/use-chat-completions/csharp.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ zone_pivot_groups: azure-ai-inference-samples
1616

1717
[!INCLUDE [Feature preview](~/reusable-content/ce-skilling/azure/includes/ai-studio/includes/feature-preview.md)]
1818

19-
This article explains how to use chat completions API with models deployed to Azure AI model inference in Azure AI services.
19+
This article explains how to use chat completions API with models deployed in Azure AI Foundry Models.
2020

2121
## Prerequisites
2222

@@ -77,7 +77,7 @@ Response<ChatCompletions> response = client.Complete(requestOptions);
7777
```
7878

7979
> [!NOTE]
80-
> Some models don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
80+
> Some models don't support system messages (`role="system"`). When you use the Foundry Models API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
8181
8282
The response is as follows, where you can see the model's usage statistics:
8383

@@ -158,7 +158,7 @@ StreamMessageAsync(client).GetAwaiter().GetResult();
158158

159159
#### Explore more parameters supported by the inference client
160160

161-
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).
161+
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Foundry Models API reference](https://aka.ms/azureai/modelinference).
162162

163163
```csharp
164164
requestOptions = new ChatCompletionsOptions()
@@ -212,7 +212,7 @@ Console.WriteLine($"Response: {response.Value.Content}");
212212

213213
### Pass extra parameters to the model
214214

215-
The Azure AI Model Inference API allows you to pass extra parameters to the model. The following code example shows how to pass the extra parameter `logprobs` to the model.
215+
The Foundry Models API allows you to pass extra parameters to the model. The following code example shows how to pass the extra parameter `logprobs` to the model.
216216

217217

218218
```csharp
@@ -230,11 +230,11 @@ response = client.Complete(requestOptions, extraParams: ExtraParameters.PassThro
230230
Console.WriteLine($"Response: {response.Value.Content}");
231231
```
232232

233-
Before you pass extra parameters to the Azure AI model inference API, make sure your model supports those extra parameters. When the request is made to the underlying model, the header `extra-parameters` is passed to the model with the value `pass-through`. This value tells the endpoint to pass the extra parameters to the model. Use of extra parameters with the model doesn't guarantee that the model can actually handle them. Read the model's documentation to understand which extra parameters are supported.
233+
Before you pass extra parameters to the Foundry Models API, make sure your model supports those extra parameters. When the request is made to the underlying model, the header `extra-parameters` is passed to the model with the value `pass-through`. This value tells the endpoint to pass the extra parameters to the model. Use of extra parameters with the model doesn't guarantee that the model can actually handle them. Read the model's documentation to understand which extra parameters are supported.
234234

235235
### Use tools
236236

237-
Some models support the use of tools, which can be an extraordinary resource when you need to offload specific tasks from the language model and instead rely on a more deterministic system or even a different language model. The Azure AI Model Inference API allows you to define tools in the following way.
237+
Some models support the use of tools, which can be an extraordinary resource when you need to offload specific tasks from the language model and instead rely on a more deterministic system or even a different language model. The Foundry Models API allows you to define tools in the following way.
238238

239239
The following code example creates a tool definition that is able to look from flight information from two different cities.
240240

@@ -363,7 +363,7 @@ response = client.Complete(requestOptions);
363363

364364
### Apply Guardrails and controls
365365

366-
The Azure AI model inference API supports [Azure AI Content Safety](https://aka.ms/azureaicontentsafety). When you use deployments with Azure AI Content Safety turned on, inputs and outputs pass through an ensemble of classification models aimed at detecting and preventing the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions.
366+
The Foundry Models API supports [Azure AI Content Safety](https://aka.ms/azureaicontentsafety). When you use deployments with Azure AI Content Safety turned on, inputs and outputs pass through an ensemble of classification models aimed at detecting and preventing the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions.
367367

368368
The following example shows how to handle events when the model detects harmful content in the input prompt.
369369

0 commit comments

Comments
 (0)