Skip to content

Commit 11dd09c

Browse files
committed
Merge branch 'release-build-ai-foundry' of github.com:MicrosoftDocs/azure-ai-docs-pr into sdg-release-dan-qs
2 parents 7c429de + 340517b commit 11dd09c

File tree

120 files changed

+3739
-870
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

120 files changed

+3739
-870
lines changed

articles/ai-foundry/.openpublishing.redirection.ai-studio.json

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -215,6 +215,11 @@
215215
"redirect_url": "/azure/ai-foundry/concepts/models-featured#mistral-ai",
216216
"redirect_document_id": false
217217
},
218+
{
219+
"source_path_from_root": "/articles/ai-foundry/how-to/model-catalog-overview.md",
220+
"redirect_url": "/azure/ai-foundry/concepts/foundry-models-overview",
221+
"redirect_document_id": false
222+
},
218223
{
219224
"source_path_from_root": "/articles/ai-studio/how-to/deploy-models-mistral-open.md",
220225
"redirect_url": "/azure/ai-foundry/how-to/deploy-models-mistral-open",

articles/ai-foundry/concepts/content-filtering.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ author: PatrickFarley
2626

2727
The content filtering system is powered by [Azure AI Content Safety](../../ai-services/content-safety/overview.md), and it works by running both the prompt input and completion output through a set of classification models designed to detect and prevent the output of harmful content. Variations in API configurations and application design might affect completions and thus filtering behavior.
2828

29-
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **serverless APIs** have content filtering enabled by default. To learn more about the default content filter enabled for serverless APIs, see [Guardrails & controls for Azure Direct Models in the model catalog](model-catalog-content-safety.md).
29+
With Azure OpenAI model deployments, you can use the default content filter or create your own content filter (described later on). Models available through **standard deployments** have content filtering enabled by default. To learn more about the default content filter enabled for standard deployments, see [Content safety for models curated by Azure AI in the model catalog](model-catalog-content-safety.md).
3030

3131
## Language support
3232

articles/ai-foundry/concepts/deployments-overview.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ The model catalog in Azure AI Foundry portal is the hub to discover and use a wi
2020
Deployment options vary depending on the model offering:
2121

2222
* **Azure OpenAI in Azure AI Foundry Models:** The latest OpenAI models that have enterprise features from Azure with flexible billing options.
23-
* **Standard deployment:** These models don't require compute quota from your subscription and are billed per token in a pay-as-you-go fashion.
23+
* **Standard deployment:** These models don't require compute quota from your subscription and are billed per token in a serverless pay per token offer.
2424
* **Open and custom models:** The model catalog offers access to a large variety of models across modalities, including models of open access. You can host open models in your own subscription with a managed infrastructure, virtual machines, and the number of instances for capacity management.
2525

2626
Azure AI Foundry offers four different deployment options:
@@ -39,7 +39,7 @@ Azure AI Foundry offers four different deployment options:
3939
| Billing bases | Token usage & [provisioned throughput units](../../ai-services/openai/concepts/provisioned-throughput.md) | Token usage | Token usage<sup>1</sup> | Compute core hours<sup>2</sup> |
4040
| Deployment instructions | [Deploy to Azure OpenAI](../how-to/deploy-models-openai.md) | [Deploy to Foundry Models](../model-inference/how-to/create-model-deployments.md) | [Deploy to Standard deployment](../how-to/deploy-models-serverless.md) | [Deploy to Managed compute](../how-to/deploy-models-managed.md) |
4141

42-
<sup>1</sup> A minimal endpoint infrastructure is billed per minute. You aren't billed for the infrastructure that hosts the model in pay-as-you-go. After you delete the endpoint, no further charges accrue.
42+
<sup>1</sup> A minimal endpoint infrastructure is billed per minute. You aren't billed for the infrastructure that hosts the model in standard deployment. After you delete the endpoint, no further charges accrue.
4343

4444
<sup>2</sup> Billing is on a per-minute basis, depending on the product tier and the number of instances used in the deployment since the moment of creation. After you delete the endpoint, no further charges accrue.
4545

articles/ai-foundry/concepts/evaluation-evaluators/agent-evaluators.md

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,12 @@ Agents are powerful productivity assistants. They can plan, make decisions, and
1919

2020
Agents emit messages, and providing the above inputs typically require parsing messages and extracting the relevant information. If you're building agents using Azure AI Agent Service, we provide native integration for evaluation that directly takes their agent messages. To learn more, see an [end-to-end example of evaluating agents in Azure AI Agent Service](https://aka.ms/e2e-agent-eval-sample).
2121

22+
Besides `IntentResolution`, `ToolCallAccuracy`, `TaskAdherence` specific to agentic workflows, you can also assess other quality as well as safety aspects of your agentic workflows, leveraging out comprehensive suite of built-in evaluators. We support this list of evaluators for Azure AI agent messages from our converter:
23+
- **Quality**: `IntentResolution`, `ToolCallAccuracy`, `TaskAdherence`, `Relevance`, `Coherence`, `Fluency`
24+
- **Safety**: `CodeVulnerabilities`, `Violence`, `Self-harm`, `Sexual`, `HateUnfairness`, `IndirectAttack`, `ProtectedMaterials`.
25+
26+
We will show examples of `IntentResolution`, `ToolCallAccuracy`, `TaskAdherence` here. See more examples in [evaluating Azure AI agents](../../how-to/develop/agent-evaluate-sdk.md#evaluate-azure-ai-agents) for other evaluators with Azure AI agent message support.
27+
2228
## Model configuration for AI-assisted evaluators
2329

2430
For reference in the following code snippets, the AI-assisted evaluators use a model configuration for the LLM-judge:
@@ -37,6 +43,9 @@ model_config = AzureOpenAIModelConfiguration(
3743
)
3844
```
3945

46+
> [!TIP]
47+
> We recommend using `o3-mini` for a balance of reasoning capability and cost efficiency.
48+
4049
## Intent resolution
4150

4251
`IntentResolutionEvaluator` measures how well the system identifies and understands a user's request, including how well it scopes the user’s intent, asks clarifying questions, and reminds end users of its scope of capabilities. Higher score means better identification of user intent.
@@ -83,6 +92,9 @@ If you're building agents outside of Azure AI Agent Serice, this evaluator accep
8392

8493
`ToolCallAccuracyEvaluator` measures an agent's ability to select appropriate tools, extract, and process correct parameters from previous steps of the agentic workflow. It detects whether each tool call made is accurate (binary) and reports back the average scores, which can be interpreted as a passing rate across tool calls made.
8594

95+
> [!NOTE]
96+
> `ToolCallAccuracyEvaluator` only supports Azure AI Agent's Function Tool evaluation, but does not support Built-in Tool evaluation. The agent messages must have at least one Function Tool actually called to be evaluated.
97+
8698
### Tool call accuracy example
8799

88100
```python

articles/ai-foundry/concepts/evaluation-evaluators/general-purpose-evaluators.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -35,6 +35,9 @@ model_config = AzureOpenAIModelConfiguration(
3535
)
3636
```
3737

38+
> [!TIP]
39+
> We recommend using `o3-mini` for a balance of reasoning capability and cost efficiency.
40+
3841
## Coherence
3942

4043
`CoherenceEvaluator` measures the logical and orderly presentation of ideas in a response, allowing the reader to easily follow and understand the writer's train of thought. A coherent response directly addresses the question with clear connections between sentences and paragraphs, using appropriate transitions and a logical sequence of ideas. Higher scores mean better coherence.

articles/ai-foundry/concepts/evaluation-evaluators/rag-evaluators.md

Lines changed: 33 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -39,6 +39,9 @@ model_config = AzureOpenAIModelConfiguration(
3939
)
4040
```
4141

42+
> [!TIP]
43+
> We recommend using `o3-mini` for a balance of reasoning capability and cost efficiency.
44+
4245
## Retrieval
4346

4447
Retrieval quality is very important given its upstream role in RAG: if the retrieval quality is poor and the response requires corpus-specific knowledge, there's less chance your LLM model gives you a satisfactory answer. `RetrievalEvaluator` measures the **textual quality** of retrieval results with an LLM without requiring ground truth (also known as query relevance judgment), which provides value compared to `DocumentRetrievalEvaluator` measuring `ndcg`, `xdcg`, `fidelity`, and other classical information retrieval metrics that require ground truth. This metric focuses on how relevant the context chunks (encoded as a string) are to address a query and how the most relevant context chunks are surfaced at the top of the list.
@@ -90,69 +93,74 @@ Retrieval quality is very important given its upstream role in RAG: if the retri
9093
```python
9194
from azure.ai.evaluation import DocumentRetrievalEvaluator
9295

96+
# these query_relevance_label are given by your human- or LLM-judges.
9397
retrieval_ground_truth = [
9498
{
9599
"document_id": "1",
96-
"query_relevance_judgement": 4
100+
"query_relevance_label": 4
97101
},
98102
{
99103
"document_id": "2",
100-
"query_relevance_judgement": 2
104+
"query_relevance_label": 2
101105
},
102106
{
103107
"document_id": "3",
104-
"query_relevance_judgement": 3
108+
"query_relevance_label": 3
105109
},
106110
{
107111
"document_id": "4",
108-
"query_relevance_judgement": 1
112+
"query_relevance_label": 1
109113
},
110114
{
111115
"document_id": "5",
112-
"query_relevance_judgement": 0
116+
"query_relevance_label": 0
113117
},
114118
]
119+
# the min and max of the label scores are inputs to document retrieval evaluator
120+
ground_truth_label_min = 0
121+
ground_truth_label_max = 4
115122

123+
# these relevance scores come from your search retrieval system
116124
retrieved_documents = [
117125
{
118126
"document_id": "2",
119-
"query_relevance_judgement": 45.1
127+
"relevance_score": 45.1
120128
},
121129
{
122130
"document_id": "6",
123-
"query_relevance_judgement": 35.8
131+
"relevance_score": 35.8
124132
},
125133
{
126134
"document_id": "3",
127-
"query_relevance_judgement": 29.2
135+
"relevance_score": 29.2
128136
},
129137
{
130138
"document_id": "5",
131-
"query_relevance_judgement": 25.4
139+
"relevance_score": 25.4
132140
},
133141
{
134142
"document_id": "7",
135-
"query_relevance_judgement": 18.8
143+
"relevance_score": 18.8
136144
},
137145
]
138146

139-
default_threshold = {
140-
"ndcg@3": 0.5,
141-
"xdcg@3": 0.5,
142-
"fidelity": 0.5,
143-
"top1_relevance": 50,
144-
"top3_max_relevance": 50,
145-
"total_retrieved_documents": 50,
146-
"total_ground_truth_documents": 50,
147-
}
148-
149-
document_retrieval_evaluator = DocumentRetrievalEvaluator(threshold=default_threshold)
147+
document_retrieval_evaluator = DocumentRetrievalEvaluator(
148+
ground_truth_label_min=ground_truth_label_min,
149+
ground_truth_label_max=ground_truth_label_max,
150+
ndcg_threshold = 0.5,
151+
xdcg_threshold = 50.0,
152+
fidelity_threshold = 0.5,
153+
top1_relevance_threshold = 50.0,
154+
top3_max_relevance_threshold = 50.0,
155+
total_retrieved_documents_threshold = 50,
156+
total_ground_truth_documents_threshold = 50
157+
)
150158
document_retrieval_evaluator(retrieval_ground_truth=retrieval_ground_truth, retrieved_documents=retrieved_documents)
151159
```
152160

153161
### Document retrieval output
154162

155-
The numerical score on a likert scale (integer 1 to 5) and a higher score is better. Given a numerical threshold (default to 3), we also output "pass" if the score <= threshold, or "fail" otherwise. Using the reason field can help you understand why the score is high or low.
163+
All numerical scores have `high_is_better=True` except for `holes` and `holes_ratio` which have `high_is_better=False`. Given a numerical threshold (default to 3), we also output "pass" if the score <= threshold, or "fail" otherwise.
156164

157165
```python
158166
{
@@ -163,15 +171,16 @@ The numerical score on a likert scale (integer 1 to 5) and a higher score is bet
163171
"top3_max_relevance": 2,
164172
"holes": 30,
165173
"holes_ratio": 0.6000000000000001,
166-
"holes_is_higher_better": False,
167-
"holes_ratio_is_higher_better": False,
174+
"holes_higher_is_better": False,
175+
"holes_ratio_higher_is_better": False,
168176
"total_retrieved_documents": 50,
169177
"total_groundtruth_documents": 1565,
170178
"ndcg@3_result": "pass",
171179
"xdcg@3_result": "pass",
172180
"fidelity_result": "fail",
173181
"top1_relevance_result": "fail",
174182
"top3_max_relevance_result": "fail",
183+
# omitting more fields ...
175184
}
176185
```
177186

articles/ai-foundry/concepts/evaluation-evaluators/textual-similarity-evaluators.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,9 @@ model_config = AzureOpenAIModelConfiguration(
3333
)
3434
```
3535

36+
> [!TIP]
37+
> We recommend using `o3-mini` for a balance of reasoning capability and cost efficiency.
38+
3639
## Similarity
3740

3841
`SimilarityEvaluator` measures the degrees of semantic similarity between the generated text and its ground truth with respect to a query. Compared to other text-similarity metrics that require ground truths, this metric focuses on semantics of a response (instead of simple overlap in tokens or n-grams) and also considers the broader context of a query.

articles/ai-foundry/concepts/fine-tuning-overview.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -84,15 +84,15 @@ It's important to call out that fine-tuning is heavily dependent on the quality
8484
## Supported models for fine-tuning
8585

8686
Now that you know when to use fine-tuning for your use case, you can go to Azure AI Foundry to find models available to fine-tune.
87-
For some models in the model catalog, fine-tuning is available by using a serverless API, or a managed compute (preview), or both.
87+
For some models in the model catalog, fine-tuning is available by using a standard deployment, or a managed compute (preview), or both.
8888

89-
Fine-tuning is available in specific Azure regions for some models that are deployed via serverless APIs. To fine-tune such models, a user must have a hub/project in the region where the model is available for fine-tuning. See [Region availability for models in serverless API endpoints](../how-to/deploy-models-serverless-availability.md) for detailed information.
89+
Fine-tuning is available in specific Azure regions for some models that are deployed via standard deployments. To fine-tune such models, a user must have a hub/project in the region where the model is available for fine-tuning. See [Region availability for models in standard deployment](../how-to/deploy-models-serverless-availability.md) for detailed information.
9090

9191
For more information on fine-tuning using a managed compute (preview), see [Fine-tune models using managed compute (preview)](../how-to/fine-tune-managed-compute.md).
9292

9393
For details about Azure OpenAI in Azure AI Foundry Models that are available for fine-tuning, see the [Azure OpenAI in Foundry Models documentation](../../ai-services/openai/concepts/models.md#fine-tuning-models) or the [Azure OpenAI models table](#fine-tuning-azure-openai-models) later in this guide.
9494

95-
For the Azure OpenAI Service models that you can fine tune, supported regions for fine-tuning include North Central US, Sweden Central, and more.
95+
For the Azure OpenAI Service models that you can fine tune, supported regions for fine-tuning include North Central US, Sweden Central, and more.
9696

9797
### Fine-tuning Azure OpenAI models
9898

@@ -102,5 +102,5 @@ For the Azure OpenAI Service models that you can fine tune, supported regions f
102102

103103
- [Fine-tune models using managed compute (preview)](../how-to/fine-tune-managed-compute.md)
104104
- [Fine-tune an Azure OpenAI model in Azure AI Foundry portal](../../ai-services/openai/how-to/fine-tuning.md?context=/azure/ai-studio/context/context)
105-
- [Fine-tune models using serverless API](../how-to/fine-tune-serverless.md)
105+
- [Fine-tune models using standard deployment](../how-to/fine-tune-serverless.md)
106106

0 commit comments

Comments
 (0)