Skip to content

Commit d1d9fd4

Browse files
Merge pull request #7752 from MicrosoftDocs/main
Auto Publish – main to live - 2025-10-21 17:09 UTC
2 parents 2966cc5 + cc53a2d commit d1d9fd4

File tree

4 files changed

+33
-27
lines changed

4 files changed

+33
-27
lines changed

articles/ai-foundry/how-to/develop/semantic-kernel.md

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,11 @@
11
---
2-
title: Develop applications with Semantic Kernel and Azure AI Foundry
2+
title: Develop Applications with Semantic Kernel and Azure AI Foundry
33
titleSuffix: Azure AI Foundry
4-
description: Develop applications with Semantic Kernel and Azure AI Foundry.
4+
description: Learn how to Develop applications with Semantic Kernel and Azure AI Foundry with models deployed from the Azure AI model catalog.
55
author: lgayhardt
66
ms.author: lagayhar
77
ms.reviewer: taochen
8-
ms.date: 02/27/2025
8+
ms.date: 10/20/2025
99
ms.topic: how-to
1010
ms.service: azure-ai-foundry
1111
---
@@ -17,27 +17,27 @@ In this article, you learn how to use [Semantic Kernel](/semantic-kernel/overvie
1717
## Prerequisites
1818

1919
- [!INCLUDE [azure-subscription](../../includes/azure-subscription.md)]
20-
- An Azure AI project as explained at [Create a project in Azure AI Foundry portal](../create-projects.md).
21-
- A model supporting the [Azure AI Model Inference API](../../../ai-foundry/model-inference/reference/reference-model-inference-api.md?tabs=python) deployed. In this example, we use a `Mistral-Large` deployment, but use any model of your preference. For using embeddings capabilities in LlamaIndex, you need an embedding model like `cohere-embed-v3-multilingual`.
20+
- An Azure AI project as explained at [Create a project for Azure AI Foundry](../create-projects.md).
21+
- A model that supports the [Azure AI Model Inference API](../../../ai-foundry/model-inference/reference/reference-model-inference-api.md?tabs=python) deployed. This article uses a `Mistral-Large` deployment. You can use any model. For using embeddings capabilities in LlamaIndex, you need an embedding model like `cohere-embed-v3-multilingual`.
2222

2323
- You can follow the instructions at [Deploy models as serverless API deployments](../deploy-models-serverless.md).
2424

2525
- Python **3.10** or later installed, including pip.
26-
- Semantic Kernel installed. You can do it with:
26+
- Semantic Kernel installed. You can use the following command:
2727

2828
```bash
2929
pip install semantic-kernel
3030
```
3131

32-
- In this example, we're working with the Model Inference API, so we need to install the relevant Azure dependencies. You can do it with:
32+
- This article uses the Model Inference API, so install the relevant Azure dependencies. You can use the following command:
3333

3434
```bash
3535
pip install semantic-kernel[azure]
3636
```
3737

3838
## Configure the environment
3939

40-
To use LLMs deployed in Azure AI Foundry portal, you need the endpoint and credentials to connect to it. Follow these steps to get the information you need from the model you want to use:
40+
To use language models deployed in Azure AI Foundry portal, you need the endpoint and credentials to connect to your project. Follow these steps to get the information you need from the model:
4141

4242
[!INCLUDE [tip-left-pane](../../includes/tip-left-pane.md)]
4343

@@ -49,14 +49,14 @@ To use LLMs deployed in Azure AI Foundry portal, you need the endpoint and crede
4949
> [!TIP]
5050
> If your model was deployed with Microsoft Entra ID support, you don't need a key.
5151

52-
In this scenario, we placed both the endpoint URL and key in the following environment variables:
52+
This example uses environment variables for both the endpoint URL and key:
5353

5454
```bash
5555
export AZURE_AI_INFERENCE_ENDPOINT="<your-model-endpoint-goes-here>"
5656
export AZURE_AI_INFERENCE_API_KEY="<your-key-goes-here>"
5757
```
5858

59-
Once configured, create a client to connect to the endpoint:
59+
After you configure the endpoint and key, create a client to connect to the endpoint:
6060

6161
```python
6262
from semantic_kernel.connectors.ai.azure_ai_inference import AzureAIInferenceChatCompletion
@@ -65,7 +65,7 @@ chat_completion_service = AzureAIInferenceChatCompletion(ai_model_id="<deploymen
6565
```
6666

6767
> [!TIP]
68-
> The client automatically reads the environment variables `AZURE_AI_INFERENCE_ENDPOINT` and `AZURE_AI_INFERENCE_API_KEY` to connect to the model. However, you can also pass the endpoint and key directly to the client via the `endpoint` and `api_key` parameters on the constructor.
68+
> The client automatically reads the environment variables `AZURE_AI_INFERENCE_ENDPOINT` and `AZURE_AI_INFERENCE_API_KEY` to connect to the model. You could instead pass the endpoint and key directly to the client by using the `endpoint` and `api_key` parameters on the constructor.
6969

7070
Alternatively, if your endpoint support Microsoft Entra ID, you can use the following code to create the client:
7171

@@ -80,7 +80,7 @@ chat_completion_service = AzureAIInferenceChatCompletion(ai_model_id="<deploymen
8080
```
8181

8282
> [!NOTE]
83-
> When using Microsoft Entra ID, make sure that the endpoint was deployed with that authentication method and that you have the required permissions to invoke it.
83+
> If you use Microsoft Entra ID, make sure that the endpoint was deployed with that authentication method and that you have the required permissions to invoke it.
8484

8585
### Azure OpenAI models
8686

@@ -104,7 +104,7 @@ chat_completion_service = AzureAIInferenceChatCompletion(
104104
105105
## Inference parameters
106106
107-
You can configure how inference is performed by using the `AzureAIInferenceChatPromptExecutionSettings` class:
107+
You can configure how to perform inference by using the `AzureAIInferenceChatPromptExecutionSettings` class:
108108
109109
```python
110110
from semantic_kernel.connectors.ai.azure_ai_inference import AzureAIInferenceChatPromptExecutionSettings
@@ -119,7 +119,7 @@ execution_settings = AzureAIInferenceChatPromptExecutionSettings(
119119
120120
## Calling the service
121121
122-
Let's first call the chat completion service with a simple chat history:
122+
First, call the chat completion service with a simple chat history:
123123
124124
> [!TIP]
125125
> Semantic Kernel is an asynchronous library, so you need to use the asyncio library to run the code.

articles/ai-foundry/openai/includes/fine-tuning-studio.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -216,6 +216,9 @@ Your job might be queued behind other jobs on the system. Training your model ca
216216

217217
When each training epoch completes a checkpoint is generated. A checkpoint is a fully functional version of a model which can both be deployed and used as the target model for subsequent fine-tuning jobs. Checkpoints can be particularly useful, as they may provide snapshots prior to overfitting. When a fine-tuning job completes you will have the three most recent versions of the model available to deploy.
218218

219+
> [!NOTE]
220+
> When copying a checkpoint from a source account, the same checkpoint name is retained in the destination account. Ensure you use this exact name for fine-tuning, deployment, or any other operation in the destination account. This checkpoint will not appear in the UI or in the `list checkpoints` API.
221+
219222
## Pause and resume
220223

221224
You can track progress in both fine-tuning views of the AI Foundry portal. You'll see your job go through the same statuses as normal fine tuning jobs (queued, running, succeeded).

articles/ai-services/language-service/concepts/model-lifecycle.md

Lines changed: 15 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -35,18 +35,18 @@ By default, API and SDK requests use the latest Generally Available model. To us
3535
3636
Use the following table to find which model versions support each feature:
3737

38-
| Feature | Supported generally available (GA) version | Latest supported preview versions |
39-
|--|--|--|
40-
| Sentiment Analysis and opinion mining | `latest` | |
41-
| Language Detection | `latest` | |
42-
| Entity Linking | `latest` | |
43-
| Named Entity Recognition (NER) | `latest` | `2025-08-01-preview` |
44-
| Personally Identifiable Information (PII) detection | `latest` | `2025-08-01-preview` |
45-
| PII detection for conversations | `latest` | `2024-11-01-preview` |
46-
| Question answering | `latest` | |
47-
| Text Analytics for health | `latest` | `2023-04-15-preview` |
48-
| Key phrase extraction | `latest` | |
49-
| Summarization | `latest` | `2025-06-10-preview` (only available for `issue` and `resolution` aspects in conversation summarization) |
38+
| Feature | Supported generally available (GA) version | Latest supported preview versions | Other supported verision |
39+
|--|--|--|--|
40+
| Sentiment Analysis and opinion mining | `latest` | | |
41+
| Language Detection | `latest` | | |
42+
| Entity Linking | `latest` | | |
43+
| Named Entity Recognition (NER) | `latest` | `2025-08-01-preview` | `2025-04-15-preview` |
44+
| Personally Identifiable Information (PII) detection | `latest` | `2025-08-01-preview` | `2025-04-15-preview` |
45+
| PII detection for conversations | `latest` | `2024-11-01-preview` | `2023-04-15-preview` |
46+
| Question answering | `latest` | | |
47+
| Text Analytics for health | `latest` | `2023-04-15-preview` | |
48+
| Key phrase extraction | `latest` | | |
49+
| Summarization | `latest` | `2025-06-10-preview` (only available for `issue` and `resolution` aspects in conversation summarization) | |
5050

5151

5252
## Custom features
@@ -90,6 +90,9 @@ Use the following table to find which API versions support each feature:
9090
| Conversational language understanding| `2022-05-01`, `2022-10-01-preview`, `2023-04-01` |`2023-04-01` |`2022-10-01-preview` |
9191
| Custom named entity recognition | `2022-05-01`, `2022-10-01-preview`, `2023-04-01`, `2023-04-15`, `2023-04-15-preview`|`2023-04-15` |`2023-04-15-preview` |
9292
| Orchestration workflow | `2022-05-01`, `2022-10-01-preview`, `2023-04-01` |`2023-04-01` |`2022-10-01-preview` |
93+
| Named Entity Recognition (NER) | `2025-05-15-preview`, `2024-11-01 (GA)`,`2024-11-15-preview` | `2024-11-01 (GA)` | `2025-05-15-preview` |
94+
| Personally Identifiable Information (PII) detection | `2025-05-15-preview`,`2024-11-01 (GA)`,`2024-11-15-preview` | `2024-11-01 (GA)` | `2025-05-15-preview` |
95+
| PII detection for conversations | `2025-05-15-preview`,`2024-11-01 (GA)`,`2024-11-15-preview` | `2024-11-01 (GA)` | `2025-05-15-preview` |
9396

9497
## Next steps
9598

articles/search/search-what-is-azure-search.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,7 +47,7 @@ Architecturally, a search service sits between the external data stores that con
4747

4848
![Azure AI Search architecture](media/search-what-is-azure-search/azure-search.svg "Azure AI Search architecture")
4949

50-
On the indexing side, if your content is on Azure, you can used indexers and skillsets for automated and AI-enriched indexing. Or, create a logic app workflow for equivalent automation over an even broader set of supported data sources.
50+
On the indexing side, if your content is on Azure, you can use indexers and skillsets for automated and AI-enriched indexing. Or, create a logic app workflow for equivalent automation over an even broader set of supported data sources.
5151

5252
On the retrieval side, your app can be an agent or tool, a bot, or any client that sends requests to a search index or knowledge agent.
5353

0 commit comments

Comments
 (0)