Skip to content

Commit d7bc455

Browse files
authored
Merge branch 'main' into 5091-5095-5100-kate-jp-joe
2 parents b772e56 + 850fa2f commit d7bc455

File tree

81 files changed

+660
-624
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

81 files changed

+660
-624
lines changed

.openpublishing.redirection.json

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -294,6 +294,21 @@
294294
"source_path_from_root": "/articles/ai-services/speech-service/text-to-speech-avatar/custom-avatar-endpoint.md",
295295
"redirect_url": "/azure/ai-services/speech-service/custom-avatar-create",
296296
"redirect_document_id": false
297+
},
298+
{
299+
"source_path_from_root": "/articles/ai-services/speech-service/migration-overview-neural-voice.md",
300+
"redirect_url": "/azure/ai-services/speech-service/custom-neural-voice",
301+
"redirect_document_id": false
302+
},
303+
{
304+
"source_path_from_root": "/articles/ai-services/speech-service/how-to-migrate-to-custom-neural-voice.md",
305+
"redirect_url": "/azure/ai-services/speech-service/custom-neural-voice",
306+
"redirect_document_id": false
307+
},
308+
{
309+
"source_path_from_root": "/articles/ai-services/speech-service/how-to-migrate-to-prebuilt-neural-voice.md",
310+
"redirect_url": "/azure/ai-services/speech-service/custom-neural-voice",
311+
"redirect_document_id": false
297312
}
298313
]
299314
}

articles/ai-foundry/how-to/connections-add.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,7 @@ Here's a table of some of the available connection types in Azure AI Foundry por
5353
|-------------------------------|:-------:|:--------------------------------------:|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
5454
| Azure AI Search | || Azure AI Search is an Azure resource that supports information retrieval over your vector and textual data stored in search indexes. |
5555
| Azure Storage | || Azure Storage is a cloud storage solution for storing unstructured data like documents, images, videos, and application installers. |
56-
| Azure Cosmos DB | || Azure Cosmos DB is a globally distributed, multi-model database service that offers low latency, high availability, and scalability across multiple geographical regions. |
56+
| Azure Cosmos DB | || Azure Cosmos DB is a globally distributed, multi-model database service that offers low latency, high availability, and scalability across multiple geographical regions. |
5757
| Azure OpenAI | | | Azure OpenAI is a service that provides access to OpenAI's models including the GPT-4o, GPT-4o mini, GPT-4, GPT-4 Turbo with Vision, GPT-3.5-Turbo, DALLE-3, and Embeddings model series with the security and enterprise capabilities of Azure. |
5858
| Application Insights | | | Azure Application Insights is a service within Azure Monitor that enables developers and DevOps teams to automatically detect performance anomalies, diagnose issues, and gain deep insights into application usage and behavior through powerful telemetry and analytics tools. |
5959
| API key | | | API Key connections handle authentication to your specified target on an individual basis. |

articles/ai-foundry/how-to/data-add.md

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.custom:
99
- build-2024
1010
- ignite-2024
1111
ms.topic: how-to
12-
ms.date: 02/11/2025
12+
ms.date: 05/21/2025
1313
ms.author: franksolomon
1414
author: fbsolo-ms1
1515
---
@@ -29,12 +29,10 @@ Data can help when you need these capabilities:
2929
> - **Lineage:** For any given data, you can view which jobs or prompt flow pipelines consume the data.
3030
> - **Ease-of-use:** An Azure AI Foundry data resembles web browser bookmarks (favorites). Instead of remembering long storage paths that *reference* your frequently-used data on Azure Storage, you can create a data *version* and then access that version of the asset with a friendly name.
3131
32-
## Prerequisites
3332

34-
To create and work with data, you need:
33+
## Prerequisites
3534

36-
- An Azure subscription. If you don't have one, create a [free account](https://azure.microsoft.com/free/).
37-
- An [Azure AI Foundry project](../how-to/create-projects.md).
35+
[!INCLUDE [hub-only-prereq](../includes/hub-only-prereq.md)]
3836

3937
## Create data
4038

articles/ai-foundry/how-to/develop/langchain.md

Lines changed: 24 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -51,8 +51,11 @@ To use LLMs deployed in Azure AI Foundry portal, you need the endpoint and crede
5151
[!INCLUDE [tip-left-pane](../../includes/tip-left-pane.md)]
5252
5353
1. Go to the [Azure AI Foundry](https://ai.azure.com/).
54+
5455
1. Open the project where the model is deployed, if it isn't already open.
56+
5557
1. Go to **Models + endpoints** and select the model you deployed as indicated in the prerequisites.
58+
5659
1. Copy the endpoint URL and the key.
5760

5861
:::image type="content" source="../../media/how-to/inference/serverless-endpoint-url-keys.png" alt-text="Screenshot of the option to copy endpoint URI and keys from an endpoint." lightbox="../../media/how-to/inference/serverless-endpoint-url-keys.png":::
@@ -63,11 +66,19 @@ To use LLMs deployed in Azure AI Foundry portal, you need the endpoint and crede
6366
In this scenario, we placed both the endpoint URL and key in the following environment variables:
6467
6568
```bash
66-
export AZURE_INFERENCE_ENDPOINT="<your-model-endpoint-goes-here>"
69+
export AZURE_INFERENCE_ENDPOINT="https://<resource>.services.ai.azure.com/models"
6770
export AZURE_INFERENCE_CREDENTIAL="<your-key-goes-here>"
6871
```
6972
70-
Once configured, create a client to connect to the endpoint. In this case, we're working with a chat completions model hence we import the class `AzureAIChatCompletionsModel`.
73+
Once configured, create a client to connect with the chat model by using the `init_chat_model`. For Azure OpenAI models, configure the client as indicated at [Using Azure OpenAI models](#using-azure-openai-models).
74+
75+
```python
76+
from langchain.chat_models import init_chat_model
77+
78+
llm = init_chat_model(model="mistral-large-2407", model_provider="azure_ai")
79+
```
80+
81+
You can also use the class `AzureAIChatCompletionsModel` directly.
7182
7283
```python
7384
import os
@@ -80,8 +91,8 @@ model = AzureAIChatCompletionsModel(
8091
)
8192
```
8293
83-
> [!TIP]
84-
> For Azure OpenAI models, configure the client as indicated at [Using Azure OpenAI models](#using-azure-openai-models).
94+
> [!CAUTION]
95+
> **Breaking change:** Parameter `model_name` was renamed `model` in version `0.1.3`.
8596
8697
You can use the following code to create the client if your endpoint supports Microsoft Entra ID:
8798
@@ -93,7 +104,7 @@ from langchain_azure_ai.chat_models import AzureAIChatCompletionsModel
93104
model = AzureAIChatCompletionsModel(
94105
endpoint=os.environ["AZURE_INFERENCE_ENDPOINT"],
95106
credential=DefaultAzureCredential(),
96-
model_name="mistral-large-2407",
107+
model="mistral-large-2407",
97108
)
98109
```
99110
@@ -111,7 +122,7 @@ from langchain_azure_ai.chat_models import AzureAIChatCompletionsModel
111122
model = AzureAIChatCompletionsModel(
112123
endpoint=os.environ["AZURE_INFERENCE_ENDPOINT"],
113124
credential=DefaultAzureCredentialAsync(),
114-
model_name="mistral-large-2407",
125+
model="mistral-large-2407",
115126
)
116127
```
117128
@@ -188,13 +199,13 @@ from langchain_azure_ai.chat_models import AzureAIChatCompletionsModel
188199
producer = AzureAIChatCompletionsModel(
189200
endpoint=os.environ["AZURE_INFERENCE_ENDPOINT"],
190201
credential=os.environ["AZURE_INFERENCE_CREDENTIAL"],
191-
model_name="mistral-large-2407",
202+
model="mistral-large-2407",
192203
)
193204
194205
verifier = AzureAIChatCompletionsModel(
195206
endpoint=os.environ["AZURE_INFERENCE_ENDPOINT"],
196207
credential=os.environ["AZURE_INFERENCE_CREDENTIAL"],
197-
model_name="mistral-small",
208+
model="mistral-small",
198209
)
199210
```
200211
@@ -271,7 +282,7 @@ from langchain_azure_ai.embeddings import AzureAIEmbeddingsModel
271282
embed_model = AzureAIEmbeddingsModel(
272283
endpoint=os.environ["AZURE_INFERENCE_ENDPOINT"],
273284
credential=os.environ['AZURE_INFERENCE_CREDENTIAL'],
274-
model_name="text-embedding-3-large",
285+
model="text-embedding-3-large",
275286
)
276287
```
277288
@@ -305,31 +316,15 @@ for doc in results:
305316
306317
## Using Azure OpenAI models
307318
308-
If you're using Azure OpenAI in Foundry Models or Foundry Models service with OpenAI models with `langchain-azure-ai` package, you might need to use `api_version` parameter to select a specific API version. The following example shows how to connect to an Azure OpenAI in Foundry Models deployment:
309-
310-
```python
311-
from langchain_azure_ai.chat_models import AzureAIChatCompletionsModel
312-
313-
llm = AzureAIChatCompletionsModel(
314-
endpoint="https://<resource>.openai.azure.com/openai/deployments/<deployment-name>",
315-
credential=os.environ["AZURE_INFERENCE_CREDENTIAL"],
316-
api_version="2024-05-01-preview",
317-
)
318-
```
319-
320-
> [!IMPORTANT]
321-
> Check which is the API version that your deployment is using. Using a wrong `api_version` or one not supported by the model results in a `ResourceNotFound` exception.
322-
323-
If the deployment is hosted in Azure AI Services, you can use the Foundry Models service:
319+
If you're using Azure OpenAI models with `langchain-azure-ai` package, use the following URL:
324320
325321
```python
326322
from langchain_azure_ai.chat_models import AzureAIChatCompletionsModel
327323
328324
llm = AzureAIChatCompletionsModel(
329-
endpoint="https://<resource>.services.ai.azure.com/models",
325+
endpoint="https://<resource>.openai.azure.com/openai/v1",
330326
credential=os.environ["AZURE_INFERENCE_CREDENTIAL"],
331-
model_name="<model-name>",
332-
api_version="2024-05-01-preview",
327+
model="gpt-4o"
333328
)
334329
```
335330
@@ -370,7 +365,7 @@ from langchain_azure_ai.chat_models import AzureAIChatCompletionsModel
370365
model = AzureAIChatCompletionsModel(
371366
endpoint=os.environ["AZURE_INFERENCE_ENDPOINT"],
372367
credential=os.environ["AZURE_INFERENCE_CREDENTIAL"],
373-
model_name="mistral-large-2407",
368+
model="mistral-large-2407",
374369
client_kwargs={"logging_enable": True},
375370
)
376371
```

articles/ai-foundry/model-inference/concepts/models.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -55,9 +55,9 @@ DeepSeek family of models includes DeepSeek-R1, which excels at reasoning tasks
5555

5656
| Model | Type | Tier | Capabilities |
5757
| ------ | ---- | ---- | ------------ |
58-
| [DeekSeek-V3-0324](https://ai.azure.com/explore/models/deepseek-v3-0324/version/1/registry/azureml-deepseek) | chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON |
59-
| [DeekSeek-R1](https://ai.azure.com/explore/models/deepseek-r1/version/1/registry/azureml-deepseek) | chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md) | Global standard | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. |
60-
| [DeekSeek-V3](https://ai.azure.com/explore/models/deepseek-v3/version/1/registry/azureml-deepseek) <br />(Legacy) | chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
58+
| [DeepSeek-V3-0324](https://ai.azure.com/explore/models/deepseek-v3-0324/version/1/registry/azureml-deepseek) | chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON |
59+
| [DeepSeek-R1](https://ai.azure.com/explore/models/deepseek-r1/version/1/registry/azureml-deepseek) | chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md) | Global standard | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. |
60+
| [DeepSeek-V3](https://ai.azure.com/explore/models/deepseek-v3/version/1/registry/azureml-deepseek) <br />(Legacy) | chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
6161

6262
For a tutorial on DeepSeek-R1, see [Tutorial: Get started with DeepSeek-R1 reasoning model in Azure AI Foundry Models](../tutorials/get-started-deepseek-r1.md).
6363

Lines changed: 22 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,22 @@
1+
---
2+
manager: nitinme
3+
ms.service: azure-ai-model-inference
4+
ms.topic: include
5+
ms.date: 1/21/2025
6+
ms.author: fasantia
7+
author: santiagxf
8+
---
9+
10+
* Install the SDK with the following command:
11+
12+
# [OpenAI API](#tab/openai)
13+
14+
```bash
15+
pip install -U openai
16+
```
17+
18+
# [Model Inference API (preview)](#tab/inference)
19+
20+
```bash
21+
pip install -U azure-ai-inference
22+
```

0 commit comments

Comments
 (0)