Skip to content

Commit 9cdf25e

Browse files
committed
Merge branch 'main' of https://github.com/MicrosoftDocs/azure-docs-pr into vnet-encryption-patch
2 parents ab256c7 + 4ba468c commit 9cdf25e

File tree

11 files changed

+78
-31
lines changed

11 files changed

+78
-31
lines changed

articles/ai-services/openai/concepts/model-retirements.md

Lines changed: 7 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI
44
description: Learn about the model deprecations and retirements in Azure OpenAI.
55
ms.service: azure-ai-openai
66
ms.topic: conceptual
7-
ms.date: 03/12/2024
7+
ms.date: 04/24/2024
88
ms.custom:
99
manager: nitinme
1010
author: mrbullwinkle
@@ -60,8 +60,8 @@ These models are currently available for use in Azure OpenAI Service.
6060

6161
| Model | Version | Retirement date |
6262
| ---- | ---- | ---- |
63-
| `gpt-35-turbo` | 0301 | No earlier than June 13, 2024 |
64-
| `gpt-35-turbo`<br>`gpt-35-turbo-16k` | 0613 | No earlier than July 13, 2024 |
63+
| `gpt-35-turbo` | 0301 | No earlier than August 1, 2024 |
64+
| `gpt-35-turbo`<br>`gpt-35-turbo-16k` | 0613 | No earlier than August 1, 2024 |
6565
| `gpt-35-turbo` | 1106 | No earlier than Nov 17, 2024 |
6666
| `gpt-35-turbo` | 0125 | No earlier than Feb 22, 2025 |
6767
| `gpt-4`<br>`gpt-4-32k` | 0314 | No earlier than July 13, 2024 |
@@ -114,6 +114,10 @@ If you're an existing customer looking for information about these models, see [
114114

115115
## Retirement and deprecation history
116116

117+
### April 24, 2024
118+
119+
Earliest retirement date for `gpt-35-turbo` 0301 and 0613 has been updated to August 1, 2024.
120+
117121
### March 13, 2024
118122

119123
We published this document to provide information about the current models, deprecated models, and upcoming retirements.

articles/ai-services/openai/concepts/models.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI
44
description: Learn about the different model capabilities that are available with Azure OpenAI.
55
ms.service: azure-ai-openai
66
ms.topic: conceptual
7-
ms.date: 04/17/2024
7+
ms.date: 04/24/2024
88
ms.custom: references_regions, build-2023, build-2023-dataai, refefences_regions
99
manager: nitinme
1010
author: mrbullwinkle #ChrisHMSFT
@@ -165,14 +165,14 @@ The following GPT-4 models are available with [Azure Government](/azure/azure-go
165165
> [!IMPORTANT]
166166
> The NEW `gpt-35-turbo (0125)` model has various improvements, including higher accuracy at responding in requested formats and a fix for a bug which caused a text encoding issue for non-English language function calls.
167167
168-
GPT-3.5 Turbo is used with the Chat Completion API. GPT-3.5 Turbo version 0301 can also be used with the Completions API. GPT-3.5 Turbo versions 0613 and 1106 only support the Chat Completions API.
168+
GPT-3.5 Turbo is used with the Chat Completion API. GPT-3.5 Turbo version 0301 can also be used with the Completions API, though this is not recommended. GPT-3.5 Turbo versions 0613 and 1106 only support the Chat Completions API.
169169

170170
GPT-3.5 Turbo version 0301 is the first version of the model released. Version 0613 is the second version of the model and adds function calling support.
171171

172172
See [model versions](../concepts/model-versions.md) to learn about how Azure OpenAI Service handles model version upgrades, and [working with models](../how-to/working-with-models.md) to learn how to view and configure the model version settings of your GPT-3.5 Turbo deployments.
173173

174174
> [!NOTE]
175-
> Version `0613` of `gpt-35-turbo` and `gpt-35-turbo-16k` will be retired no earlier than July 13, 2024. Version `0301` of `gpt-35-turbo` will be retired no earlier than June 13, 2024. See [model updates](../how-to/working-with-models.md#model-updates) for model upgrade behavior.
175+
> Version `0613` of `gpt-35-turbo` and `gpt-35-turbo-16k` will be retired no earlier than August 1, 2024. Version `0301` of `gpt-35-turbo` will be retired no earlier than August 1, 2024. See [model updates](../how-to/working-with-models.md#model-updates) for model upgrade behavior.
176176
177177
| Model ID | Max Request (tokens) | Training Data (up to) |
178178
| --------- |:------:|:----:|

articles/ai-services/openai/how-to/chat-markup-language.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ keywords: ChatGPT
1414
# Chat Markup Language ChatML (Preview)
1515

1616
> [!IMPORTANT]
17-
> Using GPT-3.5-Turbo models with the completion endpoint as described in this article remains in preview and is only possible with `gpt-35-turbo` version (0301) which is [slated for retirement as early as June 13th, 2024](../concepts/model-retirements.md#current-models). We strongly recommend using the [GA Chat Completion API/endpoint](./chatgpt.md). The Chat Completion API is the recommended method of interacting with the GPT-3.5-Turbo models. The Chat Completion API is also the only way to access the GPT-4 models.
17+
> Using GPT-3.5-Turbo models with the completion endpoint as described in this article remains in preview and is only possible with `gpt-35-turbo` version (0301) which is [slated for retirement as early as August 1, 2024](../concepts/model-retirements.md#current-models). We strongly recommend using the [GA Chat Completion API/endpoint](./chatgpt.md). The Chat Completion API is the recommended method of interacting with the GPT-3.5-Turbo models. The Chat Completion API is also the only way to access the GPT-4 models.
1818
1919
The following code snippet shows the most basic way to use the GPT-3.5-Turbo models with ChatML. If this is your first time using these models programmatically we recommend starting with our [GPT-35-Turbo & GPT-4 Quickstart](../chatgpt-quickstart.md).
2020

articles/ai-services/openai/includes/chat-completion.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -425,7 +425,7 @@ def num_tokens_from_messages(messages, model="gpt-3.5-turbo-0613"):
425425
return num_tokens_from_messages(messages, model="gpt-4-0613")
426426
else:
427427
raise NotImplementedError(
428-
f"""num_tokens_from_messages() is not implemented for model {model}. See https://github.com/openai/openai-python/blob/main/chatml.md for information on how messages are converted to tokens."""
428+
f"""num_tokens_from_messages() is not implemented for model {model}."""
429429
)
430430
num_tokens = 0
431431
for message in messages:
@@ -547,13 +547,13 @@ The token counting portion of the code demonstrated previously is a simplified v
547547

548548
Here's a troubleshooting tip.
549549

550-
### Don't use ChatML syntax with the chat completion endpoint
550+
### Don't use ChatML syntax or special tokens with the chat completion endpoint
551551

552-
Some customers try to use the [legacy ChatML syntax](../how-to/chat-markup-language.md) with the chat completion endpoints and newer models. ChatML was a preview capability that only worked with the legacy completions endpoint with the `gpt-35-turbo` version 0301 model. This model is [slated for retirement](../concepts/model-retirements.md). If you attempt to use ChatML syntax with newer models and the chat completion endpoint, it can result in errors and unexpected model response behavior. We don't recommend this use.
552+
Some customers try to use the [legacy ChatML syntax](../how-to/chat-markup-language.md) with the chat completion endpoints and newer models. ChatML was a preview capability that only worked with the legacy completions endpoint with the `gpt-35-turbo` version 0301 model. This model is [slated for retirement](../concepts/model-retirements.md). If you attempt to use ChatML syntax with newer models and the chat completion endpoint, it can result in errors and unexpected model response behavior. We don't recommend this use. This same issue can occur when using common special tokens.
553553

554554
| Error |Cause | Solution |
555555
|---|---|---|
556-
| 400 - "Failed to generate output due to special tokens in the input." | Your prompt contains legacy ChatML tokens not recognized or supported by the model/endpoint. | Ensure that your prompt/messages array doesn't contain any legacy ChatML tokens. If you're upgrading from a legacy model, exclude all special tokens before you submit an API request to the model.|
556+
| 400 - "Failed to generate output due to special tokens in the input." | Your prompt contains legacy ChatML tokens not recognized or supported by the model/endpoint. | Ensure that your prompt/messages array doesn't contain any legacy ChatML tokens/special tokens. If you're upgrading from a legacy model, exclude all special tokens before you submit an API request to the model.|
557557

558558
## Next steps
559559

articles/ai-studio/how-to/deploy-models-cohere-command.md

Lines changed: 12 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ The previously mentioned Cohere models can be deployed as a service with pay-as-
7474
- An [Azure AI hub resource](../how-to/create-azure-ai-resource.md).
7575

7676
> [!IMPORTANT]
77-
> For Cohere family models, the pay-as-you-go model deployment offering is only available with AI hubs created in EastUS, EastUS2 or Sweden Central regions.
77+
> For Cohere family models, the pay-as-you-go model deployment offering is only available with AI hubs created in EastUS2 or Sweden Central region.
7878
7979
- An [Azure AI project](../how-to/create-projects.md) in Azure AI Studio.
8080
- Azure role-based access controls (Azure RBAC) are used to grant access to operations in Azure AI Studio. To perform the steps in this article, your user account must be assigned the __Azure AI Developer role__ on the resource group. For more information on permissions, see [Role-based access control in Azure AI Studio](../concepts/rbac-ai-studio.md).
@@ -93,7 +93,7 @@ To create a deployment:
9393

9494
:::image type="content" source="../media/deploy-monitor/cohere-command/command-r-deploy-pay-as-you-go.png" alt-text="A screenshot showing how to deploy a model with the pay-as-you-go option." lightbox="../media/deploy-monitor/cohere-command/command-r-deploy-pay-as-you-go.png":::
9595

96-
1. Select the project in which you want to deploy your model. To deploy the model your project must be in the EastUS, EastUS2 or Sweden Central regions.
96+
1. Select the project in which you want to deploy your model. To deploy the model your project must be in the EastUS2 or Sweden Central region.
9797
1. In the deployment wizard, select the link to **Azure Marketplace Terms** to learn more about the terms of use.
9898
1. You can also select the **Marketplace offer details** tab to learn about pricing for the selected model.
9999
1. If this is your first time deploying the model in the project, you have to subscribe your project for the particular offering. This step requires that your account has the **Azure AI Developer role** permissions on the Resource Group, as listed in the prerequisites. Each project has its own subscription to the particular Azure Marketplace offering of the model, which allows you to control and monitor spending. Select **Subscribe and Deploy**. Currently you can have only one deployment for each model within a project.
@@ -799,6 +799,16 @@ Response:
799799
| OpenAI SDK (experimental) | [openaisdk.ipynb](https://aka.ms/samples/cohere-command/openaisdk) |
800800
| LangChain | [langchain.ipynb](https://aka.ms/samples/cohere/langchain) |
801801
| Cohere SDK | [cohere-sdk.ipynb](https://aka.ms/samples/cohere-python-sdk) |
802+
| LiteLLM SDK | [litellm.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/litellm.ipynb) |
803+
804+
##### Retrieval Augmented Generation (RAG) and tool use samples
805+
**Description** | **Sample Type** | **Sample Notebook**
806+
--|--|--
807+
Create a local Facebook AI similarity search (FAISS) vector index, using Cohere embeddings - Langchain|`langchain`, `langchain_cohere`|[cohere_faiss_langchain_embed.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/cohere_faiss_langchain_embed.ipynb)
808+
Use Cohere Command R/R+ to answer questions from data in local FAISS vector index - Langchain|`langchain`, `langchain_cohere`|[command_faiss_langchain.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/command_faiss_langchain.ipynb)
809+
Use Cohere Command R/R+ to answer questions from data in AI search vector index - Langchain|`langchain`, `langchain_cohere`|[cohere-aisearch-langchain-rag.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/cohere-aisearch-langchain-rag.ipynb)
810+
Use Cohere Command R/R+ to answer questions from data in AI search vector index - Cohere SDK| `cohere`, `azure_search_documents`|[cohere-aisearch-rag.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/cohere-aisearch-rag.ipynb)
811+
Command R+ tool/function calling, using LangChain|`cohere`, `langchain`, `langchain_cohere`|[command_tools-langchain.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/command_tools-langchain.ipynb)
802812

803813
## Cost and quotas
804814

articles/ai-studio/how-to/deploy-models-cohere-embed.md

Lines changed: 14 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ The previously mentioned Cohere models can be deployed as a service with pay-as-
5252
- An [Azure AI hub resource](../how-to/create-azure-ai-resource.md).
5353

5454
> [!IMPORTANT]
55-
> For Cohere family models, the pay-as-you-go model deployment offering is only available with AI hubs created in EastUS, EastUS2 or Sweden Central regions.
55+
> For Cohere family models, the pay-as-you-go model deployment offering is only available with AI hubs created in EastUS2 or Sweden Central region.
5656
5757
- An [Azure AI project](../how-to/create-projects.md) in Azure AI Studio.
5858
- Azure role-based access controls are used to grant access to operations in Azure AI Studio. To perform the steps in this article, your user account must be assigned the __Azure AI Developer role__ on the resource group. For more information on permissions, see [Role-based access control in Azure AI Studio](../concepts/rbac-ai-studio.md).
@@ -71,7 +71,7 @@ To create a deployment:
7171

7272
:::image type="content" source="../media/deploy-monitor/cohere-embed/embed-english-deploy-pay-as-you-go.png" alt-text="A screenshot showing how to deploy a model with the pay-as-you-go option." lightbox="../media/deploy-monitor/cohere-embed/embed-english-deploy-pay-as-you-go.png":::
7373

74-
1. Select the project in which you want to deploy your model. To deploy the model, your project must be in the EastUS, EastUS2 or Sweden Central regions.
74+
1. Select the project in which you want to deploy your model. To deploy the model, your project must be in the EastUS2 or Sweden Central region.
7575
1. In the deployment wizard, select the link to **Azure Marketplace Terms** to learn more about the terms of use.
7676
1. You can also select the **Marketplace offer details** tab to learn about pricing for the selected model.
7777
1. If it is your first time deploying the model in the project, you have to subscribe your project for the particular offering. This step requires that your account has the **Azure AI Developer role** permissions on the Resource Group, as listed in the prerequisites. Each project has its own subscription to the particular Azure Marketplace offering of the model, which allows you to control and monitor spending. Select **Subscribe and Deploy**. Currently you can have only one deployment for each model within a project.
@@ -119,15 +119,15 @@ These models can be consumed using the embed API.
119119
Content-type: application/json
120120
```
121121

122-
#### v1/emebeddings request schema
122+
#### v1/embeddings request schema
123123

124124
Cohere Embed v3 - English and Embed v3 - Multilingual accept the following parameters for a `v1/embeddings` API call:
125125

126126
| Property | Type | Default | Description |
127127
| --- | --- | --- | --- |
128128
|`input` |`array of strings` |Required |An array of strings for the model to embed. Maximum number of texts per call is 96. We recommend reducing the length of each text to be under 512 tokens for optimal quality. |
129129

130-
#### v1/emebeddings response schema
130+
#### v1/embeddings response schema
131131

132132
The response payload is a dictionary with the following fields:
133133

@@ -330,6 +330,16 @@ Response:
330330
| OpenAI SDK (experimental) | [openaisdk.ipynb](https://aka.ms/samples/cohere-embed/openaisdk) |
331331
| LangChain | [langchain.ipynb](https://aka.ms/samples/cohere-embed/langchain) |
332332
| Cohere SDK | [cohere-sdk.ipynb](https://aka.ms/samples/cohere-embed/cohere-python-sdk) |
333+
| LiteLLM SDK | [litellm.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/litellm.ipynb) |
334+
335+
##### Retrieval Augmented Generation (RAG) and tool-use samples
336+
**Description** | **Sample Type** | **Sample Notebook**
337+
--|--|--
338+
Create a local Facebook AI Similarity Search (FAISS) vector index, using Cohere embeddings - Langchain|`langchain`, `langchain_cohere`|[cohere_faiss_langchain_embed.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/cohere_faiss_langchain_embed.ipynb)
339+
Use Cohere Command R/R+ to answer questions from data in local FAISS vector index - Langchain|`langchain`, `langchain_cohere`|[command_faiss_langchain.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/command_faiss_langchain.ipynb)
340+
Use Cohere Command R/R+ to answer questions from data in AI search vector index - Langchain|`langchain`, `langchain_cohere`|[cohere-aisearch-langchain-rag.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/cohere-aisearch-langchain-rag.ipynb)
341+
Use Cohere Command R/R+ to answer questions from data in AI search vector index - Cohere SDK| `cohere`, `azure_search_documents`|[cohere-aisearch-rag.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/cohere-aisearch-rag.ipynb)
342+
Command R+ tool/function calling, using LangChain|`cohere`, `langchain`, `langchain_cohere`|[command_tools-langchain.ipynb](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/cohere/command_tools-langchain.ipynb)
333343

334344
## Cost and quotas
335345

articles/communication-services/quickstarts/chat/includes/meeting-interop-javascript.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -199,7 +199,7 @@ var chatThreadId = "";
199199

200200
async function init() {
201201
const connectionString = "<SECRET_CONNECTION_STRING>";
202-
const endpointUrl = connectionString.split(";")[0];
202+
const endpointUrl = connectionString.split(";")[0].replace("endpoint=", "");
203203

204204
const identityClient = new CommunicationIdentityClient(connectionString);
205205

articles/communication-services/quickstarts/email/includes/send-email-az-cli.md

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -42,10 +42,13 @@ You need to [sign in to Azure CLI](/cli/azure/authenticate-azure-cli). You can s
4242

4343
You can configure the `AZURE_COMMUNICATION_CONNECTION_STRING` environment variable to use Azure CLI keys operations without having to use `--connection_string` to pass in the connection string. To configure an environment variable, open a console window and select your operating system from the below tabs. Replace `<connectionString>` with your actual connection string.
4444

45+
>[!NOTE]
46+
> Don't store your connection string as an unencrypted environment variable for production environments. This is meant for testing purposes only. For production environments, you should generate new connection strings. We encourage you to encrypt connection strings and change them regularly.
47+
4548
##### [Windows](#tab/windows)
4649

4750
```console
48-
setx AZURE_COMMUNICATION_STRING "<yourConnectionString>"
51+
setx AZURE_COMMUNICATION_CONNECTION_STRING "<yourConnectionString>"
4952
```
5053

5154
After you add the environment variable, you may need to restart any running programs that will need to read the environment variable, including the console window. For example, if you're using Visual Studio as your editor, restart Visual Studio before running the example.
@@ -55,7 +58,7 @@ After you add the environment variable, you may need to restart any running prog
5558
Edit your **`.zshrc`**, and add the environment variable:
5659

5760
```bash
58-
export AZURE_COMMUNICATION_STRING="<connectionString>"
61+
export AZURE_COMMUNICATION_CONNECTION_STRING="<connectionString>"
5962
```
6063

6164
After you add the environment variable, run `source ~/.zshrc` from your console window to make the changes effective. If you created the environment variable with your IDE open, you may need to close and reopen the editor, IDE, or shell in order to access the variable.
@@ -65,7 +68,7 @@ After you add the environment variable, run `source ~/.zshrc` from your console
6568
Edit your **`.bash_profile`**, and add the environment variable:
6669

6770
```bash
68-
export AZURE_COMMUNICATION_STRING="<connectionString>"
71+
export AZURE_COMMUNICATION_CONNECTION_STRING="<connectionString>"
6972
```
7073

7174
After you add the environment variable, run `source ~/.bash_profile` from your console window to make the changes effective. If you created the environment variable with your IDE open, you may need to close and reopen the editor, IDE, or shell in order to access the variable.

0 commit comments

Comments
 (0)