You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/openai/how-to/integrate-synapseml.md
+76-7Lines changed: 76 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -35,9 +35,9 @@ This tutorial shows how to apply large language models at a distributed scale by
35
35
- To install SynapseML for your Apache Spark cluster, see [Install SynapseML](#install-synapseml).
36
36
37
37
> [!NOTE]
38
-
> This article is designed to work with the [Azure OpenAI Service legacy models](/azure/ai-services/openai/concepts/legacy-models) like `Text-Davinci-003`, which support prompt-based completions. Newer models like the current `GPT-3.5 Turbo` and `GPT-4` model series are designed to work with the new chat completion API that expects a specially formatted array of messages as input.
38
+
> The `OpenAICompletion()` transformer is designed to work with the [Azure OpenAI Service legacy models](/azure/ai-services/openai/concepts/legacy-models) like `Text-Davinci-003`, which supports prompt-based completions. Newer models like the current `GPT-3.5 Turbo` and `GPT-4` model series are designed to work with the new chat completion API that expects a specially formatted array of messages as input. If you working with embeddings or chat completion models, please check the [Chat Completion](#chat-completion) and [Generating Text Embeddings](#generating-text-embeddings) sections bellow.
39
39
>
40
-
> The Azure OpenAI SynapseML integration supports the latest models via the [OpenAIChatCompletion()](https://github.com/microsoft/SynapseML/blob/0836e40efd9c48424e91aa10c8aa3fbf0de39f31/cognitive/src/main/scala/com/microsoft/azure/synapse/ml/cognitive/openai/OpenAIChatCompletion.scala#L24) transformer, which isn't demonstrated in this article. After the [release of the GPT-3.5 Turbo Instruct model](https://techcommunity.microsoft.com/t5/azure-ai-services-blog/announcing-updates-to-azure-openai-service-models/ba-p/3866757), the newer model will be the preferred model to use with this article.
40
+
> The Azure OpenAI SynapseML integration supports the latest models via the [OpenAIChatCompletion()](https://github.com/microsoft/SynapseML/blob/0836e40efd9c48424e91aa10c8aa3fbf0de39f31/cognitive/src/main/scala/com/microsoft/azure/synapse/ml/cognitive/openai/OpenAIChatCompletion.scala#L24) transformer.
41
41
42
42
We recommend that you [create an Azure Synapse workspace](../../../synapse-analytics/get-started-create-workspace.md). However, you can also use Azure Databricks, Azure HDInsight, Spark on Kubernetes, or the Python environment with the `pyspark` package.
43
43
@@ -187,15 +187,87 @@ The following image shows example output with completions in Azure Synapse Analy
187
187
188
188
Here are some other use cases for working with Azure OpenAI Service and large datasets.
189
189
190
-
### Improve throughput with request batching
190
+
### Generating Text Embeddings
191
+
192
+
In addition to completing text, we can also embed text for use in downstream algorithms or vector retrieval architectures. Creating embeddings allows you to search and retrieve documents from large collections and can be used when prompt engineering isn't sufficient for the task. For more information on using [OpenAIEmbedding](https://mmlspark.blob.core.windows.net/docs/0.11.1/pyspark/_modules/synapse/ml/cognitive/openai/OpenAIEmbedding.html), see our [embedding guide](https://microsoft.github.io/SynapseML/docs/Explore%20Algorithms/OpenAI/Quickstart%20-%20OpenAI%20Embedding/).
193
+
194
+
from synapse.ml.services.openai import OpenAIEmbedding
195
+
196
+
```python
197
+
embedding = (
198
+
OpenAIEmbedding()
199
+
.setSubscriptionKey(key)
200
+
.setDeploymentName(deployment_name_embeddings)
201
+
.setCustomServiceName(service_name)
202
+
.setTextCol("prompt")
203
+
.setErrorCol("error")
204
+
.setOutputCol("embeddings")
205
+
)
206
+
207
+
display(embedding.transform(df))
208
+
```
209
+
210
+
### Chat Completion
211
+
Models such as ChatGPT and GPT-4 are capable of understanding chats instead of single prompts. The [OpenAIChatCompletion](https://mmlspark.blob.core.windows.net/docs/0.11.1/pyspark/_modules/synapse/ml/cognitive/openai/OpenAIChatCompletion.html) transformer exposes this functionality at scale.
212
+
213
+
```python
214
+
from synapse.ml.services.openai import OpenAIChatCompletion
215
+
from pyspark.sql import Row
216
+
from pyspark.sql.types import*
217
+
218
+
219
+
defmake_message(role, content):
220
+
return Row(role=role, content=content, name=role)
221
+
222
+
223
+
chat_df = spark.createDataFrame(
224
+
[
225
+
(
226
+
[
227
+
make_message(
228
+
"system", "You are an AI chatbot with red as your favorite color"
229
+
),
230
+
make_message("user", "Whats your favorite color"),
### Improve throughput with request batching from OpenAICompletion
191
260
192
261
You can use Azure OpenAI Service with large datasets to improve throughput with request batching. In the previous example, you make several requests to the service, one for each prompt. To complete multiple prompts in a single request, you can use batch mode.
193
262
194
-
In the `OpenAICompletion` object definition, you specify the `"batchPrompt"` value to configure the dataframe to use a **batchPrompt** column. Create the dataframe with a list of prompts for each row.
263
+
In the [OpenAItCompletion](https://mmlspark.blob.core.windows.net/docs/0.11.1/pyspark/_modules/synapse/ml/cognitive/openai/OpenAICompletion.html) object definition, you specify the `"batchPrompt"` value to configure the dataframe to use a **batchPrompt** column. Create the dataframe with a list of prompts for each row.
195
264
196
265
> [!NOTE]
197
266
> There's currently a limit of 20 prompts in a single request and a limit of 2048 tokens, or approximately 1500 words.
198
267
268
+
> [!NOTE]
269
+
> Currently, request batching is not supported by the `OpenAIChatCompletion()` transformer.
> There's currently a limit of 20 prompts in a single request and a limit of 2048 tokens, or approximately 1500 words.
232
-
233
302
### Use an automatic mini-batcher
234
303
235
304
You can use Azure OpenAI Service with large datasets to transpose the data format. If your data is in column format, you can transpose it to row format by using the SynapseML `FixedMiniBatcherTransformer` object.
Copy file name to clipboardExpand all lines: articles/ai-services/speech-service/includes/release-notes/release-notes-stt.md
+1-25Lines changed: 1 addition & 25 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -105,31 +105,7 @@ Speech to text supports two new locales as shown in the following table. Refer t
105
105
106
106
- Speech [Pronunciation Assessment](../../how-to-pronunciation-assessment.md) now supports 3 additional languages generally available in English (Canada), English (India), and French (Canada), with 3 additional languages available in preview. For more information, see the full [language list for Pronunciation Assessment](../../language-support.md?tabs=pronunciation-assessment).
107
107
108
-
| Language | Locale (BCP-47) |
109
-
|--|--|
110
-
|Arabic (Saudi Arabia)|`ar-SA`<sup>1</sup> |
111
-
|Chinese (Mandarin, Simplified)|`zh-CN`|
112
-
|English (Australia)|`en-AU`|
113
-
|English (Canada)|`en-CA`|
114
-
|English (India)|`en-IN`|
115
-
|English (United Kingdom)|`en-GB`|
116
-
|English (United States)|`en-US`|
117
-
|French (Canada)|`fr-CA`|
118
-
|French (France)|`fr-FR`|
119
-
|German (Germany)|`de-DE`|
120
-
|Italian (Italy)|`it-IT`<sup>1</sup>|
121
-
|Japanese (Japan)|`ja-JP`|
122
-
|Korean (Korea)|`ko-KR`<sup>1</sup>|
123
-
|Malay (Malaysia)|`ms-MY`<sup>1</sup>|
124
-
|Norwegian Bokmål (Norway)|`nb-NO`<sup>1</sup>|
125
-
|Portuguese (Brazil)|`pt-BR`<sup>1</sup>|
126
-
|Russian (Russia)|`ru-RU`<sup>1</sup>|
127
-
|Spanish (Mexico)|`es-MX`|
128
-
|Spanish (Spain)|`es-ES`|
129
-
|Tamil (India)|`ta-IN`<sup>1</sup> |
130
-
|Vietnamese (Vietnam)|`vi-VN`<sup>1</sup> |
131
-
132
-
<sup>1</sup> The language is in public preview for pronunciation assessment.
The table in this section summarizes the 24 locales supported for pronunciation assessment, and each language is available on all [Speech to text regions](regions.md#speech-service). Latest update extends support from English to 23 additional languages and quality enhancements to existing features, including accuracy, fluency and miscue assessment. You should specify the language that you're learning or practicing improving pronunciation. The default language is set as `en-US`. If you know your target learning language, [set the locale](how-to-pronunciation-assessment.md#get-pronunciation-assessment-results) accordingly. For example, if you're learning British English, you should specify the language as `en-GB`. If you're teaching a broader language, such as Spanish, and are uncertain about which locale to select, you can run various accent models (`es-ES`, `es-MX`) to determine the one that achieves the highest score to suit your specific scenario.
114
+
The table in this section summarizes the 25 locales supported for pronunciation assessment, and each language is available on all [Speech to text regions](regions.md#speech-service). Latest update extends support from English to 24 additional languages and quality enhancements to existing features, including accuracy, fluency and miscue assessment. You should specify the language that you're learning or practicing improving pronunciation. The default language is set as `en-US`. If you know your target learning language, [set the locale](how-to-pronunciation-assessment.md#get-pronunciation-assessment-results) accordingly. For example, if you're learning British English, you should specify the language as `en-GB`. If you're teaching a broader language, such as Spanish, and are uncertain about which locale to select, you can run various accent models (`es-ES`, `es-MX`) to determine the one that achieves the highest score to suit your specific scenario.
115
115
116
116
[!INCLUDE [Language support include](includes/language-support/pronunciation-assessment.md)]
Copy file name to clipboardExpand all lines: articles/backup/backup-support-matrix-iaas.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -191,8 +191,8 @@ Adding a disk to a protected VM | Supported.
191
191
Resizing a disk on a protected VM | Supported.
192
192
Shared storage| Backing up VMs by using Cluster Shared Volumes (CSV) or Scale-Out File Server isn't supported. CSV writers are likely to fail during backup. On restore, disks that contain CSV volumes might not come up.
193
193
[Shared disks](../virtual-machines/disks-shared-enable.md) | Not supported.
194
-
<aname="ultra-disk-backup">Ultra disks</a> | Supported with [Enhanced policy](backup-azure-vms-enhanced-policy.md). The support is currently in preview. <br><br> [Supported regions](../virtual-machines/disks-types.md#ultra-disk-limitations). <br><br> To enroll your subscription for this feature, [fill this form](https://forms.office.com/r/1GLRnNCntU). <br><br> - Configuration of Ultra disk protection is supported via Recovery Services vault only. This configuration is currently not supported via virtual machine blade. <br><br> - Cross-region restore is currently not supported for machines using Ultra disks. <br><br> - GRS type vaults cannot be used for enabling backup.
195
-
<aname="premium-ssd-v2-backup">Premium SSD v2</a> | Supported with [Enhanced policy](backup-azure-vms-enhanced-policy.md). The support is currently in preview. <br><br> [Supported regions](../virtual-machines/disks-types.md#regional-availability). <br><br> To enroll your subscription for this feature, [fill this form](https://forms.office.com/r/h56TpTc773). <br><br> - Configuration of Premium v2 disk protection is supported via Recovery Services vault only. This configuration is currently not supported via virtual machine blade. <br><br> - Cross-region restore is currently not supported for machines using Premium v2 disks. <br><br> - GRS type vaults cannot be used for enabling backup.
194
+
<aname="ultra-disk-backup">Ultra disks</a> | Supported with [Enhanced policy](backup-azure-vms-enhanced-policy.md). The support is currently in preview. <br><br> [Supported regions](../virtual-machines/disks-types.md#ultra-disk-limitations). <br><br> - The preview can be tested on any subscription and no enrollment is required. <br><br> - Configuration of Ultra disk protection is supported via Recovery Services vault and via virtual machine blade. <br><br> - Cross-region restore is currently not supported for machines using Ultra disks. <br><br> - GRS type vaults cannot be used for enabling backup. <br><br> - File-level restore is currently not supported for machines using Ultra disks.
195
+
<aname="premium-ssd-v2-backup">Premium SSD v2</a> | Supported with [Enhanced policy](backup-azure-vms-enhanced-policy.md). The support is currently in preview. <br><br> [Supported regions](../virtual-machines/disks-types.md#regional-availability). <br><br> - The preview can be tested on any subscription and no enrollment is required. <br><br> - Configuration of Premium SSD v2 disk protection is supported via Recovery Services vault and via virtual machine blade. <br><br> - Cross-region restore is currently not supported for machines using Premium v2 disks. <br><br> - GRS type vaults cannot be used for enabling backup. <br><br> - File-level restore is currently not supported for machines using Premium SSD v2 disks.
196
196
[Temporary disks](../virtual-machines/managed-disks-overview.md#temporary-disk) | Azure Backup doesn't back up temporary disks.
197
197
NVMe/[ephemeral disks](../virtual-machines/ephemeral-os-disks.md) | Not supported.
198
198
[Resilient File System (ReFS)](/windows-server/storage/refs/refs-overview) restore | Supported. Volume Shadow Copy Service (VSS) supports app-consistent backups on ReFS.
Copy file name to clipboardExpand all lines: articles/communication-services/concepts/interop/teams-user-calling.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -19,7 +19,7 @@ The Azure Communication Services Calling SDK enables Teams user devices to drive
19
19
20
20
Key features of the Calling SDK:
21
21
22
-
-**Addressing** - Azure Communication Services is using [Microsoft Entra user identifier](/powershell/module/azuread/get-azureaduser) to address communication endpoints. Clients use Microsoft Entra identities to authenticate to the service and communicate with each other. These identities are used in Calling APIs that provide clients visibility into who is connected to a call (the roster). And are also used in [Microsoft Graph API](/graph/api/user-get).
22
+
-**Addressing** - Azure Communication Services is using [Microsoft Entra user identifier](/powershell/module/microsoft.graph.users/get-mguser) to address communication endpoints. Clients use Microsoft Entra identities to authenticate to the service and communicate with each other. These identities are used in Calling APIs that provide clients visibility into who is connected to a call (the roster). And are also used in [Microsoft Graph API](/graph/api/user-get).
23
23
-**Encryption** - The Calling SDK encrypts traffic and prevents tampering on the wire.
24
24
-**Device Management and Media** - The Calling SDK provides facilities for binding to audio and video devices, encodes content for efficient transmission over the communications data plane, and renders content to output devices and views that you specify. APIs are also provided for screen and application sharing.
25
25
-**Notifications** - The Calling SDK provides APIs that allow clients to be notified of an incoming call. In situations where your app is not running in the foreground, patterns are available to [fire pop-up notifications](../notifications.md) ("toasts") to inform users of an incoming call.
0 commit comments