Skip to content

Commit a2ead7e

Browse files
committed
mc
2 parents 24573af + b1473b6 commit a2ead7e

File tree

202 files changed

+1274
-855
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

202 files changed

+1274
-855
lines changed

articles/active-directory-b2c/whats-new-docs.md

Lines changed: 10 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
22
title: "What's new in Azure Active Directory business-to-customer (B2C)"
33
description: "New and updated documentation for the Azure Active Directory business-to-customer (B2C)."
4-
ms.date: 12/01/2023
4+
ms.date: 01/05/2024
55
ms.service: active-directory
66
ms.subservice: B2C
77
ms.topic: reference
@@ -15,6 +15,15 @@ manager: CelesteDG
1515

1616
Welcome to what's new in Azure Active Directory B2C documentation. This article lists new docs that have been added and those that have had significant updates in the last three months. To learn what's new with the B2C service, see [What's new in Microsoft Entra ID](../active-directory/fundamentals/whats-new.md), [Azure AD B2C developer release notes](custom-policy-developer-notes.md) and [What's new in Microsoft Entra External ID](/entra/external-id/whats-new-docs).
1717

18+
## December 2023
19+
20+
### Updated articles
21+
22+
- [Set up a resource owner password credentials flow in Azure Active Directory B2C](add-ropc-policy.md) - Updated ROPC flow limitations
23+
- [Add user attributes and customize user input in Azure Active Directory B2C](configure-user-input.md) - Updated instructional steps
24+
- [Set up sign-up and sign-in with a Google account using Azure Active Directory B2C](identity-provider-google.md) - Editorial updates
25+
- [Localization string IDs](localization-string-ids.md) - Updated the localization string IDs
26+
1827
## November 2023
1928

2029
### Updated articles
@@ -41,25 +50,5 @@ Welcome to what's new in Azure Active Directory B2C documentation. This article
4150
- [Azure AD B2C: Frequently asked questions (FAQ)](faq.yml) - Editorial updates
4251
- [Enable JavaScript and page layout versions in Azure Active Directory B2C](javascript-and-page-layout.md) - Added breaking change on script tags
4352

44-
## September 2023
45-
46-
### Updated articles
47-
48-
- [Supported Microsoft Entra ID features](supported-azure-ad-features.md) - Editorial updates
49-
- [Publish your Azure Active Directory B2C app to the Microsoft Entra app gallery](publish-app-to-azure-ad-app-gallery.md) - Editorial updates
50-
- [Secure your API used an API connector in Azure AD B2C](secure-rest-api.md) - Editorial updates
51-
- [Azure AD B2C: Frequently asked questions (FAQ)'](faq.yml) - Editorial updates
52-
- [Define an ID token hint technical profile in an Azure Active Directory B2C custom policy](id-token-hint.md) - Editorial updates
53-
- [Set up sign-in for multitenant Microsoft Entra ID using custom policies in Azure Active Directory B2C](identity-provider-azure-ad-multi-tenant.md) - Editorial updates
54-
- [Set up sign-in for a specific Microsoft Entra organization in Azure Active Directory B2C](identity-provider-azure-ad-single-tenant.md) - Editorial updates
55-
- [Localization string IDs](localization-string-ids.md) - Editorial updates
56-
- [Define a Microsoft Entra ID multifactor authentication technical profile in an Azure AD B2C custom policy](multi-factor-auth-technical-profile.md) - Editorial updates
57-
- [Define a Microsoft Entra ID SSPR technical profile in an Azure AD B2C custom policy](aad-sspr-technical-profile.md) - Editorial updates
58-
- [Define a Microsoft Entra technical profile in an Azure Active Directory B2C custom policy](active-directory-technical-profile.md) - Editorial updates
59-
- [Monitor Azure AD B2C with Azure Monitor](azure-monitor.md) - Editorial updates
60-
- [Billing model for Azure Active Directory B2C](billing.md) - Editorial updates
61-
- [Enable custom domains for Azure Active Directory B2C](custom-domain.md) - Editorial updates
62-
- [Set up a sign-up and sign-in flow with a social account by using Azure Active Directory B2C custom policy](custom-policies-series-sign-up-or-sign-in-federation.md) - Editorial updates
63-
- [Create and read a user account by using Azure Active Directory B2C custom policy](custom-policies-series-store-user.md) - Editorial updates
6453

6554

articles/ai-services/openai/concepts/models.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ titleSuffix: Azure OpenAI
44
description: Learn about the different model capabilities that are available with Azure OpenAI.
55
ms.service: azure-ai-openai
66
ms.topic: conceptual
7-
ms.date: 01/03/2024
7+
ms.date: 01/05/2024
88
ms.custom: event-tier1-build-2022, references_regions, build-2023, build-2023-dataai, refefences_regions
99
manager: nitinme
1010
author: mrbullwinkle #ChrisHMSFT
@@ -145,7 +145,7 @@ These models can only be used with Embedding API requests.
145145
146146
| Model ID | Model Availability | Max Request (tokens) | Training Data (up to) | Output Dimensions |
147147
|---|---| :---:|:---:|:---:|
148-
| `text-embedding-ada-002` (version 2) | Australia East <br> Canada East <br> East US <br> East US2 <br> France Central <br> Japan East <br> North Central US <br> South Central US <br> Sweden Central <br> Switzerland North <br> UK South <br> West Europe |8,191 | Sep 2021 | 1,536 |
148+
| `text-embedding-ada-002` (version 2) | Australia East <br> Canada East <br> East US <br> East US2 <br> France Central <br> Japan East <br> North Central US <br> Norway East <br> South Central US <br> Sweden Central <br> Switzerland North <br> UK South <br> West Europe <br> West US |8,191 | Sep 2021 | 1,536 |
149149
| `text-embedding-ada-002` (version 1) | East US <br> South Central US <br> West Europe |2,046 | Sep 2021 | 1,536 |
150150

151151
### DALL-E models (Preview)

articles/ai-services/openai/reference.md

Lines changed: 28 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -379,9 +379,9 @@ curl -i -X POST YOUR_RESOURCE_NAME/openai/deployments/YOUR_DEPLOYMENT_NAME/exten
379379
{
380380
"type": "AzureCognitiveSearch",
381381
"parameters": {
382-
"endpoint": "'YOUR_AZURE_COGNITIVE_SEARCH_ENDPOINT'",
383-
"key": "'YOUR_AZURE_COGNITIVE_SEARCH_KEY'",
384-
"indexName": "'YOUR_AZURE_COGNITIVE_SEARCH_INDEX_NAME'"
382+
"endpoint": "YOUR_AZURE_COGNITIVE_SEARCH_ENDPOINT",
383+
"key": "YOUR_AZURE_COGNITIVE_SEARCH_KEY",
384+
"indexName": "YOUR_AZURE_COGNITIVE_SEARCH_INDEX_NAME"
385385
}
386386
}
387387
],
@@ -486,7 +486,6 @@ The following parameters can be used inside of the `parameters` field inside of
486486
|--|--|--|--|--|
487487
| `type` | string | Required | null | The data source to be used for the Azure OpenAI on your data feature. For Azure AI Search the value is `AzureCognitiveSearch`. For Azure Cosmos DB for MongoDB vCore, the value is `AzureCosmosDB`. |
488488
| `indexName` | string | Required | null | The search index to be used. |
489-
| `fieldsMapping` | dictionary | Optional for Azure AI Search. Required for Azure Cosmos DB for MongoDB vCore. | null | Index data column mapping. When using Azure Cosmos DB for MongoDB vCore, the value `vectorFields` is required, which indicates the fields that store vectors. |
490489
| `inScope` | boolean | Optional | true | If set, this value will limit responses specific to the grounding data content. |
491490
| `topNDocuments` | number | Optional | 5 | Specifies the number of top-scoring documents from your data index used to generate responses. You might want to increase the value when you have short documents or want to provide more context. This is the *retrieved documents* parameter in Azure OpenAI studio. |
492491
| `semanticConfiguration` | string | Optional | null | The semantic search configuration. Only required when `queryType` is set to `semantic` or `vectorSemanticHybrid`. |
@@ -498,13 +497,36 @@ The following parameters can be used inside of the `parameters` field inside of
498497
| `strictness` | number | Optional | 3 | Sets the threshold to categorize documents as relevant to your queries. Raising the value means a higher threshold for relevance and filters out more less-relevant documents for responses. Setting this value too high might cause the model to fail to generate responses due to limited available documents. |
499498

500499

501-
**The following parameters are used for Azure AI Search only**
500+
**The following parameters are used for Azure AI Search**
502501

503502
| Parameters | Type | Required? | Default | Description |
504503
|--|--|--|--|--|
505504
| `endpoint` | string | Required | null | Azure AI Search only. The data source endpoint. |
506505
| `key` | string | Required | null | Azure AI Search only. One of the Azure AI Search admin keys for your service. |
507506
| `queryType` | string | Optional | simple | Indicates which query option will be used for Azure AI Search. Available types: `simple`, `semantic`, `vector`, `vectorSimpleHybrid`, `vectorSemanticHybrid`. |
507+
| `fieldsMapping` | dictionary | Optional for Azure AI Search. | null | defines which [fields](./concepts/use-your-data.md?tabs=ai-search#index-field-mapping) you want to map when you add your data source. |
508+
509+
The following parameters are used inside of the `fieldsMapping` field.
510+
511+
| Parameters | Type | Required? | Default | Description |
512+
|--|--|--|--|--|
513+
| `titleField` | string | Optional | null | The field in your index that contains the original title of each document. |
514+
| `urlField` | string | Optional | null | The field in your index that contains the original URL of each document. |
515+
| `filepathField` | string | Optional | null | The field in your index that contains the original file name of each document. |
516+
| `contentFields` | dictionary | Optional | null | The fields in your index that contain the main text content of each document. |
517+
| `contentFieldsSeparator` | string | Optional | null | The separator for the your content fields. Use `\n` by default. |
518+
519+
```json
520+
"fieldsMapping": {
521+
"titleField": "myTitleField",
522+
"urlField": "myUrlField",
523+
"filepathField": "myFilePathField",
524+
"contentFields": [
525+
"myContentField"
526+
],
527+
"contentFieldsSeparator": "\n"
528+
}
529+
```
508530

509531
**The following parameters are used for Azure Cosmos DB for MongoDB vCore**
510532

@@ -516,6 +538,7 @@ The following parameters can be used inside of the `parameters` field inside of
516538
| `containerName` | string | Required | null | Azure Cosmos DB for MongoDB vCore only. The Azure Cosmos Mongo vCore container name in the database. |
517539
| `type` (found inside of`embeddingDependencyType`) | string | Required | null | Indicates the embedding model dependency. |
518540
| `deploymentName` (found inside of`embeddingDependencyType`) | string | Required | null | The embedding model deployment name. |
541+
| `fieldsMapping` | dictionary | Required for Azure Cosmos DB for MongoDB vCore. | null | Index data column mapping. When using Azure Cosmos DB for MongoDB vCore, the value `vectorFields` is required, which indicates the fields that store vectors. |
519542

520543
### Start an ingestion job
521544

Lines changed: 82 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,82 @@
1+
---
2+
title: Performance evaluations for Embedded Speech - Speech service
3+
titleSuffix: Azure AI services
4+
description: Learn how to evaluate performance of embedded speech models on your target devices.
5+
author: eric-urban
6+
manager: nitinme
7+
ms.service: azure-ai-speech
8+
ms.topic: how-to
9+
ms.date: 11/28/2023
10+
ms.author: eur
11+
---
12+
13+
# Evaluating performance of Embedded Speech
14+
15+
Embedded speech models run fully on your target devices. Understanding the performance characteristics of these models on your devices’ hardware can be critical to delivering low latency experiences within your products and applications. This guide provides information to help answer the question, "Is my device suitable to run embedded speech to text and speech translation models?"
16+
17+
## Metrics & terminology
18+
19+
**Real-time factor (RTF)** – The real-time factor (RTF) of a device measures how fast the embedded speech model can process audio input. It's the ratio of the processing time to the audio length. For example, if a device processes a 1-minute audio file in 30 seconds, the RTF is 0.5. This metric evaluates the computational power of the device for running embedded speech models. It can help identify devices that are too slow to support the models. Measurement of this metric should only be done using file-based input rather than real-time microphone input.
20+
21+
To support real-time & interactive speech experiences, the device should have an RTF of `1` or lower. An RTF value higher than `1` means that the device can't keep up with the audio input and will cause poor user experiences.
22+
23+
When measuring the RTF of a device, it's important to measure multiple samples and analyze the distribution across percentiles. This allows you to capture the effect of variations in the device's behavior like different CPU clock speeds due to thermal throttling. The predefined measurement tests outlined in [Measuring the real-time factor on your device](#measuring-the-real-time-factor-on-your-device) automatically measure the RTF for each speech recognition result, yielding a sufficiently large sample size.
24+
25+
**User-perceived latency (UPL)** – The user-perceived latency (UPL) of speech to text is the time between a word being spoken and the word being shown in the recognition results.
26+
27+
## Factors that affect performance
28+
29+
**Device specifications** – The specifications of your device play a key role in whether embedded speech models can run without performance issues. CPU clock speed, architecture (for example, x64, ARM processor, etcetera), and memory can all affect model inference speed.
30+
31+
**CPU load** – In most cases, your device is running other applications in parallel to the application where embedded speech models are integrated. The amount of CPU load your device experiences when idle and at peak can also affect performance.
32+
33+
For example, if the device is under moderate to high CPU load from all other applications running on the device, it's possible to encounter performance issues for running embedded speech in addition to the other applications, even with a powerful processor.
34+
35+
**Memory load** – An embedded speech to text model consumes between 200-300 MB of memory at runtime. If your device has less memory available than that for the embedded speech process to use, frequent fallbacks to virtual memory and paging can introduce more latencies. This can affect both the real-time factor and user-perceived latency.
36+
37+
## Built-in performance optimizations
38+
39+
All embedded speech to text models come with a Voice Activity Detector (VAD) component that aims to filter out silence and non-speech content from the audio input. The goal is to reduce the CPU load and the processing time for other speech to text model components.
40+
41+
The VAD component is always on and doesn't need any configuration from you as the developer. It works best when the audio input has non-negligible amounts of silence or non-speech content, which is common in scenarios like captioning, commanding, and dictation.
42+
43+
## Measuring the real-time factor on your device
44+
45+
For all embedded speech supported platforms, a code sample is available on GitHub that includes a performance measurement mode. In this mode, the goal is to measure the real-time factor (RTF) of your device by controlling as many variables as possible:
46+
47+
- **Model** – The English (United States) model is used for measurement. Models for all other supported locales follow similar performance characteristics, so measuring through the English (United States) model is sufficient.
48+
49+
- **Audio input** – A prebuilt audio file designed for RTF measurements is made available as a supplemental download to the sample code.
50+
51+
- **Measurement mechanism** – The start and stop markers of time measurement are preconfigured in the sample to ensure accuracy and ease of comparing results across different devices & test iterations.
52+
53+
This measurement should be done with the sample running directly on your target device(s), with no code changes other than specifying your model paths & encryption key. The device should be in a state that represents a real end-user state when embedded speech would be used (for example, other active applications, CPU & memory load, etc.).
54+
55+
Running the sample yields performance metrics outputted to the console. The full suite of metrics includes the real-time factor along with other properties like CPU usage and memory consumption. Each metric is defined and explained below.
56+
57+
### Instruction set metrics
58+
59+
| Metric | Description | Notes |
60+
|--------|-------------|-------|
61+
| AVX512Supported | True if CPU supports AVX512 instruction set. | This flag is for X64 platforms. ONNX runtime has optimizations for the various instruction sets, and having this information can help diagnose inconsistencies. |
62+
| AVXSupported | True if CPU supports AVX instruction set. | This flag is for X64 platforms. ONNX runtime has optimizations for the various instruction sets, and having this information can help diagnose inconsistencies. |
63+
| AVX2Supported | True if CPU supports AVX2 instruction set. | This flag is for X64 platforms. ONNX runtime has optimizations for the various instruction sets, and having this information can help diagnose inconsistencies. |
64+
| SSE3Available | True if CPU supports SSE3 instruction set. | This flag is for X64 platforms. ONNX runtime has optimizations for the various instruction sets, and having this information can help diagnose inconsistencies. |
65+
| NEONAvailable | True if CPU supports NEON instruction set. | This flag is for ARM processor platforms. ONNX runtime has optimizations for the various instruction sets, and having this information can help diagnose inconsistencies. |
66+
| NPU | Name of Neural Processing Unit, or N/A if none found. | This flag is for hardware acceleration. |
67+
68+
### Memory metrics
69+
70+
| Metric | Description | Notes |
71+
|--------|-------------|-------|
72+
| PagefileUsage | Amount of page file used by process. Implemented for Linux and Windows. | Values are relative to the machine configuration. |
73+
| WorkingSetSize | Amount of memory used for the process. | |
74+
| ProcessCPUUsage | Aggregate of CPU usage for the process. | Includes all threads in the process, including Speech SDK and UI threads. Aggregated across all cores. |
75+
| ThreadCPUUsage | Aggregate of CPU usage for the speech recognition or speech translation thread. | |
76+
77+
### Performance metrics
78+
79+
| Metric | Description | Notes |
80+
|--------|-------------|-------|
81+
| RealTimeFactor | Measures how much faster than real-time the embedded speech engine is processing audio. Includes audio loading time. | Values greater than `1` indicate that the engine is processing audio slower than real-time. Values less than `1` indicate the engine is processing audio faster than real-time. This value should only be analyzed in file-based input mode. It shouldn't be analyzed in streaming input mode. |
82+
| StreamingRealTimeFactor | Measures how much faster than real-time the engine is processing audio. Excludes audio loading time. | Values greater than `1` indicate that the engine is processing audio slower than real-time. Values less than `1` indicate the engine is processing audio faster than real-time. |

articles/ai-services/speech-service/embedded-speech.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ Embedded Speech is designed for on-device [speech to text](speech-to-text.md) an
2121
2222
## Platform requirements
2323

24-
Embedded speech is included with the Speech SDK (version 1.24.1 and higher) for C#, C++, and Java. Refer to the general [Speech SDK installation requirements](#embedded-speech-sdk-packages) for programming language and target platform specific details.
24+
Embedded speech is included with the Speech SDK (version 1.24.1 and higher) for C#, C++, and Java. Refer to the general [Speech SDK installation requirements](quickstarts/setup-platform.md#platform-requirements) for programming language and target platform specific details.
2525

2626
**Choose your target environment**
2727

articles/ai-services/speech-service/toc.yml

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -449,7 +449,11 @@ items:
449449
href: swagger-documentation.md
450450
displayName: rest, swagger
451451
- name: Embedded Speech
452-
href: embedded-speech.md
452+
items:
453+
- name: Overview
454+
href: embedded-speech.md
455+
- name: Performance evaluations
456+
href: embedded-speech-performance-evaluations.md
453457
- name: Power automate batch transcription
454458
href: power-automate-batch-transcription.md
455459
- name: Containers

0 commit comments

Comments
 (0)