Skip to content

Commit a2c9e17

Browse files
authored
Update overview.md
1 parent 0018a83 commit a2c9e17

File tree

1 file changed

+5
-5
lines changed
  • articles/ai-services/content-understanding/video

1 file changed

+5
-5
lines changed

articles/ai-services/content-understanding/video/overview.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -102,11 +102,11 @@ The service operates in two stages. The first stage, content extraction, involve
102102

103103
The first pass is all about extracting a first set of details—who's speaking, where are the cuts, and which faces recur. It creates a solid metadata backbone that later steps can reason over.
104104

105-
* **Transcription:** Converts conversational audio into searchable and analyzable text-based transcripts in WebVTT format. Sentence-level timestamps are available upon request. Content Understanding supports the full set of Azure AI Speech speech-to-text languages. For languages with Fast transcriptions support and for files ≤ 300 MB and/or ≤ 2 hours, transcription time is reduced substantially. Additionally, the following transcription details are important to consider:
105+
* **Transcription:** Converts conversational audio into searchable and analyzable text-based transcripts in WebVTT format. Sentence-level timestamps are available if `returnDetails=true` is set. Content Understanding supports the full set of Azure AI Speech speech-to-text languages. For sepecifics on the supported languages see [Language and region support](../language-region-support.md) .Additionally, the following transcription details are important to consider:
106106
* **Diarization:** Distinguishes between speakers in a conversation in the output, attributing parts of the transcript to specific speakers.
107-
* **Multilingual transcription:** Generates multilingual transcripts. Language/locale is applied per phrase in the transcriptPhrases output when `returnDetails=true` is set. Deviating from language detection this feature is enabled when no language/locale is specified or language is set to `auto`.
107+
* **Multilingual transcription:** Generates multilingual transcripts. Language/locale is applied per phrase in the transcript. Phrases output when `returnDetails=true` is set. Deviating from language detection this feature is enabled when no language/locale is specified or language is set to `auto`.
108108

109-
> [!WARNING]
109+
> [!NOTE]
110110
> When multilingual transcription is used, a file with an unsupported locale still produces a result. This result is based on the closest locale but most likely not correct.
111111
> This transcription behavior is known. Make sure to configure locales when not using multilingual transcription!
112112

@@ -200,11 +200,11 @@ Face identification description is an add-on that provides context to content ex
200200

201201
> [!NOTE]
202202
>
203-
> Face features incur additional cost. This feature is limited access and involves face identification and grouping; customers need to register for access at Face Recognition.
203+
> This feature is limited access and involves face identification and grouping; customers need to register for access at [Face Recognition](https://aka.ms/facerecognition). Face features incur additional cost.
204204
205205
### Content extraction: grouping and identification
206206

207-
The face add-on enables grouping and identification as output from the content extraction section.
207+
The face add-on enables grouping and identification as output from the content extraction section. To enable face capibilities set `enableFace=true` in the analyzer configuration.
208208

209209
* **Grouping:** Grouped faces appearing in a video to extract one representative face image for each person and provides segments where each one is present. The grouped face data is available as metadata and can be used to generate customized metadata fields when `returnDetails: true` for the analyzer.
210210
* **Identification:** Labels individuals in the video with names based on a Face API person directory. Customers can enable this feature by supplying a name for a Face API directory in the current resource in the `personDirectoryId` property of the analyzer.

0 commit comments

Comments
 (0)