You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/media-services/video-indexer/video-indexer-overview.md
+4Lines changed: 4 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -50,10 +50,13 @@ Following is the list of insights you can retrieve from your videos using Video
50
50
***Black frame detection**: Identifies black frames presented in the video.
51
51
***Keyframe extraction**: Detects stable keyframes in a video.
52
52
***Rolling credits**: identify the beginning and end of the rolling credits in the end of TV shows and movies.
53
+
***Animated characters detection** (preview): detection, grouping, and recognition of characters in animated content via integration with [Cognitive Services custom vision](https://azure.microsoft.com/services/cognitive-services/custom-vision-service/). For more information, see [Animated character detection](animated-characters-recognition.md).
54
+
***Editorial shot type detection**: tagging shots based on their type (such as wide shot, medium shot, close up, extreme close up, two shot, multiple people, outdoor and indoor, etc.). For more information, see [Editorial shot type detection](scenes-shots-keyframes.md#editorial-shot-type-detection).
53
55
54
56
### Audio insights
55
57
56
58
***Automatic language detection**: Automatically identifies the dominant spoken language. Supported languages include English, Spanish, French, German, Italian, Chinese (Simplified), Japanese, Russian, and Brazilian Portuguese. If the language cannot be identified with confidence, Video Indexer assumes the spoken language is English. For more information, see [Language identification model](language-identification-model.md).
59
+
***Multi-language speech identification and transcription** (preview): Automatically identifies the spoken language in different segments from audio, sending each segment of the media file to be transcribed and combine the transcription back to one unified transcription. For more information, see [Automatically identify and transcribe multi-language content](multi-language-identification-transcription.md).
57
60
***Audio transcription**: Converts speech to text in 12 languages and allows extensions. Supported languages include English, Spanish, French, German, Italian, Chinese (Simplified), Japanese, Arabic, Russian, Brazilian Portuguese, Hindi, and Korean.
58
61
***Closed captioning**: Creates closed captioning in three formats: VTT, TTML, SRT.
59
62
***Two channel processing**: Auto detects, separate transcript and merges to single timeline.
@@ -71,6 +74,7 @@ Following is the list of insights you can retrieve from your videos using Video
71
74
When indexing by one channel partial result for those models will be available
72
75
73
76
***Keywords extraction**: Extracts keywords from speech and visual text.
77
+
***Named entities extraction**: Extracts brands, locations, and people from speech and visual text via natural language processing (NLP).
74
78
***Brands extraction**: Extracts brands from speech and visual text.
75
79
***Topic inference**: Makes inference of main topics from transcripts. The 1st-level IPTC taxonomy is included.
76
80
***Artifacts**: Extracts rich set of "next level of details" artifacts for each of the models.
0 commit comments