Skip to content

Commit 5e7dab7

Browse files
authored
Merge pull request #497 from MicrosoftDocs/main
[Publishing] [Out of Band Publish] azure-ai-docs-pr - 04:30 PM PST
2 parents 00427ac + ad055eb commit 5e7dab7

26 files changed

+983
-1053
lines changed

articles/ai-services/.openpublishing.redirection.ai-services.json

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -45,6 +45,11 @@
4545
"redirect_url": "/azure/ai-services/custom-vision-service/whats-new",
4646
"redirect_document_id": false
4747
},
48+
{
49+
"source_path_from_root": "/articles/ai-services/custom-vision-service/concepts/compare-alternatives.md",
50+
"redirect_url": "/azure/ai-services/custom-vision-service/overview",
51+
"redirect_document_id": false
52+
},
4853
{
4954
"source_path_from_root": "/articles/ai-services/luis/luis-migration-authoring.md",
5055
"redirect_url": "/azure/ai-services/language-service/conversational-language-understanding/how-to/migrate-from-luis",

articles/ai-services/computer-vision/concept-describe-images-40.md

Lines changed: 12 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -8,35 +8,37 @@ manager: nitinme
88

99
ms.service: azure-ai-vision
1010
ms.topic: conceptual
11-
ms.date: 01/19/2024
11+
ms.date: 09/25/2024
1212
ms.author: pafarley
1313
---
1414

1515
# Image captions (version 4.0)
16-
Image captions in Image Analysis 4.0 are available through the **Caption** and **Dense Captions** features.
1716

18-
Caption generates a one-sentence description for all image contents. Dense Captions provides more detail by generating one-sentence descriptions of up to 10 regions of the image in addition to describing the whole image. Dense Captions also returns bounding box coordinates of the described image regions. Both these features use the latest groundbreaking Florence-based AI models.
17+
Image captions in Image Analysis 4.0 are available through the **Caption** and **Dense Captions** features.
1918

20-
At this time, image captioning is available in English only.
19+
The Caption feature generates a one-sentence description of all the image contents. Dense Captions provides more detail by generating one-sentence descriptions of up to 10 different regions of the image in addition to describing the whole image. Dense Captions also returns bounding box coordinates of the described image regions. Both of these features use the latest Florence-based AI models.
20+
21+
Image captioning is available in English only.
2122

2223
> [!IMPORTANT]
23-
> Image captioning in Image Analysis 4.0 is only available in certain Azure data center regions: see [Region availability](./overview-image-analysis.md#region-availability). You must use a Vision resource located in one of these regions to get results from Caption and Dense Captions features.
24+
> Image captioning in Image Analysis 4.0 is only available in certain Azure data center regions: see [Region availability](./overview-image-analysis.md#region-availability). You must use an Azure AI Vision resource located in one of these regions to get results from Caption and Dense Captions features.
2425
>
25-
> If you have to use a Vision resource outside these regions to generate image captions, please use [Image Analysis 3.2](concept-describing-images.md) which is available in all Azure AI Vision regions.
26+
> If you need to use a Vision resource outside these regions to generate image captions, please use [Image Analysis 3.2](concept-describing-images.md) which is available in all Azure AI Vision regions.
2627
2728
Try out the image captioning features quickly and easily in your browser using Vision Studio.
2829

2930
> [!div class="nextstepaction"]
3031
> [Try Vision Studio](https://portal.vision.cognitive.azure.com/)
3132
32-
### Gender-neutral captions
33-
Captions contain gender terms ("man", "woman", "boy" and "girl") by default. You have the option to replace these terms with "person" in your results and receive gender-neutral captions. You can do so by setting the optional API request parameter, **gender-neutral-caption** to `true` in the request URL.
33+
## Gender-neutral captions
34+
35+
By default, captions contain gender terms ("man", "woman", "boy" and "girl"). You have the option to replace these terms with "person" in your results and receive gender-neutral captions. You can do so by setting the optional API request parameter `gender-neutral-caption` to `true` in the request URL.
3436

3537
## Caption and Dense Captions examples
3638

3739
#### [Caption](#tab/image)
3840

39-
The following JSON response illustrates what the Analysis 4.0 API returns when describing the example image based on its visual features.
41+
The following JSON response illustrates what the Image Analysis 4.0 API returns when describing the example image based on its visual features.
4042

4143
![Photo of a man pointing at a screen](./Media/quickstarts/presentation.png)
4244

@@ -51,7 +53,7 @@ The following JSON response illustrates what the Analysis 4.0 API returns when d
5153

5254
#### [Dense Captions](#tab/dense)
5355

54-
The following JSON response illustrates what the Analysis 4.0 API returns when generating dense captions for the example image.
56+
The following JSON response illustrates what the Image Analysis 4.0 API returns when generating dense captions for the example image.
5557

5658
![Photo of a tractor on a farm](./Images/farm.png)
5759

articles/ai-services/computer-vision/concept-describing-images.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,15 +8,15 @@ manager: nitinme
88

99
ms.service: azure-ai-vision
1010
ms.topic: conceptual
11-
ms.date: 04/30/2024
11+
ms.date: 09/25/2024
1212
ms.author: pafarley
1313
---
1414

1515
# Image descriptions
1616

17-
Azure AI Vision can analyze an image and generate a human-readable phrase that describes its contents. The algorithm returns several descriptions based on different visual features, and each description is given a confidence score. The final output is a list of descriptions ordered from highest to lowest confidence.
17+
Azure AI Vision can analyze an image and generate a human-readable phrase that describes its contents. The service returns several descriptions based on different visual features, and each description is given a confidence score. The final output is a list of descriptions ordered from highest to lowest confidence.
1818

19-
At this time, English is the only supported language for image description.
19+
English is the only supported language for image descriptions.
2020

2121
Try out the image captioning features quickly and easily in your browser using Vision Studio.
2222

@@ -25,7 +25,7 @@ Try out the image captioning features quickly and easily in your browser using V
2525
2626
## Image description example
2727

28-
The following JSON response illustrates what the Analyze API returns when describing the example image based on its visual features.
28+
The following JSON response illustrates what the Analyze Image API returns when describing the example image based on its visual features.
2929

3030
![A black and white picture of buildings in Manhattan](./Images/bw_buildings.png)
3131

articles/ai-services/computer-vision/concept-image-retrieval.md

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,24 +1,24 @@
11
---
22
title: Multimodal embeddings concepts - Image Analysis 4.0
33
titleSuffix: Azure AI services
4-
description: Concepts related to image vectorization using the Image Analysis 4.0 API.
4+
description: Learn about concepts related to image vectorization and search/retrieval using the Image Analysis 4.0 API.
55
#services: cognitive-services
66
author: PatrickFarley
77
manager: nitinme
88

99
ms.service: azure-ai-vision
1010
ms.topic: conceptual
11-
ms.date: 02/20/2024
11+
ms.date: 09/25/2024
1212
ms.author: pafarley
1313
---
1414

1515
# Multimodal embeddings (version 4.0)
1616

17-
Multimodal embedding is the process of generating a numerical representation of an image that captures its features and characteristics in a vector format. These vectors encode the content and context of an image in a way that is compatible with text search over the same vector space.
17+
Multimodal embedding is the process of generating a vector representation of an image that captures its features and characteristics. These vectors encode the content and context of an image in a way that is compatible with text search over the same vector space.
1818

19-
Image retrieval systems have traditionally used features extracted from the images, such as content labels, tags, and image descriptors, to compare images and rank them by similarity. However, vector similarity search is gaining more popularity due to a number of benefits over traditional keyword-based search and is becoming a vital component in popular content search services.
19+
Image retrieval systems have traditionally used features extracted from the images, such as content labels, tags, and image descriptors, to compare images and rank them by similarity. However, vector similarity search offers a number of benefits over traditional keyword-based search and is becoming a vital component in popular content search services.
2020

21-
## What's the difference between vector search and keyword-based search?
21+
## Differences between vector search and keyword search
2222

2323
Keyword search is the most basic and traditional method of information retrieval. In that approach, the search engine looks for the exact match of the keywords or phrases entered by the user in the search query and compares it with the labels and tags provided for the images. The search engine then returns images that contain those exact keywords as content tags and image labels. Keyword search relies heavily on the user's ability to use relevant and specific search terms.
2424

@@ -50,18 +50,17 @@ Each dimension of the vector corresponds to a different feature or attribute of
5050

5151
The following are the main steps of the image retrieval process using Multimodal embeddings.
5252

53-
:::image type="content" source="media/image-retrieval.png" alt-text="Diagram of image retrieval process.":::
53+
:::image type="content" source="media/image-retrieval.png" alt-text="Diagram of the multimodal embedding / image retrieval process.":::
5454

5555
1. Vectorize Images and Text: the Multimodal embeddings APIs, **VectorizeImage** and **VectorizeText**, can be used to extract feature vectors out of an image or text respectively. The APIs return a single feature vector representing the entire input.
5656
> [!NOTE]
5757
> Multimodal embedding does not do any biometric processing of human faces. For face detection and identification, see the [Azure AI Face service](./overview-identity.md).
58-
5958
1. Measure similarity: Vector search systems typically use distance metrics, such as cosine distance or Euclidean distance, to compare vectors and rank them by similarity. The [Vision studio](https://portal.vision.cognitive.azure.com/) demo uses [cosine distance](./how-to/image-retrieval.md#calculate-vector-similarity) to measure similarity.
6059
1. Retrieve Images: Use the top _N_ vectors similar to the search query and retrieve images corresponding to those vectors from your photo library to provide as the final result.
6160

6261
### Relevance score
6362

64-
The image and video retrieval services return a field called "relevance." The term "relevance" denotes a measure of similarity score between a query and image or video frame embeddings. The relevance score is composed of two parts:
63+
The image and video retrieval services return a field called "relevance." The term "relevance" denotes a measure of similarity between a query and image or video frame embeddings. The relevance score is composed of two parts:
6564
1. The cosine similarity (that falls in the range of [0,1]) between the query and image or video frame embeddings.
6665
1. A metadata score, which reflects the similarity between the query and the metadata associated with the image or video frame.
6766

articles/ai-services/computer-vision/how-to/mitigate-latency.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ The quality of the input images affects both the accuracy and the latency of the
6464

6565
To achieve the optimal balance between accuracy and speed, follow these tips to optimize your input data.
6666
- For face detection and recognition operations, see [input data for face detection](../concept-face-detection.md#input-requirements) and [input data for face recognition](../concept-face-recognition.md#input-requirements).
67-
- For liveness detection, see the [tutorial](../Tutorials/liveness.md#select-a-good-reference-image).
67+
- For liveness detection, see the [tutorial](../Tutorials/liveness.md#select-a-reference-image).
6868

6969
#### Other file size tips
7070

articles/ai-services/computer-vision/language-support.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ author: PatrickFarley
77
manager: nitinme
88
ms.service: azure-ai-vision
99
ms.topic: conceptual
10-
ms.date: 03/11/2024
10+
ms.date: 09/25/2024
1111
ms.author: pafarley
1212
---
1313

@@ -125,9 +125,9 @@ The following table lists the OCR supported languages for print text by the most
125125
|Kazakh (Latin) | `kk-latn`|Zhuang | `za` |
126126
|Khaling | `klr`|Zulu | `zu` |
127127

128-
## Analyze image
128+
## Image Analysis
129129

130-
Some features of the [Analyze - Image](/rest/api/computervision/analyze-image?view=rest-computervision-v3.2) API can return results in other languages, specified with the `language` query parameter. Other actions return results in English regardless of what language is specified, and others throw an exception for unsupported languages. Actions are specified with the `visualFeatures` and `details` query parameters; see the [Overview](overview-image-analysis.md) for a list of all the actions you can do with the Analyze API, or follow the [How-to guide](/azure/ai-services/computer-vision/how-to/call-analyze-image-40) to try them out.
130+
Some features of the [Analyze - Image](/rest/api/computervision/analyze-image?view=rest-computervision-v3.2) API can return results in other languages, specified with the `language` query parameter. Other features return results in English regardless of what language is specified, and others throw an exception for unsupported languages. Features are specified with the `visualFeatures` and `details` query parameters; see the [Overview](overview-image-analysis.md) for a list of all the actions you can do with the [Analyze - Image](/rest/api/computervision/analyze-image?view=rest-computervision-v3.2) API, or follow the [How-to guide](/azure/ai-services/computer-vision/how-to/call-analyze-image-40) to try them out.
131131

132132
| Language | Language code | Categories | Tags | Description | Adult, Brands, Color, Faces, ImageType, Objects | Celebrities, Landmarks | Captions, Dense captions|
133133
|:---|:---:|:----:|:---:|:---:|:---:|:---:|:--:|

articles/ai-services/computer-vision/tutorials/liveness.md

Lines changed: 8 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -7,21 +7,19 @@ ms.service: azure-ai-vision
77
ms.custom:
88
- ignite-2023
99
ms.topic: tutorial
10-
ms.date: 11/06/2023
10+
ms.date: 09/25/2024
1111
---
1212

1313
# Tutorial: Detect liveness in faces
1414

15-
Face Liveness detection can be used to determine if a face in an input video stream is real (live) or fake (spoofed). It's an important building block in a biometric authentication system to prevent imposters from gaining access to the system using a photograph, video, mask, or other means to impersonate another person.
15+
Face Liveness detection is used to determine if a face in an input video stream is real (live) or fake (spoofed). It's an important building block in a biometric authentication system to prevent imposters from gaining access to the system using a photograph, video, mask, or other means to impersonate another person.
1616

17-
The goal of liveness detection is to ensure that the system is interacting with a physically present live person at the time of authentication. Such systems are increasingly important with the rise of digital finance, remote access control, and online identity verification processes.
17+
The goal of liveness detection is to ensure that the system is interacting with a physically present, live person at the time of authentication. These systems are increasingly important with the rise of digital finance, remote access control, and online identity verification processes.
1818

19-
The Azure AI Face liveness detection solution successfully defends against various spoof types ranging from paper printouts, 2d/3d masks, and spoof presentations on phones and laptops. Liveness detection is an active area of research, with continuous improvements being made to counteract increasingly sophisticated spoofing attacks over time. Continuous improvements will be rolled out to the client and the service components over time as the overall solution gets more robust to new types of attacks.
19+
The Azure AI Face liveness detection solution successfully defends against various spoof types ranging from paper printouts, 2D/3D masks, and spoof presentations on phones and laptops. Liveness detection is an active area of research, with continuous improvements being made to counteract increasingly sophisticated spoofing attacks. Continuous improvements are rolled out to the client and the service components over time as the overall solution gets more robust to new types of attacks.
2020

2121
[!INCLUDE [liveness-sdk-gate](../includes/liveness-sdk-gate.md)]
2222

23-
24-
2523
## Introduction
2624

2725
The liveness solution integration involves two distinct components: a frontend mobile/web application and an app server/orchestrator.
@@ -31,7 +29,7 @@ The liveness solution integration involves two distinct components: a frontend m
3129
- **Frontend application**: The frontend application receives authorization from the app server to initiate liveness detection. Its primary objective is to activate the camera and guide end-users accurately through the liveness detection process.
3230
- **App server**: The app server serves as a backend server to create liveness detection sessions and obtain an authorization token from the Face service for a particular session. This token authorizes the frontend application to perform liveness detection. The app server's objectives are to manage the sessions, to grant authorization for frontend application, and to view the results of the liveness detection process.
3331

34-
Additionally, we combine face verification with liveness detection to verify whether the person is the specific person you designated. The following table help describe details of the liveness detection features:
32+
Additionally, we combine face verification with liveness detection to verify whether the person is the specific person you designated. The following table describes details of the liveness detection features:
3533

3634
| Feature | Description |
3735
| -- |--|
@@ -40,7 +38,6 @@ Additionally, we combine face verification with liveness detection to verify whe
4038

4139
This tutorial demonstrates how to operate a frontend application and an app server to perform [liveness detection](#perform-liveness-detection) and [liveness detection with face verification](#perform-liveness-detection-with-face-verification) across various language SDKs.
4240

43-
4441
## Prerequisites
4542

4643
- Azure subscription - [Create one for free](https://azure.microsoft.com/free/cognitive-services/)
@@ -71,7 +68,7 @@ The app server/orchestrator is responsible for controlling the lifecycle of a li
7168
- For Python, follow the instructions in the [Python readme](https://github.com/Azure/azure-sdk-for-python/tree/main/sdk/face/azure-ai-vision-face/README.md)
7269
- For JavaScript, follow the instructions in the [JavaScript readme](https://github.com/Azure/azure-sdk-for-js/tree/main/sdk/face/ai-vision-face-rest/README.md)
7370

74-
#### Create environment variables
71+
### Create environment variables
7572

7673
[!INCLUDE [create environment variables](../includes/face-environment-variables.md)]
7774

@@ -415,7 +412,7 @@ There are two parts to integrating liveness with verification:
415412

416413
:::image type="content" source="../media/liveness/liveness-verify-diagram.jpg" alt-text="Diagram of the liveness-with-face-verification workflow of Azure AI Face." lightbox="../media/liveness/liveness-verify-diagram.jpg":::
417414

418-
### Select a good reference image
415+
### Select a reference image
419416

420417
Use the following tips to ensure that your input images give the most accurate recognition results.
421418

@@ -807,7 +804,7 @@ If you want to clean up and remove an Azure AI services subscription, you can de
807804
* [Azure portal](../../multi-service-resource.md?pivots=azportal#clean-up-resources)
808805
* [Azure CLI](../../multi-service-resource.md?pivots=azcli#clean-up-resources)
809806

810-
## Next steps
807+
## Related content
811808

812809
To learn about other options in the liveness APIs, see the Azure AI Vision SDK reference.
813810

0 commit comments

Comments
 (0)