You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/computer-vision/Tutorials/liveness.md
+22-23Lines changed: 22 additions & 23 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -12,15 +12,16 @@ ms.date: 11/06/2023
12
12
13
13
# Tutorial: Detect liveness in faces
14
14
15
-
Face Liveness detection can be used to determine if a face in an input video stream is real (live) or fake (spoof). It is a crucial building block in a biometric authentication system to prevent spoofing attacks from imposters trying to gain access to the system using a photograph, video, mask, or other means to impersonate another person.
15
+
Face Liveness detection can be used to determine if a face in an input video stream is real (live) or fake (spoofed). It's an important building block in a biometric authentication system to prevent imposters from gaining access to the system using a photograph, video, mask, or other means to impersonate another person.
16
16
17
-
The goal of liveness detection is to ensure that the system is interacting with a physically present live person at the time of authentication. Such systems have become increasingly important with the rise of digital finance, remote access control, and online identity verification processes.
17
+
The goal of liveness detection is to ensure that the system is interacting with a physically present live person at the time of authentication. Such systems are increasingly important with the rise of digital finance, remote access control, and online identity verification processes.
18
18
19
-
The liveness detection solution successfully defends against various spoof types ranging from paper printouts, 2d/3d masks, and spoof presentations on phones and laptops. Liveness detection is an active area of research, with continuous improvements being made to counteract increasingly sophisticated spoofing attacks over time. Continuous improvements will be rolled out to the client and the service components over time as the overall solution gets more robust to new types of attacks.
19
+
The Azure AI Face liveness detection solution successfully defends against various spoof types ranging from paper printouts, 2d/3d masks, and spoof presentations on phones and laptops. Liveness detection is an active area of research, with continuous improvements being made to counteract increasingly sophisticated spoofing attacks over time. Continuous improvements will be rolled out to the client and the service components over time as the overall solution gets more robust to new types of attacks.
The liveness solution integration involves two distinct components: a frontend mobile/web application and an app server/orchestrator.
@@ -33,16 +34,14 @@ The liveness solution integration involves two distinct components: a frontend m
33
34
Additionally, we combine face verification with liveness detection to verify whether the person is the specific person you designated. The following table help describe details of the liveness detection features:
34
35
35
36
| Feature | Description |
36
-
| -- | -- |
37
+
| -- |--|
37
38
| Liveness detection | Determine an input is real or fake, and only the app server has the authority to start the liveness check and query the result. |
38
39
| Liveness detection with face verification | Determine an input is real or fake and verify the identity of the person based on a reference image you provided. Either the app server or the frontend application can provide a reference image. Only the app server has the authority to initial the liveness check and query the result. |
39
40
40
-
41
-
## Get started
42
-
43
41
This tutorial demonstrates how to operate a frontend application and an app server to perform [liveness detection](#perform-liveness-detection) and [liveness detection with face verification](#perform-liveness-detection-with-face-verification) across various language SDKs.
44
42
45
-
### Prerequisites
43
+
44
+
## Prerequisites
46
45
47
46
- Azure subscription - [Create one for free](https://azure.microsoft.com/free/cognitive-services/)
48
47
- Your Azure account must have a **Cognitive Services Contributor** role assigned in order for you to agree to the responsible AI terms and create a resource. To get this role assigned to your account, follow the steps in the [Assign roles](/azure/role-based-access-control/role-assignments-steps) documentation, or contact your administrator.
@@ -51,32 +50,32 @@ This tutorial demonstrates how to operate a frontend application and an app serv
51
50
- You can use the free pricing tier (`F0`) to try the service, and upgrade later to a paid tier for production.
52
51
- Access to the Azure AI Vision Face Client SDK for mobile (IOS and Android) and web. To get started, you need to apply for the [Face Recognition Limited Access features](https://customervoice.microsoft.com/Pages/ResponsePage.aspx?id=v4j5cvGGr0GRqy180BHbR7en2Ais5pxKtso_Pz4b1_xUQjA5SkYzNDM4TkcwQzNEOE1NVEdKUUlRRCQlQCN0PWcu) to get access to the SDK. For more information, see the [Face Limited Access](/legal/cognitive-services/computer-vision/limited-access-identity?context=%2Fazure%2Fcognitive-services%2Fcomputer-vision%2Fcontext%2Fcontext) page.
53
52
54
-
### Setup frontend applications and app servers to perform liveness detection
53
+
##Set up frontend applications and app servers to perform liveness detection
55
54
56
-
We provide SDKs in different languages for frontend applications and app servers. See the following instructions to setup your frontend applications and app servers.
55
+
We provide SDKs in different languages for frontend applications and app servers. See the following instructions to set up your frontend applications and app servers.
57
56
58
-
#### Integrate liveness into frontend application
57
+
###Download SDK for frontend application
59
58
60
-
Once you have access to the SDK, follow instruction in the [azure-ai-vision-sdk](https://github.com/Azure-Samples/azure-ai-vision-sdk) GitHub repository to integrate the UI and the code into your native mobile application. The liveness SDK supports Java/Kotlin for Android mobile applications, Swift for iOS mobile applications and JavaScript for web applications:
59
+
Once you have access to the SDK, follow instructions in the [azure-ai-vision-sdk](https://github.com/Azure-Samples/azure-ai-vision-sdk) GitHub repository to integrate the UI and the code into your native mobile application. The liveness SDK supports Java/Kotlin for Android mobile applications, Swift for iOS mobile applications and JavaScript for web applications:
61
60
- For Swift iOS, follow the instructions in the [iOS sample](https://aka.ms/azure-ai-vision-face-liveness-client-sdk-ios-readme)
62
61
- For Kotlin/Java Android, follow the instructions in the [Android sample](https://aka.ms/liveness-sample-java)
63
62
- For JavaScript Web, follow the instructions in the [Web sample](https://aka.ms/liveness-sample-web)
64
63
65
-
Once you've added the code into your application, the SDK handles starting the camera, guiding the end-user to adjust their position, composing the liveness payload, and calling the Azure AI Face cloud service to process the liveness payload.
64
+
Once you've added the code into your application, the SDK handles starting the camera, guiding the end-user in adjusting their position, composing the liveness payload, and calling the Azure AI Face cloud service to process the liveness payload.
66
65
67
-
####Download Azure AI Face client library for an app server
66
+
### Download Azure AI Face client library for app server
68
67
69
68
The app server/orchestrator is responsible for controlling the lifecycle of a liveness session. The app server has to create a session before performing liveness detection, and then it can query the result and delete the session when the liveness check is finished. We offer a library in various languages for easily implementing your app server. Follow these steps to install the package you want:
70
69
- For C#, follow the instructions in the [dotnet readme](https://github.com/Azure/azure-sdk-for-net/tree/main/sdk/face/Azure.AI.Vision.Face/README.md)
71
70
- For Java, follow the instructions in the [Java readme](https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/face/azure-ai-vision-face/README.md)
72
71
- For Python, follow the instructions in the [Python readme](https://github.com/Azure/azure-sdk-for-python/tree/main/sdk/face/azure-ai-vision-face/README.md)
73
72
- For JavaScript, follow the instructions in the [JavaScript readme](https://github.com/Azure/azure-sdk-for-js/tree/main/sdk/face/ai-vision-face-rest/README.md)
@@ -416,15 +415,15 @@ There are two parts to integrating liveness with verification:
416
415
417
416
:::imagetype="content"source="../media/liveness/liveness-verify-diagram.jpg"alt-text="Diagram of the liveness-with-face-verification workflow of Azure AI Face."lightbox="../media/liveness/liveness-verify-diagram.jpg":::
Copy file name to clipboardExpand all lines: articles/ai-services/computer-vision/how-to/model-customization.md
-2Lines changed: 0 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -247,8 +247,6 @@ Begin by going to [Vision Studio](https://portal.vision.cognitive.azure.com/) an
247
247
248
248
Then, sign in with your Azure account and select your Vision resource. If you don't have one, you can create one from this screen.
249
249
250
-
> [!IMPORTANT]
251
-
> To train a custom model in Vision Studio, your Azure subscription needs to be approved for access. Please request access using [this form](https://aka.ms/visionaipublicpreview).
252
250
253
251
:::image type="content" source="../media/customization/select-resource.png" alt-text="Screenshot of the select resource screen.":::
Copy file name to clipboardExpand all lines: articles/ai-services/computer-vision/overview-identity.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,7 +38,7 @@ This documentation contains the following types of articles:
38
38
* The [quickstarts](./quickstarts-sdk/identity-client-library.md) are step-by-step instructions that let you make calls to the service and get results in a short period of time.
39
39
* The [how-to guides](./how-to/identity-detect-faces.md) contain instructions for using the service in more specific or customized ways.
40
40
* The [conceptual articles](./concept-face-detection.md) provide in-depth explanations of the service's functionality and features.
41
-
* The [tutorials](./enrollment-overview.md) are longer guides that show you how to use this service as a component in broader business solutions.
41
+
* The [tutorials](./Tutorials/liveness.md) are longer guides that show you how to use this service as a component in broader business solutions.
42
42
43
43
For a more structured approach, follow a Training module for Face.
44
44
*[Detect and analyze faces with the Face service](/training/modules/detect-analyze-faces/)
> [Azure Open AI On Your Data](../../openai/concepts/use-your-data.md) utilizes large language models (LLMs) to produce similar results to QnA Maker. If you wish to migrate your QnA Maker project to Azure Open AI On Your Data, please check out our [guide](../How-To/migrate-to-openai.md).
17
+
> [Azure OpenAI On Your Data](../../openai/concepts/use-your-data.md) utilizes large language models (LLMs) to produce similar results to QnA Maker. If you wish to migrate your QnA Maker project to Azure OpenAI On Your Data, please check out our [guide](../How-To/migrate-to-openai.md).
0 commit comments