Skip to content

Commit ba84250

Browse files
Merge pull request #268181 from PatrickFarley/gh-issues-address
rm postman
2 parents c34f08d + 441f26a commit ba84250

File tree

6 files changed

+12
-12
lines changed

6 files changed

+12
-12
lines changed

articles/ai-services/computer-vision/Tutorials/liveness.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ Face Liveness detection can be used to determine if a face in an input video str
1616

1717
The goal of liveness detection is to ensure that the system is interacting with a physically present live person at the time of authentication. Such systems have become increasingly important with the rise of digital finance, remote access control, and online identity verification processes.
1818

19-
The liveness detection solution successfully defends against a variety of spoof types ranging from paper printouts, 2d/3d masks, and spoof presentations on phones and laptops. Liveness detection is an active area of research, with continuous improvements being made to counteract increasingly sophisticated spoofing attacks over time. Continuous improvements will be rolled out to the client and the service components over time as the overall solution gets more robust to new types of attacks.
19+
The liveness detection solution successfully defends against various spoof types ranging from paper printouts, 2d/3d masks, and spoof presentations on phones and laptops. Liveness detection is an active area of research, with continuous improvements being made to counteract increasingly sophisticated spoofing attacks over time. Continuous improvements will be rolled out to the client and the service components over time as the overall solution gets more robust to new types of attacks.
2020

2121
[!INCLUDE [liveness-sdk-gate](../includes/liveness-sdk-gate.md)]
2222

@@ -40,7 +40,7 @@ Once you have access to the SDK, follow instruction in the [azure-ai-vision-sdk]
4040
- For Swift iOS, follow the instructions in the [iOS sample](https://aka.ms/azure-ai-vision-face-liveness-client-sdk-ios-readme)
4141
- For Kotlin/Java Android, follow the instructions in the [Android sample](https://aka.ms/liveness-sample-java)
4242

43-
Once you've added the code into your application, the SDK will handle starting the camera, guiding the end-user to adjust their position, composing the liveness payload, and calling the Azure AI Face cloud service to process the liveness payload.
43+
Once you've added the code into your application, the SDK handles starting the camera, guiding the end-user to adjust their position, composing the liveness payload, and calling the Azure AI Face cloud service to process the liveness payload.
4444

4545
### Orchestrate the liveness solution
4646

@@ -86,7 +86,7 @@ The high-level steps involved in liveness orchestration are illustrated below:
8686

8787
1. The SDK then starts the camera, guides the user to position correctly and then prepares the payload to call the liveness detection service endpoint.
8888

89-
1. The SDK calls the Azure AI Vision Face service to perform the liveness detection. Once the service responds, the SDK will notify the mobile application that the liveness check has been completed.
89+
1. The SDK calls the Azure AI Vision Face service to perform the liveness detection. Once the service responds, the SDK notifies the mobile application that the liveness check has been completed.
9090

9191
1. The mobile application relays the liveness check completion to the app server.
9292

@@ -110,7 +110,7 @@ The high-level steps involved in liveness orchestration are illustrated below:
110110
"method": "POST",
111111
"contentLength": 352568,
112112
"contentType": "multipart/form-data; boundary=--------------------------482763481579020783621915",
113-
"userAgent": "PostmanRuntime/7.34.0"
113+
"userAgent": ""
114114
},
115115
"response": {
116116
"body": {
@@ -162,12 +162,12 @@ Use the following tips to ensure that your input images give the most accurate r
162162
#### Composition requirements:
163163
- Photo is clear and sharp, not blurry, pixelated, distorted, or damaged.
164164
- Photo is not altered to remove face blemishes or face appearance.
165-
- Photo must be in an RGB color supported format (JPEG, PNG, WEBP, BMP). Recommended Face size is 200 pixels x 200 pixels. Face sizes larger than 200 pixels x 200 pixels will not result in better AI quality, and no larger than 6MB in size.
165+
- Photo must be in an RGB color supported format (JPEG, PNG, WEBP, BMP). Recommended Face size is 200 pixels x 200 pixels. Face sizes larger than 200 pixels x 200 pixels will not result in better AI quality, and no larger than 6 MB in size.
166166
- User is not wearing glasses, masks, hats, headphones, head coverings, or face coverings. Face should be free of any obstructions.
167167
- Facial jewelry is allowed provided they do not hide your face.
168168
- Only one face should be visible in the photo.
169169
- Face should be in neutral front-facing pose with both eyes open, mouth closed, with no extreme facial expressions or head tilt.
170-
- Face should be free of any shadows or red eyes. Please retake photo if either of these occur.
170+
- Face should be free of any shadows or red eyes. Retake photo if either of these occur.
171171
- Background should be uniform and plain, free of any shadows.
172172
- Face should be centered within the image and fill at least 50% of the image.
173173

@@ -243,7 +243,7 @@ The high-level steps involved in liveness with verification orchestration are il
243243
"method": "POST",
244244
"contentLength": 352568,
245245
"contentType": "multipart/form-data; boundary=--------------------------590588908656854647226496",
246-
"userAgent": "PostmanRuntime/7.34.0"
246+
"userAgent": ""
247247
},
248248
"response": {
249249
"body": {

articles/ai-services/computer-vision/how-to/shelf-analyze.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ The fastest way to start using Product Recognition is to use the built-in pretra
2525
* Once you have your Azure subscription, <a href="https://portal.azure.com/#create/Microsoft.CognitiveServicesComputerVision" title="create a Vision resource" target="_blank">create a Vision resource</a> in the Azure portal. It must be deployed in the **East US** or **West US 2** region. After it deploys, select **Go to resource**.
2626
* You'll need the key and endpoint from the resource you create to connect your application to the Azure AI Vision service. You'll paste your key and endpoint into the code below later in the guide.
2727
* An Azure Storage resource with a blob storage container. [Create one](/azure/storage/common/storage-account-create?tabs=azure-portal)
28-
* [cURL](https://curl.haxx.se/) installed. Or, you can use a different REST platform, like Postman, Swagger, or the REST Client extension for VS Code.
28+
* [cURL](https://curl.haxx.se/) installed. Or, you can use a different REST platform, like Swagger or the [REST Client](https://marketplace.visualstudio.com/items?itemName=humao.rest-client) extension for VS Code.
2929
* A shelf image. You can download our [sample image](https://github.com/Azure-Samples/cognitive-services-sample-data-files/blob/master/ComputerVision/shelf-analysis/shelf.png) or bring your own images. The maximum file size per image is 20 MB.
3030

3131
## Analyze shelf images

articles/ai-services/computer-vision/how-to/shelf-modify-images.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ This guide also shows you how to use the **Rectification API** to correct for pe
2424
* Once you have your Azure subscription, <a href="https://portal.azure.com/#create/Microsoft.CognitiveServicesComputerVision" title="create a Vision resource" target="_blank">create a Vision resource</a> in the Azure portal. It must be deployed in the **East US** or **West US 2** region. After it deploys, select **Go to resource**.
2525
* You'll need the key and endpoint from the resource you create to connect your application to the Azure AI Vision service. You'll paste your key and endpoint into the code below later in the quickstart.
2626
* An Azure Storage resource with a blob storage container. [Create one](/azure/storage/common/storage-account-create?tabs=azure-portal)
27-
* [cURL](https://curl.haxx.se/) installed. Or, you can use a different REST platform, like Postman, Swagger, or the REST Client extension for VS Code.
27+
* [cURL](https://curl.haxx.se/) installed. Or, you can use a different REST platform, like Swagger or the [REST Client](https://marketplace.visualstudio.com/items?itemName=humao.rest-client) extension for VS Code.
2828
* A set of photos that show adjacent parts of the same shelf. A 50% overlap between images is recommended. You can download and use the sample "unstitched" images from [GitHub](https://github.com/Azure-Samples/cognitive-services-sample-data-files/tree/master/ComputerVision/shelf-analysis).
2929

3030

articles/ai-services/computer-vision/how-to/shelf-planogram.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ A planogram is a diagram that indicates the correct placement of retail products
2222
2323
## Prerequisites
2424
* You must have already set up and run basic [Product Understanding analysis](./shelf-analyze.md) with the Product Understanding API.
25-
* [cURL](https://curl.haxx.se/) installed. Or, you can use a different REST platform, like Postman, Swagger, or the REST Client extension for VS Code.
25+
* [cURL](https://curl.haxx.se/) installed. Or, you can use a different REST platform, like Swagger or the [REST Client](https://marketplace.visualstudio.com/items?itemName=humao.rest-client) extension for VS Code.
2626

2727
## Prepare a planogram schema
2828

articles/ai-services/custom-vision-service/copy-move-projects.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ ms.author: pafarley
1414

1515
After you've created and trained a Custom Vision project, you may want to copy your project to another resource. If your app or business depends on a Custom Vision project, we recommend you copy your model to another Custom Vision account in another region. Then if a regional outage occurs, you can access your project in the region where it was copied.
1616

17-
The **[ExportProject](https://westus2.dev.cognitive.microsoft.com/docs/services/Custom_Vision_Training_3.3/operations/5eb0bcc6548b571998fddeb3)** and **[ImportProject](https://westus2.dev.cognitive.microsoft.com/docs/services/Custom_Vision_Training_3.3/operations/5eb0bcc7548b571998fddee3)** APIs enable this scenario by allowing you to copy projects from one Custom Vision account into others. This guide shows you how to use these REST APIs with cURL. You can also use an HTTP request service like Postman to issue the requests.
17+
The **[ExportProject](https://westus2.dev.cognitive.microsoft.com/docs/services/Custom_Vision_Training_3.3/operations/5eb0bcc6548b571998fddeb3)** and **[ImportProject](https://westus2.dev.cognitive.microsoft.com/docs/services/Custom_Vision_Training_3.3/operations/5eb0bcc7548b571998fddee3)** APIs enable this scenario by allowing you to copy projects from one Custom Vision account into others. This guide shows you how to use these REST APIs with cURL. You can also use an HTTP request service, like the [REST Client](https://marketplace.visualstudio.com/items?itemName=humao.rest-client) for Visual Studio Code, to issue the requests.
1818

1919
> [!TIP]
2020
> For an example of this scenario using the Python client library, see the [Move Custom Vision Project](https://github.com/Azure-Samples/custom-vision-move-project/tree/master/) repository on GitHub.

articles/ai-services/custom-vision-service/storage-integration.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ You can integrate your Custom Vision project with an Azure blob storage queue to
1717

1818
You can also use Azure storage to store backup copies of your published models.
1919

20-
This guide shows you how to use these REST APIs with cURL. You can also use an HTTP request service like Postman to make the requests.
20+
This guide shows you how to use these REST APIs with cURL. You can also use an HTTP request service, like the [REST Client](https://marketplace.visualstudio.com/items?itemName=humao.rest-client) for Visual Studio Code, to make the requests.
2121

2222
> [!NOTE]
2323
> Push notifications depend on the optional _notificationQueueUri_ parameter in the **CreateProject** API, and model backups require that you also use the optional _exportModelContainerUri_ parameter. This guide will use both for the full set of features.

0 commit comments

Comments
 (0)