Skip to content

Commit df1d563

Browse files
committed
Merge branch 'main' of https://github.com/MicrosoftDocs/azure-docs-pr into nw-cm-ps
2 parents 5f6616c + b338471 commit df1d563

File tree

17 files changed

+214
-241
lines changed

17 files changed

+214
-241
lines changed

.openpublishing.redirection.json

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4087,10 +4087,12 @@
40874087
"source_path_from_root": "/articles/event-grid/move-domains-across-regions.md",
40884088
"redirect_url": "/azure",
40894089
"redirect_document_id": false
4090+
},
4091+
{
4092+
"source_path_from_root": "/articles/data-factory/continuous-integration-delivery-automate-github-actions.md",
4093+
"redirect_url": "/azure",
4094+
"redirect_document_id": false
40904095
}
4091-
4092-
4093-
4094-
4096+
40954097
]
40964098
}

articles/ai-services/computer-vision/includes/identity-curl-quickstart.md

Lines changed: 81 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,7 @@ Get started with facial recognition using the Face REST API. The Face service pr
2424
* You'll need the key and endpoint from the resource you create to connect your application to the Face API. You'll paste your key and endpoint into the code below later in the quickstart.
2525
* You can use the free pricing tier (`F0`) to try the service, and upgrade later to a paid tier for production.
2626
* [PowerShell version 6.0+](/powershell/scripting/install/installing-powershell-core-on-windows), or a similar command-line application.
27-
* [cURL](https://curl.haxx.se/) installed.
28-
27+
* [cURL](https://curl.se/) installed.
2928

3029

3130

@@ -34,58 +33,134 @@ Get started with facial recognition using the Face REST API. The Face service pr
3433
> [!NOTE]
3534
> If you haven't received access to the Face service using the [intake form](https://aka.ms/facerecognition), some of these functions won't work.
3635
37-
1. First, call the Detect API on the source face. This is the face that we'll try to identify from the larger group. Copy the following command to a text editor, insert your own key, and then copy it into a shell window and run it.
36+
1. First, call the Detect API on the source face. This is the face that we'll try to identify from the larger group. Copy the following command to a text editor, insert your own key and endpoint, and then copy it into a shell window and run it.
37+
38+
#### [Windows](#tab/windows)
39+
40+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_detect":::
41+
42+
#### [Linux](#tab/linux)
3843

3944
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_detect":::
4045

46+
---
47+
4148
Save the returned face ID string to a temporary location. You'll use it again at the end.
4249

43-
1. Next you'll need to create a **LargePersonGroup**. This object will store the aggregated face data of several persons. Run the following command, inserting your own key. Optionally, change the group's name and metadata in the request body.
50+
1. Next you'll need to create a **LargePersonGroup** and give it an arbitrary ID that matches regex pattern `^[a-z0-9-_]+$`. This object will store the aggregated face data of several persons. Run the following command, inserting your own key. Optionally, change the group's name and metadata in the request body.
51+
52+
#### [Windows](#tab/windows)
53+
54+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_create_persongroup":::
55+
56+
#### [Linux](#tab/linux)
4457

4558
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_create_persongroup":::
4659

47-
Save the returned ID of the created group to a temporary location.
60+
---
61+
62+
Save the specified ID of the created group to a temporary location.
4863

4964
1. Next, you'll create **Person** objects that belong to the group. Run the following command, inserting your own key and the ID of the **LargePersonGroup** from the previous step. This command creates a **Person** named "Family1-Dad".
5065

66+
#### [Windows](#tab/windows)
67+
68+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_create_person":::
69+
70+
#### [Linux](#tab/linux)
71+
5172
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_create_person":::
5273

74+
---
75+
5376
After you run this command, run it again with different input data to create more **Person** objects: "Family1-Mom", "Family1-Son", "Family1-Daughter", "Family2-Lady", and "Family2-Man".
5477

5578
Save the IDs of each **Person** created; it's important to keep track of which person name has which ID.
5679

5780
1. Next you'll need to detect new faces and associate them with the **Person** objects that exist. The following command detects a face from the image *Family1-Dad1.jpg* and adds it to the corresponding person. You need to specify the `personId` as the ID that was returned when you created the "Family1-Dad" **Person** object. The image name corresponds to the name of the created **Person**. Also enter the **LargePersonGroup** ID and your key in the appropriate fields.
5881

82+
#### [Windows](#tab/windows)
83+
84+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_add_face":::
85+
86+
#### [Linux](#tab/linux)
87+
5988
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_add_face":::
6089

90+
---
91+
6192
Then, run the above command again with a different source image and target **Person**. The images available are: *Family1-Dad1.jpg*, *Family1-Dad2.jpg* *Family1-Mom1.jpg*, *Family1-Mom2.jpg*, *Family1-Son1.jpg*, *Family1-Son2.jpg*, *Family1-Daughter1.jpg*, *Family1-Daughter2.jpg*, *Family2-Lady1.jpg*, *Family2-Lady2.jpg*, *Family2-Man1.jpg*, and *Family2-Man2.jpg*. Be sure that the **Person** whose ID you specify in the API call matches the name of the image file in the request body.
6293

6394
At the end of this step, you should have multiple **Person** objects that each have one or more corresponding faces, detected directly from the provided images.
6495

6596
1. Next, train the **LargePersonGroup** with the current face data. The training operation teaches the model how to associate facial features, sometimes aggregated from multiple source images, to each single person. Insert the **LargePersonGroup** ID and your key before running the command.
6697

98+
#### [Windows](#tab/windows)
99+
100+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_train":::
101+
102+
#### [Linux](#tab/linux)
103+
67104
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_train":::
105+
106+
---
68107

108+
1. Check whether the training status is succeeded. If not, wait for a while and query again.
109+
110+
#### [Windows](#tab/windows)
111+
112+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_check_status":::
113+
114+
#### [Linux](#tab/linux)
115+
116+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_check_status":::
117+
118+
---
119+
69120
1. Now you're ready to call the Identify API, using the source face ID from the first step and the **LargePersonGroup** ID. Insert these values into the appropriate fields in the request body, and insert your key.
70121

122+
#### [Windows](#tab/windows)
123+
124+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_identify":::
125+
126+
#### [Linux](#tab/linux)
127+
71128
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_identify":::
72129

130+
---
131+
73132
The response should give you a **Person** ID indicating the person identified with the source face. It should be the ID that corresponds to the "Family1-Dad" person, because the source face is of that person.
74133

75134
1. To do face verification, you'll use the **Person** ID returned in the previous step, the **LargePersonGroup** ID, and also the source face ID. Insert these values into the fields in the request body, and insert your key.
76135

136+
#### [Windows](#tab/windows)
137+
138+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="verify":::
139+
140+
#### [Linux](#tab/linux)
141+
77142
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="verify":::
78143

144+
---
145+
79146
The response should give you a boolean verification result along with a confidence value.
80147

81148

82149

83150
## Clean up resources
84151

85-
To delete the **LargePersonGroup** you created in this exercise, run the LargePersonGroup - Delete call.
152+
To delete the **LargePersonGroup** you created in this exercise, run the [LargePersonGroup - Delete](/rest/api/face/person-group-operations/delete-large-person-group) call.
153+
154+
#### [Windows](#tab/windows)
155+
156+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_delete":::
157+
158+
#### [Linux](#tab/linux)
86159

87160
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_delete":::
88161

162+
---
163+
89164
If you want to clean up and remove an Azure AI services subscription, you can delete the resource or resource group. Deleting the resource group also deletes any other resources associated with it.
90165

91166
* [Portal](../../multi-service-resource.md?pivots=azportal#clean-up-resources)

articles/ai-services/document-intelligence/containers/install-run.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ ms.author: lajanuar
1919
<!-- markdownlint-disable MD051 -->
2020

2121
:::moniker range="doc-intel-2.1.0 || doc-intel-4.0.0"
22-
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` for all models and `2023-07-31 (GA)` for Read and Layout only:
22+
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` for all models and `2023-07-31 (GA)` for Read, Layout, ID Document, Receipt and Invoice models:
2323

2424
* [REST API `2022-08-31 (GA)`](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-v3.0%20(2022-08-31)&preserve-view=true&tabs=HTTP)
2525
* [REST API `2023-07-31 (GA)`](/rest/api/aiservices/document-models/analyze-document?view=rest-aiservices-v3.1%20(2023-07-31)&tabs=HTTP&preserve-view=true)
@@ -38,9 +38,9 @@ Azure AI Document Intelligence is an Azure AI service that lets you build automa
3838

3939
In this article you learn how to download, install, and run Document Intelligence containers. Containers enable you to run the Document Intelligence service in your own environment. Containers are great for specific security and data governance requirements.
4040

41-
* **Read**, and **Layout** models are supported by Document Intelligence v3.1 containers.
41+
* **Read**, **Layout**, **ID Document**, **Receipt**, and **Invoice** models are supported by Document Intelligence v3.1 containers.
4242

43-
* **Read**, **Layout**, **General Document**, **ID Document**, **Receipt**, **Invoice**, **Business Card**, and **Custom** models are supported by Document Intelligence v3.0 containers.
43+
* **Read**, **Layout**, **General Document**, **Business Card**, and **Custom** models are supported by Document Intelligence v3.0 containers.
4444

4545
* **Business Card** model is currently only supported in the [v2.1 containers](install-run.md?view=doc-intel-2.1.0&preserve-view=true).
4646

articles/ai-services/document-intelligence/studio-overview.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -53,6 +53,9 @@ Your organization can opt to disable local authentication and enforce Microsoft
5353

5454
* **Designating role assignments**. Document Intelligence Studio basic access requires the [`Cognitive Services User`](../../role-based-access-control/built-in-roles/ai-machine-learning.md#cognitive-services-user) role. For more information, *see* [Document Intelligence role assignments](quickstarts/try-document-intelligence-studio.md#azure-role-assignments) and [Document Intelligence Studio Permission](faq.yml#what-permissions-do-i-need-to-access-document-intelligence-studio-).
5555

56+
> [!IMPORTANT]
57+
> Make sure you have the Cognitive Services User role, and not the Cognitive Services Contributor role when setting up Entra authentication. In Azure concept, Contributor role can only perform actions to control and manage the resource itself, including listing the access keys. Any user accounts with "Contributor" role that is able to access the Document Intelligence service is calling with access keys. However, when setting up access with Entra ID, key-access will be disabled and Cognitive Service User role will be required for an account to use the resources.
58+
5659
## Authentication
5760

5861
Navigate to the [Document Intelligence Studio](https://formrecognizer.appliedai.azure.com/). If it's your first time logging in, a popup window appears prompting you to configure your service resource. In accordance with your organization's policy, you have one or two options:

articles/ai-services/openai/concepts/content-filter.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -989,7 +989,7 @@ data: [DONE]
989989
```
990990
991991
> [!IMPORTANT]
992-
> When content filtering is triggered for a prompt and a `"status": 400` is received as part of the response there will be a charge for this request as the prompt was evaluated by the service. [Charges will also occur](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/) when a `"status":200` is received with `"finish_reason": "content_filter"`. In this case the prompt did not have any issues, but the completion generated by the model was detected to violate the content filtering rules which results in the completion being filtered.
992+
> When content filtering is triggered for a prompt and a `"status": 400` is received as part of the response there will be a charge for this request as the prompt was evaluated by the service. Due to the asynchronous nature of the content filtering system, a charge for both the prompt and completion tokens will occur. [Charges will also occur](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/) when a `"status":200` is received with `"finish_reason": "content_filter"`. In this case the prompt did not have any issues, but the completion generated by the model was detected to violate the content filtering rules which results in the completion being filtered.
993993
994994
## Best practices
995995

articles/aks/control-plane-metrics-default-list.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -42,17 +42,20 @@ The following metrics are allow-listed with `minimalingestionprofile=true` for d
4242
- `apiserver_cache_list_returned_objects_total`
4343
- `apiserver_flowcontrol_demand_seats_average`
4444
- `apiserver_flowcontrol_current_limit_seats`
45-
- `apiserver_request_sli_duration_seconds_bucket`
45+
- ~~`apiserver_request_sli_duration_seconds_bucket`~~
4646
- `apiserver_request_sli_duration_seconds_sum`
4747
- `apiserver_request_sli_duration_seconds_count`
4848
- `process_start_time_seconds`
49-
- `apiserver_request_duration_seconds_bucket`
49+
- ~~`apiserver_request_duration_seconds_bucket`~~
5050
- `apiserver_request_duration_seconds_sum`
5151
- `apiserver_request_duration_seconds_count`
5252
- `apiserver_storage_list_fetched_objects_total`
5353
- `apiserver_storage_list_returned_objects_total`
5454
- `apiserver_current_inflight_requests`
5555

56+
> [!NOTE]
57+
`apiserver_request_sli_duration_seconds_bucket` and `apiserver_request_duration_seconds_bucket` are not collected now with a recent release. These are high cardinality metrics which may increase the number of metrics stored based on the number of custom resources in the cluster. If you would like to collect these bucket metrics, you can add it to the keep list. We highly recommend not turning off the minimal ingestion profile for the control plane components
58+
5659
**controlplane-etcd**
5760

5861
- `etcd_server_has_leader`

articles/azure-functions/flex-consumption-plan.md

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -111,14 +111,14 @@ This table shows the language stack versions that are currently supported for Fl
111111

112112
## Regional subscription memory quotas
113113

114-
Currently, each region in a given subscription has a memory limit of 512,000 MB for all instances of apps running on Flex Consumption plans in that region. This means that in a given subscription and region, you could have any of the following combinations of maximum instance sizes and counts, all of which reach the current 512,000 MB limit:
114+
Currently, each region in a given subscription has a memory limit of `512,000 MB` for all instances of apps running on Flex Consumption plans in that region. This means that in a given subscription and region, you could have any of the following combinations of maximum instance sizes and counts, all of which reach the current `512,000 MB` limit. For example:
115115

116116
| Instance memory size (MB) | Max instance counts (per region) |
117117
| ----- | ---- |
118118
| `2048 MB` | 250 |
119119
| `4096 MB` | 125 |
120120

121-
You could have any other combination of instance memory sizes and counts in a given region, as long as they stay under the 512,000 MB limit. If your apps require a larger quota, you can create a support ticket to request a quota increase.
121+
You could have any other combination of instance memory sizes and counts in a given region, as long as they stay under the `512,000 MB` limit. If your apps require a larger quota, you can create a support ticket to request a quota increase.
122122

123123
## Deprecated properties and settings
124124

@@ -135,7 +135,8 @@ Keep these other considerations in mind when using Flex Consumption plan during
135135
+ Continuous deployment using Azure DevOps Tasks (`AzureFunctionApp@2`)
136136
+ Continuous deployment using GitHub Actions (`functions-action@v1`)
137137
+ **Scale**: The lowest maximum scale in preview is `40`. The highest currently supported value is `1000`.
138-
+ **Authorization**: EasyAuth isn't currently supported. Unauthenticated callers currently aren't blocked when EasyAuth is enabled in a Flex Consumption plan app.
138+
+ **Authorization**: EasyAuth is currently not supported. Unauthenticated callers currently aren't blocked when EasyAuth is enabled in a Flex Consumption plan app.
139+
+ **CORS**: CORS settings are currently not supported. Exceptions might occur if CORS is configured for Flex Consumption apps.
139140

140141
## Related articles
141142

articles/azure-netapp-files/understand-guidelines-active-directory-domain-service-site.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -79,6 +79,7 @@ If you delete a volume in Azure NetApp Files, the associated DNS records (such a
7979
* Azure NetApp Files supports standard and secure dynamic DNS updates. If you require secure dynamic DNS updates, ensure that secure updates are configured on the DNS servers.
8080
* If dynamic DNS updates are not used, you need to manually create an A record and a PTR record for the AD DS computer account(s) created in the AD DS **Organizational Unit** (specified in the Azure NetApp Files AD connection) to support Azure NetApp Files LDAP Signing, LDAP over TLS, SMB, dual-protocol, or Kerberos NFSv4.1 volumes.
8181
* For complex or large AD DS topologies, [DNS Policies or DNS subnet prioritization may be required to support LDAP enabled NFS volumes](#ad-ds-ldap-discover).
82+
* If DNS scavenging is enabled (where stale DNS entries are automatically pruned based on timestamp/age) and dynamic DNS was used to create the DNS records for the Azure NetApp Files volume, the scavenger process might inadvertently prune the records for the volume. This pruning can lead to a service outage for name-based queries. Until this issue is resolved, manually create DNS A/AAAA and PTR entries for the Azure NetApp Files volume if DNS scavenging is enabled.
8283

8384
### Time source requirements
8485

articles/data-factory/TOC.yml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -350,8 +350,6 @@ items:
350350
href: continuous-integration-delivery.md
351351
- name: Automate with Azure Pipelines releases
352352
href: continuous-integration-delivery-automate-azure-pipelines.md
353-
- name: Automate with GitHub Actions
354-
href: continuous-integration-delivery-automate-github-actions.md
355353
- name: Manually promote a Resource Manager template for each environment
356354
href: continuous-integration-delivery-manual-promotion.md
357355
- name: Use custom parameters with a Resource Manager template

0 commit comments

Comments
 (0)