Skip to content

Commit c2811d2

Browse files
authored
Merge branch 'MicrosoftDocs:main' into heidist-june12
2 parents 386546b + a900156 commit c2811d2

File tree

277 files changed

+3101
-2697
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

277 files changed

+3101
-2697
lines changed

.openpublishing.redirection.json

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4087,10 +4087,12 @@
40874087
"source_path_from_root": "/articles/event-grid/move-domains-across-regions.md",
40884088
"redirect_url": "/azure",
40894089
"redirect_document_id": false
4090+
},
4091+
{
4092+
"source_path_from_root": "/articles/data-factory/continuous-integration-delivery-automate-github-actions.md",
4093+
"redirect_url": "/azure",
4094+
"redirect_document_id": false
40904095
}
4091-
4092-
4093-
4094-
4096+
40954097
]
40964098
}

articles/active-directory-b2c/TOC.yml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -824,8 +824,6 @@
824824
href: user-flow-versions-legacy.md
825825
- name: Resources
826826
items:
827-
- name: Azure Roadmap
828-
href: https://azure.microsoft.com/updates/?status=nowavailable,inpreview,indevelopment&category=identity,security&query=b2c
829827
- name: Frequently asked questions
830828
href: ./faq.yml
831829
displayName: FAQ

articles/ai-services/computer-vision/includes/identity-curl-quickstart.md

Lines changed: 81 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,7 @@ Get started with facial recognition using the Face REST API. The Face service pr
2424
* You'll need the key and endpoint from the resource you create to connect your application to the Face API. You'll paste your key and endpoint into the code below later in the quickstart.
2525
* You can use the free pricing tier (`F0`) to try the service, and upgrade later to a paid tier for production.
2626
* [PowerShell version 6.0+](/powershell/scripting/install/installing-powershell-core-on-windows), or a similar command-line application.
27-
* [cURL](https://curl.haxx.se/) installed.
28-
27+
* [cURL](https://curl.se/) installed.
2928

3029

3130

@@ -34,58 +33,134 @@ Get started with facial recognition using the Face REST API. The Face service pr
3433
> [!NOTE]
3534
> If you haven't received access to the Face service using the [intake form](https://aka.ms/facerecognition), some of these functions won't work.
3635
37-
1. First, call the Detect API on the source face. This is the face that we'll try to identify from the larger group. Copy the following command to a text editor, insert your own key, and then copy it into a shell window and run it.
36+
1. First, call the Detect API on the source face. This is the face that we'll try to identify from the larger group. Copy the following command to a text editor, insert your own key and endpoint, and then copy it into a shell window and run it.
37+
38+
#### [Windows](#tab/windows)
39+
40+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_detect":::
41+
42+
#### [Linux](#tab/linux)
3843

3944
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_detect":::
4045

46+
---
47+
4148
Save the returned face ID string to a temporary location. You'll use it again at the end.
4249

43-
1. Next you'll need to create a **LargePersonGroup**. This object will store the aggregated face data of several persons. Run the following command, inserting your own key. Optionally, change the group's name and metadata in the request body.
50+
1. Next you'll need to create a **LargePersonGroup** and give it an arbitrary ID that matches regex pattern `^[a-z0-9-_]+$`. This object will store the aggregated face data of several persons. Run the following command, inserting your own key. Optionally, change the group's name and metadata in the request body.
51+
52+
#### [Windows](#tab/windows)
53+
54+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_create_persongroup":::
55+
56+
#### [Linux](#tab/linux)
4457

4558
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_create_persongroup":::
4659

47-
Save the returned ID of the created group to a temporary location.
60+
---
61+
62+
Save the specified ID of the created group to a temporary location.
4863

4964
1. Next, you'll create **Person** objects that belong to the group. Run the following command, inserting your own key and the ID of the **LargePersonGroup** from the previous step. This command creates a **Person** named "Family1-Dad".
5065

66+
#### [Windows](#tab/windows)
67+
68+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_create_person":::
69+
70+
#### [Linux](#tab/linux)
71+
5172
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_create_person":::
5273

74+
---
75+
5376
After you run this command, run it again with different input data to create more **Person** objects: "Family1-Mom", "Family1-Son", "Family1-Daughter", "Family2-Lady", and "Family2-Man".
5477

5578
Save the IDs of each **Person** created; it's important to keep track of which person name has which ID.
5679

5780
1. Next you'll need to detect new faces and associate them with the **Person** objects that exist. The following command detects a face from the image *Family1-Dad1.jpg* and adds it to the corresponding person. You need to specify the `personId` as the ID that was returned when you created the "Family1-Dad" **Person** object. The image name corresponds to the name of the created **Person**. Also enter the **LargePersonGroup** ID and your key in the appropriate fields.
5881

82+
#### [Windows](#tab/windows)
83+
84+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_add_face":::
85+
86+
#### [Linux](#tab/linux)
87+
5988
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_add_face":::
6089

90+
---
91+
6192
Then, run the above command again with a different source image and target **Person**. The images available are: *Family1-Dad1.jpg*, *Family1-Dad2.jpg* *Family1-Mom1.jpg*, *Family1-Mom2.jpg*, *Family1-Son1.jpg*, *Family1-Son2.jpg*, *Family1-Daughter1.jpg*, *Family1-Daughter2.jpg*, *Family2-Lady1.jpg*, *Family2-Lady2.jpg*, *Family2-Man1.jpg*, and *Family2-Man2.jpg*. Be sure that the **Person** whose ID you specify in the API call matches the name of the image file in the request body.
6293

6394
At the end of this step, you should have multiple **Person** objects that each have one or more corresponding faces, detected directly from the provided images.
6495

6596
1. Next, train the **LargePersonGroup** with the current face data. The training operation teaches the model how to associate facial features, sometimes aggregated from multiple source images, to each single person. Insert the **LargePersonGroup** ID and your key before running the command.
6697

98+
#### [Windows](#tab/windows)
99+
100+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_train":::
101+
102+
#### [Linux](#tab/linux)
103+
67104
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_train":::
105+
106+
---
68107

108+
1. Check whether the training status is succeeded. If not, wait for a while and query again.
109+
110+
#### [Windows](#tab/windows)
111+
112+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_check_status":::
113+
114+
#### [Linux](#tab/linux)
115+
116+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_check_status":::
117+
118+
---
119+
69120
1. Now you're ready to call the Identify API, using the source face ID from the first step and the **LargePersonGroup** ID. Insert these values into the appropriate fields in the request body, and insert your key.
70121

122+
#### [Windows](#tab/windows)
123+
124+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_identify":::
125+
126+
#### [Linux](#tab/linux)
127+
71128
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_identify":::
72129

130+
---
131+
73132
The response should give you a **Person** ID indicating the person identified with the source face. It should be the ID that corresponds to the "Family1-Dad" person, because the source face is of that person.
74133

75134
1. To do face verification, you'll use the **Person** ID returned in the previous step, the **LargePersonGroup** ID, and also the source face ID. Insert these values into the fields in the request body, and insert your key.
76135

136+
#### [Windows](#tab/windows)
137+
138+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="verify":::
139+
140+
#### [Linux](#tab/linux)
141+
77142
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="verify":::
78143

144+
---
145+
79146
The response should give you a boolean verification result along with a confidence value.
80147

81148

82149

83150
## Clean up resources
84151

85-
To delete the **LargePersonGroup** you created in this exercise, run the LargePersonGroup - Delete call.
152+
To delete the **LargePersonGroup** you created in this exercise, run the [LargePersonGroup - Delete](/rest/api/face/person-group-operations/delete-large-person-group) call.
153+
154+
#### [Windows](#tab/windows)
155+
156+
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.ps1" ID="identify_delete":::
157+
158+
#### [Linux](#tab/linux)
86159

87160
:::code source="~/cognitive-services-quickstart-code/curl/face/detect.sh" ID="identify_delete":::
88161

162+
---
163+
89164
If you want to clean up and remove an Azure AI services subscription, you can delete the resource or resource group. Deleting the resource group also deletes any other resources associated with it.
90165

91166
* [Portal](../../multi-service-resource.md?pivots=azportal#clean-up-resources)

articles/ai-services/openai/assistants-reference.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ description: Learn how to use Azure OpenAI's Python & REST API with Assistants.
55
manager: nitinme
66
ms.service: azure-ai-openai
77
ms.topic: conceptual
8-
ms.date: 05/22/2024
8+
ms.date: 06/13/2024
99
author: mrbullwinkle
1010
ms.author: mbullwin
1111
recommendations: false
@@ -35,7 +35,7 @@ Create an assistant with a model and instructions.
3535
| name | string or null | Optional | The name of the assistant. The maximum length is 256 characters.|
3636
| description| string or null | Optional | The description of the assistant. The maximum length is 512 characters.|
3737
| instructions | string or null | Optional | The system instructions that the assistant uses. The maximum length is 256,000 characters.|
38-
| tools | array | Optional | Defaults to []. A list of tools enabled on the assistant. There can be a maximum of 128 tools per assistant. Tools can currently be of types `code_interpreter`, or `function`.|
38+
| tools | array | Optional | Defaults to []. A list of tools enabled on the assistant. There can be a maximum of 128 tools per assistant. Tools can currently be of types `code_interpreter`, or `function`. A `function` description can be a maximum of 1,024 characters. |
3939
| file_ids | array | Optional | Defaults to []. A list of file IDs attached to this assistant. There can be a maximum of 20 files attached to the assistant. Files are ordered by their creation date in ascending order.|
4040
| metadata | map | Optional | Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maximum of 512 characters long.|
4141
| temperature | number or null | Optional | Defaults to 1. Determines what sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. |
@@ -360,7 +360,7 @@ Modifies an assistant.
360360
| `name` | string or null | Optional | The name of the assistant. The maximum length is 256 characters. |
361361
| `description` | string or null | Optional | The description of the assistant. The maximum length is 512 characters. |
362362
| `instructions` | string or null | Optional | The system instructions that the assistant uses. The maximum length is 32768 characters. |
363-
| `tools` | array | Optional | Defaults to []. A list of tools enabled on the assistant. There can be a maximum of 128 tools per assistant. Tools can be of types code_interpreter, or function. |
363+
| `tools` | array | Optional | Defaults to []. A list of tools enabled on the assistant. There can be a maximum of 128 tools per assistant. Tools can be of types code_interpreter, or function. A `function` description can be a maximum of 1,024 characters. |
364364
| `file_ids` | array | Optional | Defaults to []. A list of File IDs attached to this assistant. There can be a maximum of 20 files attached to the assistant. Files are ordered by their creation date in ascending order. If a file was previously attached to the list but does not show up in the list, it will be deleted from the assistant. |
365365
| `metadata` | map | Optional | Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maximum of 512 characters long. |
366366

@@ -517,7 +517,7 @@ Assistants use the [same API for file upload as fine-tuning](/rest/api/azureopen
517517
| `description` | string or null | The description of the assistant. The maximum length is 512 characters.|
518518
| `model` | string | Name of the model deployment name to use.|
519519
| `instructions` | string or null | The system instructions that the assistant uses. The maximum length is 32768 characters.|
520-
| `tools` | array | A list of tool enabled on the assistant. There can be a maximum of 128 tools per assistant. Tools can be of types code_interpreter, or function.|
520+
| `tools` | array | A list of tool enabled on the assistant. There can be a maximum of 128 tools per assistant. Tools can be of types code_interpreter, or function. A `function` description can be a maximum of 1,024 characters.|
521521
| `file_ids` | array | A list of file IDs attached to this assistant. There can be a maximum of 20 files attached to the assistant. Files are ordered by their creation date in ascending order.|
522522
| `metadata` | map | Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maximum of 512 characters long.|
523523

articles/ai-services/openai/quotas-limits.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,7 @@ The following sections provide you with a quick guide to the default quotas and
2727
| OpenAI resources per region per Azure subscription | 30 |
2828
| Default DALL-E 2 quota limits | 2 concurrent requests |
2929
| Default DALL-E 3 quota limits| 2 capacity units (6 requests per minute)|
30+
| Default Whisper quota limits | 3 requests per minute |
3031
| Maximum prompt tokens per request | Varies per model. For more information, see [Azure OpenAI Service models](./concepts/models.md)|
3132
| Max fine-tuned model deployments | 5 |
3233
| Total number of training jobs per resource | 100 |
@@ -48,6 +49,7 @@ The following sections provide you with a quick guide to the default quotas and
4849
| GPT-4o max images per request (# of images in the messages array/conversation history) | 10 |
4950
| GPT-4 `vision-preview` & GPT-4 `turbo-2024-04-09` default max tokens | 16 <br><br> Increase the `max_tokens` parameter value to avoid truncated responses. GPT-4o max tokens defaults to 4096. |
5051

52+
5153
## Regional quota limits
5254

5355
[!INCLUDE [Quota](./includes/model-matrix/quota.md)]
@@ -99,7 +101,7 @@ If your Azure subscription is linked to certain [offer types](https://azure.micr
99101
|Azure for Students, Free Trials | 1 K (all models)|
100102
| Monthly credit card based accounts <sup>1</sup> | GPT 3.5 Turbo Series: 30 K <br> GPT-4 series: 8 K <br> |
101103

102-
<sup>1</sup>This currently applies to [offer type 0003P](https://azure.microsoft.com/support/legal/offer-details/)
104+
<sup>1</sup> This currently applies to [offer type 0003P](https://azure.microsoft.com/support/legal/offer-details/)
103105

104106
In the Azure portal you can view what offer type is associated with your subscription by navigating to your subscription and checking the subscriptions overview pane. Offer type corresponds to the plan field in the subscription overview.
105107

articles/ai-services/openai/whats-new.md

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ ms.custom:
1010
- ignite-2023
1111
- references_regions
1212
ms.topic: whats-new
13-
ms.date: 06/11/2024
13+
ms.date: 06/13/2024
1414
recommendations: false
1515
---
1616

@@ -20,6 +20,10 @@ This article provides a summary of the latest releases and major documentation u
2020

2121
## June 2024
2222

23+
### Token based billing for fine-tuning
24+
25+
* Azure OpenAI fine-tuning billing is now based on the number of tokens in your training file – instead of the total elapsed training time. This can result in a significant cost reduction for some training runs, and makes estimating fine-tuning costs much easier. To learn more, you can consult the [official announcement](https://techcommunity.microsoft.com/t5/ai-azure-ai-services-blog/pricing-update-token-based-billing-for-fine-tuning-training/ba-p/4164465).
26+
2327
### GPT-4o released in new regions
2428

2529
* GPT-4o is now also available in:
Lines changed: 48 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,48 @@
1+
---
2+
title: Content Credentials in Azure Text to Speech Avatar
3+
titleSuffix: Azure Text to Speech Avatar
4+
description: Learn about the content credentials feature, which lets you verify that a video was generated by the Azure text to speech avatar system.
5+
author: sally-baolian
6+
ms.author: v-baolianzou
7+
ms.service: azure-ai-speech
8+
ms.topic: conceptual
9+
ms.date: 6/11/2024
10+
manager: nitinme
11+
---
12+
13+
# Content credentials
14+
15+
The high-quality models in the Azure text to speech avatar feature generate realistic avatar videos from text input. To improve the transparency of the generated content, the Azure text to speech avatar provides content credentials, a tamper-evident way to disclose the origin and history of the content. Content credentials are based on an open technical specification from the [Coalition for Content Provenance and Authenticity (C2PA)](https://www.c2pa.org), a Joint Development Foundation project.
16+
17+
## What are content credentials?
18+
19+
Content credentials in the Azure text to speech avatar provide customers with information about the origin of an avatar video. This information is represented by a manifest attached to the video. The manifest is cryptographically signed by a certificate that traces back to Azure text to speech avatar.
20+
21+
The manifest contains several key pieces of information:
22+
23+
| Field name | Field content |
24+
| --- | --- |
25+
| `"generator"` | This field has a value of `"Microsoft Azure Text To Speech Avatar Service"` for all applicable videos, attesting to the AI-generated nature of the video. |
26+
| `"when"` | The timestamp of when the content credentials were created. |
27+
28+
Content credentials in the Azure text to speech avatar can help people understand when video content is generated by the Azure text to speech avatar system. For more information on how to responsibly build solutions with text to speech avatar models, visit the [Text to speech transparency note](/legal/cognitive-services/speech-service/text-to-speech/transparency-note?context=/azure/ai-services/speech-service/context/context).
29+
30+
## Limitations
31+
32+
The content credentials are only supported in video files generated by batch synthesis of text to speech avatar, and only `mp4` file format is supported.
33+
34+
## How do I leverage content credentials in my solution today?
35+
36+
You may leverage content credentials by:
37+
38+
- Ensuring that your Azure text to speech avatar generated video files contain content credentials
39+
40+
No additional set-up is necessary. Content credentials are automatically applied to all applicable videos generated by the Azure text to speech avatar.
41+
42+
## Verifying that a video file has content credentials
43+
44+
As for now, self-serve verification of content credentials for text to speech avatar video isn't yet available. You can contact [[email protected]](mailto:[email protected]) through email for verification of content credentials of Azure text to speech avatar generated videos.
45+
46+
## Next steps
47+
48+
* [Use batch synthesis for text to speech avatar](./batch-synthesis-avatar.md)

articles/ai-services/speech-service/toc.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -200,6 +200,8 @@ items:
200200
- name: What is text to speech avatar?
201201
href: text-to-speech-avatar/what-is-text-to-speech-avatar.md
202202
displayName: avatar
203+
- name: Content credentials
204+
href: text-to-speech-avatar/content-credentials.md
203205
- name: How to synthesize text to speech avatar
204206
items:
205207
- name: Real-time synthesis

0 commit comments

Comments
 (0)