Skip to content

Commit 91c1180

Browse files
Merge branch 'main' of https://github.com/MicrosoftDocs/azure-docs-pr into samples-auth
2 parents c45dde5 + efbfbd8 commit 91c1180

File tree

1,084 files changed

+9886
-5031
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,084 files changed

+9886
-5031
lines changed

articles/active-directory-b2c/oauth2-error-technical-profile.md

Lines changed: 52 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ manager: CelesteDG
99
ms.service: active-directory
1010

1111
ms.topic: reference
12-
ms.date: 01/11/2024
12+
ms.date: 05/07/2024
1313
ms.author: kengaderdus
1414
ms.subservice: B2C
1515

@@ -73,6 +73,43 @@ The following example shows a technical profile for `ReturnOAuth2Error`:
7373
</ClaimsProviders> -->
7474
```
7575

76+
## Define claims transformation to generate custom values of error code and error message
77+
78+
Use these steps to generate custom values of error code and error message:
79+
80+
1. Locate the `ClaimsTransformations` element, then add the following code inside it
81+
82+
```xml
83+
<!--
84+
<ClaimsTransformations> -->
85+
<ClaimsTransformation Id="GenerateErrorCode" TransformationMethod="CreateStringClaim">
86+
<InputParameters>
87+
<InputParameter Id="value" DataType="string" Value="Error_001" />
88+
</InputParameters>
89+
<OutputClaims>
90+
<OutputClaim ClaimTypeReferenceId="errorCode" TransformationClaimType="createdClaim" />
91+
</OutputClaims>
92+
</ClaimsTransformation>
93+
<ClaimsTransformation Id="GenerateErrorMessage" TransformationMethod="CreateStringClaim">
94+
<InputParameters>
95+
<InputParameter Id="value" DataType="string" Value="Insert error description." />
96+
</InputParameters>
97+
<OutputClaims>
98+
<OutputClaim ClaimTypeReferenceId="errorMessage" TransformationClaimType="createdClaim" />
99+
</OutputClaims>
100+
</ClaimsTransformation>
101+
<!--
102+
</ClaimsTransformations> -->
103+
```
104+
105+
1. Add the two claims transformations in the `OutputClaimsTransformations` element of any technical profile before Oauth2 technical that you define:
106+
107+
```xml
108+
<OutputClaimsTransformations>
109+
<OutputClaimsTransformation ReferenceId="generateErrorCode" />
110+
<OutputClaimsTransformation ReferenceId="generateErrorMessage" />
111+
</OutputClaimsTransformations>
112+
```
76113
## Input claims
77114

78115
The **InputClaims** element contains a list of claims required to return OAuth2 error.
@@ -123,6 +160,20 @@ In the following example:
123160
</UserJourney>
124161
```
125162

163+
Optionally, you can use preconditions to manipulate the Oauth2 error technical profile. For example, if there is no email claim, you can set to call Oauth2 error technical profile:
164+
165+
```xml
166+
<OrchestrationStep Order="3" Type="SendClaims" CpimIssuerTechnicalProfileReferenceId="ReturnOAuth2Error">
167+
<Preconditions>
168+
<Precondition Type="ClaimsExist" ExecuteActionsIf="false">
169+
<Value>email</Value>
170+
<Action>SkipThisOrchestrationStep</Action>
171+
</Precondition>
172+
</Preconditions>
173+
</OrchestrationStep>
174+
```
175+
176+
126177
## Next steps
127178

128179
Learn about [UserJourneys](userjourneys.md)

articles/active-directory-b2c/partner-saviynt.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -70,6 +70,9 @@ The following architecture diagram illustrates the implementation.
7070
2. Create delegated administration policies.
7171
3. Assign users the delegated administrator role.
7272

73+
>[!NOTE]
74+
>User administrator role will be required for a Service Principal Name (SPN) to integrate with Saviynt EIC system.
75+
7376
## Configure Azure AD B2C with Saviynt
7477

7578
Use the following instructions to create an application, delete users, and more.

articles/ai-services/computer-vision/how-to/find-similar-faces.md

Lines changed: 0 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -46,17 +46,6 @@ The following code uses the above method to get face data from a series of image
4646
[!code-csharp[](~/cognitive-services-quickstart-code/dotnet/Face/FaceQuickstart.cs?name=snippet_loadfaces)]
4747

4848

49-
#### [JavaScript](#tab/javascript)
50-
51-
The following face detection method is optimized for comparison operations. It doesn't extract detailed face attributes, and it uses an optimized recognition model.
52-
53-
:::code language="js" source="~/cognitive-services-quickstart-code/javascript/Face/sdk_quickstart.js" id="recognize":::
54-
55-
The following code uses the above method to get face data from a series of images.
56-
57-
:::code language="js" source="~/cognitive-services-quickstart-code/javascript/Face/sdk_quickstart.js" id="snippet_loadfaces":::
58-
59-
6049
#### [REST API](#tab/rest)
6150

6251
Copy the following cURL command and insert your key and endpoint where appropriate. Then run the command to detect one of the target faces.
@@ -89,12 +78,6 @@ The following code prints the match details to the console:
8978

9079
[!code-csharp[](~/cognitive-services-quickstart-code/dotnet/Face/FaceQuickstart.cs?name=snippet_find_similar_print)]
9180

92-
#### [JavaScript](#tab/javascript)
93-
94-
The following method takes a set of target faces and a single source face. Then, it compares them and finds all the target faces that are similar to the source face. Finally, it prints the match details to the console.
95-
96-
:::code language="js" source="~/cognitive-services-quickstart-code/javascript/Face/sdk_quickstart.js" id="find_similar":::
97-
9881

9982
#### [REST API](#tab/rest)
10083

articles/ai-services/computer-vision/includes/quickstarts-sdk/identity-javascript-sdk.md

Lines changed: 12 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ ms.author: pafarley
1515

1616
Get started with facial recognition using the Face client library for JavaScript. Follow these steps to install the package and try out the example code for basic tasks. The Face service provides you with access to advanced algorithms for detecting and recognizing human faces in images. Follow these steps to install the package and try out the example code for basic face identification using remote images.
1717

18-
[Reference documentation](/javascript/api/overview/azure/cognitiveservices-face-readme) | [Package (npm)](https://www.npmjs.com/package/@azure/cognitiveservices-face) | [Samples](/samples/browse/?products=azure&term=face&languages=javascript)
18+
[Reference documentation](https://aka.ms/azsdk-javascript-face-ref) | [Library source code](https://github.com/Azure/azure-sdk-for-js/tree/main/sdk/face/ai-vision-face-rest) | [Package (npm)](https://www.npmjs.com/package/@azure-rest/ai-vision-face) | [Samples](https://github.com/Azure/azure-sdk-for-js/tree/main/sdk/face/ai-vision-face-rest/samples)
1919

2020
## Prerequisites
2121

@@ -47,10 +47,10 @@ Get started with facial recognition using the Face client library for JavaScript
4747
npm init
4848
```
4949

50-
1. Install the `ms-rest-azure` and `azure-cognitiveservices-face` npm packages:
50+
1. Install the `@azure-rest/ai-vision-face` npm packages:
5151

5252
```console
53-
npm install @azure/cognitiveservices-face @azure/ms-rest-js uuid
53+
npm install @azure-rest/ai-vision-face
5454
```
5555

5656
Your app's `package.json` file is updated with the dependencies.
@@ -96,12 +96,18 @@ Add face to the person group person: (Family2-Lady) from image: Family2-Lady1.jp
9696
Done adding faces to person group.
9797

9898
Training person group: c08484e0-044b-4610-8b7e-c957584e5d2d.
99-
Waiting 10 seconds...
10099
Training status: succeeded.
101-
100+
No persons identified for face with ID 259dd648-be70-499c-9942-3512594e21eb
102101
Person: Family1-Mom is identified for face in: identification1.jpg with ID: b7f7f542-c338-4a40-ad52-e61772bc6e14. Confidence: 0.96921.
103102
Person: Family1-Son is identified for face in: identification1.jpg with ID: 600dc1b4-b2c4-4516-87de-edbbdd8d7632. Confidence: 0.92886.
104103
Person: Family1-Dad is identified for face in: identification1.jpg with ID: e83b494f-9ad2-473f-9d86-3de79c01e345. Confidence: 0.96725.
104+
Verification result between face bb7f7f542-c338-4a40-ad52-e61772bc6e14 and person de1d7dea-a393-4f69-9062-10cb66d4cf17: true with confidence: 0.96921
105+
Verification result between face 600dc1b4-b2c4-4516-87de-edbbdd8d7632 and person 05fd84e4-41b0-4716-b767-4376e33fa207: true with confidence: 0.92886
106+
Verification result between face e83b494f-9ad2-473f-9d86-3de79c01e345 and person c5124fe2-39dd-47ba-9163-1ed2998fdeb2: true with confidence: 0.96725
107+
108+
Deleting person group: c08484e0-044b-4610-8b7e-c957584e5d2d
109+
110+
Done.
105111
```
106112

107113

@@ -120,4 +126,4 @@ In this quickstart, you learned how to use the Face client library for JavaScrip
120126
> [Specify a face detection model version](../../how-to/specify-detection-model.md)
121127
122128
* [What is the Face service?](../../overview-identity.md)
123-
* More extensive sample code can be found on [GitHub](https://github.com/Azure-Samples/cognitive-services-quickstart-code/blob/master/javascript/Face/sdk_quickstart.js).
129+
* More extensive sample code can be found on [GitHub](https://github.com/Azure/azure-sdk-for-js/tree/main/sdk/face/ai-vision-face-rest/samples).

articles/ai-services/openai/concepts/model-retirements.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -66,9 +66,9 @@ These models are currently available for use in Azure OpenAI Service.
6666
| `gpt-35-turbo` | 0125 | No earlier than Feb 22, 2025 |
6767
| `gpt-4`<br>`gpt-4-32k` | 0314 | No earlier than July 13, 2024 |
6868
| `gpt-4`<br>`gpt-4-32k` | 0613 | No earlier than Sep 30, 2024 |
69-
| `gpt-4` | 1106-preview | To be upgraded to `gpt-4` Version: `turbo-2024-04-09`, starting on June 10, 2024, or later **<sup>1</sup>** |
70-
| `gpt-4` | 0125-preview |To be upgraded to `gpt-4` Version: `turbo-2024-04-09`, starting on June 10, 2024, or later **<sup>1</sup>** |
71-
| `gpt-4` | vision-preview | To be upgraded to `gpt-4` Version: `turbo-2024-04-09`, starting on June 10, 2024, or later **<sup>1</sup>** |
69+
| `gpt-4` | 1106-preview | To be upgraded to `gpt-4` Version: `turbo-2024-04-09`, starting on July 15, 2024, or later **<sup>1</sup>** |
70+
| `gpt-4` | 0125-preview |To be upgraded to `gpt-4` Version: `turbo-2024-04-09`, starting on July 15, 2024, or later **<sup>1</sup>** |
71+
| `gpt-4` | vision-preview | To be upgraded to `gpt-4` Version: `turbo-2024-04-09`, starting on July 15, 2024, or later **<sup>1</sup>** |
7272
| `gpt-3.5-turbo-instruct` | 0914 | No earlier than Sep 14, 2025 |
7373
| `text-embedding-ada-002` | 2 | No earlier than April 3, 2025 |
7474
| `text-embedding-ada-002` | 1 | No earlier than April 3, 2025 |

articles/ai-services/openai/concepts/use-your-data.md

Lines changed: 12 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -78,7 +78,9 @@ You might want to consider using an Azure AI Search index when you either want t
7878
* Reuse an index created before by ingesting data from other data sources.
7979

8080
> [!NOTE]
81-
> To use an existing index, it must have at least one searchable field.
81+
> * To use an existing index, it must have at least one searchable field.
82+
> * Set the CORS **Allow Origin Type** option to `all` and the **Allowed origins** option to `*`.
83+
8284

8385
### Search types
8486

@@ -135,6 +137,7 @@ If you want to implement additional value-based criteria for query execution, yo
135137

136138
[!INCLUDE [ai-search-ingestion](../includes/ai-search-ingestion.md)]
137139

140+
[!INCLUDE [authentication](../includes/on-your-data-authentication.md)]
138141

139142
# [Vector Database in Azure Cosmos DB for MongoDB](#tab/mongo-db)
140143

@@ -202,6 +205,8 @@ To modify the schedule, you can use the [Azure portal](https://portal.azure.com/
202205

203206
[!INCLUDE [ai-search-ingestion](../includes/ai-search-ingestion.md)]
204207

208+
[!INCLUDE [authentication](../includes/on-your-data-authentication.md)]
209+
205210
# [Upload files (preview)](#tab/file-upload)
206211

207212
Using Azure OpenAI Studio, you can upload files from your machine to try Azure OpenAI On Your Data. You also have the option to create a new Azure Blob Storage account and Azure AI Search resource. The service then stores the files to an Azure storage container and performs ingestion from the container. You can use the [quickstart](../use-your-data-quickstart.md) article to learn how to use this data source option.
@@ -210,6 +215,8 @@ Using Azure OpenAI Studio, you can upload files from your machine to try Azure O
210215

211216
[!INCLUDE [ai-search-ingestion](../includes/ai-search-ingestion.md)]
212217

218+
[!INCLUDE [authentication](../includes/on-your-data-authentication.md)]
219+
213220
# [URL/Web address (preview)](#tab/web-pages)
214221

215222
You can paste URLs and the service will store the webpage content, using it when generating responses from the model. The content in URLs/web addresses that you use need to have the following characteristics to be properly ingested:
@@ -224,6 +231,10 @@ You can paste URLs and the service will store the webpage content, using it when
224231

225232
Once you have added the URL/web address for data ingestion, the web pages from your URL are fetched and saved to Azure Blob Storage with a container name: `webpage-<index name>`. Each URL will be saved into a different container within the account. Then the files are indexed into an Azure AI Search index, which is used for retrieval when you’re chatting with the model.
226233

234+
[!INCLUDE [ai-search-ingestion](../includes/ai-search-ingestion.md)]
235+
236+
[!INCLUDE [authentication](../includes/on-your-data-authentication.md)]
237+
227238
# [Elasticsearch (preview)](#tab/elasticsearch)
228239

229240
You can connect to your [Elasticsearch vector database](https://www.elastic.co/guide/en/elasticsearch/reference/current/elasticsearch-intro.html) and chat with your data.

articles/ai-services/openai/faq.yml

Lines changed: 7 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -248,6 +248,10 @@ sections:
248248
Can I bring my own data store to use with Assistants?
249249
answer: |
250250
No. Currently Assistants supports only local files uploaded to the Assistants-managed storage. You cannot use your private storage account with Assistants.
251+
- question: |
252+
Does Assistants support customer-managed key encryption (CMK)?
253+
answer: |
254+
Today we support CMK for Threads and Files in Assistants. See the [What's new page](./whats-new.md#customer-managed-key-cmk-support-for-assistants) for available regions for this feature.
251255
- question: |
252256
Is my data used by Microsoft for training models?
253257
answer: |
@@ -259,9 +263,9 @@ sections:
259263
- question: |
260264
How am I charged for Assistants?
261265
answer: |
262-
Currently, when you use Assistants API, you're billed for the following:
263-
- Inference cost (input and output) of the base model you're using for each Assistant (for example gpt-4-0125). If you've created multiple Assistants, you will be charged for the base model attached to each Assistant.
264-
- If you've enabled the Code Interpreter tool. For example if your assistant calls Code Interpreter simultaneously in two different threads, this would create two Code Interpreter sessions, each of which would be charged. Each session is active by default for one hour, which means that you would only pay this fee once if your user keeps giving instructions to Code Interpreter in the same thread for up to one hour.
266+
* Inference cost (input and output) of the base model you're using for each Assistant (for example gpt-4-0125). If you've created multiple Assistants, you will be charged for the base model attached to each Assistant.
267+
* If you've enabled the Code Interpreter tool. For example if your assistant calls Code Interpreter simultaneously in two different threads, this would create two Code Interpreter sessions, each of which would be charged. Each session is active by default for one hour, which means that you would only pay this fee once if your user keeps giving instructions to Code Interpreter in the same thread for up to one hour.
268+
* File search is currently not billed.
265269
266270
For more information, see the [pricing page](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/).
267271
- question: |

articles/ai-services/openai/includes/assistants-ai-studio.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.author: mbullwin
1212
author: mrbullwinkle
1313
---
1414

15-
[!INCLUDE [Feature preview](../../../ai-studio/includes/feature-preview.md)]
15+
[!INCLUDE [Feature preview](~/reusable-content/ce-skilling/azure/includes/ai-studio/includes/feature-preview.md)]
1616

1717
## Prerequisites
1818

articles/ai-services/openai/includes/fine-tuning-openai-in-ai-studio.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
ms.custom: include, build-2024
1010
---
1111

12-
[!INCLUDE [Feature preview](../../../ai-studio/includes/feature-preview.md)]
12+
[!INCLUDE [Feature preview](~/reusable-content/ce-skilling/azure/includes/ai-studio/includes/feature-preview.md)]
1313

1414
## Prerequisites
1515

Lines changed: 20 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,20 @@
1+
---
2+
manager: nitinme
3+
author: aahill
4+
ms.author: aahi
5+
ms.service: azure-ai-openai
6+
ms.topic: include
7+
ms.date: 03/27/2024
8+
---
9+
10+
## Data connection
11+
12+
You need to select how you want to authenticate the connection from Azure OpenAI, Azure AI Search, and Azure blob storage. You can choose a *System assigned managed identity* or an *API key*. By selecting *API key* as the authentication type, the system will automatically populate the API key for you to connect with your Azure AI Search, Azure OpenAI, and Azure Blob Storage resources. By selecting *System assigned managed identity*, the authentication will be based on the [role assignment](../how-to/use-your-data-securely.md#role-assignments) you have. *System assigned managed identity* is selected by default for security.
13+
14+
15+
:::image type="content" source="../media/use-your-data/data-connection-authentication.png" alt-text="A screenshot showing the managed identity option in Azure OpenAI Studio." lightbox="../media/use-your-data/data-connection-authentication.png":::
16+
17+
Once you select the **next** button, it will automatically validate your setup to use the selected authentication method. If you encounter an error, see the [role assignments article](../how-to/use-your-data-securely.md#role-assignments) to update your setup.
18+
19+
Once you have fixed the setup, select **next** again to validate and proceed. API users can also [configure authentication](../references/azure-search.md#api-key-authentication-options) with assigned managed identity and API keys.
20+

0 commit comments

Comments
 (0)