Skip to content

Commit 0ebb3d5

Browse files
author
David Curwin
committed
Merge branch 'main' of https://github.com/MicrosoftDocs/azure-docs-pr into wi-207920-new-agent-feb11-2024
2 parents 3a1d304 + a61cfe1 commit 0ebb3d5

File tree

140 files changed

+1514
-908
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

140 files changed

+1514
-908
lines changed

.openpublishing.redirection.json

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -10640,11 +10640,6 @@
1064010640
"redirect_url": "/azure/orbital/overview",
1064110641
"redirect_document_id": false
1064210642
},
10643-
{
10644-
"source_path_from_root": "/articles/load-balancer/cross-region-overview.md",
10645-
"redirect_url": "/azure/reliability/reliability-load-balancer",
10646-
"redirect_document_id": false
10647-
},
1064810643
{
1064910644
"source_path_from_root": "/articles/load-balancer/load-balancer-standard-availability-zones.md",
1065010645
"redirect_url": "/azure/reliability/reliability-load-balancer",

articles/active-directory-b2c/custom-policies-series-validate-user-input.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ Azure Active Directory B2C (Azure AD B2C) custom policy not only allows you to m
4949

5050
## Step 1 - Validate user input by limiting user input options
5151

52-
If you know all the possible values that a user can enter for a given input, you can provide a finite set of values that a user must select from. You can use *DropdownSinglSelect*, *CheckboxMultiSelect*, and *RadioSingleSelect* [UserInputType](claimsschema.md#userinputtype) for this purpose. In this article, you'll use a *RadioSingleSelect* input type:
52+
If you know all the possible values that a user can enter for a given input, you can provide a finite set of values that a user must select from. You can use *DropdownSingleSelect*, *CheckboxMultiSelect*, and *RadioSingleSelect* [UserInputType](claimsschema.md#userinputtype) for this purpose. In this article, you'll use a *RadioSingleSelect* input type:
5353

5454
1. In VS Code, open the file `ContosoCustomPolicy.XML`.
5555

articles/active-directory-b2c/session-behavior.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,9 @@ With single sign-on, users sign in once with a single account and get access to
2525

2626
When the user initially signs in to an application, Azure AD B2C persists a cookie-based session. Upon subsequent authentication requests, Azure AD B2C reads and validates the cookie-based session, and issues an access token without prompting the user to sign in again. If the cookie-based session expires or becomes invalid, the user is prompted to sign-in again.
2727

28+
>[!NOTE]
29+
>If the user uses a browser that blocks third-party cookies, there are limitations with SSO due to limited access to the cookie-based session. The most user-visible impact is that there are more interactions required for sign-in. Additionally, the front channel sign-out doesn't immediately clear authentication state from federated applications. Check our recommended ways about [how to handle third-party cookie blocking in browsers](/entra/identity-platform/reference-third-party-cookies-spas).
30+
2831
## Prerequisites
2932

3033
[!INCLUDE [active-directory-b2c-customization-prerequisites](../../includes/active-directory-b2c-customization-prerequisites.md)]

articles/ai-services/document-intelligence/concept-read.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ ms.author: lajanuar
3131

3232
> [!NOTE]
3333
>
34-
> For extracting text from external images like labels, street signs, and posters, use the [Azure AI Vision v4.0 preview Read](../../ai-services/Computer-vision/concept-ocr.md) feature optimized for general, non-document images with a performance-enhanced synchronous API that makes it easier to embed OCR in your user experience scenarios.
34+
> For extracting text from external images like labels, street signs, and posters, use the [Azure AI Image Analysis v4.0 Read](../../ai-services/Computer-vision/concept-ocr.md) feature optimized for general, non-document images with a performance-enhanced synchronous API that makes it easier to embed OCR in your user experience scenarios.
3535
>
3636
3737
Document Intelligence Read Optical Character Recognition (OCR) model runs at a higher resolution than Azure AI Vision Read and extracts print and handwritten text from PDF documents and scanned images. It also includes support for extracting text from Microsoft Word, Excel, PowerPoint, and HTML documents. It detects paragraphs, text lines, words, locations, and languages. The Read model is the underlying OCR engine for other Document Intelligence prebuilt models like Layout, General Document, Invoice, Receipt, Identity (ID) document, Health insurance card, W2 in addition to custom models.

articles/ai-services/language-service/question-answering/how-to/azure-openai-integration.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ ms.service: azure-ai-language
66
author: jboback
77
ms.author: jboback
88
ms.topic: how-to
9-
ms.date: 12/19/2023
9+
ms.date: 02/09/2024
1010
---
1111

1212
# Connect Custom Question Answering with Azure OpenAI on your data
@@ -72,7 +72,7 @@ At the same time, customers often require a custom answer authoring experience t
7272

7373
:::image type="content" source="../media/question-answering/chat-playground.png" alt-text="A screenshot of the playground page of the Azure OpenAI Studio with sections highlighted." lightbox="../media/question-answering/chat-playground.png":::
7474

75-
You can now start exploring Azure OpenAI capabilities with a no-code approach through the chat playground. It's simply a text box where you can submit a prompt to generate a completion. From this page, you can quickly iterate and experiment with the capabilities. You can also launch a [web app](../../..//openai/concepts/use-your-data.md#using-the-web-app) to chat with the model over the web.
75+
You can now start exploring Azure OpenAI capabilities with a no-code approach through the chat playground. It's simply a text box where you can submit a prompt to generate a completion. From this page, you can quickly iterate and experiment with the capabilities. You can also launch a [web app](../../../openai/how-to/use-web-app.md) to chat with the model over the web.
7676

7777
## Next steps
7878
* [Using Azure OpenAI on your data](../../../openai/concepts/use-your-data.md)

0 commit comments

Comments
 (0)