Skip to content

Commit 7edd97c

Browse files
authored
Merge branch 'MicrosoftDocs:main' into haim-na/sentinel-logstash-plugin-jan24
2 parents 6abf14d + afc5cf2 commit 7edd97c

File tree

522 files changed

+2367
-3272
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

522 files changed

+2367
-3272
lines changed

articles/active-directory-b2c/custom-policies-series-validate-user-input.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -379,7 +379,7 @@ Follow the steps in [Upload custom policy file](custom-policies-series-hello-wor
379379

380380
## Step 7 - Validate user input by using validation technical profiles
381381

382-
The validation techniques we've used in step 1, step 2 and step 3 aren't applicable for all scenarios. If your business rules are complex to be defined at claim declaration level, you can configure a [Validation Technical](validation-technical-profile.md), and then call it from a [Self-Asserted Technical Profile](self-asserted-technical-profile.md).
382+
The validation techniques we've used in step 1, step 2 and step 3 aren't applicable for all scenarios. If your business rules are too complex to be defined at claim declaration level, you can configure a [Validation Technical](validation-technical-profile.md), and then call it from a [Self-Asserted Technical Profile](self-asserted-technical-profile.md).
383383

384384
> [!NOTE]
385385
> Only self-asserted technical profiles can use validation technical profiles. Learn more about [validation technical profile](validation-technical-profile.md)

articles/ai-services/document-intelligence/containers/configuration.md

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -16,22 +16,24 @@ ms.author: lajanuar
1616

1717
# Configure Document Intelligence containers
1818

19-
:::moniker range="doc-intel-2.1.0 || doc-intel-3.1.0||doc-intel-4.0.0"
19+
:::moniker range="doc-intel-2.1.0 || doc-intel-4.0.0"
2020

21-
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` only:
21+
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` for all models and `2023-07-31 (GA)` for Read and Layout only:
2222

2323
* [REST API `2022-08-31 (GA)`](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-2022-08-31/operations/AnalyzeDocument)
24+
* [REST API `2023-07-31 (GA)`](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-2023-07-31/operations/AnalyzeDocument)
2425
* [SDKs targeting `REST API 2022-08-31 (GA)`](../sdk-overview-v3-0.md)
26+
* [SDKs targeting `REST API 2023-07-31 (GA)`](../sdk-overview-v3-1.md)
2527

2628
✔️ See [**Configure Document Intelligence v3.0 containers**](?view=doc-intel-3.0.0&preserve-view=true) for supported container documentation.
2729

2830
:::moniker-end
2931

30-
:::moniker range="doc-intel-3.0.0"
32+
:::moniker range="doc-intel-3.0.0 || doc-intel-3.1.0"
3133

32-
**This content applies to:** ![checkmark](../media/yes-icon.png) **v3.0 (GA)**
34+
**This content applies to:** ![checkmark](../media/yes-icon.png) **v3.0 (GA)** ![checkmark](../media/yes-icon.png) **v3.1 (GA)**
3335

34-
With Document Intelligence containers, you can build an application architecture optimized to take advantage of both robust cloud capabilities and edge locality. Containers provide a minimalist, isolated environment that can be easily deployed on-premises and in the cloud. In this article, we show you how to configure the Document Intelligence container run-time environment by using the `docker compose` command arguments. Document Intelligence features are supported by six Document Intelligence feature containers—**Layout**, **Business Card**,**ID Document**, **Receipt**, **Invoice**, **Custom**. These containers have both required and optional settings. For a few examples, see the [Example docker-compose.yml file](#example-docker-composeyml-file) section.
36+
With Document Intelligence containers, you can build an application architecture optimized to take advantage of both robust cloud capabilities and edge locality. Containers provide a minimalist, isolated environment that can be easily deployed on-premises and in the cloud. In this article, we show you how to configure the Document Intelligence container run-time environment by using the `docker compose` command arguments. Document Intelligence features are supported by seven Document Intelligence feature containers—**Read**, **Layout**, **Business Card**,**ID Document**, **Receipt**, **Invoice**, **Custom**. These containers have both required and optional settings. For a few examples, see the [Example docker-compose.yml file](#example-docker-composeyml-file) section.
3537

3638
## Configuration settings
3739

articles/ai-services/document-intelligence/containers/disconnected.md

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -15,20 +15,22 @@ ms.author: lajanuar
1515

1616
# Containers in disconnected (offline) environments
1717

18-
:::moniker range="doc-intel-2.1.0 || doc-intel-3.1.0||doc-intel-4.0.0"
18+
:::moniker range="doc-intel-2.1.0 || doc-intel-4.0.0"
1919

20-
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)`:
20+
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` for all models and `2023-07-31 (GA)` for Read and Layout only:
2121

2222
* [REST API `2022-08-31 (GA)`](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-2022-08-31/operations/AnalyzeDocument)
23+
* [REST API `2023-07-31 (GA)`](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-2023-07-31/operations/AnalyzeDocument)
2324
* [SDKs targeting `REST API 2022-08-31 (GA)`](../sdk-overview-v3-0.md)
25+
* [SDKs targeting `REST API 2023-07-31 (GA)`](../sdk-overview-v3-1.md)
2426

2527
✔️ See [**Document Intelligence v3.0 containers in disconnected environments**](?view=doc-intel-3.0.0&preserve-view=true) for supported container documentation.
2628

2729
:::moniker-end
2830

29-
:::moniker range="doc-intel-3.0.0"
31+
:::moniker range="doc-intel-3.0.0 || doc-intel-3.1.0"
3032

31-
**This content applies to:** ![checkmark](../media/yes-icon.png) **v3.0 (GA)**
33+
**This content applies to:** ![checkmark](../media/yes-icon.png) **v3.0 (GA)** ![checkmark](../media/yes-icon.png) **v3.1 (GA)**
3234

3335
## What are disconnected containers?
3436

articles/ai-services/document-intelligence/containers/image-tags.md

Lines changed: 22 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,12 +16,14 @@ ms.author: lajanuar
1616

1717
<!-- markdownlint-disable MD051 -->
1818

19-
:::moniker range="doc-intel-2.1.0 || doc-intel-3.1.0||doc-intel-4.0.0"
19+
:::moniker range="doc-intel-2.1.0 || doc-intel-4.0.0"
2020

21-
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` only:
21+
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` for all models and `2023-07-31 (GA)` for Read and Layout only:
2222

2323
* [REST API `2022-08-31 (GA)`](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-2022-08-31/operations/AnalyzeDocument)
24+
* [REST API `2023-07-31 (GA)`](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-2023-07-31/operations/AnalyzeDocument)
2425
* [SDKs targeting `REST API 2022-08-31 (GA)`](../sdk-overview-v3-0.md)
26+
* [SDKs targeting `REST API 2023-07-31 (GA)`](../sdk-overview-v3-1.md)
2527

2628
✔️ See [**Document Intelligence container image tags**](?view=doc-intel-3.0.0&preserve-view=true) for supported container documentation.
2729

@@ -51,6 +53,24 @@ The following containers support DocumentIntelligence v3.0 models and features:
5153

5254
::: moniker-end
5355

56+
::: moniker range="doc-intel-3.1.0"
57+
58+
**This content applies to:** ![checkmark](../media/yes-icon.png) **v3.1 (GA)**
59+
60+
## Microsoft container registry (MCR)
61+
62+
Document Intelligence container images can be found within the [**Microsoft Artifact Registry** (also know as Microsoft Container Registry(MCR))](https://mcr.microsoft.com/catalog?search=document%20intelligence), the primary registry for all Microsoft published container images.
63+
64+
The following containers support DocumentIntelligence v3.0 models and features:
65+
66+
| Container name |image |
67+
|---|---|
68+
|[**Document Intelligence Studio**](https://mcr.microsoft.com/product/azure-cognitive-services/form-recognizer/studio/tags)| `mcr.microsoft.com/azure-cognitive-services/form-recognizer/studio:latest`|
69+
| [**Read 3.1**](https://mcr.microsoft.com/product/azure-cognitive-services/form-recognizer/read-3.1/tags) |`mcr.microsoft.com/azure-cognitive-services/form-recognizer/read-3.1:latest`|
70+
| [**Layout 3.1**](https://mcr.microsoft.com/en-us/product/azure-cognitive-services/form-recognizer/layout-3.1/tags) |`mcr.microsoft.com/azure-cognitive-services/form-recognizer/layout-3.1:latest`|
71+
72+
::: moniker-end
73+
5474
:::moniker range="doc-intel-2.1.0"
5575

5676
> [!IMPORTANT]

articles/ai-services/document-intelligence/containers/install-run.md

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -18,25 +18,29 @@ ms.author: lajanuar
1818
<!-- markdownlint-disable MD024 -->
1919
<!-- markdownlint-disable MD051 -->
2020

21-
:::moniker range="doc-intel-2.1.0 || doc-intel-3.1.0||doc-intel-4.0.0"
21+
:::moniker range="doc-intel-2.1.0 || doc-intel-4.0.0"
2222

23-
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` only:
23+
Support for containers is currently available with Document Intelligence version `2022-08-31 (GA)` for all models and `2023-07-31 (GA)` for Read and Layout only:
2424

2525
* [REST API `2022-08-31 (GA)`](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-2022-08-31/operations/AnalyzeDocument)
26+
* [REST API `2023-07-31 (GA)`](https://westus.dev.cognitive.microsoft.com/docs/services/form-recognizer-api-2023-07-31/operations/AnalyzeDocument)
2627
* [SDKs targeting `REST API 2022-08-31 (GA)`](../sdk-overview-v3-0.md)
28+
* [SDKs targeting `REST API 2023-07-31 (GA)`](../sdk-overview-v3-1.md)
2729

2830
✔️ See [**Install and run Document Intelligence v3.0 containers**](?view=doc-intel-3.0.0&preserve-view=true) for supported container documentation.
2931

3032
:::moniker-end
3133

32-
:::moniker range="doc-intel-3.0.0"
34+
:::moniker range="doc-intel-3.0.0 || doc-intel-3.1.0"
3335

34-
**This content applies to:** ![checkmark](../media/yes-icon.png) **v3.0 (GA)**
36+
**This content applies to:** ![checkmark](../media/yes-icon.png) **v3.0 (GA)** ![checkmark](../media/yes-icon.png) **v3.1 (GA)**
3537

3638
Azure AI Document Intelligence is an Azure AI service that lets you build automated data processing software using machine-learning technology. Document Intelligence enables you to identify and extract text, key/value pairs, selection marks, table data, and more from your documents. The results are delivered as structured data that ../includes the relationships in the original file.
3739

3840
In this article you learn how to download, install, and run Document Intelligence containers. Containers enable you to run the Document Intelligence service in your own environment. Containers are great for specific security and data governance requirements.
3941

42+
* **Read**, and **Layout** models are supported by Document Intelligence v3.1 containers.
43+
4044
* **Read**, **Layout**, **General Document**, **ID Document**, **Receipt**, **Invoice**, **Business Card**, and **Custom** models are supported by Document Intelligence v3.0 containers.
4145

4246
* **Business Card** model is currently only supported in the [v2.1 containers](install-run.md?view=doc-intel-2.1.0&preserve-view=true).

articles/ai-services/document-intelligence/faq.yml

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -556,6 +556,20 @@ sections:
556556
557557
FOTT has a version that uses local storage. The version needs to be installed on a Windows machine. You can install it from [this location](https://github.com/microsoft/OCR-Form-Tools/releases/download/v2.1-ga/oflt-2.1.3-win32.exe). On the project page, specify the Label folder URI as /shared or /shared/sub-dir if your labeling files are in a sub directory. All other Document Intelligence Sample Labeling Tool behavior is the same as the hosted service.
558558
559+
- question: |
560+
What is the best practice for scaling up?
561+
answer: |
562+
563+
- For asynchronous calls, you can run multiple containers with shared storages. The container processing the POST Analyze call stores the output in the storage. Then, any other container can fetch the results from the storage and serve the GET calls. Thus, the request ID is not tied to a container.
564+
565+
- For synchronous calls, multiple containers can be run and only one container will serve a request. As it is a blocking call, any container from the pool can serve the request and send the response. Here, only one container will be tied to a request at a time and no polling is required.
566+
567+
- question: |
568+
How can I set up containers with shared storage?
569+
answer: |
570+
571+
The containers use ‘Mounts:Shared’ property while starting up for specifying the shared storage to store the processing files. You can refer the following [documentation](https://learn.microsoft.com/en-us/azure/ai-services/document-intelligence/containers/install-run?view=doc-intel-3.0.0&tabs=custom#create-a-folder-for-storing-internal-processing-shared-between-the-containers) to see the use of this property.
572+
559573
- name: Security and Privacy
560574
questions:
561575
- question: |

articles/ai-services/openai/concepts/use-your-data.md

Lines changed: 15 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -232,15 +232,26 @@ When you want to reuse the same URL/web address, you can select [Azure AI Search
232232
---
233233

234234

235-
## Custom parameters
235+
## Ingestion parameters
236236

237-
You can modify the following additional settings in the **Data parameters** section in Azure OpenAI Studio and [the API](../reference.md#completions-extensions).
237+
You can use the following parameter to change how your data is ingested in Azure OpenAI Studio, Azure AI Studio, and the ingestion API. Changing the parameter requires re-ingesting your data into Azure Search.
238+
239+
|Parameter name | Description |
240+
|---------|---------|
241+
| **Chunk size** | Azure OpenAI on your data processes your documents by splitting them into chunks before indexing them in Azure Search. The chunk size is the maximum number of tokens for any chunk in the search index. The default chunk size is 1024 tokens. However, given the uniqueness of your data, you may find a different chunk size (such as 256, 512, or 1536 tokens for example) more effective. Adjusting the chunk size can enhance the performance of the chat bot. While finding the optimal chunk size requires some trial and error, start by considering the nature of your dataset. A smaller chunk size is generally better for datasets with direct facts and less context, while a larger chunk size might be beneficial for more contextual information, though it can affect retrieval performance. This is the `chunkSize` parameter in the API.|
242+
243+
244+
## Runtime parameters
245+
246+
You can modify the following additional settings in the **Data parameters** section in Azure OpenAI Studio and [the API](../reference.md#completions-extensions). You do not need to re-ingest your your data when you update these parameters.
238247

239248

240249
|Parameter name | Description |
241250
|---------|---------|
242-
|**Retrieved documents** | Specifies the number of top-scoring documents from your data index used to generate responses. You might want to increase the value when you have short documents or want to provide more context. The default value is 5. This is the `topNDocuments` parameter in the API. |
243-
| **Strictness** | Sets the threshold to categorize documents as relevant to your queries. Raising the value means a higher threshold for relevance and filters out more less-relevant documents for responses. Setting this value too high might cause the model to fail to generate responses due to limited available documents. The default value is 3. |
251+
| **Limit responses to your data** | This flag configures the chatbot's approach to handling queries unrelated to the data source or when search documents are insufficient for a complete answer. When this setting is disabled, the model supplements its responses with its own knowledge in addition to your documents. When this setting is enabled, the model attempts to only rely on your documents for responses. This is the `inScope` parameter in the API. |
252+
|**Top K Documents** | This parameter is an integer that can be set to 3, 5, 10, or 20, and controls the number of document chunks provided to the large language model for formulating the final response. By default, this is set to 5. The search process can be noisy and sometimes, due to chunking, relevant information may be spread across multiple chunks in the search index. Selecting a top-K number, like 5, ensures that the model can extract relevant information, despite the inherent limitations of search and chunking. However, increasing the number too high can potentially distract the model. Additionally, the maximum number of documents that can be effectively used depends on the version of the model, as each has a different context size and capacity for handling documents. If you find that responses are missing important context, try increasing this parameter. Conversely, if you think the model is providing irrelevant information alongside useful data, consider decreasing it. When experimenting with the [chunk size](#ingestion-parameters), we recommend adjusting the top-K parameter to achieve the best performance. Usually, it is beneficial to change the top-K value in the opposite direction of your chunk size adjustment. For example, if you decrease the chunk size from the default of 1024, you might want to increase the top-K value to 10 or 20. This ensures a similar amount of information is provided to the model, as reducing the chunk size decreases the amount of information in the 5 documents given to the model. This is the `topNDocuments` parameter in the API. |
253+
| **Strictness** | Determines the system's aggressiveness in filtering search documents based on their similarity scores. The system queries Azure Search or other document stores, then decides which documents to provide to large language models like ChatGPT. Filtering out irrelevant documents can significantly enhance the performance of the end-to-end chatbot. Some documents are excluded from the top-K results if they have low similarity scores before forwarding them to the model. This is controlled by an integer value ranging from 1 to 5. Setting this value to 1 means that the system will minimally filter documents based on search similarity to the user query. Conversely, a setting of 5 indicates that the system will aggressively filter out documents, applying a very high similarity threshold. If you find that the chatbot omits relevant information, lower the filter's strictness (set the value closer to 1) to include more documents. Conversely, if irrelevant documents distract the responses, increase the threshold (set the value closer to 5). This is the `strictness` parameter in the API. |
254+
244255

245256
## Document-level access control
246257

articles/ai-services/openai/whats-new.md

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,6 +43,9 @@ Azure OpenAI Service now supports text to speech APIs with OpenAI's voices. Get
4343
- [Fine-tuning & function calling](./how-to/fine-tuning-functions.md)
4444
- [`gpt-35-turbo 1106` support](./concepts/models.md#fine-tuning-models)
4545

46+
### Chunk size parameter for Azure OpenAI on your data
47+
48+
- You can now set the [chunk size](./concepts/use-your-data.md#ingestion-parameters) parameter when your data is ingested. Adjusting the chunk size can enhance the model's responses by setting the maximum number of tokens for any given chunk of your data in the search index.
4649

4750
## December 2023
4851

@@ -118,7 +121,7 @@ Try out DALL-E 3 by following a [quickstart](./dall-e-quickstart.md).
118121

119122
### Azure OpenAI on your data
120123

121-
- New [custom parameters](./concepts/use-your-data.md#custom-parameters) for determining the number of retrieved documents and strictness.
124+
- New [custom parameters](./concepts/use-your-data.md#runtime-parameters) for determining the number of retrieved documents and strictness.
122125
- The strictness setting sets the threshold to categorize documents as relevant to your queries.
123126
- The retrieved documents setting specifies the number of top-scoring documents from your data index used to generate responses.
124127
- You can see data ingestion/upload status in the Azure OpenAI Studio.

articles/ai-services/security-controls-policy.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
22
title: Azure Policy Regulatory Compliance controls for Azure AI services
33
description: Lists Azure Policy Regulatory Compliance controls available for Azure AI services. These built-in policy definitions provide common approaches to managing the compliance of your Azure resources.
4-
ms.date: 01/22/2024
4+
ms.date: 02/06/2024
55
ms.topic: sample
66
author: PatrickFarley
77
ms.author: pafarley

0 commit comments

Comments
 (0)