Skip to content

Commit 5f40309

Browse files
Merge pull request #6174 from MicrosoftDocs/main
Auto Publish – main to live - 2025-07-23 05:05 UTC
2 parents e9c44be + a8d34f6 commit 5f40309

17 files changed

+46
-1167
lines changed

.openpublishing.redirection.json

Lines changed: 20 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -489,6 +489,26 @@
489489
"source_path": "articles/ai-foundry/model-inference/index.yml",
490490
"redirect_url": "../foundry-models/index",
491491
"redirect_document_id": false
492+
},
493+
{
494+
"source_path": "articles/machine-learning/concept-package-models.md",
495+
"redirect_url": "/azure/machine-learning/concept-endpoints",
496+
"redirect_document_id": false
497+
},
498+
{
499+
"source_path": "articles/machine-learning/how-to-package-models-moe.md",
500+
"redirect_url": "/azure/machine-learning/concept-endpoints",
501+
"redirect_document_id": false
502+
},
503+
{
504+
"source_path": "articles/machine-learning/how-to-package-models-app-service.md",
505+
"redirect_url": "/azure/machine-learning/concept-endpoints",
506+
"redirect_document_id": false
507+
},
508+
{
509+
"source_path": "articles/machine-learning/how-to-package-models.md",
510+
"redirect_url": "/azure/machine-learning/concept-endpoints",
511+
"redirect_document_id": false
492512
}
493513
]
494514
}

articles/ai-foundry/how-to/configure-private-link.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -48,8 +48,6 @@ You get several hub default resources in your resource group. You need to config
4848
> [!IMPORTANT]
4949
> We don't recommend using the 172.17.0.0/16 IP address range for your VNet. This is the default subnet range used by the Docker bridge network on-premises.
5050
51-
* Disable network policies for private endpoints before adding the private endpoint.
52-
5351
:::zone pivot="fdp-project"
5452

5553
## Create a Foundry project that uses a private endpoint

articles/ai-services/content-safety/overview.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -153,7 +153,7 @@ For more information, see [Language support](/azure/ai-services/content-safety/l
153153

154154
To use the Content Safety APIs, you must create your Azure AI Content Safety resource in a supported region. Currently, the Content Safety features are available in the following Azure regions with different API versions:
155155

156-
| Region | Custom Category | Groundedness | Image | Multimodal(Image with Tex) | Incident Response | Prompt Shield | Protected Material (Text) | Protected Material (Code) | Text |
156+
| Region | Custom Category (standard) | Groundedness | Image | Multimodal(Image with Tex) | Custom Category (rapid) | Prompt Shield | Protected Material (Text) | Protected Material (Code) | Text |
157157
|--------------------|--------------------|--------------------|-------|-----------------------------|-------------------|---------------|---------------------------|---------------------------|------|
158158
| Australia East || || ||||||
159159
| Canada East | | || ||||||

articles/ai-services/language-service/conversational-language-understanding/concepts/entity-components.md

Lines changed: 11 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -6,24 +6,24 @@ author: laujan
66
manager: nitinme
77
ms.service: azure-ai-language
88
ms.topic: conceptual
9-
ms.date: 06/04/2025
9+
ms.date: 07/22/2025
1010
ms.author: lajanuar
1111
ms.custom: language-service-clu
1212
---
1313

1414
# Entity components
1515

16-
In conversational language understanding, entities are relevant pieces of information that are extracted from your utterances. An entity can be extracted by different methods. They can be learned through context, matched from a list, or detected by a prebuilt recognized entity. Every entity in your project is composed of one or more of these methods, which are defined as your entity's components.
16+
In conversational language understanding, entities are relevant pieces of information that are extracted from your utterances. You can extract an entity using several different methods. Entities can be detected through context, matched from a list, or detected by a prebuilt recognized entity. Every entity in your project is composed of one or more of these methods, which are defined as your entity's components.
1717

18-
When an entity is defined by more than one component, their predictions can overlap. You can determine the behavior of an entity prediction when its components overlap by using a fixed set of options in the *entity options*.
18+
When more than one component defines an entity, predictions can overlap. You can determine the behavior of an entity prediction when its components overlap by using a fixed set of options in the *entity options*.
1919

2020
## Component types
2121

2222
An entity component determines a way that you can extract the entity. An entity can contain one component, which determines the only method to be used to extract the entity. An entity can also contain multiple components to expand the ways in which the entity is defined and extracted.
2323

2424
### Learned component
2525

26-
The learned component uses the entity tags you label your utterances with to train a machine-learned model. The model learns to predict where the entity is based on the context within the utterance. Your labels provide examples of where the entity is expected to be present in an utterance, based on the meaning of the words around it and as the words that were labeled.
26+
The learned component uses the entity tags you label your utterances with to train a machine-learned model. The model learns to predict where the entity is based on the context within the utterance. Your labels provide examples of where the entity is expected to be present in an utterance. This determination is based on the meaning of the words around it and as the words that were labeled.
2727

2828
This component is only defined if you add labels by tagging utterances for the entity. If you don't tag any utterances with the entity, it doesn't have a learned component.
2929

@@ -53,13 +53,13 @@ In multilingual projects, you can specify a different expression for each langua
5353

5454
## Entity options
5555

56-
When multiple components are defined for an entity, their predictions might overlap. When an overlap occurs, each entity's final prediction is determined by one of the following options.
56+
If multiple components define an entity, their predictions may overlap. When overlap happens, one of the following options determines each entity's final prediction:
5757

5858
### Combine components
5959

6060
Combine components as one entity when they overlap by taking the union of all the components.
6161

62-
Use this option to combine all components when they overlap. When components are combined, you get all the extra information that's tied to a list or prebuilt component when they're present.
62+
Use this option to combine all components when they overlap. When components are combined, you get all the extra information associated with a list or prebuilt component if present.
6363

6464
#### Example
6565

@@ -71,7 +71,7 @@ By using combined components, the entity returns with the full context as "Prose
7171

7272
:::image type="content" source="../media/union-overlap-example-1-part-2.svg" alt-text="Screenshot that shows the result of a combined component." lightbox="../media/union-overlap-example-1-part-2.svg":::
7373

74-
Suppose you had the same utterance, but only "OS 9" was predicted by the learned component:
74+
Suppose you had the same utterance, but only "OS 9" predicts the learned component:
7575

7676
:::image type="content" source="../media/union-overlap-example-2.svg" alt-text="Screenshot that shows an utterance with O S 9 predicted by the learned component." lightbox="../media/union-overlap-example-2.svg":::
7777

@@ -95,7 +95,7 @@ When you don't combine components, the entity returns twice:
9595

9696
### Required components
9797

98-
Sometimes an entity can be defined by multiple components but requires one or more of them to be present. Every component can be set as *required*, which means the entity *won't* be returned if that component wasn't present. For example, if you have an entity with a list component and a required learned component, it's guaranteed that any returned entity includes a learned component. If it doesn't, the entity isn't returned.
98+
Sometimes, you can define an entity using multiple components, but the entity requires at least one or more of them to be present. You can mark any component as *required*, which means the system *doesn't* return the entity unless that component is present. For example, if an entity has a list component and a *required* learned component, the system guarantees that any returned entity includes a learned component. If an entity doesn't have the required component, the system doesn't return it.
9999

100100
Required components are most frequently used with learned components because they can restrict the other component types to a specific context, which is commonly associated to *roles*. You can also require all components to make sure that every component is present for an entity.
101101

@@ -120,8 +120,9 @@ Other times, you might be interested in extracting an entity through context, su
120120
When you don't combine components, you allow every component to act as an independent entity extractor. One way of using this option is to separate the entities extracted from a list to the ones extracted through the learned or prebuilt components to handle and treat them differently.
121121

122122
> [!NOTE]
123-
> Previously during the public preview of the service, there were four available options: **Longest overlap**, **Exact overlap**, **Union overlap**, and **Return all separately**. **Longest overlap** and **Exact overlap** are deprecated and are only supported for projects that previously had those options selected. **Union overlap** has been renamed to **Combine components**, while **Return all separately** has been renamed to **Do not combine components**.
123+
> Previously during the public preview of the service, there were four available options: **Longest overlap**, **Exact overlap**, **Union overlap**, and **Return all separately**. **Longest overlap** and **Exact overlap** are deprecated and are only supported for projects that previously had those options selected. **Union overlap** is renamed to **Combine components**, while **Return all separately** is renamed to **Do not combine components**.
124124
125125
## Related content
126126

127-
- [Supported prebuilt components](../prebuilt-component-reference.md)
127+
[Supported prebuilt components](../prebuilt-component-reference.md)
128+

articles/ai-services/language-service/toc.yml

Lines changed: 8 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -839,8 +839,14 @@ items:
839839
items:
840840
- name: C#
841841
items:
842-
- name: Text analysis
843-
href: /dotnet/api/overview/azure/ai.textanalytics-readme?view=azure-dotnet-preview&preserve-view=true
842+
- name: Language Text
843+
href: /dotnet/api/overview/azure/ai.language.text-readme?view=azure-dotnet-preview&preserve-view=true
844+
- name: Conversational Language Understanding
845+
href: /dotnet/api/overview/azure/ai.language.conversations-readme?view=azure-dotnet-preview&preserve-view=true
846+
- name: Conversations Authoring
847+
href: /dotnet/api/overview/azure/ai.language.conversations.authoring-readme?view=azure-dotnet-preview&preserve-view=true
848+
- name: Text Authoring
849+
href: /dotnet/api/overview/azure/ai.language.text.authoring-readme?view=azure-dotnet-preview&preserve-view=true
844850
- name: Custom question answering
845851
href: /dotnet/api/overview/azure/ai.language.questionanswering-readme?view=azure-dotnet-preview&preserve-view=true
846852
- name: Python

articles/machine-learning/concept-endpoints.md

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -124,7 +124,6 @@ The following table shows a summary of the different features available to stand
124124
| Deployment types | Models | Models | Models and Pipeline components |
125125
| MLflow model deployment | No, only specific models in the catalog | Yes | Yes |
126126
| Custom model deployment | No, only specific models in the catalog | Yes, with scoring script | Yes, with scoring script |
127-
| Model package deployment <sup>2</sup> | Built-in | Yes (preview) | No |
128127
| Inference server <sup>3</sup> | Azure AI Model Inference API | - Azure Machine Learning Inferencing Server<br /> - Triton<br /> - Custom (using BYOC) | Batch Inference |
129128
| Compute resource consumed | None (serverless) | Instances or granular resources | Cluster instances |
130129
| Compute type | None (serverless) | Managed compute and Kubernetes | Managed compute and Kubernetes |
@@ -135,13 +134,12 @@ The following table shows a summary of the different features available to stand
135134
| Cost basis<sup>5</sup> | Per token | Per deployment: compute instances running | Per job: compute instanced consumed in the job (capped to the maximum number of instances of the cluster) |
136135
| Local testing of deployments | No | Yes | No |
137136

138-
<sup>2</sup> Deploying MLflow models to endpoints without outbound internet connectivity or private networks requires [packaging the model](concept-package-models.md) first.
139137

140-
<sup>3</sup> *Inference server* refers to the serving technology that takes requests, processes them, and creates responses. The inference server also dictates the format of the input and the expected outputs.
138+
<sup>2</sup> *Inference server* refers to the serving technology that takes requests, processes them, and creates responses. The inference server also dictates the format of the input and the expected outputs.
141139

142-
<sup>4</sup> *Autoscaling* is the ability to dynamically scale up or scale down the deployment's allocated resources based on its load. Online and batch deployments use different strategies for autoscaling. While online deployments scale up and down based on the resource utilization (like CPU, memory, requests, etc.), batch endpoints scale up or down based on the number of jobs created.
140+
<sup>3</sup> *Autoscaling* is the ability to dynamically scale up or scale down the deployment's allocated resources based on its load. Online and batch deployments use different strategies for autoscaling. While online deployments scale up and down based on the resource utilization (like CPU, memory, requests, etc.), batch endpoints scale up or down based on the number of jobs created.
143141

144-
<sup>5</sup> Both online and batch deployments charge by the resources consumed. In online deployments, resources are provisioned at deployment time. In batch deployment, resources aren't consumed at deployment time but at the time that the job runs. Hence, there's no cost associated with the batch deployment itself. Likewise, queued jobs don't consume resources either.
142+
<sup>4</sup> Both online and batch deployments charge by the resources consumed. In online deployments, resources are provisioned at deployment time. In batch deployment, resources aren't consumed at deployment time but at the time that the job runs. Hence, there's no cost associated with the batch deployment itself. Likewise, queued jobs don't consume resources either.
145143

146144
## Developer interfaces
147145

0 commit comments

Comments
 (0)