You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/foundry-models/concepts/default-safety-policies.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,12 +16,12 @@ reviewer: santiagxf
16
16
17
17
Azure AI Foundry Models includes default safety applied to all models, excluding Azure OpenAI Whisper. These configurations provide you with a responsible experience by default.
18
18
19
-
Default safety aims to mitigate risks such as hate and fairness, sexual, violence, self-harm, protected material content, and user prompt injection attacks. To learn more about content filtering, read [our documentation describing categories and severity levels](content-filter.md).
19
+
Default safety aims to mitigate risks such as hate and fairness, sexual, violence, self-harm, protected material content, and user prompt injection attacks. To learn more about content filtering, read [our documentation describing categories and severity levels](../../model-inference/concepts/content-filter.md).
20
20
21
21
This document describes the default configuration.
22
22
23
23
> [!TIP]
24
-
> By default, all model deployments use the default configuration. However, you can configure content filtering per model deployment as explained at [Configuring content filtering](../how-to/configure-content-filters.md).
24
+
> By default, all model deployments use the default configuration. However, you can configure content filtering per model deployment as explained at [Configuring content filtering](../../model-inference/how-to/configure-content-filters.md).
25
25
26
26
## Text models
27
27
@@ -80,4 +80,4 @@ Content filters apply only to text prompts and completions. Images aren't subjec
80
80
81
81
## Next steps
82
82
83
-
*[Configure content filters in Azure AI Foundry Models](../how-to/configure-content-filters.md)
83
+
*[Configure content filters in Azure AI Foundry Models](../../model-inference/how-to/configure-content-filters.md)
Copy file name to clipboardExpand all lines: articles/ai-foundry/foundry-models/concepts/deployment-types.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -33,7 +33,7 @@ Azure AI Foundry Models provides customers with choices on the hosting structure
33
33
34
34
Global deployments are available in the same Azure AI Foundry Services as non-global deployment types but allow you to leverage Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global standard provides the highest default quota and eliminates the need to load balance across multiple resources.
35
35
36
-
Customers with high consistent volume may experience greater latency variability. The threshold is set per model. See the [quotas page to learn more](../quotas-limits.md). For applications that require the lower latency variance at large workload usage, we recommend purchasing provisioned throughput.
36
+
Customers with high consistent volume may experience greater latency variability. The threshold is set per model. See the [quotas page to learn more](../../model-inference/quotas-limits.md). For applications that require the lower latency variance at large workload usage, we recommend purchasing provisioned throughput.
37
37
38
38
## Global provisioned
39
39
@@ -119,5 +119,5 @@ Administrators can control which model deployment types are available to their u
Copy file name to clipboardExpand all lines: articles/ai-foundry/foundry-models/concepts/endpoints.md
+9-9Lines changed: 9 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -36,7 +36,7 @@ Deployments capture:
36
36
37
37
An Azure AI Foundry resource can have as many model deployments as needed and they don't incur in cost unless inference is performed for those models. Deployments are Azure resources and hence they're subject to Azure policies.
38
38
39
-
To learn more about how to create deployments see [Add and configure model deployments](../how-to/create-model-deployments.md).
39
+
To learn more about how to create deployments see [Add and configure model deployments](../../model-inference/how-to/create-model-deployments.md).
40
40
41
41
## Endpoints
42
42
@@ -46,7 +46,7 @@ Azure AI Foundry Services (formerly known Azure AI Services) expose multiple end
46
46
> * Azure AI inference endpoint (usually with the form `https://<resource-name>.services.ai.azure.com/models`)
47
47
> * Azure OpenAI endpoint (usually with the form `https://<resource-name>.openai.azure.com`)
48
48
49
-
The **Azure AI inference endpoint** allows customers to use a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. All the models support this capability. This endpoint follows the [Azure AI Model Inference API](.././reference/reference-model-inference-api.md).
49
+
The **Azure AI inference endpoint** allows customers to use a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. All the models support this capability. This endpoint follows the [Azure AI Model Inference API](../../model-inference/reference/reference-model-inference-api.md).
50
50
51
51
The **Azure OpenAI API** exposes the full capabilities of OpenAI models and supports more features like assistants, threads, files, and batch inference. Non-OpenAI models may also be exposed in this route.
52
52
@@ -60,21 +60,21 @@ The inference endpoint routes requests to a given deployment by matching the par
60
60
61
61
For example, if you create a deployment named `Mistral-large`, then such deployment can be invoked as:
If you specify a model name that doesn't match any given model deployment, you get an error that the model doesn't exist. You can control which models are available for users by creating model deployments as explained at [add and configure model deployments](../how-to/create-model-deployments.md).
69
+
If you specify a model name that doesn't match any given model deployment, you get an error that the model doesn't exist. You can control which models are available for users by creating model deployments as explained at [add and configure model deployments](../../model-inference/how-to/create-model-deployments.md).
70
70
71
71
## Key-less authentication
72
72
73
73
Models deployed to Azure AI Foundry Models in Azure AI Services support key-less authorization using Microsoft Entra ID. Key-less authorization enhances security, simplifies the user experience, reduces operational complexity, and provides robust compliance support for modern development. It makes it a strong choice for organizations adopting secure and scalable identity management solutions.
74
74
75
-
To use key-less authentication, [configure your resource and grant access to users](../how-to/configure-entra-id.md) to perform inference. Once configured, then you can authenticate as follows:
75
+
To use key-less authentication, [configure your resource and grant access to users](../../model-inference/how-to/configure-entra-id.md) to perform inference. Once configured, then you can authenticate as follows:
Copy file name to clipboardExpand all lines: articles/ai-foundry/foundry-models/concepts/models.md
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -61,10 +61,10 @@ DeepSeek family of models includes DeepSeek-R1, which excels at reasoning tasks
61
61
| ------ | ---- | ---- | ------------ |
62
62
|[DeepSeek-R1-0528]() <br /> | chat-completion | Global standard | - **Input:** text (163,840 tokens) <br /> - **Output:** text (163,840 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text |
63
63
|[DeepSeek-V3-0324](https://ai.azure.com/explore/models/deepseek-v3-0324/version/1/registry/azureml-deepseek)| chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON |
64
-
|[DeepSeek-R1](https://ai.azure.com/explore/models/deepseek-r1/version/1/registry/azureml-deepseek)| chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md)| Global standard | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. |
64
+
|[DeepSeek-R1](https://ai.azure.com/explore/models/deepseek-r1/version/1/registry/azureml-deepseek)| chat-completion <br /> [(with reasoning content)](../../model-inference/how-to/use-chat-reasoning.md)| Global standard | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. |
65
65
|[DeepSeek-V3](https://ai.azure.com/explore/models/deepseek-v3/version/1/registry/azureml-deepseek) <br />(Legacy) | chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
66
66
67
-
For a tutorial on DeepSeek-R1, see [Tutorial: Get started with DeepSeek-R1 reasoning model in Azure AI Foundry Models](../tutorials/get-started-deepseek-r1.md).
67
+
For a tutorial on DeepSeek-R1, see [Tutorial: Get started with DeepSeek-R1 reasoning model in Azure AI Foundry Models](../../model-inference/tutorials/get-started-deepseek-r1.md).
68
68
69
69
See [this model collection in Azure AI Foundry portal](https://ai.azure.com/explore/models?&selectedCollection=deepseek).
70
70
@@ -75,7 +75,7 @@ Microsoft models include various model groups such as MAI models, Phi models, he
75
75
76
76
| Model | Type | Tier | Capabilities |
77
77
| ------ | ---- | --- | ------------ |
78
-
|[MAI-DS-R1](https://ai.azure.com/explore/models/MAI-DS-R1/version/1/registry/azureml)| chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md)| Global standard | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. |
78
+
|[MAI-DS-R1](https://ai.azure.com/explore/models/MAI-DS-R1/version/1/registry/azureml)| chat-completion <br /> [(with reasoning content)](../../model-inference/how-to/use-chat-reasoning.md)| Global standard | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. |
79
79
80
80
### Mistral AI
81
81
@@ -112,7 +112,7 @@ xAI's Grok 3 and Grok 3 Mini models are designed to excel in various enterprise
112
112
113
113
Models from Partners and Community available for deployment with pay-as-you-go billing (for example, Cohere models) are offered by the model provider but hosted in Microsoft-managed Azure infrastructure and accessed via API in the Azure AI Foundry. Model providers define the license terms and set the price for use of their models, while Azure AI Foundry manages the hosting infrastructure.
114
114
115
-
Models from Partners and Community are offered through Azure Marketplace and [requires additional configuration for enabling](../how-to/configure-marketplace.md).
115
+
Models from Partners and Community are offered through Azure Marketplace and [requires additional configuration for enabling](../../model-inference/how-to/configure-marketplace.md).
116
116
117
117
### AI21 Labs
118
118
@@ -229,4 +229,4 @@ Those models can be of open access or IP protected. In both cases, you have to d
229
229
230
230
## Next steps
231
231
232
-
- Get started today and [deploy your fist model in Azure AI Foundry Models](../how-to/create-model-deployments.md)
232
+
- Get started today and [deploy your fist model in Azure AI Foundry Models](../../model-inference/how-to/create-model-deployments.md)
0 commit comments