Skip to content

Commit e145d27

Browse files
Merge pull request #2755 from santiagxf/santiagxf-patch-1
Update quickstart-github-models.md
2 parents 6fa3d4b + 31ba210 commit e145d27

8 files changed

+24
-24
lines changed

articles/ai-foundry/model-inference/how-to/inference.md

Lines changed: 9 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,6 +48,14 @@ For a chat model, you can create a request as follows:
4848

4949
If you specify a model name that doesn't match any given model deployment, you get an error that the model doesn't exist. You can control which models are available for users by creating model deployments as explained at [add and configure model deployments](create-model-deployments.md).
5050

51+
## Key-less authentication
52+
53+
Models deployed to Azure AI model inference in Azure AI Services support key-less authorization using Microsoft Entra ID. Key-less authorization enhances security, simplifies the user experience, reduces operational complexity, and provides robust compliance support for modern development. It makes it a strong choice for organizations adopting secure and scalable identity management solutions.
54+
55+
To use key-less authentication, [configure your resource and grant access to users](configure-entra-id.md) to perform inference. Once configured, then you can authenticate as follows:
56+
57+
[!INCLUDE [code-create-chat-client-entra](../includes/code-create-chat-client-entra.md)]
58+
5159
## Limitations
5260

5361
* Azure OpenAI Batch can't be used with the Azure AI model inference endpoint. You have to use the dedicated deployment URL as explained at [Batch API support in Azure OpenAI documentation](../../../ai-services/openai/how-to/batch.md#api-support).
@@ -56,4 +64,4 @@ If you specify a model name that doesn't match any given model deployment, you g
5664
## Next steps
5765

5866
* [Use embedding models](use-embeddings.md)
59-
* [Use chat completion models](use-chat-completions.md)
67+
* [Use chat completion models](use-chat-completions.md)

articles/ai-foundry/model-inference/how-to/quickstart-github-models.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -85,6 +85,7 @@ Use the parameter `model="<deployment-name>` to route your request to this deplo
8585
Azure AI model inference supports additional features not available in GitHub Models, including:
8686

8787
* [Explore the model catalog](https://ai.azure.com/github/models) to see additional models not available in GitHub Models.
88+
* Configure [key-less authentication](configure-entra-id.md).
8889
* Configure [content filtering](configure-content-filters.md).
8990
* Configure rate limiting (for specific models).
9091
* Explore additional [deployment SKUs (for specific models)](../concepts/deployment-types.md).
@@ -97,4 +98,4 @@ See the [FAQ section](../faq.yml) to explore more help.
9798
## Next steps
9899

99100
* [Explore the model catalog](https://ai.azure.com/github/models) in Azure AI studio.
100-
* [Add more models](create-model-deployments.md) to your endpoint.
101+
* [Add more models](create-model-deployments.md) to your endpoint.

articles/ai-foundry/model-inference/includes/code-create-chat-client-entra.md

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -12,12 +12,9 @@ author: santiagxf
1212
Install the package `azure-ai-inference` using your package manager, like pip:
1313

1414
```bash
15-
pip install azure-ai-inference>=1.0.0b5
15+
pip install azure-ai-inference
1616
```
1717

18-
> [!WARNING]
19-
> Azure AI Services resource requires the version `azure-ai-inference>=1.0.0b5` for Python.
20-
2118
Then, you can use the package to consume the model. The following example shows how to create a client to consume chat completions with Entra ID:
2219

2320
```python

articles/ai-foundry/model-inference/includes/code-create-chat-client.md

Lines changed: 3 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -12,12 +12,9 @@ author: santiagxf
1212
Install the package `azure-ai-inference` using your package manager, like pip:
1313

1414
```bash
15-
pip install azure-ai-inference>=1.0.0b5
15+
pip install azure-ai-inference
1616
```
1717

18-
> [!WARNING]
19-
> Azure AI Services resource requires the version `azure-ai-inference>=1.0.0b5` for Python.
20-
2118
Then, you can use the package to consume the model. The following example shows how to create a client to consume chat completions:
2219

2320
```python
@@ -115,7 +112,7 @@ __Request__
115112

116113
```HTTP/1.1
117114
POST https://<resource>.services.ai.azure.com/models/chat/completions?api-version=2024-05-01-preview
118-
Authorization: Bearer <bearer-token>
115+
api-key: <api-key>
119116
Content-Type: application/json
120117
```
121-
---
118+
---

articles/ai-foundry/model-inference/includes/code-create-chat-completion.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -78,7 +78,7 @@ __Request__
7878

7979
```HTTP/1.1
8080
POST https://<resource>.services.ai.azure.com/models/chat/completions?api-version=2024-05-01-preview
81-
Authorization: Bearer <bearer-token>
81+
api-key: <api-key>
8282
Content-Type: application/json
8383
```
8484

@@ -98,4 +98,4 @@ Content-Type: application/json
9898
}
9999
```
100100

101-
---
101+
---

articles/ai-foundry/model-inference/includes/code-create-embeddings-client.md

Lines changed: 3 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -12,12 +12,9 @@ author: santiagxf
1212
Install the package `azure-ai-inference` using your package manager, like pip:
1313

1414
```bash
15-
pip install azure-ai-inference>=1.0.0b5
15+
pip install azure-ai-inference
1616
```
1717

18-
> [!WARNING]
19-
> Azure AI Services resource requires the version `azure-ai-inference>=1.0.0b5` for Python.
20-
2118
Then, you can use the package to consume the model. The following example shows how to create a client to consume chat completions:
2219

2320
```python
@@ -132,7 +129,7 @@ __Request__
132129

133130
```HTTP/1.1
134131
POST https://<resource>.services.ai.azure.com/models/embeddings?api-version=2024-05-01-preview
135-
Authorization: Bearer <bearer-token>
132+
api-key: <api-key>
136133
Content-Type: application/json
137134
```
138-
---
135+
---

articles/ai-foundry/model-inference/includes/code-create-embeddings.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ __Request__
5454

5555
```HTTP/1.1
5656
POST https://<resource>.services.ai.azure.com/models/embeddings?api-version=2024-05-01-preview
57-
Authorization: Bearer <bearer-token>
57+
api-key: <api-key>
5858
Content-Type: application/json
5959
```
6060

@@ -100,4 +100,4 @@ __Response__
100100
}
101101
```
102102

103-
---
103+
---

articles/ai-foundry/model-inference/includes/code-manage-content-filtering.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -122,8 +122,8 @@ try {
122122
__Request__
123123

124124
```HTTP/1.1
125-
POST /chat/completions?api-version=2024-05-01-preview
126-
Authorization: Bearer <bearer-token>
125+
POST https://<resource>.services.ai.azure.com/models/chat/completions?api-version=2024-05-01-preview
126+
api-key: <api-key>
127127
Content-Type: application/json
128128
```
129129

0 commit comments

Comments
 (0)