Skip to content

Commit f467c74

Browse files
committed
acrolyx
1 parent 37ae4ad commit f467c74

File tree

6 files changed

+29
-29
lines changed

6 files changed

+29
-29
lines changed

articles/ai-studio/ai-services/concepts/endpoints.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -14,13 +14,13 @@ ms.custom: ignite-2024, github-universe-2024
1414

1515
# Use the Azure AI model inference endpoint
1616

17-
Azure AI inference service in Azure AI services allow customers to consume the most powerful models from flagship model providers using a single endpoint and credentials. This means that you can switch between models and consume them from your application without changing a single line of code.
17+
Azure AI inference service in Azure AI services allows customers to consume the most powerful models from flagship model providers using a single endpoint and credentials. This means that you can switch between models and consume them from your application without changing a single line of code.
1818

1919
The article explains how models are organized inside of the service and how to use the inference endpoint to invoke them.
2020

2121
## Deployments
2222

23-
Azure AI model inference service make models available using the **deployment** concept. **Deployments** are a way to give a model a name under certain configurations. Then, you can invoke such model configuration by indicating its name on your requests.
23+
Azure AI model inference service makes models available using the **deployment** concept. **Deployments** are a way to give a model a name under certain configurations. Then, you can invoke such model configuration by indicating its name on your requests.
2424

2525
Deployments capture:
2626

@@ -39,15 +39,15 @@ To learn more about how to create deployments see [Add and configure model deplo
3939

4040
## Azure AI inference endpoint
4141

42-
The Azure AI inference endpoint allow customers to use a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. This endpoint follows the [Azure AI model inference API](../../reference/reference-model-inference-api.md) which is supported by all the models in Azure AI model inference service.
42+
The Azure AI inference endpoint allows customers to use a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. This endpoint follows the [Azure AI model inference API](../../reference/reference-model-inference-api.md) which is supported by all the models in Azure AI model inference service.
4343

44-
You can see the endpoint URL and credentials in the **Overview** section:
44+
You can see the endpoint URL and credentials in the **Overview** section. The endpoint usually has the form `https://<resource-name>.services.ai.azure.com/models`:
4545

4646
:::image type="content" source="../../media/ai-services/overview/overview-endpoint-and-key.png" alt-text="An screenshot showing how to get the URL and key associated with the resource." lightbox="../../media/ai-services/overview/overview-endpoint-and-key.png":::
4747

4848
### Routing
4949

50-
The inference endpoint routes requests to a given deployment by matching the parameter `name` inside of the request to the name of the deployment. This means that *deployments work as an alias of a given model under certain configurations*. This flexibility allow you to deploy a given model multiple times in the service but under different configurations if needed.
50+
The inference endpoint routes requests to a given deployment by matching the parameter `name` inside of the request to the name of the deployment. This means that *deployments work as an alias of a given model under certain configurations*. This flexibility allows you to deploy a given model multiple times in the service but under different configurations if needed.
5151

5252
:::image type="content" source="../../media/ai-services/endpoint/endpoint-routing.png" alt-text="An illustration showing how routing works for a Meta-llama-3.2-8b-instruct model by indicating such name in the parameter 'model' inside of the payload request." lightbox="../../media/ai-services/endpoint/endpoint-routing.png":::
5353

@@ -71,9 +71,9 @@ All models deployed in Azure AI model inference service support the [Azure AI mo
7171

7272
## Azure OpenAI inference endpoint
7373

74-
Azure OpenAI models also support the Azure OpenAI API. This API exposes the full capabilities of OpenAI models and support additional features like assistants, threads, files, and batch inference.
74+
Azure OpenAI models also support the Azure OpenAI API. This API exposes the full capabilities of OpenAI models and supports additional features like assistants, threads, files, and batch inference.
7575

76-
Azure OpenAI inference endpoints are used per-deployment and they have they own URL that is associated with only one deployment. However, the same authentication mechanism can be used to consume it. Learn more in the reference page for [Azure OpenAI API](../../../ai-services/openai/reference.md)
76+
Each OpenAI model deployment has its own URL associated with such deployment under the Azure OpenAI inference endpoint. However, the same authentication mechanism can be used to consume it. URLs are usually in the form of `https://<resource-name>.openai.azure.com/openai/deployments/<model-deployment-name>`. Learn more in the reference page for [Azure OpenAI API](../../../ai-services/openai/reference.md)
7777

7878
:::image type="content" source="../../media/ai-services/endpoint/endpoint-openai.png" alt-text="An illustration showing how Azure OpenAI deployments contain a single URL for each deployment." lightbox="../../media/ai-services/endpoint/endpoint-openai.png":::
7979

articles/ai-studio/ai-services/faq.yml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@ sections:
4040
- question: |
4141
What's the difference between Azure AI model inference service and Serverless API model deployments in Azure AI studio?
4242
answer: |
43-
Both technologies allow you to deploy models without requiring compute resources as they are based on the Models as a Service idea. [Serverless API model deployments](../how-to/deploy-models-serverless.md) allow you to deploy a single models under a unique endpoint and credentials. You need to create a different endpoint for each model you want to deploy. On top of that, they are always created in the context of the project and while they can be shared by creating connections from other projects, they live in the context of a given project.
43+
Both technologies allow you to deploy models without requiring compute resources as they are based on the Models as a Service idea. [Serverless API model deployments](../how-to/deploy-models-serverless.md) allow you to deploy a single model under a unique endpoint and credentials. You need to create a different endpoint for each model you want to deploy. On top of that, they are always created in the context of the project and while they can be shared by creating connections from other projects, they live in the context of a given project.
4444
4545
Azure AI model inference service allows you to deploy multiple models under the same endpoint and credentials. You can switch between models without changing your code. They are also in the context of a shared resource, the Azure AI Services resource, which implies you can connect the resource to any project or hub that requires to consume the models you made available. Azure AI model inference service comes with a built-in model routing capability that routes the request to the right model based on the model name you pass in the request.
4646
@@ -79,7 +79,7 @@ sections:
7979
- question: |
8080
I'm making a request for a model that Azure AI model inference service supports, but I'm getting a 404 error. What should I do?
8181
answer: |
82-
Ensure you created a deployment for the given model and that the deployment name matches **exactly** the value you're passing in `model` parameter. Although routing isn't case sensitive, ensure there's no special punctuation or spaces as they're common mistakes.
82+
Ensure you created a deployment for the given model and that the deployment name matches **exactly** the value you're passing in `model` parameter. Although routing isn't case sensitive, ensure there's no special punctuation or spaces typos.
8383
- question: |
8484
I'm using the `azure-ai-inference` package for Python and I get a 401 error when I try to authenticate using keys. What should I do?
8585
answer: |
@@ -115,5 +115,5 @@ sections:
115115
- question: |
116116
Do you use my company data to train any of the models?
117117
answer: |
118-
Azure AI model inference don't use customer data to retrain models. Your data is never shared with model providers.
118+
Azure AI model inference doesn't use customer data to retrain models. Your data is never shared with model providers.
119119
additionalContent: |

articles/ai-studio/ai-services/how-to/create-model-deployments.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@ To use it:
4040

4141
:::image type="content" source="../../media/ai-services/add-model-deployments/models-deploy-endpoint-url.png" alt-text="An screenshot showing how to get the URL and key associated with the deployment." lightbox="../../media/ai-services/add-model-deployments/models-deploy-endpoint-url.png":::
4242

43-
2. Use the model inference endpoint URL and the keys from before when constructing your client. The following examples uses the Azure AI Inference package:
43+
2. Use the model inference endpoint URL and the keys from before when constructing your client. The following example uses the Azure AI Inference package:
4444

4545
[!INCLUDE [code-create-chat-client](../../includes/ai-services/code-create-chat-client.md)]
4646

articles/ai-studio/ai-services/how-to/quickstart-github-models.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ recommendations: false
1414

1515
# Upgrade from GitHub Models to the Azure AI model inference service
1616

17-
If you want to develop a generative AI application, you can use [GitHub Models](https://docs.github.com/en/github-models/) to find and experiment with AI models for free. The playground and free API usage are [rate limited](https://docs.github.com/en/github-models/prototyping-with-ai-models#rate-limits) by requests per minute, requests per day, tokens per request, and concurrent requests. If you get rate limited, you will need to wait for the rate limit that you hit to reset before you can make more requests.
17+
If you want to develop a generative AI application, you can use [GitHub Models](https://docs.github.com/en/github-models/) to find and experiment with AI models for free. The playground and free API usage are [rate limited](https://docs.github.com/en/github-models/prototyping-with-ai-models#rate-limits) by requests per minute, requests per day, tokens per request, and concurrent requests. If you get rate limited, you need to wait for the rate limit that you hit to reset before you can make more requests.
1818

1919
Once you're ready to bring your application to production, you can upgrade your experience by deploying an Azure AI Services resource in an Azure subscription and start using the Azure AI model inference service. You don't need to change anything else in your code.
2020

@@ -52,18 +52,18 @@ To obtain the key and endpoint:
5252

5353
:::image type="content" source="../../media/ai-services/add-model-deployments/models-deploy-endpoint-url.png" alt-text="An screenshot showing how to get the URL and key associated with the deployment." lightbox="../../media/ai-services/add-model-deployments/models-deploy-endpoint-url.png":::
5454

55-
At this point, the model you selected will be ready to consume.
55+
At this point, the model you selected is ready to consume.
5656

5757
> [!TIP]
5858
> Use the parameter `model="<deployment-name>` to route your request to this deployment. *Deployments work as an alias of a given model under certain configurations*. See [Routing](../concepts/endpoints.md#routing) concept page to learn how Azure AI Services route deployments.
5959
6060
## Upgrade your code to use the new endpoint
6161

62-
Once your Azure AI Services resource is configured, you can start consuming it from your code. You will need the endpoint URL and key for it, which can be found in the **Overview** section:
62+
Once your Azure AI Services resource is configured, you can start consuming it from your code. You need the endpoint URL and key for it, which can be found in the **Overview** section:
6363

6464
:::image type="content" source="../../media/ai-services/overview/overview-endpoint-and-key.png" alt-text="An screenshot showing how to get the URL and key associated with the resource." lightbox="../../media/ai-services/overview/overview-endpoint-and-key.png":::
6565

66-
You can use any of the supported SDKs to get predictions out from the endpoint. The following SDKs are officially supported:
66+
You can use any of the supported SDK's to get predictions out from the endpoint. The following SDK's are officially supported:
6767

6868
* OpenAI SDK
6969
* Azure OpenAI SDK
@@ -77,14 +77,14 @@ Generate your first chat completion:
7777

7878
[!INCLUDE [code-create-chat-completion](../../includes/ai-services/code-create-chat-completion.md)]
7979

80-
## Explore additional features
80+
## Explore more features
8181

82-
Azure AI model inference supports additional features not available in GitHub Models, including:
82+
Azure AI model inference supports more features not available in GitHub Models, including:
8383

84-
* [Explore the model catalog](https://ai.azure.com/github/models) to see additional models not available in GitHub Models.
84+
* [Explore the model catalog](https://ai.azure.com/github/models) to see other models not available in GitHub Models.
8585
* Configure [content filtering](../../concepts/content-filtering.md).
8686
* Configure rate limiting (for specific models).
87-
* Explore additional [deployment SKUs (for specific models)](../concepts/deployment-types.md).
87+
* Explore more [deployment SKUs (for specific models)](../concepts/deployment-types.md).
8888
* Configure [private networking](../../../ai-services/cognitive-services-virtual-networks.md?context=/azure/ai-studio/context/context).
8989

9090
## Got troubles?

0 commit comments

Comments
 (0)