Skip to content

Commit 397a64e

Browse files
committed
fixes
1 parent 86eae1a commit 397a64e

File tree

6 files changed

+57
-77
lines changed

6 files changed

+57
-77
lines changed

articles/ai-studio/ai-services/concepts/endpoints.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@ All models deployed in Azure AI model inference service support the [Azure AI mo
7171

7272
## Azure OpenAI inference endpoint
7373

74-
Azure OpenAI models deployed to AI services also support the Azure OpenAI API. This API exposes the full capabilities of OpenAI models and support additional features like assistants, threads, files, and batch inference.
74+
Azure OpenAI models also support the Azure OpenAI API. This API exposes the full capabilities of OpenAI models and support additional features like assistants, threads, files, and batch inference.
7575

7676
Azure OpenAI inference endpoints are used per-deployment and they have they own URL that is associated with only one deployment. However, the same authentication mechanism can be used to consume it. Learn more in the reference page for [Azure OpenAI API](../../../ai-services/openai/reference.md)
7777

articles/ai-studio/ai-services/faq.yml

Lines changed: 15 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -34,15 +34,23 @@ sections:
3434
3535
Learn more about the [Azure OpenAI service](../../ai-services/openai/overview.md).
3636
- question: |
37-
What's the difference between Azure AI services and Azure AI studio?
37+
What's the difference between Azure AI model inference and Azure AI studio?
3838
answer: |
39-
Azure AI services are a suite of AI services that provide prebuilt APIs for common AI scenarios. Azure AI studio is a web-based tool that allows you to build, train, and deploy machine learning models. Azure AI services can be used in Azure AI studio to enhance your models with prebuilt AI capabilities.
39+
Azure AI services are a suite of AI services that provide prebuilt APIs for common AI scenarios. One of them is Azure AI model inference service which focuses on inference service of different state-of-the-art models. Azure AI studio is a web-based tool that allows you to build, train, and deploy machine learning models. Azure AI services can be used in Azure AI studio to enhance your models with prebuilt AI capabilities.
40+
- question: |
41+
What's the difference between Azure AI model inference service and Serverless API model deployments in Azure AI studio?
42+
answer: |
43+
Both technologies allow you to deploy models without requiring compute resources as they are based on the Models as a Service idea. [Serverless API model deployments](../how-to/deploy-models-serverless.md) allow you to deploy a single models under a unique endpoint and credentials. You need to create a different endpoint for each model you want to deploy. On top of that, they are always created in the context of the project and while they can be shared by creating connections from other projects, they live in the context of a given project.
44+
45+
Azure AI model inference service allows you to deploy multiple models under the same endpoint and credentials. You can switch between models without changing your code. They are also in the context of a shared resource, the Azure AI Services resource, which implies you can connect the resource to any project or hub that requires to consume the models you made available. Azure AI model inference service comes with a built-in model routing capability that routes the request to the right model based on the model name you pass in the request.
46+
47+
These two model deployment options have some differences in terms of their capabilities too. You can read about them at [../concepts/deployment-overview.md]
4048
- name: Models
4149
questions:
4250
- question: |
43-
Why aren't all the models in the Azure AI model catalog supported in Azure AI services?
51+
Why aren't all the models in the Azure AI model catalog supported in Azure AI model inference in Azure AI Services?
4452
answer: |
45-
The Azure AI model inference service in AI services supports all the models in the Azure AI catalog having pay-as-you-go billing. For more information, see [the Models section](model-inference.md#models).
53+
The Azure AI model inference service in AI services supports all the models in the Azure AI catalog with pay-as-you-go billing (per-token). For more information, see [the Models section](model-inference.md#models).
4654
4755
The Azure AI model catalog contains a wider list of models, however, those models require compute quota from your subscription. They also need to have a project or AI hub where to host the deployment. For more information, see [deployment options in Azure AI studio](../concepts/deployments-overview.md).
4856
- question: |
@@ -57,11 +65,11 @@ sections:
5765
Which are the supported SDKs and programming languages for Azure AI model inference service?
5866
answer: |
5967
You can use Azure Inference SDK with any model that is supported by:
60-
* The Azure AI model inference service in Azure AI services
68+
* The Azure AI Inference SDK
6169
* The `AzureOpenAI` class in OpenAI SDK
6270
* The Azure OpenAI SDK
6371
64-
Cohere SDK, Mistral SDK, and model provider-specific SDKs are not supported when connected to Azure AI services.
72+
Cohere SDK, Mistral SDK, and model provider-specific SDKs are not supported when connected to Azure AI model inference service.
6573
6674
For more information, see [supported SDKs and programming languages](concepts/endpoints.md).
6775
- question: |
@@ -107,11 +115,5 @@ sections:
107115
- question: |
108116
Do you use my company data to train any of the models?
109117
answer: |
110-
Azure AI services don't use customer data to retrain models. Your data is never shared with model providers.
111-
- name: Customer Copyright Commitment
112-
questions:
113-
- question: |
114-
How do I obtain coverage under the Customer Copyright Commitment?
115-
answer:
116-
The Customer Copyright Commitment is a provision to be included in the December 1, 2023, Microsoft Product Terms that describes Microsoft’s obligation to defend customers against certain third-party intellectual property claims relating to Output Content. If the subject of the claim is Output Content generated from the Azure OpenAI Service (or any other Covered Product that allows customers to configure the safety systems), then to receive coverage, customer must have implemented all mitigations required by the Azure OpenAI Service documentation in the offering that delivered the Output Content. The required mitigations are documented [here](/legal/cognitive-services/openai/customer-copyright-commitment?context=/azure/ai-services/openai/context/context) and updated on an ongoing basis. For new services, features, models, or use cases, new CCC requirements will be posted and take effect at or following the launch of such service, feature, model, or use case. Otherwise, customers will have six months from the time of publication to implement new mitigations to maintain coverage under the CCC. If a customer tenders a claim, the customer will be required to demonstrate compliance with the relevant requirements. These mitigations are required for Covered Products that allow customers to configure the safety systems, including Azure OpenAI Service; they do not impact coverage for customers using other Covered Products.
118+
Azure AI model inference don't use customer data to retrain models. Your data is never shared with model providers.
117119
additionalContent: |

articles/ai-studio/ai-services/how-to/create-model-deployments.md

Lines changed: 7 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -40,11 +40,16 @@ To use it:
4040

4141
:::image type="content" source="../../media/ai-services/add-model-deployments/models-deploy-endpoint-url.png" alt-text="An screenshot showing how to get the URL and key associated with the deployment." lightbox="../../media/ai-services/add-model-deployments/models-deploy-endpoint-url.png":::
4242

43-
2. When constructing your request, indicate the parameter `model` and insert the model deployment name you created.
43+
2. Use the model inference endpoint URL and the keys from before when constructing your client. The following examples uses the Azure AI Inference package:
4444

45+
[!INCLUDE [code-create-chat-client](../../includes/ai-services/code-create-chat-client.md)]
46+
47+
3. When constructing your request, indicate the parameter `model` and insert the model deployment name you created.
48+
4549
[!INCLUDE [code-create-chat-completion](../../includes/ai-services/code-create-chat-completion.md)]
4650

47-
3. When using the endpoint, you can change the `model` parameter to any available model deployment in your resource.
51+
> [!TIP]
52+
> When using the endpoint, you can change the `model` parameter to any available model deployment in your resource.
4853
4954
Additionally, Azure OpenAI models can be consumed using the [Azure OpenAI service endpoint](../../../ai-services/openai/supported-languages.md) in the resource. This endpoint is exclusive for each model deployment and has its own URL.
5055

articles/ai-studio/ai-services/model-inference.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ recommendations: false
1616

1717
Azure AI models inference service provides access to the most powerful models available in the **Azure AI model catalog**. Coming from the key model providers in the industry including OpenAI, Microsoft, Meta, Mistral, Cohere, G42, and AI21 Labs; these models can be integrated with software solutions to deliver a wide range of tasks including content generation, summarization, image understanding, semantic search, and code generation.
1818

19-
The Azure AI model inference service in Azure AI Services provides a way to **consume models as APIs without hosting them on your infrastructure**. Models are hosted in a Microsoft-managed infrastructure, which enables API-based access to the model provider's model. API-based access can dramatically reduce the cost of accessing a model and simplify the provisioning experience.
19+
The Azure AI model inference service provides a way to **consume models as APIs without hosting them on your infrastructure**. Models are hosted in a Microsoft-managed infrastructure, which enables API-based access to the model provider's model. API-based access can dramatically reduce the cost of accessing a model and simplify the provisioning experience.
2020

2121
## Models
2222

@@ -33,7 +33,7 @@ You can get access to the key model providers in the industry including OpenAI,
3333

3434
You can [decide and configure which models are available for inference](how-to/create-model-deployments.md) in the created resource. When a given model is configured, you can then generate predictions from it by indicating its model name or deployment name on your requests. No further changes are required in your code to use it.
3535

36-
To learn how to add models to the AI Services resource and use them read [Add and configure models to Azure AI Models in Azure AI Services](how-to/create-model-deployments.md).
36+
To learn how to add models to the Azure AI model inference resource and use them read [Add and configure models to Azure AI Models in Azure AI model inference](how-to/create-model-deployments.md).
3737

3838
## Pricing
3939

@@ -43,4 +43,4 @@ Models that are offered by Microsoft (for example, Phi-3 models and Azure OpenAI
4343

4444
## Next steps
4545

46-
* [Create your first model deployment in Azure AI Services](how-to/create-model-deployments.md)
46+
* [Create your first model deployment in Azure AI model inference](how-to/create-model-deployments.md)

0 commit comments

Comments
 (0)