You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-studio/ai-services/concepts/endpoints.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -71,7 +71,7 @@ All models deployed in Azure AI model inference service support the [Azure AI mo
71
71
72
72
## Azure OpenAI inference endpoint
73
73
74
-
Azure OpenAI models deployed to AI services also support the Azure OpenAI API. This API exposes the full capabilities of OpenAI models and support additional features like assistants, threads, files, and batch inference.
74
+
Azure OpenAI models also support the Azure OpenAI API. This API exposes the full capabilities of OpenAI models and support additional features like assistants, threads, files, and batch inference.
75
75
76
76
Azure OpenAI inference endpoints are used per-deployment and they have they own URL that is associated with only one deployment. However, the same authentication mechanism can be used to consume it. Learn more in the reference page for [Azure OpenAI API](../../../ai-services/openai/reference.md)
Copy file name to clipboardExpand all lines: articles/ai-studio/ai-services/faq.yml
+15-13Lines changed: 15 additions & 13 deletions
Original file line number
Diff line number
Diff line change
@@ -34,15 +34,23 @@ sections:
34
34
35
35
Learn more about the [Azure OpenAI service](../../ai-services/openai/overview.md).
36
36
- question: |
37
-
What's the difference between Azure AI services and Azure AI studio?
37
+
What's the difference between Azure AI model inference and Azure AI studio?
38
38
answer: |
39
-
Azure AI services are a suite of AI services that provide prebuilt APIs for common AI scenarios. Azure AI studio is a web-based tool that allows you to build, train, and deploy machine learning models. Azure AI services can be used in Azure AI studio to enhance your models with prebuilt AI capabilities.
39
+
Azure AI services are a suite of AI services that provide prebuilt APIs for common AI scenarios. One of them is Azure AI model inference service which focuses on inference service of different state-of-the-art models. Azure AI studio is a web-based tool that allows you to build, train, and deploy machine learning models. Azure AI services can be used in Azure AI studio to enhance your models with prebuilt AI capabilities.
40
+
- question: |
41
+
What's the difference between Azure AI model inference service and Serverless API model deployments in Azure AI studio?
42
+
answer: |
43
+
Both technologies allow you to deploy models without requiring compute resources as they are based on the Models as a Service idea. [Serverless API model deployments](../how-to/deploy-models-serverless.md) allow you to deploy a single models under a unique endpoint and credentials. You need to create a different endpoint for each model you want to deploy. On top of that, they are always created in the context of the project and while they can be shared by creating connections from other projects, they live in the context of a given project.
44
+
45
+
Azure AI model inference service allows you to deploy multiple models under the same endpoint and credentials. You can switch between models without changing your code. They are also in the context of a shared resource, the Azure AI Services resource, which implies you can connect the resource to any project or hub that requires to consume the models you made available. Azure AI model inference service comes with a built-in model routing capability that routes the request to the right model based on the model name you pass in the request.
46
+
47
+
These two model deployment options have some differences in terms of their capabilities too. You can read about them at [../concepts/deployment-overview.md]
40
48
- name: Models
41
49
questions:
42
50
- question: |
43
-
Why aren't all the models in the Azure AI model catalog supported in Azure AI services?
51
+
Why aren't all the models in the Azure AI model catalog supported in Azure AI model inference in Azure AI Services?
44
52
answer: |
45
-
The Azure AI model inference service in AI services supports all the models in the Azure AI catalog having pay-as-you-go billing. For more information, see [the Models section](model-inference.md#models).
53
+
The Azure AI model inference service in AI services supports all the models in the Azure AI catalog with pay-as-you-go billing (per-token). For more information, see [the Models section](model-inference.md#models).
46
54
47
55
The Azure AI model catalog contains a wider list of models, however, those models require compute quota from your subscription. They also need to have a project or AI hub where to host the deployment. For more information, see [deployment options in Azure AI studio](../concepts/deployments-overview.md).
48
56
- question: |
@@ -57,11 +65,11 @@ sections:
57
65
Which are the supported SDKs and programming languages for Azure AI model inference service?
58
66
answer: |
59
67
You can use Azure Inference SDK with any model that is supported by:
60
-
* The Azure AI model inference service in Azure AI services
68
+
* The Azure AI Inference SDK
61
69
* The `AzureOpenAI` class in OpenAI SDK
62
70
* The Azure OpenAI SDK
63
71
64
-
Cohere SDK, Mistral SDK, and model provider-specific SDKs are not supported when connected to Azure AI services.
72
+
Cohere SDK, Mistral SDK, and model provider-specific SDKs are not supported when connected to Azure AI model inference service.
65
73
66
74
For more information, see [supported SDKs and programming languages](concepts/endpoints.md).
67
75
- question: |
@@ -107,11 +115,5 @@ sections:
107
115
- question: |
108
116
Do you use my company data to train any of the models?
109
117
answer: |
110
-
Azure AI services don't use customer data to retrain models. Your data is never shared with model providers.
111
-
- name: Customer Copyright Commitment
112
-
questions:
113
-
- question: |
114
-
How do I obtain coverage under the Customer Copyright Commitment?
115
-
answer:
116
-
The Customer Copyright Commitment is a provision to be included in the December 1, 2023, Microsoft Product Terms that describes Microsoft’s obligation to defend customers against certain third-party intellectual property claims relating to Output Content. If the subject of the claim is Output Content generated from the Azure OpenAI Service (or any other Covered Product that allows customers to configure the safety systems), then to receive coverage, customer must have implemented all mitigations required by the Azure OpenAI Service documentation in the offering that delivered the Output Content. The required mitigations are documented [here](/legal/cognitive-services/openai/customer-copyright-commitment?context=/azure/ai-services/openai/context/context) and updated on an ongoing basis. For new services, features, models, or use cases, new CCC requirements will be posted and take effect at or following the launch of such service, feature, model, or use case. Otherwise, customers will have six months from the time of publication to implement new mitigations to maintain coverage under the CCC. If a customer tenders a claim, the customer will be required to demonstrate compliance with the relevant requirements. These mitigations are required for Covered Products that allow customers to configure the safety systems, including Azure OpenAI Service; they do not impact coverage for customers using other Covered Products.
118
+
Azure AI model inference don't use customer data to retrain models. Your data is never shared with model providers.
Copy file name to clipboardExpand all lines: articles/ai-studio/ai-services/how-to/create-model-deployments.md
+7-2Lines changed: 7 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -40,11 +40,16 @@ To use it:
40
40
41
41
:::image type="content" source="../../media/ai-services/add-model-deployments/models-deploy-endpoint-url.png" alt-text="An screenshot showing how to get the URL and key associated with the deployment." lightbox="../../media/ai-services/add-model-deployments/models-deploy-endpoint-url.png":::
42
42
43
-
2.When constructing your request, indicate the parameter `model` and insert the model deployment name you created.
43
+
2.Use the model inference endpoint URL and the keys from before when constructing your client. The following examples uses the Azure AI Inference package:
3. When using the endpoint, you can change the `model` parameter to any available model deployment in your resource.
51
+
> [!TIP]
52
+
> When using the endpoint, you can change the `model` parameter to any available model deployment in your resource.
48
53
49
54
Additionally, Azure OpenAI models can be consumed using the [Azure OpenAI service endpoint](../../../ai-services/openai/supported-languages.md) in the resource. This endpoint is exclusive for each model deployment and has its own URL.
Copy file name to clipboardExpand all lines: articles/ai-studio/ai-services/model-inference.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,7 +16,7 @@ recommendations: false
16
16
17
17
Azure AI models inference service provides access to the most powerful models available in the **Azure AI model catalog**. Coming from the key model providers in the industry including OpenAI, Microsoft, Meta, Mistral, Cohere, G42, and AI21 Labs; these models can be integrated with software solutions to deliver a wide range of tasks including content generation, summarization, image understanding, semantic search, and code generation.
18
18
19
-
The Azure AI model inference service in Azure AI Services provides a way to **consume models as APIs without hosting them on your infrastructure**. Models are hosted in a Microsoft-managed infrastructure, which enables API-based access to the model provider's model. API-based access can dramatically reduce the cost of accessing a model and simplify the provisioning experience.
19
+
The Azure AI model inference service provides a way to **consume models as APIs without hosting them on your infrastructure**. Models are hosted in a Microsoft-managed infrastructure, which enables API-based access to the model provider's model. API-based access can dramatically reduce the cost of accessing a model and simplify the provisioning experience.
20
20
21
21
## Models
22
22
@@ -33,7 +33,7 @@ You can get access to the key model providers in the industry including OpenAI,
33
33
34
34
You can [decide and configure which models are available for inference](how-to/create-model-deployments.md) in the created resource. When a given model is configured, you can then generate predictions from it by indicating its model name or deployment name on your requests. No further changes are required in your code to use it.
35
35
36
-
To learn how to add models to the AI Services resource and use them read [Add and configure models to Azure AI Models in Azure AI Services](how-to/create-model-deployments.md).
36
+
To learn how to add models to the Azure AI model inference resource and use them read [Add and configure models to Azure AI Models in Azure AI model inference](how-to/create-model-deployments.md).
37
37
38
38
## Pricing
39
39
@@ -43,4 +43,4 @@ Models that are offered by Microsoft (for example, Phi-3 models and Azure OpenAI
43
43
44
44
## Next steps
45
45
46
-
*[Create your first model deployment in Azure AI Services](how-to/create-model-deployments.md)
46
+
*[Create your first model deployment in Azure AI model inference](how-to/create-model-deployments.md)
0 commit comments