You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/machine-learning/concept-data-privacy.md
+8-8Lines changed: 8 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -26,25 +26,25 @@ When you deploy models in Azure Machine Learning, the following types of data ar
26
26
27
27
## Generate inferencing outputs with real-time endpoints
28
28
29
-
Deploying models to managed online endpoints deploys model weights to dedicated Virtual Machines and exposes a REST API for real-time inference. Learn more about deploying models from the [Model Catalog to real-time endpoints](concept-model-catalog.md). You manage the infrastructure for these real-time endpoints, and Azure’s data, privacy, and security commitments apply. Learn more about [Azure compliance offerings](https://servicetrust.microsoft.com/DocumentPage/7adf2d9e-d7b5-4e71-bad8-713e6a183cf3) applicable to Azure Machine Learning.
29
+
Deploying models to managed compute deploys model weights to dedicated Virtual Machines and exposes a REST API for real-time inference. Learn more about deploying models from the [Model Catalog to real-time endpoints](concept-model-catalog.md). You manage the infrastructure for these real-time endpoints, and Azure's data, privacy, and security commitments apply. Learn more about [Azure compliance offerings](https://servicetrust.microsoft.com/DocumentPage/7adf2d9e-d7b5-4e71-bad8-713e6a183cf3) applicable to Azure Machine Learning.
30
30
31
-
Although containers for models “Curated by Azure AI” have been scanned for vulnerabilities that could exfiltrate data, not all models available through the Model Catalog have been scanned. To reduce the risk of data exfiltration, you can protect your deployment using virtual networks. Follow this link to [learn more](./how-to-network-isolation-model-catalog.md). You can also use [Azure Policy](./how-to-regulate-registry-deployments.md) to regulate the models that can be deployed by your users.
31
+
Although containers for models "Curated by Azure AI" have been scanned for vulnerabilities that could exfiltrate data, not all models available through the model catalog have been scanned. To reduce the risk of data exfiltration, you can protect your deployment using virtual networks. Follow this link to [learn more](./how-to-network-isolation-model-catalog.md). You can also use [Azure Policy](./how-to-regulate-registry-deployments.md) to regulate the models that can be deployed by your users.
32
32
33
33
:::image type="content" source="media/concept-data-privacy/platform-service.png" alt-text="A diagram showing the platform service life cycle." lightbox="media/concept-data-privacy/platform-service.png":::
34
34
35
-
## Generate inferencing outputs with pay-as-you-go deployments (Models-as-a-Service)
35
+
## Generate inferencing outputs with serverless APIs (Models-as-a-Service)
36
36
37
-
When you deploy a model from the Model Catalog (base or finetuned) using pay-as-you-go deployments for inferencing, an API is provisioned giving you access to the model hosted and managed by the Azure Machine Learning Service. Learn more about [Models-as-a-Service](concept-model-catalog.md). The model processes your input prompts and generates outputs based on the functionality of the model, as described in the model details provided for the model. While the model is provided by the model provider, and your use of the model (and the model provider’s accountability for the model and its outputs) is subject to the license terms provided with the model, Microsoft provides and manages the hosting infrastructure and API endpoint. The models hosted in Models-as-a-Service are subject to Azure’s data, privacy, and security commitments. Learn more about Azure compliance offerings applicable to Azure Machine Learning [here](https://servicetrust.microsoft.com/DocumentPage/7adf2d9e-d7b5-4e71-bad8-713e6a183cf3).
37
+
When you deploy a model from the model catalog (base or finetuned) as a serverless API for inferencing, an API is provisioned giving you access to the model hosted and managed by the Azure Machine Learning Service. Learn more about [Models-as-a-Service](concept-model-catalog.md). The model processes your input prompts and generates outputs based on the functionality of the model, as described in the model details provided for the model. While the model is provided by the model provider, and your use of the model (and the model provider's accountability for the model and its outputs) is subject to the license terms provided with the model, Microsoft provides and manages the hosting infrastructure and API endpoint. The models hosted in Models-as-a-Service are subject to Azure's data, privacy, and security commitments. Learn more about Azure compliance offerings applicable to Azure Machine Learning [here](https://servicetrust.microsoft.com/DocumentPage/7adf2d9e-d7b5-4e71-bad8-713e6a183cf3).
38
38
39
-
Microsoft acts as the data processor for prompts and outputs sent to and generated by a model deployed for pay-as-you-go inferencing (MaaS). Microsoft does not share these prompts and outputs with the model provider, and Microsoft does not use these prompts and outputs to train or improve Microsoft’s, the model provider’s, or any third party’s models. Models are stateless and no prompts or outputs are stored in the model. If content filtering is enabled, prompts and outputs are screened for certain categories of harmful content by the Azure AI Content Safety service in real time; learn more about how Azure AI Content Safety processes data [here](/legal/cognitive-services/content-safety/data-privacy). Prompts and outputs are processed within the geography specified during deployment but may be processed between regions within the geography for operational purposes (including performance and capacity management).
39
+
Microsoft acts as the data processor for prompts and outputs sent to and generated by a model deployed for pay-as-you-go inferencing (MaaS). Microsoft does not share these prompts and outputs with the model provider, and Microsoft does not use these prompts and outputs to train or improve Microsoft's, the model provider's, or any third party's models. Models are stateless and no prompts or outputs are stored in the model. If content filtering is enabled, prompts and outputs are screened for certain categories of harmful content by the Azure AI Content Safety service in real time; learn more about how Azure AI Content Safety processes data [here](/legal/cognitive-services/content-safety/data-privacy). Prompts and outputs are processed within the geography specified during deployment but may be processed between regions within the geography for operational purposes (including performance and capacity management).
40
40
41
41
:::image type="content" source="media/concept-data-privacy/model-publisher-cycle.png" alt-text="A diagram showing model publisher service cycle." lightbox="media/concept-data-privacy/model-publisher-cycle.png":::
42
42
43
43
As explained during the deployment process for Models-as-a-Service, Microsoft may share customer contact information and transaction details (including usage volume associated with the offering) with the model publisher so that they can contact customers regarding the model. Learn more about information available to model publishers, [follow this link](/partner-center/analytics).
44
44
45
-
## Finetune a model for pay-as-you-go deployment (Models-as-a-Service)
45
+
## Finetune a model with serverless APIs (Models-as-a-Service)
46
46
47
-
If a model available for pay-as-you-go deployment (MaaS) supports finetuning, you can upload data to (or designate data already in) an [Azure Machine Learning Datastore](./concept-data.md) to finetune the model. You can then create a pay-as-you-go deployment for the finetuned model. The finetuned model can't be downloaded, but the finetuned model:
47
+
If a model available for serverless API deployment supports finetuning, you can upload data to (or designate data already in) an [Azure Machine Learning Datastore](./concept-data.md) to finetune the model. You can then create a serverless API for the finetuned model. The finetuned model can't be downloaded, but the finetuned model:
48
48
49
49
* Is available exclusively for your use;
50
50
@@ -56,7 +56,7 @@ Training data uploaded for finetuning isn't used to train, retrain, or improve a
56
56
57
57
## Data processing for downloaded models
58
58
59
-
If you download a model from the Model Catalog, you choose where to deploy the model, and you're responsible for how data is processed when you use the model.
59
+
If you download a model from the model catalog, you choose where to deploy the model, and you're responsible for how data is processed when you use the model.
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-deploy-models-cohere-command.md
+9-9Lines changed: 9 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -15,7 +15,7 @@ ms.custom: [references_regions]
15
15
#This functionality is also available in Azure AI Studio: /azure/ai-studio/how-to/deploy-models-cohere.md
16
16
---
17
17
# How to deploy Cohere Command models with Azure Machine Learning studio
18
-
Cohere offers two Command models in Azure Machine Learning studio. These models are available with pay-as-you-go tokenbased billing with Models as a Service.
18
+
Cohere offers two Command models in Azure Machine Learning studio. These models are available as serverless APIs with pay-as-you-go token-based billing.
19
19
20
20
* Cohere Command R
21
21
* Cohere Command R+
@@ -24,7 +24,7 @@ You can browse the Cohere family of models in the model catalog by filtering on
24
24
25
25
## Models
26
26
27
-
In this article, you learn how to use Azure Machine Learning studio to deploy the Cohere Command models as a service with pay-as you go billing.
27
+
In this article, you learn how to use Azure Machine Learning studio to deploy the Cohere Command models as a serverless API with pay-as you go billing.
28
28
29
29
### Cohere Command R
30
30
Command R is a highly performant generative large language model, optimized for a variety of use cases including reasoning, summarization, and question answering.
@@ -61,11 +61,11 @@ Pre-training data additionally included the following 13 languages: Russian, Pol
Certain models in the model catalog can be deployed as a service with pay-as-you-go, providing a way to consume them as an API without hosting them on your subscription, while keeping the enterprise security and compliance organizations need. This deployment option doesn't require quota from your subscription.
66
+
Certain models in the model catalog can be deployed as a serverless API with pay-as-you-go billing. This method of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance organizations need. This deployment option doesn't require quota from your subscription.
67
67
68
-
The previously mentioned Cohere models can be deployed as a service with pay-as-you-go, and are offered by Cohere through the Microsoft Azure Marketplace. Cohere can change or update the terms of use and pricing of this model.
68
+
The previously mentioned Cohere models can be deployed as a serverless API with pay-as-you-go, and are offered by Cohere through the Microsoft Azure Marketplace. Cohere can change or update the terms of use and pricing of this model.
69
69
70
70
### Prerequisites
71
71
@@ -84,12 +84,12 @@ The previously mentioned Cohere models can be deployed as a service with pay-as-
84
84
To create a deployment:
85
85
86
86
1. Go to [Azure Machine Learning studio](https://ml.azure.com/home).
87
-
1. Select the workspace in which you want to deploy your models. To use the pay-as-you-go model deployment offering, your workspace must belong to the EastUS2 or Sweden Central region.
87
+
1. Select the workspace in which you want to deploy your models. To use the serverless API deployment offering, your workspace must belong to the EastUS2 or Sweden Central region.
88
88
1. Choose the model you want to deploy from the [model catalog](https://ml.azure.com/model/catalog).
89
89
90
90
Alternatively, you can initiate deployment by going to your workspace and selecting **Endpoints** > **Serverless endpoints** > **Create**.
91
91
92
-
1. On the model's overview page in the model catalog, select **Deploy** and then **Pay-as-you-go**.
92
+
1. On the model's overview page in the model catalog, select **Deploy**.
93
93
94
94
:::image type="content" source="media/how-to-deploy-models-cohere-command/command-r-deploy-pay-as-you-go.png" alt-text="A screenshot showing how to deploy a model with the pay-as-you-go option." lightbox="media/how-to-deploy-models-cohere-command/command-r-deploy-pay-as-you-go.png":::
95
95
@@ -115,7 +115,7 @@ To create a deployment:
115
115
116
116
To learn about billing for models deployed with pay-as-you-go, see [Cost and quota considerations for Cohere models deployed as a service](#cost-and-quota-considerations-for-models-deployed-as-a-service).
117
117
118
-
### Consume the models as a service
118
+
### Consume the Cohere models as a service
119
119
120
120
The previously mentioned Cohere models can be consumed using the chat API.
121
121
@@ -126,7 +126,7 @@ The previously mentioned Cohere models can be consumed using the chat API.
126
126
127
127
For more information on using the APIs, see the [reference](#reference-for-cohere-models-deployed-as-a-service) section.
128
128
129
-
## Reference for Cohere models deployed as a service
129
+
## Reference for Cohere models deployed as a serverless API
130
130
131
131
Cohere Command R and Command R+ models accept both the [Azure AI Model Inference API](reference-model-inference-api.md) on the route `/chat/completions` and the native [Cohere Chat API](#cohere-chat-api) on `/v1/chat`.
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-deploy-models-cohere-embed.md
+7-8Lines changed: 7 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,7 +16,7 @@ ms.custom: [references_regions]
16
16
---
17
17
18
18
# How to deploy Cohere Embed models with Azure Machine Learning studio
19
-
Cohere offers two Embed models in Azure Machine Learning studio. These models are available with pay-as-you-go tokenbased billing with Models as a Service.
19
+
Cohere offers two Embed models in Azure Machine Learning studio. These models are available as serverless APIs with pay-as-you-go, token-based billing.
20
20
21
21
* Cohere Embed v3 - English
22
22
* Cohere Embed v3 - Multilingual
@@ -25,7 +25,7 @@ You can browse the Cohere family of models in the model catalog by filtering on
25
25
26
26
## Models
27
27
28
-
In this article, you learn how to use Azure Machine Learning studio to deploy the Cohere models as a service with pay-as you go billing.
28
+
In this article, you learn how to use Azure Machine Learning studio to deploy the Cohere models as a serverless API with pay-as you go billing.
29
29
30
30
### Cohere Embed v3 - English
31
31
Cohere Embed English is the market's leading text representation model used for semantic search, retrieval-augmented generation (RAG), classification, and clustering. Embed English has top performance on the HuggingFace MTEB benchmark and performs well on various industries such as Finance, Legal, and General-Purpose Corpora.
@@ -41,9 +41,8 @@ Cohere Embed Multilingual is the market's leading text representation model used
Certain models in the model catalog can be deployed as a service with pay-as-you-go, providing a way to consume them as an API without hosting them on your subscription, while keeping the enterprise security and compliance organizations need. This deployment option doesn't require quota from your subscription.
44
+
## Deploy as a serverless API
45
+
Certain models in the model catalog can be deployed as a serverless API with pay-as-you-go billing, providing a way to consume them as an API without hosting them on your subscription, while keeping the enterprise security and compliance organizations need. This deployment option doesn't require quota from your subscription.
47
46
48
47
The previously mentioned Cohere models can be deployed as a service with pay-as-you-go, and are offered by Cohere through the Microsoft Azure Marketplace. Cohere can change or update the terms of use and pricing of this model.
49
48
@@ -69,7 +68,7 @@ To create a deployment:
69
68
70
69
Alternatively, you can initiate deployment by going to your workspace and selecting **Endpoints** > **Serverless endpoints** > **Create**.
71
70
72
-
1. On the model's overview page in the model catalog, select **Deploy** and then **Pay-as-you-go**.
71
+
1. On the model's overview page in the model catalog, select **Deploy**.
73
72
74
73
:::image type="content" source="media/how-to-deploy-models-cohere-embed/embed-english-deploy-pay-as-you-go.png" alt-text="A screenshot showing how to deploy a model with the pay-as-you-go option." lightbox="media/how-to-deploy-models-cohere-embed/embed-english-deploy-pay-as-you-go.png":::
75
74
@@ -95,7 +94,7 @@ To create a deployment:
95
94
96
95
To learn about billing for models deployed with pay-as-you-go, see [Cost and quota considerations for Cohere models deployed as a service](#cost-and-quota-considerations-for-models-deployed-as-a-service).
97
96
98
-
### Consume the models as a service
97
+
### Consume the models deployed as a serverless API
99
98
100
99
The previously mentioned Cohere models can be consumed using the chat API.
101
100
@@ -106,7 +105,7 @@ The previously mentioned Cohere models can be consumed using the chat API.
106
105
107
106
For more information on using the APIs, see the [reference](#embed-api-reference-for-cohere-embed-models-deployed-as-a-service) section.
108
107
109
-
## Embed API reference for Cohere Embed models deployed as a service
108
+
## Embed API reference for Cohere Embed models deployed as a serverless API
110
109
111
110
Cohere Embed v3 - English and Embed v3 - Multilingual accept both the [Azure AI Model Inference API](reference-model-inference-api.md) on the route `/embeddings` (for text) and `/images/embeddings` (for images), and the native [Cohere Embed v3 API](#cohere-embed-v3) on `/embed`.
0 commit comments