You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
| Compute type | None (serverless) | Managed compute and Kubernetes | Managed compute and Kubernetes |
@@ -135,13 +134,12 @@ The following table shows a summary of the different features available to stand
135
134
| Cost basis<sup>5</sup> | Per token | Per deployment: compute instances running | Per job: compute instanced consumed in the job (capped to the maximum number of instances of the cluster) |
136
135
| Local testing of deployments | No | Yes | No |
137
136
138
-
<sup>2</sup> Deploying MLflow models to endpoints without outbound internet connectivity or private networks requires [packaging the model](concept-package-models.md) first.
139
137
140
-
<sup>3</sup> *Inference server* refers to the serving technology that takes requests, processes them, and creates responses. The inference server also dictates the format of the input and the expected outputs.
138
+
<sup>2</sup> *Inference server* refers to the serving technology that takes requests, processes them, and creates responses. The inference server also dictates the format of the input and the expected outputs.
141
139
142
-
<sup>4</sup> *Autoscaling* is the ability to dynamically scale up or scale down the deployment's allocated resources based on its load. Online and batch deployments use different strategies for autoscaling. While online deployments scale up and down based on the resource utilization (like CPU, memory, requests, etc.), batch endpoints scale up or down based on the number of jobs created.
140
+
<sup>3</sup> *Autoscaling* is the ability to dynamically scale up or scale down the deployment's allocated resources based on its load. Online and batch deployments use different strategies for autoscaling. While online deployments scale up and down based on the resource utilization (like CPU, memory, requests, etc.), batch endpoints scale up or down based on the number of jobs created.
143
141
144
-
<sup>5</sup> Both online and batch deployments charge by the resources consumed. In online deployments, resources are provisioned at deployment time. In batch deployment, resources aren't consumed at deployment time but at the time that the job runs. Hence, there's no cost associated with the batch deployment itself. Likewise, queued jobs don't consume resources either.
142
+
<sup>4</sup> Both online and batch deployments charge by the resources consumed. In online deployments, resources are provisioned at deployment time. In batch deployment, resources aren't consumed at deployment time but at the time that the job runs. Hence, there's no cost associated with the batch deployment itself. Likewise, queued jobs don't consume resources either.
- An Azure subscription. If you don't have an Azure subscription, create a [free account](https://azure.microsoft.com/free/) before you begin.
@@ -419,19 +417,6 @@ version = registered_model.version
419
417
)
420
418
```
421
419
422
-
Alternatively, if your endpoint doesn't have egress connectivity, use [model packaging (preview)](how-to-package-models.md) by including the argument `with_package=True`:
423
-
424
-
```python
425
-
blue_deployment = ManagedOnlineDeployment(
426
-
name="blue",
427
-
endpoint_name=endpoint_name,
428
-
model=model,
429
-
instance_type="Standard_F4s_v2",
430
-
instance_count=1,
431
-
with_package=True,
432
-
)
433
-
```
434
-
435
420
# [Python (MLflow SDK)](#tab/mlflow)
436
421
437
422
```python
@@ -473,10 +458,8 @@ version = registered_model.version
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-deploy-mlflow-models-online-progressive.md
-13Lines changed: 0 additions & 13 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -319,19 +319,6 @@ So far, the endpoint is empty. There are no deployments on it. Let's create the
319
319
instance_count=1,
320
320
)
321
321
```
322
-
323
-
If your endpoint doesn't have egress connectivity, use [model packaging (preview)](how-to-package-models.md) by including the argument `with_package=True`:
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-deploy-mlflow-models.md
+2-5Lines changed: 2 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -32,8 +32,6 @@ For no-code deployment, Azure Machine Learning:
32
32
- Packages required for Azure Machine Learning to perform inference, including [`mlflow-skinny`](https://github.com/mlflow/mlflow/blob/master/skinny/README_SKINNY.md).
Azure Machine Learning automatically generates environments to run inference on MLflow models. To build the environments, Azure Machine Learning reads the conda dependencies that are specified in the MLflow model and adds any packages that are required to run the inferencing server. These extra packages vary depending on deployment type.
@@ -238,11 +236,10 @@ Each tool has different capabilities, particularly for which type of compute it
238
236
| Deploy to web services like Azure Container Instances or Azure Kubernetes Service (AKS) | Legacy support<sup>2</sup> | Not supported<sup>2</sup> |
239
237
| Deploy to web services like Container Instances or AKS with a scoring script | Not supported<sup>3</sup> | Legacy support<sup>2</sup> |
240
238
241
-
<sup>1</sup> Deployment to online endpoints that are in workspaces with private link enabled requires you to [package models before deployment (preview)](how-to-package-models.md).
242
239
243
-
<sup>2</sup> Switch to [managed online endpoints](concept-endpoints.md) if possible.
240
+
<sup>1</sup> Switch to [managed online endpoints](concept-endpoints.md) if possible.
244
241
245
-
<sup>3</sup> Open-source MLflow doesn't have the concept of a scoring script and doesn't support batch execution.
242
+
<sup>2</sup> Open-source MLflow doesn't have the concept of a scoring script and doesn't support batch execution.
0 commit comments