Skip to content

Commit 632c660

Browse files
committed
Address review comments: Add Azure to Machine Learning instances
1 parent b08ecc9 commit 632c660

File tree

1 file changed

+17
-17
lines changed

1 file changed

+17
-17
lines changed

articles/machine-learning/how-to-access-data-batch-endpoints-jobs.md

Lines changed: 17 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ ms.custom:
2020

2121
# Create jobs and input data for batch endpoints
2222

23-
When you use batch endpoints in Azure Machine Learning, you can perform long batch operations over large amounts of data. The data can be located in different places, such as across disperse regions. Certain types of batch endpoints can also receive literal parameters as inputs.
23+
When you use batch endpoints in Azure Machine Learning, you can perform long batch operations over large amounts of input data. The data can be located in different places, such as across different regions. Certain types of batch endpoints can also receive literal parameters as inputs.
2424

2525
This article describes how to specify parameter inputs for batch endpoints and create deployment jobs. The process supports working with data from various sources, such as data assets, data stores, storage accounts, and local files.
2626

@@ -36,7 +36,7 @@ This article describes how to specify parameter inputs for batch endpoints and c
3636
3737
## Establish authentication
3838

39-
To invoke an endpoint, you need a valid Microsoft Entra token. When you invoke an endpoint, Machine Learning creates a batch deployment job under the identity that's associated with the token.
39+
To invoke an endpoint, you need a valid Microsoft Entra token. When you invoke an endpoint, Azure Machine Learning creates a batch deployment job under the identity that's associated with the token.
4040

4141
- If you use the Azure Machine Learning CLI (v2) or the Azure Machine Learning SDK for Python (v2) to invoke endpoints, you don't need to get the Microsoft Entra token manually. During sign in, the system authenticates your user identity. It also retrieves and passes the token for you.
4242
- If you use the REST API to invoke endpoints, you need to get the token manually.
@@ -53,7 +53,7 @@ az login
5353

5454
# [Python](#tab/sdk)
5555

56-
Use the Machine Learning SDK for Python to sign in:
56+
Use the Azure Machine Learning SDK for Python to sign in:
5757

5858
```python
5959
from azure.ai.ml import MLClient
@@ -62,7 +62,7 @@ from azure.identity import DefaultAzureCredential
6262
ml_client = MLClient.from_config(DefaultAzureCredential())
6363
```
6464

65-
If your configuration runs outside a Machine Learning compute instance, you need to specify the workspace where the endpoint is deployed:
65+
If your configuration runs outside an Azure Machine Learning compute instance, you need to specify the workspace where the endpoint is deployed:
6666

6767
```python
6868
from azure.ai.ml import MLClient
@@ -92,7 +92,7 @@ For more information about various types of credentials, see [How to run jobs us
9292

9393
## Create basic jobs
9494

95-
To create a job from a batch endpoint, you invoke the endpoint. Invocation can be done by using the Machine Learning CLI, the Machine Learning SDK for Python, or a REST API call.
95+
To create a job from a batch endpoint, you invoke the endpoint. Invocation can be done by using the Azure Machine Learning CLI, the Azure Machine Learning SDK for Python, or a REST API call.
9696

9797
The following examples show invocation basics for a batch endpoint that receives a single input data folder for processing. For examples that involve various inputs and outputs, see [Understand inputs and outputs](#understand-inputs-and-outputs).
9898

@@ -298,7 +298,7 @@ Batch endpoints provide a durable API that consumers can use to create batch job
298298

299299
Batch endpoints support two types of inputs:
300300

301-
- [Data inputs](#explore-data-inputs), or pointers to a specific storage location or Machine Learning asset
301+
- [Data inputs](#explore-data-inputs), or pointers to a specific storage location or Azure Machine Learning asset
302302
- [Literal inputs](#explore-literal-inputs), or literal values like numbers or strings that you want to pass to the job
303303

304304
The number and type of inputs and outputs depend on the [type of batch deployment](concept-endpoints-batch.md#batch-deployments). Model deployments always require one data input and produce one data output. Literal inputs aren't supported in model deployments. In contrast, pipeline component deployments provide a more general construct for building endpoints. In a pipeline component deployment, you can specify any number of data inputs, literal inputs, and outputs.
@@ -319,15 +319,15 @@ Data inputs refer to inputs that point to a location where data is placed. Becau
319319

320320
Batch endpoints can read files that are located in the following types of storage:
321321

322-
- [Machine Learning data assets](#use-input-data-from-a-data-asset), including the folder (`uri_folder`) and file (`uri_file`) types.
323-
- [Machine Learning data stores](#use-input-data-from-a-data-store), including Azure Blob Storage, Azure Data Lake Storage Gen1, and Azure Data Lake Storage Gen2.
322+
- [Azure Machine Learning data assets](#use-input-data-from-a-data-asset), including the folder (`uri_folder`) and file (`uri_file`) types.
323+
- [Azure Machine Learning data stores](#use-input-data-from-a-data-store), including Azure Blob Storage, Azure Data Lake Storage Gen1, and Azure Data Lake Storage Gen2.
324324
- [Azure Storage accounts](#use-input-data-from-a-storage-account), including Blob Storage, Data Lake Storage Gen1, and Data Lake Storage Gen2.
325-
- Local data folders and files, when you use the Machine Learning CLI or the Machine Learning SDK for Python to invoke endpoints. But the local data gets uploaded to the default data store of your Machine Learning workspace.
325+
- Local data folders and files, when you use the Azure Machine Learning CLI or the Azure Machine Learning SDK for Python to invoke endpoints. But the local data gets uploaded to the default data store of your Azure Machine Learning workspace.
326326

327327
> [!IMPORTANT]
328328
> **Deprecation notice**: Data assets of type `FileDataset` (V1) are deprecated and will be retired in the future. Existing batch endpoints that rely on this functionality will continue to work. But there's no support for V1 datasets in batch endpoints that are created with:
329329
>
330-
> - Versions of the Machine Learning CLI v2 that are generally available (2.4.0 and newer).
330+
> - Versions of the Azure Machine Learning CLI v2 that are generally available (2.4.0 and newer).
331331
> - Versions of the REST API that are generally available (2022-05-01 and newer).
332332
333333
### Explore literal inputs
@@ -343,7 +343,7 @@ Literal inputs are only supported in pipeline component deployments. To see how
343343

344344
### Explore data outputs
345345

346-
Data outputs refer to the location where the results of a batch job are placed. Each output has an identifiable name, and Machine Learning automatically assigns a unique path to each named output. You can specify another path if you need to.
346+
Data outputs refer to the location where the results of a batch job are placed. Each output has an identifiable name, and Azure Machine Learning automatically assigns a unique path to each named output. You can specify another path if you need to.
347347

348348
> [!IMPORTANT]
349349
> Batch endpoints only support writing outputs in Blob Storage data stores. If you need to write to a storage account with hierarchical namespaces enabled, such as Data Lake Storage Gen2, you can register the storage service as a Blob Storage data store, because the services are fully compatible. In this way, you can write outputs from batch endpoints to Data Lake Storage Gen2.
@@ -354,7 +354,7 @@ The following examples show how to create jobs while taking data inputs from [da
354354

355355
### Use input data from a data asset
356356

357-
Machine Learning data assets (formerly known as datasets) are supported as inputs for jobs. Follow these steps to run a batch endpoint job that uses input data that's stored in a registered data asset in Machine Learning.
357+
Azure Machine Learning data assets (formerly known as datasets) are supported as inputs for jobs. Follow these steps to run a batch endpoint job that uses input data that's stored in a registered data asset in Azure Machine Learning.
358358

359359
> [!WARNING]
360360
> Data assets of type table (`MLTable`) aren't currently supported.
@@ -414,7 +414,7 @@ Machine Learning data assets (formerly known as datasets) are supported as input
414414

415415
# [REST](#tab/rest)
416416

417-
Use the Machine Learning CLI or the Machine Learning SDK for Python to create the data asset.
417+
Use the Azure Machine Learning CLI or the Azure Machine Learning SDK for Python to create the data asset.
418418

419419
---
420420

@@ -536,13 +536,13 @@ Machine Learning data assets (formerly known as datasets) are supported as input
536536

537537
### Use input data from a data store
538538

539-
Your batch deployment jobs can directly reference data that's in Machine Learning registered data stores. In this example, you first upload some data to a data store in your Machine Learning workspace. Then you run a batch deployment on that data.
539+
Your batch deployment jobs can directly reference data that's in Azure Machine Learning registered data stores. In this example, you first upload some data to a data store in your Azure Machine Learning workspace. Then you run a batch deployment on that data.
540540

541-
This example uses the default data store, but you can use a different data store. In any Machine Learning workspace, the name of the default blob data store is **workspaceblobstore**. If you want to use a different data store in the following steps, replace `workspaceblobstore` with the name of your preferred data store.
541+
This example uses the default data store, but you can use a different data store. In any Azure Machine Learning workspace, the name of the default blob data store is **workspaceblobstore**. If you want to use a different data store in the following steps, replace `workspaceblobstore` with the name of your preferred data store.
542542

543543
1. Upload sample data to the data store. The sample data is available in the [azureml-examples](https://github.com/Azure/azureml-examples) repository. You can find the data in the [sdk/python/endpoints/batch/deploy-models/heart-classifier-mlflow/data](https://github.com/Azure/azureml-examples/tree/main/sdk/python/endpoints/batch/deploy-models/heart-classifier-mlflow/data) folder of that repository.
544544

545-
1. In Machine Learning studio, open the data assets page for your default blob data store, and then look up the name of its blob container.
545+
1. In Azure Machine Learning studio, open the data assets page for your default blob data store, and then look up the name of its blob container.
546546
1. Use a tool like Azure Storage Explorer or AzCopy to upload the sample data to a folder named heart-disease-uci-unlabeled in that container.
547547

548548
1. Set up the input information:
@@ -672,7 +672,7 @@ This example uses the default data store, but you can use a different data store
672672

673673
### Use input data from a Storage account
674674

675-
Machine Learning batch endpoints can read data from cloud locations in Storage accounts, both public and private. Use the following steps to run a batch endpoint job with data in a storage account.
675+
Azure Machine Learning batch endpoints can read data from cloud locations in Storage accounts, both public and private. Use the following steps to run a batch endpoint job with data in a storage account.
676676

677677
For more information about extra required configurations for reading data from storage accounts, see [Configure compute clusters for data access](how-to-authenticate-batch-endpoint.md#configure-compute-clusters-for-data-access).
678678

0 commit comments

Comments
 (0)