Skip to content

Commit 91c6f3c

Browse files
committed
PM feedback
1 parent 0c5e634 commit 91c6f3c

File tree

1 file changed

+2
-3
lines changed

1 file changed

+2
-3
lines changed

articles/machine-learning/how-to-create-your-first-pipeline.md

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -337,6 +337,8 @@ pipeline1 = Pipeline(workspace=ws, steps=steps)
337337

338338
To use either a `TabularDataset` or `FileDataset` in your pipeline, you need to turn it into a [DatasetConsumptionConfig](https://docs.microsoft.com/python/api/azureml-core/azureml.data.dataset_consumption_config.datasetconsumptionconfig?view=azure-ml-py) object by calling [as_named_input(name)](https://docs.microsoft.com/python/api/azureml-core/azureml.data.abstract_dataset.abstractdataset?view=azure-ml-py#as-named-input-name-). You pass this `DatasetConsumptionConfig` object as one of the `inputs` to your pipeline step.
339339

340+
Datasets created from Azure Blob storage, Azure Files, Azure Data Lake Storage Gen1, Azure Data Lake Storage Gen2, Azure SQL Database, and Azure Database for PostgreSQL can be used as input to any pipeline step. With the exception of [DataTransferStep](https://docs.microsoft.com/python/api/azureml-pipeline-steps/azureml.pipeline.steps.datatransferstep?view=azure-ml-py) and [DatabricksStep](https://docs.microsoft.com/python/api/azureml-pipeline-steps/azureml.pipeline.steps.databricks_step.databricksstep?view=azure-ml-py), output data ([PipelineData](https://docs.microsoft.com/python/api/azureml-pipeline-core/azureml.pipeline.core.pipelinedata?view=azure-ml-py)) can only be written to Azure Blob and Azure File share datastores.
341+
340342
```python
341343
dataset_consuming_step = PythonScriptStep(
342344
script_name="iris_train.py",
@@ -357,9 +359,6 @@ iris_dataset = run_context.input_datasets['iris_data']
357359
dataframe = iris_dataset.to_pandas_dataframe()
358360
```
359361

360-
>[!NOTE]
361-
> All types of datasets (Blob, File Share, ADLS Gen 2, etc. ) can be used as input to any pipeline step, and output can be used in the DataTransferStep. However, writing output data (PipelineData) back to your ADLS Gen 2 dataset is not supported.
362-
363362
For more information, see the [azure-pipeline-steps package](https://docs.microsoft.com/python/api/azureml-pipeline-steps/?view=azure-ml-py) and [Pipeline class](https://docs.microsoft.com/python/api/azureml-pipeline-core/azureml.pipeline.core.pipeline%28class%29?view=azure-ml-py) reference.
364363

365364
## Submit the pipeline

0 commit comments

Comments
 (0)