|
| 1 | +# Web Service Input/Output |
| 2 | + |
| 3 | +This article describes **Web Service Input** module and **Web Service Output** module in Azure Machine Learning designer (preview). |
| 4 | + |
| 5 | +**Web Service Input** module can only connect to input port with type **DataFrameDirectory**. And **Web Service Output** module can only be connected from output port with type **DataFrameDirectory**. The two modules can be found in the module tree, under **Web Service** category. |
| 6 | + |
| 7 | +**Web Service Input** module is used to indicate where user data enters the pipeline and **Web Service Output** module is used to indicate where user data is returned in a Real-time inference Pipeline. |
| 8 | + |
| 9 | +## How to use Web Service Input/Output |
| 10 | + |
| 11 | +1. When you create a real-time inference pipeline from your training pipeline, **Web Service Input** and **Web Service Output** module will be automatically added to show where user data enters the pipeline and where data is returned. |
| 12 | + |
| 13 | + Learn more about [create a real-time inference pipeline](https://docs.microsoft.com/en-us/azure/machine-learning/tutorial-designer-automobile-price-deploy#create-a-real-time-inference-pipeline). |
| 14 | + |
| 15 | + > Note: |
| 16 | + > |
| 17 | + > Automatically generating real-time inference pipeline is a rule-based best-effort process, there is no guarantee for the correctness. You can manually add or remove **Web Service Input/Output** modules to satisfy your requirements. Make sure there is at least one **Web Service Input** module and one **Web Service Output** module in your real-time inference pipeline. If you have multiple **Web Service Input** or **Web Service Output** modules, make sure they have unique names, which you can input the name in the right panel of the module. |
| 18 | +
|
| 19 | +2. You can also manually create a real-time inference pipeline by adding **Web Service Input** and **Web Service Output** modules to your unsubmitted pipeline. |
| 20 | + |
| 21 | + > Note: |
| 22 | + > |
| 23 | + > The pipeline type will be determined at the first time you submit it. So be sure to add **Web Service Input** and **Web Service Output** module before you submit for the first time if you want to create a real-time inference pipeline. |
| 24 | +
|
| 25 | + Below example shows how to manually create real-time inference pipeline from **Execute Python Script** module. |
| 26 | + |
| 27 | +  |
| 28 | + |
| 29 | + After you submit the pipeline and the run completes successfully, you will be able to deploy the real-time endpoint. |
| 30 | + |
| 31 | + > Note: |
| 32 | + > |
| 33 | + > In the above example, **Enter Data Manually** provides the data schema for web service input and is necessary for deploying the real-time endpoint. Generally, you should always connect a module or dataset to the port which **Web Service Input** is connected to provide the data schema. |
| 34 | + |
| 35 | +## Next steps |
| 36 | +Learn more about [deploy the real-time endpoint](https://docs.microsoft.com/en-us/azure/machine-learning/tutorial-designer-automobile-price-deploy#deploy-the-real-time-endpoint). |
| 37 | +See the [set of modules available](module-reference.md) to Azure Machine Learning. |
0 commit comments