You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The batch client takes a yaml configuration file that specifies the on-prem container endpoints. The following example can be written to `/mnt/my_nfs/config.yaml`, which is used in the examples below.
45
+
The batch client takes a yaml configuration file that specifies the on-premises container endpoints. The following example can be written to `/mnt/my_nfs/config.yaml`, which is used in the examples below.
46
46
47
47
48
48
@@ -64,13 +64,20 @@ MyContainer3:
64
64
rtf: 4
65
65
```
66
66
67
-
This yaml example specifies three speech containers on three hosts. The first host is specified by a IPv4 address, the second is running on the same VM as the batch-client, and the third container is specified by the DNS hostname of another VM. The `concurrency` value specifies the maximum concurrent file transcriptions that can run on the same container. The `rtf` (Real-Time Factor) value is optional, and can be used to tune performance.
67
+
This yaml example specifies three speech containers on three hosts. The first host is specified by a IPv4 address, the second is running on the same VM as the batch-client, and the third container is specified by the DNS hostname of another VM. The `concurrency` value specifies the maximum concurrent file transcriptions that can run on the same container. The `rtf` (Real-Time Factor) value is optional and can be used to tune performance.
68
+
68
69
The batch client can dynamically detect if an endpoint becomes unavailable (for example, due to a container restart or networking issue), and when it becomes available again. Transcription requests will not be sent to containers that are unavailable, and the client will continue using other available containers. You can add, remove, or edit endpoints at any time without interrupting the progress of your batch.
69
70
70
71
71
72
72
73
73
74
75
+
76
+
77
+
78
+
79
+
80
+
74
81
## Run the batch processing container
75
82
76
83
> [!NOTE]
@@ -83,6 +90,10 @@ Use the Docker `run` command to start the container. This will start an interact
The client will start running. If an audio file has already been transcribed in a previous run, the client will automatically skip the file. Files are sent with an automatic retry if transient errors occur, and you can differentiate between which errors you want to the client to retry on. On a transcription error, the client will continue transcription, and can retry without losing progress.
109
127
110
128
## Run modes
@@ -141,18 +159,17 @@ The batch processing kit offers three modes, using the `--run-mode` parameter.
141
159
142
160
#### [REST](#tab/rest)
143
161
144
-
`REST`mode is an API server mode that provides a basic set of HTTP endpoints for audio file batch submission, status checking, and long polling. Also enables programmatic consumption using a Python module extension, or importing as a submodule.
162
+
`REST`mode is an API server mode that provides a basic set of HTTP endpoints for audio file batch submission, status checking, and long polling. Also enables programmatic consumption using a Python module extension or importing as a submodule.
145
163
146
164
:::image type="content" source="media/containers/batch-rest-api-mode.png" alt-text="A diagram showing the batch-kit container processing files in REST mode.":::
147
165
148
166
1. Define the Speech container endpoints that the batch client will use in the `config.yaml` file.
149
-
2. Send an HTTP request request to one of the API server's endpoints.
150
-
167
+
1. Send an HTTP request to one of the API server's endpoints.
151
168
|Endpoint |Description |
152
-
|---------|---------|
153
-
|`/submit` | Endpoint for creating new batch requests. |
154
-
|`/status` | Endpoint for checking the status of a batch request. The connection will stay open until the batch completes. |
155
-
|`/watch` | Endpoint for using HTTP long polling until the batch completes. |
169
+
|---------|---------|
170
+
|`/submit` | Endpoint for creating new batch requests. |
171
+
|`/status` | Endpoint for checking the status of a batch request. The connection will stay open until the batch completes. |
172
+
|`/watch` | Endpoint for using HTTP long polling until the batch completes. |
156
173
157
174
3. Audio files are uploaded from the input directory. If the audio file has already been transcribed in a previous run with the same output directory (same file name and checksum), the client will skip the file.
158
175
4. If a request is sent to the `/submit` endpoint, the files are dispatched to the container endpoints from step 1.
@@ -173,3 +190,4 @@ The output directory specified by `-output_folder` will contain a *run_summary.j
173
190
* [How to install and run containers](speech-container-howto.md)
0 commit comments