You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/document-intelligence/prebuilt/read.md
+43-43Lines changed: 43 additions & 43 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -26,7 +26,7 @@ ms.author: lajanuar
26
26
27
27
> [!NOTE]
28
28
>
29
-
> For extracting text from external images like labels, street signs, and posters, use the [Azure AI Image Analysis v4.0 Read](../../Computer-vision/concept-ocr.md) feature optimized for general, non-document images with a performance-enhanced synchronous API that makes it easier to embed OCR in real-time user experience scenarios.
29
+
> To extract text from external images like labels, street signs, and posters, use the [Azure AI Image Analysis v4.0 Read](../../Computer-vision/concept-ocr.md) feature optimized for general (not document) images with a performance-enhanced synchronous API. This capability makes it easier to embed OCR in real-time user experience scenarios.
30
30
>
31
31
32
32
Document Intelligence Read Optical Character Recognition (OCR) model runs at a higher resolution than Azure AI Vision Read and extracts print and handwritten text from PDF documents and scanned images. It also includes support for extracting text from Microsoft Word, Excel, PowerPoint, and HTML documents. It detects paragraphs, text lines, words, locations, and languages. The Read model is the underlying OCR engine for other Document Intelligence prebuilt models like Layout, General Document, Invoice, Receipt, Identity (ID) document, Health insurance card, W2 in addition to custom models.
@@ -83,11 +83,11 @@ See our [Language Support—document analysis models](../language-support/ocr.md
83
83
## Data extraction (v4)
84
84
85
85
> [!NOTE]
86
-
> Microsoft Word and HTML file are supported in v4.0. Compared with PDF and images, below features are not supported:
86
+
> Microsoft Word and HTML file are supported in v4.0. The following capabilities are currently not supported:
87
87
>
88
-
> *There are no angle, width/height and unit with each page object.
89
-
> *For each object detected, there is no bounding polygon or bounding region.
90
-
> *Page range (`pages`) is not supported as a parameter.
88
+
> *No angle, width/height, and unit returned with each page object.
89
+
> *No bounding polygon or bounding region for each object detected.
90
+
> *No page range (`pages`) as a parameter returned.
91
91
> * No `lines` object.
92
92
93
93
## Searchable PDFs
@@ -96,8 +96,8 @@ The searchable PDF capability enables you to convert an analog PDF, such as scan
96
96
97
97
> [!IMPORTANT]
98
98
>
99
-
> * Currently, the searchable PDF capability is only supported by Read OCR model `prebuilt-read`. When using this feature, please specify the `modelId` as `prebuilt-read`, as other model types will return error for this preview version.
100
-
> * Searchable PDF is included with the 2024-11-30 GA `prebuilt-read` model with no additional cost for generating a searchable PDF output.
99
+
> * Currently, only the Read OCR model `prebuilt-read` supports the searchable PDF capability. When using this feature, specify the `modelId` as `prebuilt-read`. Other model types return an error for this preview version.
100
+
> * Searchable PDF is included with the `2024-11-30` GA `prebuilt-read` model with no added cost for generating a searchable PDF output.
101
101
102
102
### Use searchable PDFs
103
103
@@ -124,41 +124,41 @@ Upon successful completion, the PDF can be retrieved and downloaded as `applicat
124
124
// Upon successful completion, retrieve the PDF as application/pdf.
125
125
GET {endpoint}/documentintelligence/documentModels/prebuilt-read/analyzeResults/{resultId}/pdf?api-version=2024-11-30
126
126
URI Parameters
127
-
Name In RequiredTypeDescription
128
-
endpointpathTrue
127
+
Name In RequiredTypeDescription
128
+
endpointpathTrue
129
129
string
130
130
131
-
uri
131
+
uri
132
132
The Document Intelligence service endpoint.
133
133
134
-
modelIdpathTrue
134
+
modelIdpathTrue
135
135
string
136
136
137
137
Unique document model name.
138
138
139
139
Regex pattern: ^[a-zA-Z0-9][a-zA-Z0-9._~-]{1,63}$
140
140
141
-
resultIdpathTrue
141
+
resultIdpathTrue
142
142
string
143
143
144
-
uuid
144
+
uuid
145
145
Analyze operation result ID.
146
146
147
-
api-versionqueryTrue
147
+
api-versionqueryTrue
148
148
string
149
149
150
150
The API version to use for this operation.
151
151
152
152
Responses
153
-
NameTypeDescription
154
-
200 OK
153
+
NameTypeDescription
154
+
200 OK
155
155
file
156
156
157
157
The request has succeeded.
158
158
159
159
Media Types: "application/pdf", "application/json"
An object containing more specific information about the error.
207
207
208
208
DocumentIntelligenceError
209
209
The error object.
210
210
211
-
NameTypeDescription
212
-
code
211
+
NameTypeDescription
212
+
code
213
213
string
214
214
215
215
One of a server-defined set of error codes.
216
216
217
-
details
217
+
details
218
218
DocumentIntelligenceError[]
219
219
220
220
An array of details about specific errors that led to this reported error.
221
221
222
-
innererror
222
+
innererror
223
223
DocumentIntelligenceInnerError
224
224
225
225
An object containing more specific information than the current object about the error.
226
226
227
-
message
227
+
message
228
228
string
229
229
230
230
A human-readable representation of the error.
231
231
232
-
target
232
+
target
233
233
string
234
234
235
235
The target of the error.
236
236
237
237
DocumentIntelligenceErrorResponse
238
238
Error response object.
239
239
240
-
NameTypeDescription
241
-
error
240
+
NameTypeDescription
241
+
error
242
242
DocumentIntelligenceError
243
243
244
244
Error info.
245
245
246
246
DocumentIntelligenceInnerError
247
247
An object containing more specific information about the error.
248
248
249
-
NameTypeDescription
250
-
code
249
+
NameTypeDescription
250
+
code
251
251
string
252
252
253
253
One of a server-defined set of error codes.
254
254
255
-
innererror
255
+
innererror
256
256
DocumentIntelligenceInnerError
257
257
258
258
Inner error.
259
259
260
-
message
260
+
message
261
261
string
262
262
263
263
A human-readable representation of the error.
@@ -439,7 +439,7 @@ Find more samples on GitHub:
439
439
440
440
> [!NOTE]
441
441
>
442
-
> For extracting text from external images like labels, street signs, and posters, use the [Azure AI Image Analysis v4.0 Read](../..//Computer-vision/concept-ocr.md) feature optimized for general, non-document images with a performance-enhanced synchronous API that makes it easier to embed OCR in your user experience scenarios.
442
+
> To extract text from external images like labels, street signs, and posters, use the [Azure AI Image Analysis v4.0 Read](../../Computer-vision/concept-ocr.md) feature optimized for general (not document) images with a performance-enhanced synchronous API. This capability makes it easier to embed OCR in real-time user experience scenarios.
443
443
>
444
444
445
445
Document Intelligence Read Optical Character Recognition (OCR) model runs at a higher resolution than Azure AI Vision Read and extracts print and handwritten text from PDF documents and scanned images. It also includes support for extracting text from Microsoft Word, Excel, PowerPoint, and HTML documents. It detects paragraphs, text lines, words, locations, and languages. The Read model is the underlying OCR engine for other Document Intelligence prebuilt models like Layout, General Document, Invoice, Receipt, Identity (ID) document, Health insurance card, W2 in addition to custom models.
@@ -513,11 +513,11 @@ See our [Language Support—document analysis models](../language-support/ocr.md
513
513
## Data extraction
514
514
515
515
> [!NOTE]
516
-
> Microsoft Word and HTML file are supported inv3.1 and later versions. Compared with PDF and images, below features are not supported:
516
+
> Microsoft Word and HTML file are supported inv4.0. The following capabilities are currently not supported:
517
517
>
518
-
>*There are no angle, width/height and unit with each page object.
519
-
>*For each object detected, there is no bounding polygon or bounding region.
520
-
>*Page range (`pages`) is not supported as a parameter.
518
+
>*No angle, width/height, and unit returned with each page object.
519
+
>*No bounding polygon or bounding regionfor each object detected.
520
+
>*No page range (`pages`) as a parameter returned.
521
521
>* No `lines` object.
522
522
523
523
## Searchable PDF
@@ -526,9 +526,9 @@ The searchable PDF capability enables you to convert an analog PDF, such as scan
526
526
527
527
> [!IMPORTANT]
528
528
>
529
-
> * Currently, the searchable PDF capability is only supported by Read OCR model `prebuilt-read`. When using this feature, please specify the `modelId` as `prebuilt-read`, as other model types will return an error.
530
-
> * Searchable PDF is included with the 2024-11-30 `prebuilt-read` model with no additional cost for generating a searchable PDF output.
531
-
> * Searchable PDF currently only supports PDF files as input. Support for other file types, such as image files, will be available later.
529
+
> * Currently, only Read OCR model `prebuilt-read` supports the searchable PDF capability. When using this feature, specify the `modelId` as `prebuilt-read`. Other model types return an error.
530
+
> * Searchable PDF is included with the `2024-11-30` `prebuilt-read` model with no added cost for generating a searchable PDF output.
531
+
> * Searchable PDF currently only supports PDF files as input.
0 commit comments