Skip to content

Commit d748418

Browse files
author
ecfan
committed
Update reference info
1 parent a2f62a3 commit d748418

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

articles/logic-apps/parse-document-chunk-text.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -143,25 +143,25 @@ Now, when you add other actions that expect and use tokenized input, such as the
143143

144144
#### Parameters
145145

146-
| Name | Value | Data type | Description | Limit |
146+
| Name | Value | Data type | Description | Limits |
147147
|------|-------|-----------|-------------|-------|
148-
| **Chunking Strategy** | **FixedLength** or **TokenSize** | String enum | **FixedLength**: Split the content, based on the number of characters <br><br>**TokenSize**: Split the content, based on the number of tokens. <br><br>Default: **FixedLength** ||
148+
| **Chunking Strategy** | **FixedLength** or **TokenSize** | String enum | **FixedLength**: Split the content, based on the number of characters <br><br>**TokenSize**: Split the content, based on the number of tokens. <br><br>Default: **FixedLength** | Not applicable |
149149
| **Text** | <*content-to-chunk*> | Any | The content to chunk. | See [Limits and configuration reference guide](logic-apps-limits-and-config.md#character-limits) |
150150

151151
For **Chunking Strategy** set to **FixedLength**:
152152

153-
| Name | Value | Data type | Description | Limit |
153+
| Name | Value | Data type | Description | Limits |
154154
|------|-------|-----------|-------------|-------|
155155
| **MaxPageLength** | <*max-char-per-chunk*> | Integer | The maximum number of characters per content chunk. <br><br>Default: **5000** | Minimum: **1** |
156156
| **PageOverlapLength** | <*number-of-overlapping-characters*> | Integer | The number of characters from the end of the previous chunk to include in the next chunk. This setting helps you avoid losing important information when splitting content into chunks and preserves continuity and context across chunks. <br><br>Default: **0** - No overlapping characters exist. | Minimum: **0** |
157157
| **Language** | <*language*> | String | The [language](/azure/ai-services/language-service/language-detection/language-support) to use for the resulting chunks. <br><br>Default: **en-us** | Not applicable |
158158

159159
For **Chunking Strategy** set to **TokenSize**:
160160

161-
| Name | Value | Data type | Description | Limit |
161+
| Name | Value | Data type | Description | Limits |
162162
|------|-------|-----------|-------------|-------|
163-
| **TokenSize** | <*max-tokens-per-chunk*> | Integer | The maximum number of tokens per content chunk. <br><br>Default: None | - Minimum: 1 <br><br>- Maximum: 8000 |
164-
| **Encoding model** | <*encoding-method*> | String enum | The [encoding method]() to use: **cl100k_base**, **cl200k_base**, **p50k_base**, **p50k_edit**, **r50k_base** <br><br>Default: None | Not applicable |
163+
| **TokenSize** | <*max-tokens-per-chunk*> | Integer | The maximum number of tokens per content chunk. <br><br>Default: None | Minimum: **1** <br>Maximum: **8000** |
164+
| **Encoding model** | <*encoding-method*> | String enum | The encoding model to use: <br><br>- Default: **cl100k_base (gpt4, gpt-3.5-turbo, gpt-35-turbo)** <br><br>- **r50k_base (gpt-3)** <br><br>- **p50k_base (gpt-3)** <br><br>- **p50k_edit (gpt-3)** <br><br>- **cl200k_base (gpt-4o)** <br><br>For more information, see [OpenAI - Models overview](https://platform.openai.com/docs/models/overview). | Not applicable |
165165

166166
> [!TIP]
167167
>

0 commit comments

Comments
 (0)