Skip to content

Commit 2056edb

Browse files
authored
Merge branch 'main' into streams-274-md
2 parents 87908b8 + 5b0bb52 commit 2056edb

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

explore-analyze/elastic-inference/inference-api.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -88,7 +88,7 @@ The behavior of allocations depends on several factors:
8888
If you enable adaptive allocations and set the `min_number_of_allocations` to a value greater than `0`, you will be charged for the machine learning resources, even if no inference requests are sent.
8989

9090
However, setting the `min_number_of_allocations` to a value greater than `0` keeps the model always available without scaling delays. Choose the configuration that best fits your workload and availability needs.
91-
::::
91+
::::
9292

9393
For more information about adaptive allocations and resources, refer to the [trained model autoscaling](/deploy-manage/autoscaling/trained-model-autoscaling.md) documentation.
9494

@@ -105,9 +105,9 @@ By default, documents are split into sentences and grouped in sections up to 250
105105

106106
### Chunking strategies
107107

108-
Several strategies are available for chunking:
108+
Several strategies are available for chunking:
109109

110-
#### `sentence`
110+
#### `sentence`
111111

112112
The `sentence` strategy splits the input text at sentence boundaries. Each chunk contains one or more complete sentences ensuring that the integrity of sentence-level context is preserved, except when a sentence causes a chunk to exceed a word count of `max_chunk_size`, in which case it will be split across chunks. The `sentence_overlap` option defines the number of sentences from the previous chunk to include in the current chunk which is either `0` or `1`.
113113

@@ -134,7 +134,7 @@ The default chunking strategy is `sentence`.
134134

135135
#### `word`
136136

137-
The `word` strategy splits the input text on individual words up to the `max_chunk_size` limit. The `overlap` option is the number of words from the previous chunk to include in the current chunk.
137+
The `word` strategy splits the input text on individual words up to the `max_chunk_size` limit. The `overlap` option is the number of words from the previous chunk to include in the current chunk.
138138

139139
The following example creates an {{infer}} endpoint with the `elasticsearch` service that deploys the ELSER model and configures the chunking behavior with the `word` strategy, setting a maximum of 120 words per chunk and an overlap of 40 words between chunks.
140140

@@ -158,7 +158,7 @@ PUT _inference/sparse_embedding/word_chunks
158158
#### `recursive`
159159

160160
```{applies_to}
161-
stack: ga 9.1`
161+
stack: ga 9.1
162162
```
163163

164164
The `recursive` strategy splits the input text based on a configurable list of separator patterns (for example, newlines or Markdown headers). The chunker applies these separators in order, recursively splitting any chunk that exceeds the `max_chunk_size` word limit. If no separator produces a small enough chunk, the strategy falls back to sentence-level splitting.
@@ -215,7 +215,7 @@ PUT _inference/sparse_embedding/recursive_custom_chunks
215215
#### `none`
216216

217217
```{applies_to}
218-
stack: ga 9.1`
218+
stack: ga 9.1
219219
```
220220

221221
The `none` strategy disables chunking and processes the entire input text as a single block, without any splitting or overlap. When using this strategy, you can instead [pre-chunk](https://www.elastic.co/docs/reference/elasticsearch/mapping-reference/semantic-text#auto-text-chunking) the input by providing an array of strings, where each element acts as a separate chunk to be sent directly to the inference service without further chunking.

0 commit comments

Comments
 (0)