You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/reference/connector/docs/connectors-content-extraction.asciidoc
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ The logic for content extraction is defined in {connectors-python}/connectors/ut
8
8
While intended primarily for PDF and Microsoft Office formats, you can use any of the <<es-connectors-content-extraction-supported-file-types, supported formats>>.
9
9
10
10
Enterprise Search uses an {ref}/ingest.html[Elasticsearch ingest pipeline^] to power the web crawler's binary content extraction.
11
-
The default pipeline, `ent-search-generic-ingestion`, is automatically created when Enterprise Search first starts.
11
+
The default pipeline, `content-generic-ingestion`, is automatically created when Enterprise Search first starts.
12
12
13
13
You can {ref}/ingest.html#create-manage-ingest-pipelines[view^] this pipeline in Kibana.
14
14
Customizing your pipeline usage is also an option.
Copy file name to clipboardExpand all lines: docs/reference/connector/docs/connectors-filter-extract-transform.asciidoc
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -13,7 +13,7 @@ The following diagram provides an overview of how content extraction, sync rules
13
13
[.screenshot]
14
14
image::images/pipelines-extraction-sync-rules.png[Architecture diagram of data pipeline with content extraction, sync rules, and ingest pipelines]
15
15
16
-
By default, only the connector specific logic (2) and the default `ent-search-generic-ingestion` pipeline (6) extract and transform your data, as configured in your deployment.
16
+
By default, only the connector specific logic (2) and the default `search-default-ingestion` pipeline (6) extract and transform your data, as configured in your deployment.
17
17
18
18
The following tools are available for more advanced use cases:
19
19
@@ -50,4 +50,4 @@ Use ingest pipelines for data enrichment, normalization, and more.
50
50
51
51
Elastic connectors use a default ingest pipeline, which you can copy and customize to meet your needs.
52
52
53
-
Refer to {ref}/ingest-pipeline-search.html[ingest pipelines in Search] in the {es} documentation.
53
+
Refer to {ref}/ingest-pipeline-search.html[ingest pipelines in Search] in the {es} documentation.
Copy file name to clipboardExpand all lines: docs/reference/ingest/search-ingest-pipelines.asciidoc
+18-15Lines changed: 18 additions & 15 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -40,7 +40,7 @@ Considerations such as error handling, conditional execution, sequencing, versio
40
40
To this end, when you create indices for search use cases, (including {enterprise-search-ref}/crawler.html[Elastic web crawler], <<es-connectors,connectors>>.
41
41
, and API indices), each index already has a pipeline set up with several processors that optimize your content for search.
42
42
43
-
This pipeline is called `ent-search-generic-ingestion`.
43
+
This pipeline is called `search-default-ingestion`.
44
44
While it is a "managed" pipeline (meaning it should not be tampered with), you can view its details via the Kibana UI or the Elasticsearch API.
45
45
You can also <<ingest-pipeline-search-details-generic-reference,read more about its contents below>>.
46
46
@@ -56,14 +56,14 @@ This will not effect existing indices.
56
56
57
57
Each index also provides the capability to easily create index-specific ingest pipelines with customizable processing.
58
58
If you need that extra flexibility, you can create a custom pipeline by going to your pipeline settings and choosing to "copy and customize".
59
-
This will replace the index's use of `ent-search-generic-ingestion` with 3 newly generated pipelines:
59
+
This will replace the index's use of `search-default-ingestion` with 3 newly generated pipelines:
60
60
61
61
1. `<index-name>`
62
62
2. `<index-name>@custom`
63
63
3. `<index-name>@ml-inference`
64
64
65
-
Like `ent-search-generic-ingestion`, the first of these is "managed", but the other two can and should be modified to fit your needs.
66
-
You can view these pipelines using the platform tools (Kibana UI, Elasticsearch API), and can also
65
+
Like `search-default-ingestion`, the first of these is "managed", but the other two can and should be modified to fit your needs.
66
+
You can view these pipelines using the platform tools (Kibana UI, Elasticsearch API), and can also
67
67
<<ingest-pipeline-search-details-specific,read more about their content below>>.
You can access this pipeline with the <<get-pipeline-api, Elasticsearch Ingest Pipelines API>> or via Kibana's <<create-manage-ingest-pipelines,Stack Management > Ingest Pipelines>> UI.
129
129
@@ -149,7 +149,7 @@ If you want to make customizations, we recommend you utilize index-specific pipe
The `ent-search-generic-ingestion` pipeline does not always run all processors.
152
+
The `search-default-ingestion` pipeline does not always run all processors.
153
153
It utilizes a feature of ingest pipelines to <<conditionally-run-processor,conditionally run processors>> based on the contents of each individual document.
154
154
155
155
* `_extract_binary_content` - if this field is present and has a value of `true` on a source document, the pipeline will attempt to run the `attachment`, `set_body`, and `remove_replacement_chars` processors.
@@ -167,8 +167,8 @@ See <<ingest-pipeline-search-pipeline-settings>>.
167
167
==== Index-specific ingest pipelines
168
168
169
169
In the Kibana UI for your index, by clicking on the Pipelines tab, then *Settings > Copy and customize*, you can quickly generate 3 pipelines which are specific to your index.
170
-
These 3 pipelines replace `ent-search-generic-ingestion` for the index.
171
-
There is nothing lost in this action, as the `<index-name>` pipeline is a superset of functionality over the `ent-search-generic-ingestion` pipeline.
170
+
These 3 pipelines replace `search-default-ingestion` for the index.
171
+
There is nothing lost in this action, as the `<index-name>` pipeline is a superset of functionality over the `search-default-ingestion` pipeline.
172
172
173
173
[IMPORTANT]
174
174
====
@@ -179,7 +179,7 @@ Refer to the Elastic subscriptions pages for https://www.elastic.co/subscription
This pipeline looks and behaves a lot like the <<ingest-pipeline-search-details-generic-reference,`ent-search-generic-ingestion` pipeline>>, but with <<ingest-pipeline-search-details-specific-reference-processors,two additional processors>>.
182
+
This pipeline looks and behaves a lot like the <<ingest-pipeline-search-details-generic-reference,`search-default-ingestion` pipeline>>, but with <<ingest-pipeline-search-details-specific-reference-processors,two additional processors>>.
183
183
184
184
[WARNING]
185
185
=========================
@@ -197,7 +197,7 @@ If you want to make customizations, we recommend you utilize <<ingest-pipeline-s
In addition to the processors inherited from the <<ingest-pipeline-search-details-generic-reference,`ent-search-generic-ingestion` pipeline>>, the index-specific pipeline also defines:
200
+
In addition to the processors inherited from the <<ingest-pipeline-search-details-generic-reference,`search-default-ingestion` pipeline>>, the index-specific pipeline also defines:
201
201
202
202
* `index_ml_inference_pipeline` - this uses the <<pipeline-processor, Pipeline>> processor to run the `<index-name>@ml-inference` pipeline.
203
203
This processor will only be run if the source document includes a `_run_ml_inference` field with the value `true`.
@@ -206,7 +206,7 @@ In addition to the processors inherited from the <<ingest-pipeline-search-detail
Like the `ent-search-generic-ingestion` pipeline, the `<index-name>` pipeline does not always run all processors.
209
+
Like the `search-default-ingestion` pipeline, the `<index-name>` pipeline does not always run all processors.
210
210
In addition to the `_extract_binary_content` and `_reduce_whitespace` control flow parameters, the `<index-name>` pipeline also supports:
211
211
212
212
* `_run_ml_inference` - if this field is present and has a value of `true` on a source document, the pipeline will attempt to run the `index_ml_inference_pipeline` processor.
@@ -220,7 +220,7 @@ See <<ingest-pipeline-search-pipeline-settings>>.
220
220
===== `<index-name>@ml-inference` Reference
221
221
222
222
This pipeline is empty to start (no processors), but can be added to via the Kibana UI either through the Pipelines tab of your index, or from the *Stack Management > Ingest Pipelines* page.
223
-
Unlike the `ent-search-generic-ingestion` pipeline and the `<index-name>` pipeline, this pipeline is NOT "managed".
223
+
Unlike the `search-default-ingestion` pipeline and the `<index-name>` pipeline, this pipeline is NOT "managed".
224
224
225
225
It's possible to add one or more ML inference pipelines to an index in the *Content* UI.
226
226
This pipeline will serve as a container for all of the ML inference pipelines configured for the index.
@@ -241,7 +241,7 @@ The `monitor_ml` Elasticsearch cluster permission is required in order to manage
241
241
242
242
This pipeline is empty to start (no processors), but can be added to via the Kibana UI either through the Pipelines
243
243
tab of your index, or from the *Stack Management > Ingest Pipelines* page.
244
-
Unlike the `ent-search-generic-ingestion` pipeline and the `<index-name>` pipeline, this pipeline is NOT "managed".
244
+
Unlike the `search-default-ingestion` pipeline and the `<index-name>` pipeline, this pipeline is NOT "managed".
245
245
246
246
You are encouraged to make additions and edits to this pipeline, provided its name remains the same.
247
247
This provides a convenient hook from which to add custom processing and transformations for your data.
@@ -272,9 +272,12 @@ extraction.
272
272
These changes should be re-applied to each index's `<index-name>@custom` pipeline in order to ensure a consistent data processing experience.
273
273
In 8.5+, the <<ingest-pipeline-search-pipeline-settings, index setting to enable binary content>> is required *in addition* to the configurations mentioned in the {enterprise-search-ref}/crawler-managing.html#crawler-managing-binary-content[Elastic web crawler Guide].
274
274
275
-
* `ent-search-generic-ingestion` - Since 8.5, Native Connectors, Connector Clients, and new (>8.4) Elastic web crawler indices will all make use of this pipeline by default.
275
+
* `ent-search-generic-ingestion` - Since 8.5, Native Connectors, Connector Clients, and new (>8.4) Elastic web crawler indices all made use of this pipeline by default.
276
+
This pipeline evolved into the `search-default-ingestion` pipeline.
277
+
278
+
* `search-default-ingestion` - Since 9.0, Connectors have made use of this pipeline by default.
276
279
You can <<ingest-pipeline-search-details-generic-reference, read more about this pipeline>> above.
277
-
As this pipeline is "managed", any modifications that were made to `app_search_crawler` and/or `ent_search_crawler` should NOT be made to `ent-search-generic-ingestion`.
280
+
As this pipeline is "managed", any modifications that were made to `app_search_crawler` and/or `ent_search_crawler` should NOT be made to `search-default-ingestion`.
278
281
Instead, if such customizations are desired, you should utilize <<ingest-pipeline-search-details-specific>>, placing all modifications in the `<index-name>@custom` pipeline(s).
Copy file name to clipboardExpand all lines: x-pack/plugin/core/template-resources/src/main/resources/entsearch/connector/elastic-connectors-mappings.json
Copy file name to clipboardExpand all lines: x-pack/plugin/ent-search/src/main/java/org/elasticsearch/xpack/application/connector/ConnectorTemplateRegistry.java
-10Lines changed: 0 additions & 10 deletions
Original file line number
Diff line number
Diff line change
@@ -46,10 +46,6 @@ public class ConnectorTemplateRegistry extends IndexTemplateRegistry {
Copy file name to clipboardExpand all lines: x-pack/plugin/ent-search/src/test/java/org/elasticsearch/xpack/application/connector/ConnectorIngestPipelineTests.java
+1-1Lines changed: 1 addition & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -50,7 +50,7 @@ public void testToXContent() throws IOException {
0 commit comments