Skip to content

Commit cc0f8df

Browse files
committed
Addresses feedback.
1 parent 0914720 commit cc0f8df

File tree

3 files changed

+16
-3
lines changed

3 files changed

+16
-3
lines changed

explore-analyze/elastic-inference.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ navigation_title: Elastic Inference
99

1010
## Overview
1111

12-
{{infer-cap}} is a process of using an LLM or a {{ml}} trained model to make predictions or operations - such as text embedding, completion, or reranking - on your data.
12+
{{infer-cap}} is a process of using a {{ml}} trained model to make predictions or operations - such as text embedding, or reranking - on your data.
1313
You can use {{infer}} during ingest time (for example, to create embeddings from textual data you ingest) or search time (to perform [semantic search](/solutions/search/semantic-search.md)).
1414
There are several ways to perform {{infer}} in the {{stack}}:
1515

explore-analyze/elastic-inference/eis.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,9 +19,11 @@ Instead, you can use {{ml}} models for ingest, search, and chat independently of
1919

2020
## Region and hosting [eis-regions]
2121

22-
The EIS requests are currently proxying to AWS Bedrock in AWS US regions, beginning with `us-east-1`.
22+
Requests through the Elastic Managed LLM are currently proxying to AWS Bedrock in AWS US regions, beginning with `us-east-1`.
2323
The request routing does not restrict the location of your deployments.
2424

25+
ELSER requests are managed by Elastic own EIS infrastructure.
26+
2527
## ELSER via Elastic {{infer-cap}} Service (ELSER on EIS)
2628

2729
{applies_to}`stack: preview 9.1` {applies_to}`serverless: preview`

explore-analyze/elastic-inference/inference-api.md

Lines changed: 12 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,9 +20,20 @@ You can use the default {{infer}} endpoints your deployment contains or create a
2020

2121
## Default {{infer}} endpoints [default-enpoints]
2222

23-
Your {{es}} deployment contains preconfigured {{infer}} endpoints which makes them easier to use when defining `semantic_text` fields or using {{infer}} processors. The following list contains the default {{infer}} endpoints listed by `inference_id`:
23+
Your {{es}} deployment contains preconfigured {{infer}} endpoints, which makes them easier to use when defining `semantic_text` fields or using {{infer}} processors. These endpoints come in two forms:
24+
25+
- **Elastic Inference Service (EIS) endpoints**, which provide {{infer}} as a managed service and do not consume resources from your own nodes.
26+
27+
- **ML node-based endpoints**, which run on your dedicated {{ml}} nodes.
28+
29+
The following section lists the default {{infer}} endpoints, identified by their `inference_id`, grouped by whether they are EIS- or ML node–based.
30+
31+
### Default endpoints for Elastic {{infer-cap}} Service (EIS)
2432

2533
- {applies_to}`stack: preview 9.1` {applies_to}`serverless: preview` `.elser-2-elastic`: uses the [ELSER](/explore-analyze/machine-learning/nlp/ml-nlp-elser.md) trained model as an Elastic {{infer-cap}} Service for `sparse_embedding` tasks (recommended for English language text). The `model_id` is `.elser_model_2`.
34+
35+
### Default endpoints used on ML-nodes
36+
2637
- `.elser-2-elasticsearch`: uses the [ELSER](/explore-analyze/machine-learning/nlp/ml-nlp-elser.md) built-in trained model for `sparse_embedding` tasks (recommended for English language text). The `model_id` is `.elser_model_2_linux-x86_64`.
2738
- `.multilingual-e5-small-elasticsearch`: uses the [E5](../../explore-analyze/machine-learning/nlp/ml-nlp-e5.md) built-in trained model for `text_embedding` tasks (recommended for non-English language texts). The `model_id` is `.e5_model_2_linux-x86_64`.
2839

0 commit comments

Comments
 (0)