Skip to content

Commit 5b750ae

Browse files
szabosteveseanhandleyleemthompomaxjakobflorent-leborgne
authored andcommitted
[E&A] Marks ELSER on EIS as GA (elastic#3014)
## Overview Related to elastic/search-team#10758 and elastic#3008 This PR: * amends the `applies_to` tags relevant to ELSER on EIS to mark the feature as GA * adds a Pricing section that links to the pricing page * removes limitations that are not relevant anymore ### Important note **Serverless:** After this PR is merged, the docs will state that ELSER on EIS is GA on Serverless. **Only merge once the feature is actually released as GA on Serverless.** **Stack:** If the PR is merged before the 9.2 release, the docs will show the GA status on the stack as Planned, without referencing a specific version. After the release, the badge will update dynamically to show the feature as GA in 9.2. --------- Co-authored-by: Sean Handley <[email protected]> Co-authored-by: Liam Thompson <[email protected]> Co-authored-by: Max Jakob <[email protected]> Co-authored-by: florent-leborgne <[email protected]>
1 parent a86675e commit 5b750ae

File tree

1 file changed

+8
-24
lines changed
  • explore-analyze/elastic-inference

1 file changed

+8
-24
lines changed

explore-analyze/elastic-inference/eis.md

Lines changed: 8 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -9,64 +9,48 @@ applies_to:
99

1010
# Elastic {{infer-cap}} Service [elastic-inference-service-eis]
1111

12-
The Elastic {{infer-cap}} Service (EIS) enables you to leverage AI-powered search as a service without deploying a model in your cluster.
12+
The Elastic {{infer-cap}} Service (EIS) enables you to leverage AI-powered search as a service without deploying a model in your environment.
1313
With EIS, you don't need to manage the infrastructure and resources required for {{ml}} {{infer}} by adding, configuring, and scaling {{ml}} nodes.
1414
Instead, you can use {{ml}} models for ingest, search, and chat independently of your {{es}} infrastructure.
1515

1616
## AI features powered by EIS [ai-features-powered-by-eis]
1717

1818
* Your Elastic deployment or project comes with a default [`Elastic Managed LLM` connector](https://www.elastic.co/docs/reference/kibana/connectors-kibana/elastic-managed-llm). This connector is used in the AI Assistant, Attack Discovery, Automatic Import and Search Playground.
1919

20-
* You can use [ELSER](/explore-analyze/machine-learning/nlp/ml-nlp-elser.md) to perform semantic search as a service (ELSER on EIS). {applies_to}`stack: preview 9.1` {applies_to}`serverless: preview`
20+
* You can use [ELSER](/explore-analyze/machine-learning/nlp/ml-nlp-elser.md) to perform semantic search as a service (ELSER on EIS). {applies_to}`stack: preview 9.1, ga 9.2` {applies_to}`serverless: ga`
2121

2222
## Region and hosting [eis-regions]
2323

2424
Requests through the `Elastic Managed LLM` are currently proxying to AWS Bedrock in AWS US regions, beginning with `us-east-1`.
2525
The request routing does not restrict the location of your deployments.
2626

27-
ELSER requests are managed by Elastic's own EIS infrastructure and are also hosted in AWS US regions, beginning with `us-east-1`. All Elastic Cloud hosted deployments and serverless projects in any CSP and region can access the endpoint. As we expand the service to Azure and GCP and more regions, we will automatically route requests to the same CSP and closest region the Elaticsearch cluster is hosted on.
2827

28+
ELSER requests are managed by Elastic's own EIS infrastructure and are also hosted in AWS US regions, beginning with `us-east-1`. All Elastic Cloud hosted deployments and serverless projects in any CSP and region can access the endpoint. As we expand the service to Azure and GCP and more regions, we will automatically route requests to the same CSP and closest region the Elaticsearch cluster is hosted on.
2929

3030
## ELSER via Elastic {{infer-cap}} Service (ELSER on EIS) [elser-on-eis]
3131

3232
```{applies_to}
33-
stack: preview 9.1
34-
serverless: preview
33+
stack: preview 9.1, ga 9.2
34+
serverless: ga
3535
```
3636

37-
ELSER on EIS enables you to use the ELSER model on GPUs, without having to manage your own ML nodes. We expect significantly better performance for throughput and consistent search latency as compared to ML nodes, and will continue to benchmark, remove limitations and address concerns as we move towards General Availability.
37+
ELSER on EIS enables you to use the ELSER model on GPUs, without having to manage your own ML nodes. We expect better performance for ingest throughput than ML nodes and equivalent performance for search latency. We will continue to benchmark, remove limitations and address concerns.
3838

3939
### Using the ELSER on EIS endpoint
4040

4141
You can now use `semantic_text` with the new ELSER endpoint on EIS. To learn how to use the `.elser-2-elastic` inference endpoint, refer to [Using ELSER on EIS](elasticsearch://reference/elasticsearch/mapping-reference/semantic-text.md#using-elser-on-eis).
4242

4343
#### Get started with semantic search with ELSER on EIS
44-
[Semantic Search with `semantic_text`](/solutions/search/semantic-search/semantic-search-semantic-text.md) has a detailed tutorial on using the `semantic_text` field and using the ELSER endpoint on EIS instead of the default endpoint. This is a great way to get started and try the new endpoint.
45-
46-
### Limitations
47-
48-
While we do encourage experimentation, we do not recommend implementing production use cases on top of this feature while it is in Technical Preview.
4944

50-
#### Uptime
51-
52-
There are no uptime guarantees during the Technical Preview.
53-
While Elastic will address issues promptly, the feature may be unavailable for extended periods.
54-
55-
#### Throughput and latency
45+
[Semantic Search with `semantic_text`](/solutions/search/semantic-search/semantic-search-semantic-text.md) has a detailed tutorial on using the `semantic_text` field and using the ELSER endpoint on EIS instead of the default endpoint. This is a great way to get started and try the new endpoint.
5646

57-
{{infer-cap}} throughput via this endpoint is expected to exceed that of {{infer}} operations on an ML node.
58-
However, throughput and latency are not guaranteed.
59-
Performance may vary during the Technical Preview.
47+
### Limitations
6048

6149
#### Batch size
6250

6351
Batches are limited to a maximum of 16 documents.
6452
This is particularly relevant when using the [_bulk API](https://www.elastic.co/docs/api/doc/elasticsearch/v9/operation/operation-bulk) for data ingestion.
6553

66-
#### Rate limits
67-
68-
Rate limit for search and ingest is currently at 500 requests per minute. This allows you to ingest approximately 8000 documents per minute at 16 documents per request.
69-
7054
## Pricing
7155

7256
All models on EIS incur a charge per million tokens. The pricing details are at our [Pricing page](https://www.elastic.co/pricing/serverless-search) for the Elastic Managed LLM and ELSER.

0 commit comments

Comments
 (0)