Skip to content

Commit 68d3b99

Browse files
committed
[E&A] Marks ELSER on EIS as GA.
1 parent 5e76db5 commit 68d3b99

File tree

1 file changed

+11
-16
lines changed
  • explore-analyze/elastic-inference

1 file changed

+11
-16
lines changed

explore-analyze/elastic-inference/eis.md

Lines changed: 11 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -7,15 +7,15 @@ applies_to:
77

88
# Elastic {{infer-cap}} Service [elastic-inference-service-eis]
99

10-
The Elastic {{infer-cap}} Service (EIS) enables you to leverage AI-powered search as a service without deploying a model in your cluster.
10+
The Elastic {{infer-cap}} Service (EIS) enables you to leverage AI-powered search as a service without deploying a model in your environment.
1111
With EIS, you don't need to manage the infrastructure and resources required for {{ml}} {{infer}} by adding, configuring, and scaling {{ml}} nodes.
1212
Instead, you can use {{ml}} models for ingest, search, and chat independently of your {{es}} infrastructure.
1313

1414
## AI features powered by EIS [ai-features-powered-by-eis]
1515

1616
* Your Elastic deployment or project comes with a default [`Elastic Managed LLM` connector](https://www.elastic.co/docs/reference/kibana/connectors-kibana/elastic-managed-llm). This connector is used in the AI Assistant, Attack Discovery, Automatic Import and Search Playground.
1717

18-
* You can use [ELSER](/explore-analyze/machine-learning/nlp/ml-nlp-elser.md) to perform semantic search as a service (ELSER on EIS). {applies_to}`stack: preview 9.1` {applies_to}`serverless: preview`
18+
* You can use [ELSER](/explore-analyze/machine-learning/nlp/ml-nlp-elser.md) to perform semantic search as a service (ELSER on EIS). {applies_to}`stack: preview 9.1, ga 9.2` {applies_to}`serverless: ga`
1919

2020
## Region and hosting [eis-regions]
2121

@@ -27,25 +27,20 @@ ELSER requests are managed by Elastic's own EIS infrastructure.
2727
## ELSER via Elastic {{infer-cap}} Service (ELSER on EIS) [elser-on-eis]
2828

2929
```{applies_to}
30-
stack: preview 9.1
31-
serverless: preview
30+
stack: preview 9.1, ga 9.2
31+
serverless: ga
3232
```
3333

34-
ELSER on EIS enables you to use the ELSER model on GPUs, without having to manage your own ML nodes. We expect better performance for throughput and latency than ML nodes, and will continue to benchmark, remove limitations and address concerns as we move towards General Availability.
34+
ELSER on EIS enables you to use the ELSER model on GPUs, without having to manage your own ML nodes. We expect better performance for throughput and latency than ML nodes, and will continue to benchmark, remove limitations and address concerns.
3535

36-
### Limitations
37-
38-
While we do encourage experimentation, we do not recommend implementing production use cases on top of this feature while it is in Technical Preview.
36+
### Pricing
3937

40-
#### Access
38+
ELSER on EIS usage is billed separately from your other Elastic deployment resources.
39+
For details about request-based pricing and billing dimensions, refer to the [ELSER on GPU item on the pricing page](https://www.elastic.co/pricing/serverless-search).
4140

42-
This feature is being gradually rolled out to Serverless and Cloud Hosted customers.
43-
It may not be available to all users at launch.
44-
45-
#### Uptime
41+
### Limitations
4642

47-
There are no uptime guarantees during the Technical Preview.
48-
While Elastic will address issues promptly, the feature may be unavailable for extended periods.
43+
Elastic is continuously working to remove these constraints and further improve performance and scalability.
4944

5045
#### Throughput and latency
5146

@@ -58,6 +53,6 @@ Performance may vary during the Technical Preview.
5853
Batches are limited to a maximum of 16 documents.
5954
This is particularly relevant when using the [_bulk API](https://www.elastic.co/docs/api/doc/elasticsearch/v9/operation/operation-bulk) for data ingestion.
6055

61-
#### Rate Limits
56+
#### Rate Limits
6257

6358
Rate limit for search and ingest is currently at 2000 requests per minute.

0 commit comments

Comments
 (0)