You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: explore-analyze/elastic-inference/eis.md
+10-12Lines changed: 10 additions & 12 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,46 +8,45 @@ navigation_title: Elastic Inference Service (EIS)
8
8
# Elastic {{infer-cap}} Service [elastic-inference-service-eis]
9
9
10
10
The Elastic {{infer-cap}} Service (EIS) enables you to leverage AI-powered search as a service without deploying a model in your cluster.
11
-
With EIS, you don't need to manage the infrastructure and resources required for large language models (LLMs) by adding, configuring, and scaling {{ml}} nodes.
12
-
Instead, you can use {{ml}} models in high-throughput, low-latency scenarios independently of your {{es}} infrastructure.
11
+
With EIS, you don't need to manage the infrastructure and resources required for {{ml}} {{infer}} by adding, configuring, and scaling {{ml}} nodes.
12
+
Instead, you can use {{ml}} models for ingest, search and chat independently of your {{es}} infrastructure.
13
13
14
14
% TO DO: Link to the EIS inference endpoint reference docs when it's added to the OpenAPI spec. (Comming soon) %
15
15
16
16
## Available task types
17
17
18
18
EIS offers the following {{infer}} task types to perform:
19
19
20
-
*Chat completion
20
+
*`chat_completion`
21
21
22
-
## How to use EIS [using-eis]
22
+
## AI features powered by EIS [ai-features-powered-by-eis]
23
23
24
-
Your Elastic deployment comes with default endpoints for EIS that you can use performing {{infer}} tasks.
25
-
You can either do it by calling the {{infer}} API or using the default `Elastic LLM` model in the AI Assistant, Attack Discovery UI, and Search Playground.
24
+
Your Elastic deployment or project comes with a default `Elastic LLM` connector. This connector is used in the AI Assistant, Attack Discovery, Automatic Import and Search Playground.
26
25
27
26
% TO DO: Link to the EIS inference endpoint reference docs when it's added to the OpenAPI spec. (Comming soon) %
28
27
29
28
## Default EIS endpoints [default-eis-inference-endpoints]
30
29
31
30
Your {{es}} deployment includes a preconfigured EIS endpoint, making it easier to use chat completion via the {{infer}} API:
32
31
33
-
*`rainbow-sprinkles-elastic`
32
+
*`.rainbow-sprinkles-elastic`
34
33
35
34
::::{note}
36
35
37
-
*The model appears as `Elastic LLM` in the AI Assistant, Attack Discovery UI, preconfigured connectors list, and the Search Playground.
36
+
*This endpoint is used by the `Elastic LLM`AI connector, which in turn powers the AI Assistant, Attack Discovery, Automatic Import, and the Search Playground.
38
37
39
38
::::
40
39
41
40
% TO DO: Link to the AI assistant documentation in the different solutions and possibly connector docs. %
42
41
43
42
## Regions [eis-regions]
44
43
45
-
EIS runs on AWS in the following regions:
44
+
All EIS requests are handled by one of these AWS regions:
46
45
47
46
*`us-east-1`
48
47
*`us-west-2`
49
48
50
-
For more details on AWS regions, refer to the [AWS Global Infrastructure](https://aws.amazon.com/about-aws/global-infrastructure/regions_az/) and the [supported cross-region {{infer}} profiles](https://docs.aws.amazon.com/bedrock/latest/userguide/inference-profiles-support.html) documentation.
49
+
For more details on AWS regions, refer to the [AWS Global Infrastructure](https://aws.amazon.com/about-aws/global-infrastructure/regions_az/).
51
50
52
51
## Examples
53
52
@@ -67,15 +66,14 @@ POST /_inference/chat_completion/.rainbow-sprinkles-elastic/_stream
67
66
}
68
67
```
69
68
70
-
The request returns the following response:
69
+
The request returns the following response as a stream:
0 commit comments