You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: explore-analyze/elastic-inference/eis.md
+31-8Lines changed: 31 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,20 +11,30 @@ The Elastic {{infer-cap}} Service (EIS) enables you to leverage AI-powered searc
11
11
With EIS, you don't need to manage the infrastructure and resources required for large language models (LLMs) by adding, configuring, and scaling {{ml}} nodes.
12
12
Instead, you can use {{ml}} models in high-throughput, low-latency scenarios independently of your {{es}} infrastructure.
13
13
14
-
Currently, you can perform chat completion tasks through EIS using the {{infer}} API.
14
+
% TO DO: Link to the EIS inference endpoint reference docs when it's added to the OpenAPI spec. (Comming soon) %
15
+
16
+
## Available task types
17
+
18
+
EIS offers the following {{infer}} task types to perform:
19
+
20
+
* Chat completion
21
+
22
+
## How to use EIS [using-eis]
23
+
24
+
Your Elastic deployment comes with default endpoints for EIS that you can use performing {{infer}} tasks.
25
+
You can either do it by calling the {{infer}} API or using the default `Elastic LLM` model in the AI Assistant, Attack Discovery UI, and Search Playground.
15
26
16
27
% TO DO: Link to the EIS inference endpoint reference docs when it's added to the OpenAPI spec. (Comming soon) %
17
28
18
29
## Default EIS endpoints [default-eis-inference-endpoints]
19
30
20
31
Your {{es}} deployment includes a preconfigured EIS endpoint, making it easier to use chat completion via the {{infer}} API:
21
32
22
-
*`rainbow-sprinkles-elastic`: uses Anthropic's Claude Sonnet 3.5 model for chat completion {{infer}} tasks.
33
+
*`rainbow-sprinkles-elastic`
23
34
24
35
::::{note}
25
36
26
37
* The model appears as `Elastic LLM` in the AI Assistant, Attack Discovery UI, preconfigured connectors list, and the Search Playground.
27
-
* To fine-tune prompts sent to `rainbow-sprinkles-elastic`, optimize them for Claude Sonnet 3.5.
28
38
29
39
::::
30
40
@@ -39,10 +49,6 @@ EIS runs on AWS in the following regions:
39
49
40
50
For more details on AWS regions, refer to the [AWS Global Infrastructure](https://aws.amazon.com/about-aws/global-infrastructure/regions_az/) and the [supported cross-region {{infer}} profiles](https://docs.aws.amazon.com/bedrock/latest/userguide/inference-profiles-support.html) documentation.
41
51
42
-
## LLM hosts [llm-hosts]
43
-
44
-
The LLM used with EIS is hosted by [Amazon Bedrock](https://aws.amazon.com/bedrock/).
45
-
46
52
## Examples
47
53
48
54
The following example demonstrates how to perform a `chat_completion` task through EIS by using the `.rainbow-sprinkles-elastic` default {{infer}} endpoint.
@@ -58,6 +64,23 @@ POST /_inference/chat_completion/.rainbow-sprinkles-elastic/_stream
0 commit comments