Skip to content

Commit 6937860

Browse files
szabostevemaxjakob
andauthored
Apply suggestions from code review
Co-authored-by: Max Jakob <[email protected]>
1 parent 9d02479 commit 6937860

File tree

1 file changed

+10
-12
lines changed
  • explore-analyze/elastic-inference

1 file changed

+10
-12
lines changed

explore-analyze/elastic-inference/eis.md

Lines changed: 10 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -8,46 +8,45 @@ navigation_title: Elastic Inference Service (EIS)
88
# Elastic {{infer-cap}} Service [elastic-inference-service-eis]
99

1010
The Elastic {{infer-cap}} Service (EIS) enables you to leverage AI-powered search as a service without deploying a model in your cluster.
11-
With EIS, you don't need to manage the infrastructure and resources required for large language models (LLMs) by adding, configuring, and scaling {{ml}} nodes.
12-
Instead, you can use {{ml}} models in high-throughput, low-latency scenarios independently of your {{es}} infrastructure.
11+
With EIS, you don't need to manage the infrastructure and resources required for {{ml}} {{infer}} by adding, configuring, and scaling {{ml}} nodes.
12+
Instead, you can use {{ml}} models for ingest, search and chat independently of your {{es}} infrastructure.
1313

1414
% TO DO: Link to the EIS inference endpoint reference docs when it's added to the OpenAPI spec. (Comming soon) %
1515

1616
## Available task types
1717

1818
EIS offers the following {{infer}} task types to perform:
1919

20-
* Chat completion
20+
* `chat_completion`
2121

22-
## How to use EIS [using-eis]
22+
## AI features powered by EIS [ai-features-powered-by-eis]
2323

24-
Your Elastic deployment comes with default endpoints for EIS that you can use performing {{infer}} tasks.
25-
You can either do it by calling the {{infer}} API or using the default `Elastic LLM` model in the AI Assistant, Attack Discovery UI, and Search Playground.
24+
Your Elastic deployment or project comes with a default `Elastic LLM` connector. This connector is used in the AI Assistant, Attack Discovery, Automatic Import and Search Playground.
2625

2726
% TO DO: Link to the EIS inference endpoint reference docs when it's added to the OpenAPI spec. (Comming soon) %
2827

2928
## Default EIS endpoints [default-eis-inference-endpoints]
3029

3130
Your {{es}} deployment includes a preconfigured EIS endpoint, making it easier to use chat completion via the {{infer}} API:
3231

33-
* `rainbow-sprinkles-elastic`
32+
* `.rainbow-sprinkles-elastic`
3433

3534
::::{note}
3635

37-
* The model appears as `Elastic LLM` in the AI Assistant, Attack Discovery UI, preconfigured connectors list, and the Search Playground.
36+
* This endpoint is used by the `Elastic LLM` AI connector, which in turn powers the AI Assistant, Attack Discovery, Automatic Import, and the Search Playground.
3837

3938
::::
4039

4140
% TO DO: Link to the AI assistant documentation in the different solutions and possibly connector docs. %
4241

4342
## Regions [eis-regions]
4443

45-
EIS runs on AWS in the following regions:
44+
All EIS requests are handled by one of these AWS regions:
4645

4746
* `us-east-1`
4847
* `us-west-2`
4948

50-
For more details on AWS regions, refer to the [AWS Global Infrastructure](https://aws.amazon.com/about-aws/global-infrastructure/regions_az/) and the [supported cross-region {{infer}} profiles](https://docs.aws.amazon.com/bedrock/latest/userguide/inference-profiles-support.html) documentation.
49+
For more details on AWS regions, refer to the [AWS Global Infrastructure](https://aws.amazon.com/about-aws/global-infrastructure/regions_az/).
5150

5251
## Examples
5352

@@ -67,15 +66,14 @@ POST /_inference/chat_completion/.rainbow-sprinkles-elastic/_stream
6766
}
6867
```
6968

70-
The request returns the following response:
69+
The request returns the following response as a stream:
7170

7271
```json
7372
(...)
7473
{
7574
"role" : "assistant",
7675
"content": "Yes",
7776
"model" : "rainbow-sprinkles",
78-
"object" : "chat.completion.chunk",
7977
"usage" : {
8078
"completion_tokens" : 4,
8179
"prompt_tokens" : 13,

0 commit comments

Comments
 (0)