Skip to content

Commit c46505e

Browse files
authored
Merge branch 'main' into 2025/10/27/unsafely_incompatible_with_s3_conditional_writes
2 parents e8d5ada + 02bea7f commit c46505e

File tree

24 files changed

+357
-114
lines changed

24 files changed

+357
-114
lines changed

deploy-manage/deploy/elastic-cloud/elastic-cloud-hosted-planning.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ We recommend that you use at least two availability zones for production and thr
2828
Increasing the number of zones should not be used to add more resources. The concept of zones is meant for High Availability (2 zones) and Fault Tolerance (3 zones), but neither will work if the cluster relies on the resources from those zones to be operational.
2929
:::
3030

31-
With multiple {{es}} nodes in multiple availability zones you have the recommended hardware. The next step is to ensure proper index replication. Each index, with the exception of searchable snapshot indices, should have one or more replicas. Use the [index settings API](https://www.elastic.co/docs/api/doc/elasticsearch/v8/operation/operation-indices-get-settings-1) to find any indices without replicas:
31+
With multiple {{es}} nodes in multiple availability zones you have the recommended hardware. The next step is to ensure proper index replication. Each index, with the exception of searchable snapshot indices, should have one or more replicas. Use the [index settings API]({{es-apis}}operation/operation-indices-get-settings) to find any indices without replicas:
3232

3333
```sh
3434
GET _all/_settings/index.number_of_replicas

explore-analyze/elastic-inference/eis.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ You can now use `semantic_text` with the new ELSER endpoint on EIS. To learn how
4949
#### Batch size
5050

5151
Batches are limited to a maximum of 16 documents.
52-
This is particularly relevant when using the [_bulk API](https://www.elastic.co/docs/api/doc/elasticsearch/v9/operation/operation-bulk) for data ingestion.
52+
This is particularly relevant when using the [_bulk API]({{es-apis}}operation/operation-bulk) for data ingestion.
5353

5454
## Pricing
5555

explore-analyze/elastic-inference/inference-api.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,11 +11,11 @@ products:
1111

1212
# {{infer-cap}} integrations
1313

14-
{{es}} provides a machine learning [{{infer}} API](https://www.elastic.co/docs/api/doc/elasticsearch/v9/group/endpoint-inference) to create and manage {{infer}} endpoints that integrate with services such as {{es}} (for built-in NLP models like [ELSER](/explore-analyze/machine-learning/nlp/ml-nlp-elser.md) and [E5](/explore-analyze/machine-learning/nlp/ml-nlp-e5.md)), as well as popular third-party services like Amazon Bedrock, Anthropic, Azure AI Studio, Cohere, Google AI, Mistral, OpenAI, Hugging Face, and more.
14+
{{es}} provides a machine learning [{{infer}} API]({{es-apis}}group/endpoint-inference) to create and manage {{infer}} endpoints that integrate with services such as {{es}} (for built-in NLP models like [ELSER](/explore-analyze/machine-learning/nlp/ml-nlp-elser.md) and [E5](/explore-analyze/machine-learning/nlp/ml-nlp-e5.md)), as well as popular third-party services like Amazon Bedrock, Anthropic, Azure AI Studio, Cohere, Google AI, Mistral, OpenAI, Hugging Face, and more.
1515

1616
You can use the default {{infer}} endpoints your deployment contains or create a new {{infer}} endpoint:
1717

18-
- using the [Create an inference endpoint API](https://www.elastic.co/docs/api/doc/elasticsearch/v9/operation/operation-inference-put)
18+
- using the [Create an inference endpoint API]({{es-apis}}operation/operation-inference-put)
1919
- through the [Inference endpoints UI](#add-inference-endpoints).
2020

2121
## Default {{infer}} endpoints [default-enpoints]

explore-analyze/machine-learning/nlp/ml-nlp-deploy-model.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ You can optimize your deplyoment for typical use cases, such as search and inges
2424
Each deployment will be fine-tuned automatically based on its specific purpose you choose.
2525

2626
::::{note}
27-
Since eland uses APIs to deploy the models, you cannot see the models in {{kib}} until the saved objects are synchronized. You can follow the prompts in {{kib}}, wait for automatic synchronization, or use the [sync {{ml}} saved objects API](https://www.elastic.co/docs/api/doc/kibana/v8/group/endpoint-ml).
27+
Since eland uses APIs to deploy the models, you cannot see the models in {{kib}} until the saved objects are synchronized. You can follow the prompts in {{kib}}, wait for automatic synchronization, or use the [sync {{ml}} saved objects API]({{kib-apis}}group/endpoint-ml).
2828
::::
2929

3030
You can define the resource usage level of the NLP model during model deployment. The resource usage levels behave differently depending on [adaptive resources](../../../deploy-manage/autoscaling/trained-model-autoscaling.md#enabling-autoscaling-through-apis-adaptive-allocations) being enabled or disabled. When adaptive resources are disabled but {{ml}} autoscaling is enabled, vCPU usage of Cloud deployments derived from the Cloud console and functions as follows:

explore-analyze/machine-learning/nlp/ml-nlp-ner-example.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ You need to provide an administrator username and its password and replace the `
4949

5050
Since the `--start` option is used at the end of the Eland import command, {{es}} deploys the model ready to use. If you have multiple models and want to select which model to deploy, you can use the **{{ml-app}} > Model Management** user interface in {{kib}} to manage the starting and stopping of models.
5151

52-
Go to the **{{ml-app}} > Trained Models** page and synchronize your trained models. A warning message is displayed at the top of the page that says *"ML job and trained model synchronization required"*. Follow the link to *"Synchronize your jobs and trained models."* Then click **Synchronize**. You can also wait for the automatic synchronization that occurs in every hour, or use the [sync {{ml}} objects API](https://www.elastic.co/docs/api/doc/kibana/v8/group/endpoint-ml).
52+
Go to the **{{ml-app}} > Trained Models** page and synchronize your trained models. A warning message is displayed at the top of the page that says *"ML job and trained model synchronization required"*. Follow the link to *"Synchronize your jobs and trained models."* Then click **Synchronize**. You can also wait for the automatic synchronization that occurs in every hour, or use the [sync {{ml}} objects API]({{kib-apis}}group/endpoint-ml).
5353

5454
## Test the NER model [ex-ner-test]
5555

explore-analyze/machine-learning/nlp/ml-nlp-text-emb-vector-search-example.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,7 @@ You need to provide an administrator username and password and replace the `$CLO
5353

5454
Since the `--start` option is used at the end of the Eland import command, {{es}} deploys the model ready to use. If you have multiple models and want to select which model to deploy, you can use the **{{ml-app}} > Model Management** user interface in {{kib}} to manage the starting and stopping of models.
5555

56-
Go to the **{{ml-app}} > Trained Models** page and synchronize your trained models. A warning message is displayed at the top of the page that says *"ML job and trained model synchronization required"*. Follow the link to *"Synchronize your jobs and trained models."* Then click **Synchronize**. You can also wait for the automatic synchronization that occurs in every hour, or use the [sync {{ml}} objects API](https://www.elastic.co/docs/api/doc/kibana/v8/group/endpoint-ml).
56+
Go to the **{{ml-app}} > Trained Models** page and synchronize your trained models. A warning message is displayed at the top of the page that says *"ML job and trained model synchronization required"*. Follow the link to *"Synchronize your jobs and trained models."* Then click **Synchronize**. You can also wait for the automatic synchronization that occurs in every hour, or use the [sync {{ml}} objects API]({{kib-apis}}group/endpoint-ml).
5757

5858
## Test the text embedding model [ex-text-emb-test]
5959

explore-analyze/manage-access-to-ai-assistant.md

Lines changed: 13 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -9,16 +9,14 @@ products:
99
- id: cloud-serverless
1010
---
1111

12-
# Manage access to AI Features
13-
14-
This page describes how to use the GenAI Settings page to control access to AI-powered features in your deployments in the following ways:
12+
# Manage access to AI features
1513

14+
The GenAI Settings page lets you control access to AI-powered features in the following ways:
1615

1716
- Manage which AI connectors are available in your environment.
1817
- Enable or disable AI Assistant and other AI-powered features in your environment.
1918
- {applies_to}`stack: ga 9.2` {applies_to}`serverless: unavailable` Specify in which Elastic solutions the `AI Assistant for Observability and Search` and the `AI Assistant for Security` appear.
2019

21-
2220
## Requirements
2321

2422
- To access the **GenAI Settings** page, you need the `Actions and connectors: all` or `Actions and connectors: read` [{{kib}} privilege](/deploy-manage/users-roles/cluster-or-deployment-auth/kibana-privileges.md).
@@ -37,10 +35,16 @@ To manage these settings, go to the **GenAI Settings** page by using the navigat
3735

3836
The **GenAI Settings** page has the following settings:
3937

40-
- **Default AI Connector**: Use this setting to specify which connector is selected by default. This affects all AI-powered features, not just AI Assistant.
41-
- **Disallow all other connectors**: Enable this setting to prevent connectors other than the default connector specified above from being used in your space. This affects all AI-powered features, not just AI Assistant.
42-
- **AI feature visibility**: This button opens the current Space's settings page. Here you can specify which features should appear in your environment, including AI-powered features.
43-
- **AI Assistant visibility**: This setting allows you to choose which AI Assistants are available to use and where. You can choose to only show the AI Assistants in their native solutions, in other {{kib}} pages (for example, Discover, Dashboards, and Stack Management), or select **Hide all assistants** to disable AI Assistant throughout {{kib}}.
38+
- **Default AI Connector**: Use this setting to specify which connector is selected by default when you access AI-powered features. Default setting: **No default connector**.
39+
- If **No default connector** is selected, AI-powered features will default to the connector that was most recently used in your environment.
40+
- If **Elastic Managed LLM** or a custom LLM connector is selected, AI-powered features will default to that connector regardless of which connector was used most recently in your environment.
41+
- **Disallow all other connectors**: When this setting is disabled, whenever you use an AI-powered feature you can select which connector should power it. Enable it to prevent connectors other than the default connector from being used in your space. Default setting: disabled.
42+
- **AI feature visibility**: This button opens the current Space's settings page, where you can specify which features are enabled in your environment, including AI-powered features.
43+
- **AI Assistant visibility**: This setting allows you to choose which AI Assistants are available to use and where. There are several options:
44+
- **Only in their solutions** (default): The Security AI Assistant appears in {{elastic-sec}}, and the {{obs-ai-assistant}} appears in {{es}} and {{observability}}.
45+
- **{{obs-ai-assistant}} in other apps**: The {{obs-ai-assistant}} appears throughout {{kib}} regardless of solution. The Security AI Assistant does not appear anywhere.
46+
- **Security AI Assistant in other apps**: The Security AI Assistant appears throughout {{kib}} regardless of solution. The {{obs-ai-assistant}} does not appear anywhere.
47+
- **Hide all assistants**: Disables AI Assistant throughout {{kib}}.
4448

4549
:::
4650

@@ -51,7 +55,7 @@ The **GenAI Settings** page has the following settings:
5155
The **GenAI Settings** page has the following settings:
5256

5357
- **Default AI Connector**: Click **Manage connectors** to open the **Connectors** page, where you can create or delete AI connectors. To update these settings, you need the `Actions and connectors: all` [{{kib}} privilege](/deploy-manage/users-roles/cluster-or-deployment-auth/kibana-privileges.md).
54-
- **AI feature visibility**: Click **Go to Permissions tab** to access the active {{kib}} space's settings page, where you can specify which features each custom user role has access to in your environment. This includes AI-powered features such as AI Assistant.
58+
- **AI feature visibility**: Click **Go to Permissions tab** to access the active {{kib}} space's settings page, where you can specify which features each [user role](/deploy-manage/users-roles/cluster-or-deployment-auth/kibana-privileges.md) has access to in your environment. This includes AI-powered features.
5559

5660
:::
5761

Lines changed: 45 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,45 @@
1+
Elastic {{observability}} provides unified observability across applications and infrastructure. It combines logs, metrics, application traces, user experience data, and more into a single, integrated platform.
2+
This consolidation allows for powerful, cross-referenced analysis, enabling teams to move from detecting issues to understanding their root causes quickly and efficiently.
3+
By leveraging the search and analytics capabilities of {{es}}, it offers a holistic view of system behavior.
4+
5+
Elastic {{observability}} embraces open standards like OpenTelemetry for flexible data collection, and offers scalable, cost-efficient data retention with tiered storage.
6+
7+
For a complete overview, refer to [](/solutions/observability/get-started/what-is-elastic-observability.md).
8+
9+
## Use cases [observability-use-cases]
10+
11+
Apply {{observability}} to various scenarios to improve operational awareness and system reliability.
12+
13+
:::{dropdown} Use cases
14+
:open:
15+
* **Log monitoring and analytics:** Centralize and analyze petabytes of log data from any source. This enables quick searching, ad-hoc queries with ES|QL, and visualization with prebuilt dashboards to diagnose issues.
16+
* **Application Performance Monitoring (APM):** Gain code-level visibility into application performance. By collecting and analyzing traces with native OTel support, teams can identify bottlenecks, track errors, and optimize the end-user experience.
17+
* **Infrastructure monitoring:** Monitor metrics from servers, virtual machines, containers, and serverless environments with over 400 out-of-the-box integrations, including OpenTelemetry. This provides deep insights into resource utilization and overall system health.
18+
* **AI-powered log analysis with Streams**: Ingest raw logs in any format directly to a single endpoint without the need for complex agent management or manual parsing pipelines. Streams leverages AI to automatically parse, structure, and analyze log data on the fly.
19+
* **Digital experience monitoring:**
20+
* **Real User Monitoring (RUM):** Capture and analyze data on how real users interact with web applications to improve perceived performance.
21+
* **Synthetic monitoring:** Proactively simulate user journeys and API calls to test application availability and functionality.
22+
* **Uptime monitoring:** Continuously check the status of services and applications to ensure they are available.
23+
* **Universal Profiling:** Gain visibility into system performance and identify expensive lines of code without application instrumentation, helping to increase CPU efficiency and reduce cloud spend.
24+
* **LLM Observability:** Gain deep insights into the performance, usage, and costs of Large Language Model (LLM) prompts and responses.
25+
* **Incident response and management:** Investigate operational incidents by correlating data from multiple sources, accelerating root cause analysis and resolution.
26+
:::
27+
28+
To start your {{observability}} journey, read the [**Get started**](/solutions/observability/get-started.md) guide, which presents all the essential steps, with links to valuable resources. You can also browse the {{observability}} [**Quickstart guides**](/solutions/observability/get-started/quickstarts.md).
29+
30+
## Core concepts [observability-concepts]
31+
32+
At the heart of Elastic {{observability}} are several key components that enable its capabilities.
33+
34+
:::{dropdown} Concepts
35+
:open:
36+
* The three pillars of {{observability}} are:
37+
38+
* [**Logs:**](/solutions/observability/logs.md) Timestamped records of events that provide detailed, contextual information.
39+
* [**Metrics:**](/solutions/observability/infra-and-hosts/analyze-infrastructure-host-metrics.md) Numerical measurements of system performance and health over time.
40+
* [**Traces:**](/solutions/observability/apm/traces.md) Representations of end-to-end journeys of requests as they travel through distributed systems.
41+
* [**OpenTelemetry:**](/solutions/observability/apm/opentelemetry/index.md) {{Observability}} offers first-class, production-grade support for OpenTelemetry. This allows organizations to use vendor-neutral instrumentation and stream native OTel data without proprietary agents, leveraging the Elastic Distribution of OpenTelemetry (EDOT).
42+
* [**AIOps and AI Assistant:**](/solutions/observability/observability-ai-assistant.md) Leverages predictive analytics and an LLM-powered AI Assistant to reduce the time required to detect, investigate, and resolve incidents. This includes zero-config anomaly detection, pattern analysis, and the ability to surface correlations and root causes.
43+
* **[Alerting](/solutions/observability/incident-management/alerting.md) and [Cases](/solutions/observability/incident-management/cases.md):** Allows you to create rules to detect complex conditions and perform actions. Cases allows teams to stay aware of potential issues and track investigation details, assign tasks, and collaborate on resolutions.
44+
* [**Service Level Objectives (SLOs):**](/solutions/observability/incident-management/service-level-objectives-slos.md) A framework for defining and monitoring the reliability of a service. Elastic {{observability}} allows for creating and tracking SLOs to ensure that performance targets are being met.
45+
:::

0 commit comments

Comments
 (0)