You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: deploy-manage/production-guidance.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -64,7 +64,7 @@ Other sections of the documentation provide important guidance for running {{sta
64
64
65
65
### Plan your data structure, availability, and formatting [ec_plan_your_data_structure_availability_and_formatting]
66
66
67
-
* Build a [data architecture](/manage-data/lifecycle/data-tiers.md) that best fits your needs. Based on your own access and retention policies, you can add warm, cold, frozen data tiers, and automated deletion of old data.
67
+
* Build a [data architecture](/manage-data/lifecycle/data-tiers.md) that best fits your needs. Based on your own access and retention policies, you can add warm, cold, and frozen data tiers, and automate deletion of old data.
68
68
* Make your data [highly available](/deploy-manage/tools.md) for production environments or otherwise critical data stores, and take regular [backup snapshots](./tools/snapshot-and-restore.md), or consider [](./tools/cross-cluster-replication.md) to replicate indices across clusters.
69
69
* Normalize event data to better analyze, visualize, and correlate your events by adopting the [Elastic Common Schema](asciidocalypse://docs/ecs/docs/reference/ecs-getting-started.md) (ECS). Elastic integrations use ECS out-of-the-box. If you are writing your own integrations, ECS is recommended.
Copy file name to clipboardExpand all lines: deploy-manage/production-guidance/availability-and-resilience.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -14,11 +14,11 @@ applies_to:
14
14
15
15
Distributed systems like {{es}} are designed to keep working even if some of their components have failed. As long as there are enough well-connected nodes to take over their responsibilities, an {{es}} cluster can continue operating normally if some of its nodes are unavailable or disconnected.
16
16
17
-
{{es}} implements high availability at three key levels:
17
+
{{es}} implements high availability (HA) at three key levels:
18
18
19
-
* Node level – Running multiple nodes within the cluster to avoid single points of failure and maintain operational stability.
20
-
* Cluster level – Ensuring redundancy by distributing nodes across availability zones to prevent failures from affecting the entire cluster.
21
-
* Index level – Configuring shard replication to protect against data loss and improve search performance by distributing queries across multiple copies.
19
+
***Node level**: Running multiple nodes within the cluster to avoid single points of failure and maintain operational stability.
20
+
***Cluster level**: Ensuring redundancy by distributing nodes across availability zones to prevent failures from affecting the entire cluster.
21
+
***Index level**: Configuring shard replication to protect against data loss and improve search performance by distributing queries across multiple copies.
22
22
23
23
Each of these HA mechanisms contributes to {{es}}’s resilience and scalability. The appropriate strategy depends on factors such as data criticality, query patterns, and infrastructure constraints. It is up to you to determine the level of resiliency and high availability that best fits your use case. This sections provides detailed guidance on designing a production-ready {{es}} deployment that balances availability, performance, and scalability.
Copy file name to clipboardExpand all lines: deploy-manage/production-guidance/availability-and-resilience/resilience-in-ech.md
+1-3Lines changed: 1 addition & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -29,17 +29,15 @@ With multiple {{es}} nodes in multiple availability zones you have the recommend
29
29
GET _all/_settings/index.number_of_replicas
30
30
```
31
31
32
-
Moreover, a high availability (HA) cluster requires at least three master-eligible nodes. For clusters that have fewer than six {{es}} nodes, any data node in the hot tier will also be a master-eligible node. You can achieve this by having hot nodes (serving as both data and master-eligible nodes) in three availability zones, or by having data nodes in two zones and a tiebreaker (will be automatically added if you choose two zones). For clusters that have six {{es}} nodes and beyond, dedicated master-eligible nodes are introduced. When your cluster grows, consider separating dedicated master-eligible nodes from dedicated data nodes. We recommend using at least 4GB RAM for dedicated master nodes.
32
+
A high availability (HA) cluster requires at least three master-eligible nodes. For clusters that have fewer than six {{es}} nodes, any data node in the hot tier will also be a master-eligible node. You can achieve this by having hot nodes (serving as both data and master-eligible nodes) in three availability zones, or by having data nodes in two zones and a tiebreaker (will be automatically added if you choose two zones). For clusters that have six {{es}} nodes and beyond, dedicated master-eligible nodes are introduced. When your cluster grows, consider separating dedicated master-eligible nodes from dedicated data nodes. We recommend using at least 4GB RAM for dedicated master nodes.
33
33
34
34
The data in your {{es}} clusters is also backed up every 30 minutes, 4 hours, or 24 hours, depending on which snapshot interval you choose. These regular intervals provide an extra level of redundancy. We do support [snapshot and restore](../../../deploy-manage/tools/snapshot-and-restore.md), regardless of whether you use one, two, or three availability zones. However, with only a single availability zone and in the event of an outage, it might take a while for your cluster come back online. Using a single availability zone also leaves your cluster exposed to the risk of data loss, if the backups you need are not useable (failed or partial snapshots missing the indices to restore) or no longer available by the time that you realize that you might need the data (snapshots have a retention policy).
35
35
36
36
## Important considerations
37
37
38
-
::::{warning}
39
38
* Clusters that use only one availability zone are not highly available and are at risk of data loss. To safeguard against data loss, you must use at least two availability zones.
40
39
* Indices with no replica, except for searchable snapshot indices, are not highly available. You should use replicas to mitigate against possible data loss.
41
40
* Clusters that only have one master node are not highly available and are at risk of data loss. You must have three master-eligible nodes.
Copy file name to clipboardExpand all lines: deploy-manage/production-guidance/kibana-in-production-environments.md
+6-4Lines changed: 6 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -22,7 +22,7 @@ Additionally, the task manager enables distributed coordination across multiple
22
22
23
23
::::{important}
24
24
* {{kib}} does not support rolling [upgrades](/deploy-manage/upgrade/deployment-or-cluster/kibana.md), and deploying mixed versions of {{kib}} can result in data loss or upgrade failures. Please shut down all instances of {{kib}} before performing an upgrade, and ensure all running {{kib}} instances have matching versions.
25
-
* While {{kib}} isn’t terribly resource intensive, we still recommend running {{kib}} separate from your {{es}} data or master nodes.
25
+
* While {{kib}} isn’t resource intensive, we still recommend running {{kib}} separate from your {{es}} data or master nodes.
26
26
::::
27
27
28
28
## Section overview
@@ -54,7 +54,7 @@ deployment:
54
54
self: all
55
55
```
56
56
57
-
To run multiple {{kib}} instances connected to the same {{es}} cluster, you need to adjust the configuration. See [Configuring {{kib}}](../deploy/self-managed/configure-kibana.md) for details on each setting.
57
+
To run multiple {{kib}} instances connected to the same {{es}} cluster, you need to adjust the configuration. See the [{{kib}} configuration reference](kibana://reference/configuration-reference) for details on each setting.
58
58
59
59
::::{note}
60
60
When adding multiple {{kib}} instances to the same deployment in {{ech}}, {{ece}}, or {{eck}}, the orchestrator applies the necessary configuration, requiring no manual setup.
@@ -132,10 +132,12 @@ elasticsearch.hosts:
132
132
- http://elasticsearch2:9200
133
133
```
134
134
135
-
Related configurations include `elasticsearch.sniffInterval`, `elasticsearch.sniffOnStart`, and `elasticsearch.sniffOnConnectionFault`. These can be used to automatically update the list of hosts as a cluster is resized. Parameters can be found on the [settings page](kibana://reference/configuration-reference/general-settings.md).
135
+
Related configurations include `elasticsearch.sniffInterval`, `elasticsearch.sniffOnStart`, and `elasticsearch.sniffOnConnectionFault`. These can be used to automatically update the list of hosts as a cluster is resized. Parameters can be found in the [{{kib}} configuration reference](kibana://reference/configuration-reference/general-settings.md).
136
136
137
137
::::{note}
138
-
The previous configuration can be useful when there is no load balancer or reverse proxy in front of {{es}}. If a load balancer is in place to distribute traffic among {{es}} instances, Kibana should be configured to connect to it instead. In orchestrated deployments, {{kib}} is automatically configured to connect to {{es}} through a load-balanced service.
138
+
The previous configuration can be useful when there is no load balancer or reverse proxy in front of {{es}}. If a load balancer is in place to distribute traffic among {{es}} instances, Kibana should be configured to connect to it instead.
139
+
140
+
In [orchestrated deployments](/deploy-manage/deploy.md#about-orchestration), {{kib}} is automatically configured to connect to {{es}} through a load-balanced service.
Copy file name to clipboardExpand all lines: deploy-manage/production-guidance/kibana-reporting-production-considerations.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -10,7 +10,7 @@ applies_to:
10
10
# Kibana reporting production considerations [reporting-production-considerations]
11
11
12
12
13
-
To generate reports, {{kib}} uses the Chromium web browser, which runs on the server in headless mode. Chromium is an open-source project not related to Elastic, and is embedded into {{kib}}. Chromium may require [additional OS dependencies](/deploy-manage/kibana-reporting-configuration.md#install-reporting-packages) to run properly.
13
+
To generate [reports](/explore-analyze/report-and-share.md), {{kib}} uses the Chromium web browser, which runs on the server in headless mode. Chromium is an open-source project not related to Elastic, and is embedded into {{kib}}. Chromium may require [additional OS dependencies](/deploy-manage/deploy/kibana-reporting-configuration.md#install-reporting-packages) to run properly.
14
14
15
15
::::{note}
16
16
Chromium is not embedded into {{kib}} for the Darwin (Mac OS) architecture. When running on Darwin, Reporting will download Chromium into the proper area of the {{kib}} installation path the first time the server starts. To separately download and install the browser, see [Manually install the Chromium browser for Darwin](../../explore-analyze/report-and-share/reporting-troubleshooting-pdf.md#reporting-manual-chromium-install).
Copy file name to clipboardExpand all lines: deploy-manage/reference-architectures/hotfrozen-high-availability.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -47,7 +47,7 @@ We use an Availability Zone (AZ) concept in the architecture above. When running
47
47
::::
48
48
49
49
50
-
The diagram illustrates an {{es}} cluster deployed across 3 availability zones (AZ). For production we recommend a minimum of 2 availability zones and 3 availability zones for mission critical applications. See [Plan for production](/deploy-manage/production-guidance.md) for more details. A cluster that is running in {{ecloud}} that has data nodes in only two AZs will create a third master-eligible node in a third AZ. High availability cannot be achieved without three zones for any distributed computing technology.
50
+
The diagram illustrates an {{es}} cluster deployed across 3 availability zones (AZ). For production we recommend a minimum of 2 availability zones and 3 availability zones for mission critical applications. See [](/deploy-manage/production-guidance/availability-and-resilience/resilience-in-ech.md) for more details. A cluster that is running in {{ecloud}} that has data nodes in only two AZs will create a third master-eligible node in a third AZ. High availability cannot be achieved without three zones for any distributed computing technology.
51
51
52
52
The number of data nodes shown for each tier (hot and frozen) is illustrative and would be scaled up depending on ingest volume and retention period. Hot nodes contain both primary and replica shards. By default, primary and replica shards are always guaranteed to be in different availability zones in {{ech}}, but when self-deploying [shard allocation awareness](../distributed-architecture/shard-allocation-relocation-recovery/shard-allocation-awareness.md) would need to be configured. Frozen nodes act as a large high-speed cache and retrieve data from the snapshot store as needed.
0 commit comments