Skip to content

Commit ae3db60

Browse files
authored
[8.x] [DOCS] Concept cleanup 2 - ES settings (#119373) (#119642)
1 parent 8a14c14 commit ae3db60

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

51 files changed

+959
-856
lines changed

docs/plugins/discovery-ec2.asciidoc

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -241,7 +241,7 @@ The `discovery-ec2` plugin can automatically set the `aws_availability_zone`
241241
node attribute to the availability zone of each node. This node attribute
242242
allows you to ensure that each shard has copies allocated redundantly across
243243
multiple availability zones by using the
244-
{ref}/modules-cluster.html#shard-allocation-awareness[Allocation Awareness]
244+
{ref}/shard-allocation-awareness.html#[Allocation Awareness]
245245
feature.
246246

247247
In order to enable the automatic definition of the `aws_availability_zone`
@@ -333,7 +333,7 @@ labelled as `Moderate` or `Low`.
333333

334334
* It is a good idea to distribute your nodes across multiple
335335
https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/using-regions-availability-zones.html[availability
336-
zones] and use {ref}/modules-cluster.html#shard-allocation-awareness[shard
336+
zones] and use {ref}/shard-allocation-awareness.html[shard
337337
allocation awareness] to ensure that each shard has copies in more than one
338338
availability zone.
339339

docs/reference/cat/nodeattrs.asciidoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ console. They are _not_ intended for use by applications. For application
1717
consumption, use the <<cluster-nodes-info,nodes info API>>.
1818
====
1919

20-
Returns information about custom node attributes.
20+
Returns information about <<custom-node-attributes,custom node attributes>>.
2121

2222
[[cat-nodeattrs-api-request]]
2323
==== {api-request-title}

docs/reference/cluster.asciidoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ one of the following:
3535
master-eligible nodes, all data nodes, all ingest nodes, all voting-only
3636
nodes, all machine learning nodes, and all coordinating-only nodes.
3737
* a pair of patterns, using `*` wildcards, of the form `attrname:attrvalue`,
38-
which adds to the subset all nodes with a custom node attribute whose name
38+
which adds to the subset all nodes with a <<custom-node-attributes,custom node attribute>> whose name
3939
and value match the respective patterns. Custom node attributes are
4040
configured by setting properties in the configuration file of the form
4141
`node.attr.attrname: attrvalue`.

docs/reference/commands/node-tool.asciidoc

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -23,8 +23,8 @@ bin/elasticsearch-node repurpose|unsafe-bootstrap|detach-cluster|override-versio
2323
This tool has a number of modes:
2424

2525
* `elasticsearch-node repurpose` can be used to delete unwanted data from a
26-
node if it used to be a <<data-node,data node>> or a
27-
<<master-node,master-eligible node>> but has been repurposed not to have one
26+
node if it used to be a <<data-node-role,data node>> or a
27+
<<master-node-role,master-eligible node>> but has been repurposed not to have one
2828
or other of these roles.
2929

3030
* `elasticsearch-node remove-settings` can be used to remove persistent settings

docs/reference/data-management.asciidoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ Data older than this period can be deleted by {es} at a later time.
4343

4444
**Elastic Curator** is a tool that allows you to manage your indices and snapshots using user-defined filters and predefined actions. If ILM provides the functionality to manage your index lifecycle, and you have at least a Basic license, consider using ILM in place of Curator. Many stack components make use of ILM by default. {curator-ref-current}/ilm.html[Learn more].
4545

46-
NOTE: <<xpack-rollup,Data rollup>> is a deprecated Elasticsearch feature that allows you to manage the amount of data that is stored in your cluster, similar to the downsampling functionality of {ilm-init} and data stream lifecycle. This feature should not be used for new deployments.
46+
NOTE: <<xpack-rollup,Data rollup>> is a deprecated {es} feature that allows you to manage the amount of data that is stored in your cluster, similar to the downsampling functionality of {ilm-init} and data stream lifecycle. This feature should not be used for new deployments.
4747

4848
[TIP]
4949
====

docs/reference/data-management/migrate-index-allocation-filters.asciidoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
[[migrate-index-allocation-filters]]
33
== Migrate index allocation filters to node roles
44

5-
If you currently use custom node attributes and
5+
If you currently use <<custom-node-attributes,custom node attributes>> and
66
<<shard-allocation-filtering, attribute-based allocation filters>> to
77
move indices through <<data-tiers, data tiers>> in a
88
https://www.elastic.co/blog/implementing-hot-warm-cold-in-elasticsearch-with-index-lifecycle-management[hot-warm-cold architecture],

docs/reference/data-store-architecture.asciidoc

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,10 +9,16 @@ from any node.
99
The topics in this section provides information about the architecture of {es} and how it stores and retrieves data:
1010

1111
* <<nodes-shards,Nodes and shards>>: Learn about the basic building blocks of an {es} cluster, including nodes, shards, primaries, and replicas.
12+
* <<node-roles-overview,Node roles>>: Learn about the different roles that nodes can have in an {es} cluster.
1213
* <<docs-replication,Reading and writing documents>>: Learn how {es} replicates read and write operations across shards and shard copies.
1314
* <<shard-allocation-relocation-recovery,Shard allocation, relocation, and recovery>>: Learn how {es} allocates and balances shards across nodes.
15+
** <<shard-allocation-awareness,Shard allocation awareness>>: Learn how to use custom node attributes to distribute shards across different racks or availability zones.
16+
* <<shard-request-cache,Shard request cache>>: Learn how {es} caches search requests to improve performance.
1417
--
1518
1619
include::nodes-shards.asciidoc[]
20+
include::node-roles.asciidoc[]
1721
include::docs/data-replication.asciidoc[leveloffset=-1]
18-
include::modules/shard-ops.asciidoc[]
22+
include::modules/shard-ops.asciidoc[]
23+
include::modules/cluster/allocation_awareness.asciidoc[leveloffset=+1]
24+
include::shard-request-cache.asciidoc[leveloffset=-1]

docs/reference/data-streams/downsampling.asciidoc

Lines changed: 39 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -72,6 +72,45 @@ the granularity of `cold` archival data to monthly or less.
7272
.Downsampled metrics series
7373
image::images/data-streams/time-series-downsampled.png[align="center"]
7474

75+
[discrete]
76+
[[downsample-api-process]]
77+
==== The downsampling process
78+
79+
The downsampling operation traverses the source TSDS index and performs the
80+
following steps:
81+
82+
. Creates a new document for each value of the `_tsid` field and each
83+
`@timestamp` value, rounded to the `fixed_interval` defined in the downsample
84+
configuration.
85+
. For each new document, copies all <<time-series-dimension,time
86+
series dimensions>> from the source index to the target index. Dimensions in a
87+
TSDS are constant, so this is done only once per bucket.
88+
. For each <<time-series-metric,time series metric>> field, computes aggregations
89+
for all documents in the bucket. Depending on the metric type of each metric
90+
field a different set of pre-aggregated results is stored:
91+
92+
** `gauge`: The `min`, `max`, `sum`, and `value_count` are stored; `value_count`
93+
is stored as type `aggregate_metric_double`.
94+
** `counter`: The `last_value` is stored.
95+
. For all other fields, the most recent value is copied to the target index.
96+
97+
[discrete]
98+
[[downsample-api-mappings]]
99+
==== Source and target index field mappings
100+
101+
Fields in the target, downsampled index are created based on fields in the
102+
original source index, as follows:
103+
104+
. All fields mapped with the `time-series-dimension` parameter are created in
105+
the target downsample index with the same mapping as in the source index.
106+
. All fields mapped with the `time_series_metric` parameter are created
107+
in the target downsample index with the same mapping as in the source
108+
index. An exception is that for fields mapped as `time_series_metric: gauge`
109+
the field type is changed to `aggregate_metric_double`.
110+
. All other fields that are neither dimensions nor metrics (that is, label
111+
fields), are created in the target downsample index with the same mapping
112+
that they had in the source index.
113+
75114
[discrete]
76115
[[running-downsampling]]
77116
=== Running downsampling on time series data

docs/reference/datatiers.asciidoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -190,7 +190,7 @@ tier].
190190
[[configure-data-tiers-on-premise]]
191191
==== Self-managed deployments
192192

193-
For self-managed deployments, each node's <<data-node,data role>> is configured
193+
For self-managed deployments, each node's <<data-node-role,data role>> is configured
194194
in `elasticsearch.yml`. For example, the highest-performance nodes in a cluster
195195
might be assigned to both the hot and content tiers:
196196

docs/reference/high-availability/cluster-design.asciidoc

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -87,7 +87,7 @@ the same thing, but it's not necessary to use this feature in such a small
8787
cluster.
8888

8989
We recommend you set only one of your two nodes to be
90-
<<master-node,master-eligible>>. This means you can be certain which of your
90+
<<master-node-role,master-eligible>>. This means you can be certain which of your
9191
nodes is the elected master of the cluster. The cluster can tolerate the loss of
9292
the other master-ineligible node. If you set both nodes to master-eligible, two
9393
nodes are required for a master election. Since the election will fail if either
@@ -164,12 +164,12 @@ cluster that is suitable for production deployments.
164164
[[high-availability-cluster-design-three-nodes]]
165165
==== Three-node clusters
166166

167-
If you have three nodes, we recommend they all be <<data-node,data nodes>> and
167+
If you have three nodes, we recommend they all be <<data-node-role,data nodes>> and
168168
every index that is not a <<searchable-snapshots,searchable snapshot index>>
169169
should have at least one replica. Nodes are data nodes by default. You may
170170
prefer for some indices to have two replicas so that each node has a copy of
171171
each shard in those indices. You should also configure each node to be
172-
<<master-node,master-eligible>> so that any two of them can hold a master
172+
<<master-node-role,master-eligible>> so that any two of them can hold a master
173173
election without needing to communicate with the third node. Nodes are
174174
master-eligible by default. This cluster will be resilient to the loss of any
175175
single node.
@@ -188,8 +188,8 @@ service provides such a load balancer.
188188

189189
Once your cluster grows to more than three nodes, you can start to specialise
190190
these nodes according to their responsibilities, allowing you to scale their
191-
resources independently as needed. You can have as many <<data-node,data
192-
nodes>>, <<ingest,ingest nodes>>, <<ml-node,{ml} nodes>>, etc. as needed to
191+
resources independently as needed. You can have as many <<data-node-role,data
192+
nodes>>, <<ingest,ingest nodes>>, <<ml-node-role,{ml} nodes>>, etc. as needed to
193193
support your workload. As your cluster grows larger, we recommend using
194194
dedicated nodes for each role. This allows you to independently scale resources
195195
for each task.

0 commit comments

Comments
 (0)