You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: deploy-manage/deploy/self-managed.md
+2-1Lines changed: 2 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -33,7 +33,8 @@ For a comparison of these deployment options, refer to [Choosing your deployment
33
33
34
34
This section focuses on deploying {{es}} and {{kib}} without an orchestrator.
35
35
36
-
Depending on your use case, you might need to deploy other components, such as APM, Fleet, or Logstash. Deploying those components is not covered in this section. [Learn more about optional components](/get-started/the-stack.md).
36
+
Depending on your use case, you might need to deploy other components, such as APM, Fleet, or Logstash.
37
+
Deploying those components is not covered in this section. [Learn more about optional components](/get-started/the-stack.md).
{{es}} requires very little configuration to get started, but there are a number of items which **must** be considered before using your cluster in production:
14
15
15
16
*[Path settings](#path-settings)
16
-
*[Cluster name setting](elasticsearch://reference/elasticsearch/configuration-reference/miscellaneous-cluster-settings.md#cluster-name)
@@ -241,3 +243,7 @@ In a disaster, [snapshots](../../tools/snapshot-and-restore.md) can prevent perm
241
243
**Taking a snapshot is the only reliable and supported way to back up a cluster.** You cannot back up an {{es}} cluster by making copies of the data directories of its nodes. There are no supported methods to restore any data from a file system-level backup. If you try to restore a cluster from such a backup, it may fail with reports of corruption or missing files or other data inconsistencies, or it may appear to have succeeded having silently lost some of your data.
242
244
243
245
::::
246
+
247
+
## DNS cache settings [networkaddress-cache-ttl]
248
+
249
+
{{es}} runs with a security manager in place. With a security manager in place, the JVM defaults to caching positive hostname resolutions indefinitely and defaults to caching negative hostname resolutions for ten seconds. {{es}} overrides this behavior with default values to cache positive lookups for sixty seconds, and to cache negative lookups for ten seconds. These values should be suitable for most environments, including environments where DNS resolutions vary with time. If not, you can edit the values `es.networkaddress.cache.ttl` and `es.networkaddress.cache.negative.ttl` in the [JVM options](elasticsearch://reference/elasticsearch/jvm-settings.md#set-jvm-options). Note that the values [`networkaddress.cache.ttl=<timeout>`](https://docs.oracle.com/javase/8/docs/technotes/guides/net/properties.md) and [`networkaddress.cache.negative.ttl=<timeout>`](https://docs.oracle.com/javase/8/docs/technotes/guides/net/properties.md) in the [Java security policy](https://docs.oracle.com/javase/8/docs/technotes/guides/security/PolicyFiles.md) are ignored by {{es}} unless you remove the settings for `es.networkaddress.cache.ttl` and `es.networkaddress.cache.negative.ttl`.
Copy file name to clipboardExpand all lines: deploy-manage/distributed-architecture/discovery-cluster-formation/discovery-hosts-providers.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -25,7 +25,7 @@ Refer to [Troubleshooting discovery](../../../troubleshoot/elasticsearch/discove
25
25
26
26
By default the cluster formation module offers two seed hosts providers to configure the list of seed nodes: a *settings*-based and a *file*-based seed hosts provider. It can be extended to support cloud environments and other forms of seed hosts providers via [discovery plugins](elasticsearch://reference/elasticsearch-plugins/discovery-plugins.md). Seed hosts providers are configured using the `discovery.seed_providers` setting, which defaults to the *settings*-based hosts provider. This setting accepts a list of different providers, allowing you to make use of multiple ways to find the seed hosts for your cluster.
27
27
28
-
Each seed hosts provider yields the IP addresses or hostnames of the seed nodes. If it returns any hostnames then these are resolved to IP addresses using a DNS lookup. If a hostname resolves to multiple IP addresses then {{es}} tries to find a seed node at all of these addresses. If the hosts provider does not explicitly give the TCP port of the node by then, it will implicitly use the first port in the port range given by `transport.profiles.default.port`, or by `transport.port` if `transport.profiles.default.port` is not set. The number of concurrent lookups is controlled by `discovery.seed_resolver.max_concurrent_resolvers` which defaults to `10`, and the timeout for each lookup is controlled by `discovery.seed_resolver.timeout` which defaults to `5s`. Note that DNS lookups are subject to [JVM DNS caching](../../deploy/self-managed/networkaddress-cache-ttl.md).
28
+
Each seed hosts provider yields the IP addresses or hostnames of the seed nodes. If it returns any hostnames then these are resolved to IP addresses using a DNS lookup. If a hostname resolves to multiple IP addresses then {{es}} tries to find a seed node at all of these addresses. If the hosts provider does not explicitly give the TCP port of the node by then, it will implicitly use the first port in the port range given by `transport.profiles.default.port`, or by `transport.port` if `transport.profiles.default.port` is not set. The number of concurrent lookups is controlled by `discovery.seed_resolver.max_concurrent_resolvers` which defaults to `10`, and the timeout for each lookup is controlled by `discovery.seed_resolver.timeout` which defaults to `5s`. Note that DNS lookups are subject to [JVM DNS caching](/deploy-manage/deploy/self-managed/important-settings-configuration.md#networkaddress-cache-ttl).
Copy file name to clipboardExpand all lines: deploy-manage/maintenance/ece/perform-ece-hosts-maintenance.md
+76-2Lines changed: 76 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -19,13 +19,19 @@ These steps show how you can safely perform maintenance on hosts in your ECE ins
19
19
20
20
You can perform these maintenance actions on the hosts in your ECE installation using one of these methods:
21
21
22
-
*[By disabling the Docker daemon (nondestructive)](#ece-perform-host-maintenance-docker-disable)
22
+
*[By disabling the container services (nondestructive)](#ece-perform-host-maintenance-container-engine-disable):
23
+
*[For Docker-based installations: disable the Docker service](#ece-perform-host-maintenance-docker-disable)
24
+
*[For Podman-based installations: disable the Podman-related services](#ece-perform-host-maintenance-podman-disable)
23
25
*[By deleting the host (destructive)](#ece-perform-host-maintenance-delete-runner)
24
26
*[By shutting down the host (less destructive)](#ece-perform-host-maintenance-delete-runner)
25
27
26
28
Which method you choose depends on how invasive your host maintenance needs to be. If your host maintenance could affect ECE, use the destructive method that first deletes the host from your installation. These methods include a step that moves any hosted {{es}} clusters and {{kib}} instances off the affected hosts and are generally considered safe, provided that your ECE installation still has sufficient resources available to operate after the host has been removed.
27
29
28
-
## By disabling the Docker daemon [ece-perform-host-maintenance-docker-disable]
30
+
## By disabling the container services (nondestructive) [ece-perform-host-maintenance-container-engine-disable]
31
+
32
+
The way that you disable container services differs based on the platform you used to deploy your ECE hosts.
33
+
34
+
### For Docker-based installations: disable the Docker service [ece-perform-host-maintenance-docker-disable]
29
35
30
36
This method lets you perform maintenance actions on hosts without first removing the associated host from your {{ece}} installation. It works by disabling the Docker daemon. The host remains a part of your ECE installation throughout these steps but will be offline and the resources it provides will not be available.
31
37
@@ -71,6 +77,74 @@ To perform host maintenance:
71
77
72
78
After the host shows a green status in the Cloud UI, it is fully functional again and can be used as before.
73
79
80
+
### For Podman-based installations: disable the Podman-related services [ece-perform-host-maintenance-podman-disable]
81
+
82
+
This method lets you perform maintenance actions on hosts without first removing the associated host from your {{ece}} installation. It works by disabling the Podman related services. The host remains a part of your ECE installation throughout these steps but will be offline and the resources it provides will not be available.
83
+
84
+
To perform host maintenance:
85
+
86
+
1. Recommended: If the host holds the allocator role and you have enough spare capacity:
87
+
1. [Enable maintenance mode](enable-maintenance-mode.md) on the allocator.
88
+
2. [Move all nodes off the allocator](move-nodes-instances-from-allocators.md) and to other allocators in your installation. Moving all nodes lets you retain the same level of redundancy for highly available {{es}} clusters and ensures that other clusters without high availability remain available.
89
+
::::{important}
90
+
Skipping Step 1 will affect the availability of clusters with nodes on the allocator.
91
+
::::
92
+
93
+
2. Disable the Podman service, Podman socket, and Podman restart service:
94
+
95
+
```sh
96
+
sudo systemctl disable podman.service
97
+
sudo systemctl disable podman.socket
98
+
sudo systemctl disable podman-restart.service
99
+
```
100
+
101
+
3. Reboot the host:
102
+
103
+
```sh
104
+
sudo reboot
105
+
```
106
+
107
+
4. After rebooting, confirm there are no running containers by running the following command. The output should be empty.
108
+
```sh
109
+
sudo podman ps
110
+
```
111
+
112
+
If an `frc-*` or `fac-*` container is returned in the output, stop it:
4. Perform your maintenance on the host, such as patching the operating system.
119
+
5. Re-enable the Podman related services:
120
+
121
+
```sh
122
+
sudo systemctl enable podman.service
123
+
sudo systemctl enable podman.socket
124
+
sudo systemctl enable podman-restart.service
125
+
```
126
+
127
+
6. Reboot the host again:
128
+
129
+
```sh
130
+
sudo reboot
131
+
```
132
+
133
+
7. Confirm the containers have started:
134
+
135
+
```sh
136
+
sudo podman ps -a
137
+
```
138
+
139
+
The use `-a` flag ensures that no containers are overlooked.
140
+
141
+
142
+
8. If you enabled maintenance mode in Step 1, take the allocator out of maintenance mode.
143
+
9. Optional for allocators: ECE will start using the allocator again as you create new or change existing clusters, but it will not automatically redistribute nodes to an allocator after it becomes available. If you want to move nodes back to the same allocator after host maintenance, you need to manually [move the nodes](move-nodes-instances-from-allocators.md) and specify the allocator as a target.
144
+
10. Verify that all ECE services and deployments are back up by checking that the host shows a green status in the Cloud UI.
145
+
146
+
After the host shows a green status in the Cloud UI, it is fully functional again and can be used as before.
147
+
74
148
## By deleting the host (destructive) [ece-perform-host-maintenance-delete-runner]
75
149
76
150
This method lets you perform potentially destructive maintenance actions on hosts. It works by deleting the associated host, which removes the host from your {{ece}} installation. To add the host to your ECE installation again after host maintenance is complete, you must reinstall ECE.
Copy file name to clipboardExpand all lines: deploy-manage/monitor/autoops/cc-cloud-connect-autoops-faq.md
+15-3Lines changed: 15 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -24,9 +24,10 @@ Find answers to your questions about AutoOps for ECE, ECK, and self-managed clus
24
24
*[Can I use macOS to install {{agent}} for this feature?](#macos-install)
25
25
*[Do I have to define an Elastic IP address to enable the agent to send data to {{ecloud}}?](#elastic-ip-address)
26
26
27
-
**Questions about collected metrics**
27
+
**Questions about collected metrics and data**
28
28
*[Where are AutoOps metrics stored?](#autoops-metrics-storage)
29
29
*[What information does {{agent}} extract from my cluster?](#extracted-info)
30
+
*[How does AutoOps gather data from my cluster and ensure its security?](#data-gathering)
30
31
31
32
## General questions
32
33
$$$why-autoops$$$**Why should I use AutoOps for my clusters?**
@@ -42,7 +43,7 @@ $$$autoops-metrics-cost$$$ **Is there an added cost for shipping metrics data to
42
43
You can [choose the CSP region where your data is stored](#autoops-metrics-storage).
43
44
44
45
$$$es-versions$$$**Which versions of {{es}} does AutoOps support?**
45
-
: AutoOps is compatible with all [supported {{es}} versions](https://www.elastic.co/support/eol).
46
+
: AutoOps is compatible with [supported {{es}} versions](https://www.elastic.co/support/eol) (7.17.x and above).
46
47
47
48
$$$deployment-types$$$**Which deployment types can be connected to AutoOps?**
48
49
: You can connect to AutoOps on a standalone {{stack}}, ECE ({{ece}}), or ECK ({{eck}}) deployment.
@@ -65,7 +66,7 @@ $$$elastic-ip-address$$$ **Do I have to define an Elastic IP address to enable t
65
66
66
67
: For more information, refer to [](/deploy-manage/security/elastic-cloud-static-ips.md).
67
68
68
-
## Questions about collected metrics
69
+
## Questions about collected metrics and data
69
70
$$$autoops-metrics-storage$$$**Where are AutoOps metrics stored?**
70
71
: You can choose where to store your metrics from the following AWS regions:
71
72
@@ -88,3 +89,14 @@ $$$extracted-info$$$ **What information does {{agent}} extract from my cluster?*
88
89
| [_template](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-indices-get-template) | Retrieves legacy index templates | Similar to composable index templates but in older format |
89
90
| [_resolve/index/*](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-indices-resolve-index) | Resolves index, data stream, and alias names to their current definitions | Mappings between names and underlying data objects |
90
91
92
+
$$$data-gathering$$$**How does AutoOps gather data from my cluster and ensure its security?**
93
+
: AutoOps gathers data from your cluster using two protocols:
94
+
* **HTTP request**: Made to our Cloud Connected API to register your cluster with {{ecloud}} and gather registration-related data.
95
+
* **OpenTelemetry Protocol (OTLP)**: Used to gather all other operational data.
96
+
97
+
Each channel is authenticated through an API key or token to ensure your data's security. The following table offers more details:
98
+
99
+
| Protocol | Data extracted | Port | Authentication method |
100
+
| --- | --- | --- | --- |
101
+
| HTTP | Basic cluster information from the `/` endpoint <br><br> License information from the `/_license` endpoint | **443**: standard HTTPS port | Uses an {{ecloud}} API key which is limited for use with Cloud Connect only. |
102
+
| OTLP | Operational information | **4318**: standard OTLP HTTP port <br><br> This service will be exposed on port 443 in the future. | Uses an AutoOps token which is functionally equivalent to an API key. |
Copy file name to clipboardExpand all lines: deploy-manage/monitor/autoops/cc-cloud-connect-autoops-troubleshooting.md
+2-1Lines changed: 2 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -49,4 +49,5 @@ The following table shows the errors you might encounter if something goes wrong
49
49
|`LICENSE_EXPIRED`| Elastic license is expired | Contact [sales](https://www.elastic.co/contact#sales) to renew your license. |
50
50
|`LICENSE_USED_BY_ANOTHER_ACCOUNT`| License key connected to another account | A license key can only be connected to one {{ecloud}} organization. Contact [Elastic support](https://support.elastic.co/) for help. |
51
51
|`VERSION_MISMATCH`| {{es}} version is unsupported | Upgrade your cluster to a [supported version](https://www.elastic.co/support/eol). |
52
-
|`UNKNOWN_ERROR`| Installation failed | {{agent}} couldn't be installed due to an unknown issue. Consult the troubleshooting guide or contact [Elastic support](https://support.elastic.co/) for more help. |
52
+
|`UNKNOWN_ERROR`| Installation failed | {{agent}} couldn't be installed due to an unknown issue. Consult the troubleshooting guide or contact [Elastic support](https://support.elastic.co/) for more help. |
53
+
| `` | Failed to register Cloud Connected Mode: cluster license type is not supported | The cluster you are trying to connect doesn't have the required license to connect to AutoOps. For more information, refer to the [prerequisites](/deploy-manage/monitor/autoops/cc-connect-self-managed-to-autoops.md#prerequisites). |
0 commit comments