You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/infrastructure/components/elastic.agent.md
+17Lines changed: 17 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,16 @@
1
1
# Elastic Agent
2
2
3
+
## Table of contents
4
+
5
+
-[Overview](#overview)
6
+
-[Fleet](#fleet)
7
+
-[Installation](#installation)
8
+
-[Configuration](#configuration)
9
+
-[When to update](#when-to-update)
10
+
-[Investigating the metrics](#investigating-the-metrics)
11
+
12
+
## Overview
13
+
3
14
The agent runs as a DaemonSet and collects:
4
15
5
16
- Kubernetes logs (not collected in this setup)
@@ -33,6 +44,12 @@ From the standard configuration, the following changes have been made:
33
44
- Collect Kubernetes container logs has been de-activated. We already collect these logs using Fluent Bit and we want to avoid duplication.
34
45
- Collect Kubernetes events from Kubernetes API Server has been de-activated. We already collect these events using Event exporter and we want to avoid duplication.
35
46
47
+
## When to update
48
+
49
+
- Update Elastic Agent whenever the Elastic Stack (e.g. Elastic Cloud) is upgraded to keep versions aligned.
50
+
- Perform updates via Helm (code change), not directly in Elastic Fleet, to avoid configuration drift.
51
+
- Also update for critical fixes or security advisories from Elastic.
Copy file name to clipboardExpand all lines: docs/infrastructure/disaster.recovery.md
-4Lines changed: 0 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -54,7 +54,6 @@ If any of the cluster infrastructure exists but is not functional, see the above
54
54
```
55
55
56
56
3. Deploy Kubernetes components:
57
-
58
57
1. Connect AWS CLI to the new cluster: `aws eks update-kubeconfig --name=Workflows`.
59
58
2. Create the Argo Workflows configuration files: `npx cdk8s synth`.
60
59
3. (ONLY IF [RECREATING DATABASE](#rds-database)) Remove the `persistence` section of `dist/0005-argo-workflows.k8s.yaml` to disable workflow archiving to database. For example:
@@ -127,7 +126,6 @@ If there is any issue on the RDS instance that can't be recovered, we might have
127
126
2. [Deploy the EKS cluster](#deployment-of-new-cluster)
128
127
129
128
3. Create a temporary RDS database from [the manual snapshot created](#update-database-version-if-necessary):
130
-
131
129
1. Get details of the new cluster database: `aws rds describe-db-instances --query="DBInstances[?DBName=='argo'].{EndpointAddress: Endpoint.Address, DBSubnetGroupName: DBSubnetGroup.DBSubnetGroupName, VpcSecurityGroupIds: VpcSecurityGroups[].VpcSecurityGroupId}"`.
132
130
2. Go to https://ap-southeast-2.console.aws.amazon.com/rds/home?region=ap-southeast-2#db-snapshot:engine=postgres;id=ID, replacing "ID" with the `DBSnapshotIdentifier` of the manual snapshot.
133
131
3. Click on _Actions_ → _Restore snapshot_.
@@ -140,7 +138,6 @@ If there is any issue on the RDS instance that can't be recovered, we might have
140
138
10. Wait for the temporary DB to get to the "Available" state.
141
139
142
140
4. Dump the temporary database to the new Argo database:
143
-
144
141
1. Submit a ["sleep" workflow](../../workflows/test/sleep.yml) to the new Argo Workflows installation to spin up a pod:
145
142
`argo submit --namespace=argo workflows/test/sleep.yml`. This will be used to connect to RDS to dump the database to a file.
146
143
2. Connect to the sleep pod (it can take a while for the pod to spin up, so you might have to retry the second command):
@@ -165,7 +162,6 @@ If there is any issue on the RDS instance that can't be recovered, we might have
165
162
You will be prompted for a password, get the password from the [AWS Systems Manager Parameter Store](https://ap-southeast-2.console.aws.amazon.com/systems-manager/parameters/%252Feks%252Fargo%252Fpostgres%252Fpassword/description?region=ap-southeast-2&tab=Table).
166
163
167
164
5. Redeploy the cluster configuration files to enable the connection to the database and turn on workflow archiving:
168
-
169
165
1. Run `npx cdk8s synth` to recreate the `persistence` section in `dist/0005-argo-workflows.k8s.yaml`.
170
166
2. Redeploy the Argo config file: `kubectl replace --filename=dist/0005-argo-workflows.k8s.yaml`.
171
167
3. Restart the workflow controller and the server:
0 commit comments