You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: contribute-docs/api-docs/kibana-api-docs-quickstart.md
+18-4Lines changed: 18 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -202,7 +202,7 @@ responses:
202
202
:sync: code-generated
203
203
204
204
:::{note}
205
-
**This step is optional.** CI will automatically capture the snapshot when you push your `.ts` changes. Running this locally is useful for validating changes before pushing or debugging issues. See [`capture_oas_snapshot.sh`](https://github.com/elastic/kibana/blob/main/.buildkite/scripts/steps/checks/capture_oas_snapshot.sh) for the full list of paths captured in CI.
205
+
**This step is optional.** CI will automatically capture the snapshot when you push your `.ts` changes. Running this locally is useful for validating changes before pushing or debugging issues.
206
206
:::
207
207
208
208
This step captures the OpenAPI specification that {{kib}} generates at runtime from your route definitions. It spins up a local {{es}} and {{kib}} cluster with your code changes. This generates the following output files in the `oas_docs` directory:
@@ -214,13 +214,27 @@ This step captures the OpenAPI specification that {{kib}} generates at runtime f
214
214
- [Docker](https://docs.docker.com/get-docker/) must be running
215
215
- If you're an Elastician, ensure you're logged into Docker with your Elastic account
216
216
217
-
**Capture all API paths** (recommended):
217
+
To capture all the documented API paths, copy the command from [`capture_oas_snapshot.sh`](https://github.com/elastic/kibana/blob/main/.buildkite/scripts/steps/checks/capture_oas_snapshot.sh). For example:
218
218
219
219
```bash
220
-
node scripts/capture_oas_snapshot --update
220
+
node scripts/capture_oas_snapshot \
221
+
--include-path /api/status \
222
+
--include-path /api/alerting/rule/ \
223
+
--include-path /api/alerting/rules \
224
+
--include-path /api/actions \
225
+
--include-path /api/security/role \
226
+
--include-path /api/spaces \
227
+
--include-path /api/streams \
228
+
--include-path /api/fleet \
229
+
--include-path /api/saved_objects/_import \
230
+
--include-path /api/saved_objects/_export \
231
+
--include-path /api/maintenance_window \
232
+
--include-path /api/agent_builder
233
+
--update
221
234
```
222
235
223
-
**For faster iteration**, capture the specific paths you're working on:
236
+
For faster iteration, you can capture the specific paths you're working on, though this minimized output should not be included in your pull request.
Copy file name to clipboardExpand all lines: deploy-manage/deploy/cloud-enterprise/resize-deployment.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -34,7 +34,7 @@ To resize a deployment:
34
34
::::
35
35
36
36
RAM per instance
37
-
: Node and instance capacity should be sufficient to sustain your search workload, even if you lose an availability zone. Currently, half of the memory is assigned to the JVM heap. For example, on an {{es}} cluster node with 32 GB RAM, 16 GB would be allotted to heap. Up to 64 GB RAM and 1 TB storage per node are supported.
37
+
: Node and instance capacity should be sufficient to sustain your search workload, even if you lose an availability zone. For instances up to 64 GB of RAM, half the memory is assigned to the JVM heap. For instances larger than 64 GB, the heap size is capped at 32 GB. For example, on an {{es}} cluster node with 32 GB RAM, 16 GB would be allotted to heap, while on a 128 GB node, 32 GB would be allotted to heap. Up to 256 GB RAM and 1 TB storage per node are supported.
38
38
39
39
Before finalizing your changes, you can review the **Architecture** summary, which shows the total number of instances per zone, with each circle color representing a different type of instance.
Copy file name to clipboardExpand all lines: deploy-manage/deploy/elastic-cloud/ec-customize-deployment-components.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -31,7 +31,7 @@ To change a cluster’s topology, from deployment management, select **Edit depl
31
31
32
32
For trials, larger sizes are not available until you [add a credit card](../../cloud-organization/billing/add-billing-details.md).
33
33
34
-
Currently, half the memory is assigned to the JVM heap (a bit less when monitoring is activated). For example, on a 32 GB cluster, 16 GB are allotted to heap. The disk-to-RAM ratio currently is 1:24, meaning that you get 24 GB of storage space for each 1 GB of RAM. All clusters are backed by SSD drives.
34
+
For instances up to 64 GB of RAM, half the memory is assigned to the JVM heap (a bit less when monitoring is activated). For instances larger than 64 GB, the heap size is capped at 32 GB. For example, on a 32 GB instance, 16 GB are allotted to heap, while on a 128 GB instance, 32 GB are allotted to heap. Up to 256 GB RAM per instance is supported. The disk-to-RAM ratio currently is 1:24, meaning that you get 24 GB of storage space for each 1 GB of RAM. All clusters are backed by SSD drives.
35
35
36
36
::::{tip}
37
37
For production systems, each {{es}} instance in your cluster should have at least 4 GB of RAM, which assigns 2 GB to the JVM heap. Review [Minimum size recommendations for production use](elastic-cloud-hosted-planning.md#ec-minimum-recommendations) for more details.
Copy file name to clipboardExpand all lines: deploy-manage/tools/cross-cluster-replication/ccr-getting-started-prerequisites.md
+5-4Lines changed: 5 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,8 +16,9 @@ products:
16
16
17
17
To complete this tutorial, you need:
18
18
19
-
* The `manage` cluster privilege on the local cluster.
20
-
* A license on both clusters that includes {{ccr}}. [Activate a free 30-day trial](../../license/manage-your-license-in-self-managed-cluster.md).
21
-
* An index on the remote cluster that contains the data you want to replicate. This tutorial uses the sample eCommerce orders data set. [Load sample data](../../../explore-analyze/index.md#gs-get-data-into-kibana).
22
-
* In the local cluster, all nodes with the `master`[node role](elasticsearch://reference/elasticsearch/configuration-reference/node-settings.md#node-roles) must also have the [`remote_cluster_client`](../../distributed-architecture/clusters-nodes-shards/node-roles.md#remote-node) role. The local cluster must also have at least one node with both a data role and the [`remote_cluster_client`](../../distributed-architecture/clusters-nodes-shards/node-roles.md#remote-node) role. Individual tasks for coordinating replication scale based on the number of data nodes with the `remote_cluster_client` role in the local cluster.
19
+
- The `manage` cluster privilege on the local cluster.
20
+
- Both clusters must have the same [license](/deploy-manage/license.md) type, and that license must support {{ccr}}.
21
+
-[Activate a free 30-day trial](../../license/manage-your-license-in-self-managed-cluster.md).
22
+
- An index on the remote cluster that contains the data you want to replicate. This tutorial uses the sample eCommerce orders data set. [Load sample data](../../../explore-analyze/index.md#gs-get-data-into-kibana).
23
+
- In the local cluster, all nodes with the `master`[node role](elasticsearch://reference/elasticsearch/configuration-reference/node-settings.md#node-roles) must also have the [`remote_cluster_client`](../../distributed-architecture/clusters-nodes-shards/node-roles.md#remote-node) role. The local cluster must also have at least one node with both a data role and the [`remote_cluster_client`](../../distributed-architecture/clusters-nodes-shards/node-roles.md#remote-node) role. Individual tasks for coordinating replication scale based on the number of data nodes with the `remote_cluster_client` role in the local cluster.
0 commit comments