Skip to content

Commit ea2f29b

Browse files
Merge branch 'main' into ts-obs-applies
2 parents f7d68d2 + 1414409 commit ea2f29b

File tree

135 files changed

+1881
-6509
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

135 files changed

+1881
-6509
lines changed

deploy-manage/autoscaling/trained-model-autoscaling.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -32,6 +32,7 @@ The available resources of self-managed deployments are static, so trained model
3232
{{serverless-full}} Security and Observability projects are only charged for data ingestion and retention. They are not charged for processing power (VCU usage), which is used for more complex operations, like running advanced search models. For example, in Search projects, models such as ELSER require significant processing power to provide more accurate search results.
3333

3434
## Enabling autoscaling through APIs - adaptive allocations [enabling-autoscaling-through-apis-adaptive-allocations]
35+
3536
$$$nlp-model-adaptive-resources$$$
3637

3738
Model allocations are independent units of work for NLP tasks. If you set the numbers of threads and allocations for a model manually, they remain constant even when not all the available resources are fully used or when the load on the model requires more resources. Instead of setting the number of allocations manually, you can enable adaptive allocations to set the number of allocations based on the load on the process. This can help you to manage performance and cost more easily. (Refer to the [pricing calculator](https://cloud.elastic.co/pricing) to learn more about the possible costs.)
@@ -55,6 +56,7 @@ When you create inference endpoints on {{serverless-short}} using Kibana, adapti
5556
:::
5657

5758
### Optimizing for typical use cases [optimizing-for-typical-use-cases]
59+
5860
You can optimize your model deployment for typical use cases, such as search and ingest. When you optimize for ingest, the throughput will be higher, which increases the number of {{infer}} requests that can be performed in parallel. When you optimize for search, the latency will be lower during search processes.
5961

6062
* If you want to optimize for ingest, set the number of threads to `1` (`"threads_per_allocation": 1`).
@@ -86,6 +88,11 @@ The used resources for trained model deployments depend on three factors:
8688
* the use case you optimize the model deployment for (ingest or search)
8789
* whether model autoscaling is enabled with adaptive allocations/resources to have dynamic resources, or disabled for static resources
8890

91+
::::{note}
92+
On {{serverless-short}}, VCUs for {{ml}} are based on the amount of vCPU and memory consumed. For {{ml}}, `1` VCU equals `0.125` of vCPU and `1GB` of memory, where vCPUs are measured by allocations multiplied by threads, and where memory is the amount consumed by trained models or {{ml}} jobs.
93+
As a math formula, `VCUs = 8 * allocations * threads`, or `1` VCU for every `1GB` of memory consumed, whichever is greater.
94+
::::
95+
8996
If you use a self-managed cluster or ECK, vCPUs level ranges are derived from the `total_ml_processors` and `max_single_ml_node_processors` values. Use the [get {{ml}} info API](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-ml-info) to check these values.
9097

9198
The following tables show you the number of allocations, threads, and vCPUs available in ECE and ECH when adaptive resources are enabled or disabled.

deploy-manage/deploy/elastic-cloud/add-plugins-extensions.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,11 +16,11 @@ Plugins extend the core functionality of {{es}}. There are many suitable plugins
1616
* Analysis plugins, to provide analyzers targeted at languages other than English.
1717
* Scripting plugins, to provide additional scripting languages.
1818

19-
Plugins can come from different sources: the official ones created or at least maintained by Elastic, community-sourced plugins from other users, and plugins that you provide. Some of the official plugins are always provided with our service, and can be [enabled per deployment](elasticsearch://reference/elasticsearch-plugins/cloud/ec-adding-elastic-plugins.md).
19+
Plugins can come from different sources: the official ones created or at least maintained by Elastic, community-sourced plugins from other users, and plugins that you provide. Some of the official plugins are always provided with our service, and can be [enabled per deployment](elasticsearch://reference/elasticsearch-plugins/plugin-management.md).
2020

2121
There are two ways to add plugins to a hosted deployment in {{ecloud}}:
2222

23-
* [Enable one of the official plugins already available in {{ecloud}}](elasticsearch://reference/elasticsearch-plugins/cloud/ec-adding-elastic-plugins.md).
23+
* [Enable one of the official plugins already available in {{ecloud}}](elasticsearch://reference/elasticsearch-plugins/plugin-management.md).
2424
* [Upload a custom plugin and then enable it per deployment](upload-custom-plugins-bundles.md).
2525

2626
Custom plugins can include the official {{es}} plugins not provided with {{ecloud}}, any of the community-sourced plugins, or [plugins that you write yourself](elasticsearch://extend/index.md). Uploading custom plugins is available only to Gold, Platinum, and Enterprise subscriptions. For more information, check [Upload custom plugins and bundles](upload-custom-plugins-bundles.md).

0 commit comments

Comments
 (0)