Skip to content

Commit b9d3ec6

Browse files
TM autoscale on-prem
1 parent 41b1ce5 commit b9d3ec6

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

deploy-manage/autoscaling/trained-model-autoscaling.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -22,6 +22,8 @@ To fully leverage model autoscaling in {{ech}}, {{ece}}, and {{eck}}, it is high
2222

2323
Trained model autoscaling is available for both {{serverless-short}} and Cloud deployments. In serverless deployments, processing power is managed differently across Search, Observability, and Security projects, which impacts their costs and resource limits.
2424

25+
Trained model autoscaling is unavailable for on-premises. The used resources for trained model deployment are static, your available resources are segmented (low, medium, high) based on your settings.
26+
2527
Security and Observability projects are only charged for data ingestion and retention. They are not charged for processing power (VCU usage), which is used for more complex operations, like running advanced search models. For example, in Search projects, models such as ELSER require significant processing power to provide more accurate search results.
2628

2729
## Enabling autoscaling through APIs - adaptive allocations [enabling-autoscaling-through-apis-adaptive-allocations]

0 commit comments

Comments
 (0)