Skip to content

Commit c424d94

Browse files
Merge branch 'main' into use-version-variables
2 parents 8043457 + a64409b commit c424d94

File tree

11 files changed

+33
-19
lines changed

11 files changed

+33
-19
lines changed

.github/CODEOWNERS

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -47,8 +47,6 @@
4747
/solutions/observability/get-started/ @elastic/ski-docs
4848
/solutions/search/ @elastic/developer-docs
4949
/solutions/security/ @elastic/experience-docs
50-
/solutions/security/get-started/ @elastic/ingest-docs @elastic/experience-docs
51-
/solutions/security/cloud/ @elastic/ingest-docs
5250

5351
/troubleshoot/ @elastic/docs
5452
/troubleshoot/deployments/ @elastic/admin-docs

deploy-manage/autoscaling/trained-model-autoscaling.md

Lines changed: 6 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -22,11 +22,13 @@ There are two ways to enable autoscaling:
2222
* through APIs by enabling adaptive allocations
2323
* in {{kib}} by enabling adaptive resources
2424

25+
For {{serverless-short}} projects, trained model autoscaling is automatically enabled and cannot be disabled.
26+
2527
::::{important}
2628
To fully leverage model autoscaling in {{ech}}, {{ece}}, and {{eck}}, it is highly recommended to enable [{{es}} deployment autoscaling](../../deploy-manage/autoscaling.md).
2729
::::
2830

29-
Trained model autoscaling is available for {{serverless-short}}, {{ech}}, {{ece}}, and {{eck}} deployments. In serverless deployments, processing power is managed differently across Search, Observability, and Security projects, which impacts their costs and resource limits.
31+
Trained model autoscaling is available for {{serverless-short}}, {{ech}}, {{ece}}, and {{eck}} deployments. In {{serverless-short}} projects, processing power is managed differently across Search, Observability, and Security projects, which impacts their costs and resource limits.
3032

3133
:::{admonition} Trained model auto-scaling for self-managed deployments
3234
The available resources of self-managed deployments are static, so trained model autoscaling is not applicable. However, available resources are still segmented based on the settings described in this section.
@@ -54,10 +56,6 @@ You can enable adaptive allocations by using:
5456

5557
If the new allocations fit on the current {{ml}} nodes, they are immediately started. If more resource capacity is needed for creating new model allocations, then your {{ml}} node will be scaled up if {{ml}} autoscaling is enabled to provide enough resources for the new allocation. The number of model allocations can be scaled down to 0. They cannot be scaled up to more than 32 allocations, unless you explicitly set the maximum number of allocations to more. Adaptive allocations must be set up independently for each deployment and [{{infer}} endpoint](https://www.elastic.co/docs/api/doc/elasticsearch/group/endpoint-inference).
5658

57-
:::{note}
58-
When you create inference endpoints on {{serverless-short}} using {{kib}}, adaptive allocations are automatically turned on, and there is no option to disable them.
59-
:::
60-
6159
### Optimizing for typical use cases [optimizing-for-typical-use-cases]
6260

6361
You can optimize your model deployment for typical use cases, such as search and ingest. When you optimize for ingest, the throughput will be higher, which increases the number of {{infer}} requests that can be performed in parallel. When you optimize for search, the latency will be lower during search processes.
@@ -73,16 +71,16 @@ You can choose from three levels of resource usage for your trained model deploy
7371

7472
Refer to the tables in the [Model deployment resource matrix](#model-deployment-resource-matrix) section to find out the settings for the level you selected.
7573

76-
:::{image} /deploy-manage/images/machine-learning-ml-nlp-deployment-id-elser-v2.png
74+
The image below shows the process of starting a trained model on an {{ech}} deployment. In {{serverless-short}} projects, the **Adaptive resources** toggle is not available when starting trained model deployments, as adaptive allocations are always enabled and cannot be disabled.
75+
76+
:::{image} /deploy-manage/images/ml-nlp-deployment-id-elser.png
7777
:alt: ELSER deployment with adaptive resources enabled.
7878
:screenshot:
7979
:width: 500px
8080
:::
8181

8282
In {{serverless-full}}, Search projects are given access to more processing resources, while Security and Observability projects have lower limits. This difference is reflected in the UI configuration: Search projects have higher resource limits compared to Security and Observability projects to accommodate their more complex operations.
8383

84-
On {{serverless-short}}, adaptive allocations are automatically enabled for all project types.
85-
8684
## Model deployment resource matrix [model-deployment-resource-matrix]
8785

8886
The used resources for trained model deployments depend on three factors:
@@ -100,10 +98,6 @@ If you use a self-managed cluster or ECK, vCPUs level ranges are derived from th
10098

10199
The following tables show you the number of allocations, threads, and vCPUs available in ECE and ECH when adaptive resources are enabled or disabled.
102100

103-
::::{note}
104-
On {{serverless-short}}, adaptive allocations are automatically enabled for all project types. However, the "Adaptive resources" control is not displayed in {{kib}} for Observability and Security projects.
105-
::::
106-
107101
### Ingest optimized
108102

109103
In case of ingest-optimized deployments, we maximize the number of model allocations.
Binary file not shown.
185 KB
Loading

deploy-manage/security/secure-settings.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ products:
2222

2323
Some settings are sensitive, and relying on filesystem permissions to protect their values is not sufficient. For this use case, {{es}} and {{kib}} provide secure keystores to store sensitive configuration values such as passwords, API keys, and tokens.
2424

25-
Secure settings are often referred to as **keystore settings**, since they must be added to the product-specific keystore rather than the standard [`elasticsearch.yml` or `kibana.yml files](/deploy-manage/stack-settings.md). Unlike regular settings, they are encrypted and protected at rest, and they cannot be read or modified through the usual configuration files or environment variables.
25+
Secure settings are often referred to as **keystore settings**, since they must be added to the product-specific keystore rather than the standard [`elasticsearch.yml` or `kibana.yml` files](/deploy-manage/stack-settings.md). Unlike regular settings, they are encrypted and protected at rest, and they cannot be read or modified through the usual configuration files or environment variables.
2626

2727
Keystore settings must be handled using a specific tool or method depending on the deployment type. The following table summarizes how {{es}} and {{kib}} keystores are supported and managed across different deployment models:
2828

Binary file not shown.
185 KB
Loading

explore-analyze/machine-learning/nlp/ml-nlp-deploy-model.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ You can deploy a model multiple times by assigning a unique deployment ID when s
1616

1717
You can optimize your deplyoment for typical use cases, such as search and ingest. When you optimize for ingest, the throughput will be higher, which increases the number of {{infer}} requests that can be performed in parallel. When you optimize for search, the latency will be lower during search processes. When you have dedicated deployments for different purposes, you ensure that the search speed remains unaffected by ingest workloads, and vice versa. Having separate deployments for search and ingest mitigates performance issues resulting from interactions between the two, which can be hard to diagnose.
1818

19-
:::{image} /explore-analyze/images/machine-learning-ml-nlp-deployment-id-elser-v2.png
19+
:::{image} /explore-analyze/images/ml-nlp-deployment-id-elser.png
2020
:alt: Model deployment on the Trained Models UI.
2121
:screenshot:
2222
:::

explore-analyze/machine-learning/nlp/ml-nlp-elser.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -107,7 +107,7 @@ You can also download and deploy ELSER either from **{{ml-app}}** > **Trained Mo
107107
3. After the download is finished, start the deployment by clicking the **Start deployment** button.
108108
4. Provide a deployment ID, select the priority, and set the number of allocations and threads per allocation values.
109109

110-
:::{image} /explore-analyze/images/machine-learning-ml-nlp-deployment-id-elser-v2.png
110+
:::{image} /explore-analyze/images/ml-nlp-deployment-id-elser.png
111111
:alt: Deploying ELSER
112112
:screenshot:
113113
:::

explore-analyze/scripting/script-fields-api.md

Lines changed: 23 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,12 @@ Use the `field` API to access document fields:
2525
field('my_field').get(<default_value>)
2626
```
2727

28+
Alternatively use the shortcut of `$` to get a field.
29+
30+
```painless
31+
$('my_field', <default_value>)
32+
```
33+
2834
This API fundamentally changes how you access documents in Painless. Previously, you had to access the `doc` map with the field name that you wanted to access:
2935

3036
```painless
@@ -77,7 +83,6 @@ ZonedDateTime end = field('end').get(null);
7783
return start == null || end == null ? -1 : ChronoUnit.MILLIS.between(start, end)
7884
```
7985

80-
8186
## Supported mapped field types [_supported_mapped_field_types]
8287

8388
The following table indicates the mapped field types that the `field` API supports. For each supported type, values are listed that are returned by the `field` API (from the `get` and `as<Type>` methods) and the `doc` map (from the `getValue` and `get` methods).
@@ -112,3 +117,20 @@ The `fields` API currently doesn’t support some fields, but you can still acce
112117
| `wildcard` | `String` | - | `String` | `String` |
113118
| `flattened` | `String` | - | `String` | `String` |
114119

120+
## Manipulation of the fields data
121+
122+
The field API provides a `set(<value>)` operation that will take the field name and create the necessary structure. Calling this inside an ingest pipelines script processor context:
123+
124+
```painless
125+
field("foo.bar").set("abc")
126+
```
127+
128+
leads to the generation of this JSON representation.
129+
130+
```json
131+
{
132+
"foo": {
133+
"bar": "abc"
134+
}
135+
}
136+
```

0 commit comments

Comments
 (0)