Skip to content

Commit 0c992ca

Browse files
Merge remote-tracking branch 'origin/main' into feature/hugging-face-chat-completion-integration
# Conflicts: # output/schema/schema-serverless.json # output/schema/schema.json
2 parents 3729546 + f15ccea commit 0c992ca

File tree

9 files changed

+161
-158
lines changed

9 files changed

+161
-158
lines changed

docs/modeling-guide.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -671,6 +671,9 @@ class Request {
671671
}
672672
```
673673
674+
You can see the existing tag values in [elasticsearch-shared-overlays.yaml](https://github.com/elastic/elasticsearch-specification/blob/main/docs/overlays/elasticsearch-shared-overlays.yaml).
675+
If you add a new tag value in your specification, you must also add it to this file.
676+
674677
NOTE: In the OpenAPI specification, operations can have multiple tags. However, we currently support only a single tag.
675678
676679

output/openapi/elasticsearch-openapi.json

Lines changed: 3 additions & 1 deletion
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

output/openapi/elasticsearch-serverless-openapi.json

Lines changed: 2 additions & 0 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

output/schema/schema-serverless.json

Lines changed: 61 additions & 73 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

output/schema/schema.json

Lines changed: 64 additions & 76 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

output/typescript/types.ts

Lines changed: 1 addition & 1 deletion
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

specification/_json_spec/indices.put_data_stream_settings.json

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,11 @@
2525
]
2626
},
2727
"params": {
28+
"dry_run": {
29+
"type": "boolean",
30+
"description": "Perform a dry run but do not actually change any settings",
31+
"default": false
32+
},
2833
"timeout": {
2934
"type": "time",
3035
"description": "Specify timeout for acknowledging the cluster state update"

specification/inference/_types/CommonTypes.ts

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -585,6 +585,8 @@ export class CohereServiceSettings {
585585
api_key: string
586586
/**
587587
* For a `text_embedding` task, the types of embeddings you want to get back.
588+
* Use `binary` for binary embeddings, which are encoded as bytes with signed int8 precision.
589+
* Use `bit` for binary embeddings, which are encoded as bytes with signed int8 precision (this is a synonym of `binary`).
588590
* Use `byte` for signed int8 embeddings (this is a synonym of `int8`).
589591
* Use `float` for the default float embeddings.
590592
* Use `int8` for signed int8 embeddings.
@@ -625,6 +627,8 @@ export enum CohereServiceType {
625627
}
626628

627629
export enum CohereEmbeddingType {
630+
binary,
631+
bit,
628632
byte,
629633
float,
630634
int8

specification/snapshot/repository_analyze/SnapshotAnalyzeRepositoryRequest.ts

Lines changed: 18 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -24,23 +24,31 @@ import { Duration } from '@_types/Time'
2424

2525
/**
2626
* Analyze a snapshot repository.
27-
* Analyze the performance characteristics and any incorrect behaviour found in a repository.
2827
*
29-
* The response exposes implementation details of the analysis which may change from version to version.
30-
* The response body format is therefore not considered stable and may be different in newer versions.
28+
* Performs operations on a snapshot repository in order to check for incorrect behaviour.
3129
*
3230
* There are a large number of third-party storage systems available, not all of which are suitable for use as a snapshot repository by Elasticsearch.
33-
* Some storage systems behave incorrectly, or perform poorly, especially when accessed concurrently by multiple clients as the nodes of an Elasticsearch cluster do. This API performs a collection of read and write operations on your repository which are designed to detect incorrect behaviour and to measure the performance characteristics of your storage system.
31+
* Some storage systems behave incorrectly, or perform poorly, especially when accessed concurrently by multiple clients as the nodes of an Elasticsearch cluster do.
32+
* This API performs a collection of read and write operations on your repository which are designed to detect incorrect behaviour and to measure the performance characteristics of your storage system.
3433
*
3534
* The default values for the parameters are deliberately low to reduce the impact of running an analysis inadvertently and to provide a sensible starting point for your investigations.
3635
* Run your first analysis with the default parameter values to check for simple problems.
37-
* If successful, run a sequence of increasingly large analyses until you encounter a failure or you reach a `blob_count` of at least `2000`, a `max_blob_size` of at least `2gb`, a `max_total_data_size` of at least `1tb`, and a `register_operation_count` of at least `100`.
36+
* Some repositories may behave correctly when lightly loaded but incorrectly under production-like workloads.
37+
* If the first analysis is successful, run a sequence of increasingly large analyses until you encounter a failure or you reach a `blob_count` of at least `2000`, a `max_blob_size` of at least `2gb`, a `max_total_data_size` of at least `1tb`, and a `register_operation_count` of at least `100`.
3838
* Always specify a generous timeout, possibly `1h` or longer, to allow time for each analysis to run to completion.
39+
* Some repositories may behave correctly when accessed by a small number of Elasticsearch nodes but incorrectly when accessed concurrently by a production-scale cluster.
3940
* Perform the analyses using a multi-node cluster of a similar size to your production cluster so that it can detect any problems that only arise when the repository is accessed by many nodes at once.
4041
*
4142
* If the analysis fails, Elasticsearch detected that your repository behaved unexpectedly.
4243
* This usually means you are using a third-party storage system with an incorrect or incompatible implementation of the API it claims to support.
4344
* If so, this storage system is not suitable for use as a snapshot repository.
45+
* Repository analysis triggers conditions that occur only rarely when taking snapshots in a production system.
46+
* Snapshotting to unsuitable storage may appear to work correctly most of the time despite repository analysis failures.
47+
* However your snapshot data is at risk if you store it in a snapshot repository that does not reliably pass repository analysis.
48+
* You can demonstrate that the analysis failure is due to an incompatible storage implementation by verifying that Elasticsearch does not detect the same problem when analysing the reference implementation of the storage protocol you are using.
49+
* For instance, if you are using storage that offers an API which the supplier claims to be compatible with AWS S3, verify that repositories in AWS S3 do not fail repository analysis.
50+
* This allows you to demonstrate to your storage supplier that a repository analysis failure must only be caused by an incompatibility with AWS S3 and cannot be attributed to a problem in Elasticsearch.
51+
* Please do not report Elasticsearch issues involving third-party storage systems unless you can demonstrate that the same issue exists when analysing a repository that uses the reference implementation of the same storage protocol.
4452
* You will need to work with the supplier of your storage system to address the incompatibilities that Elasticsearch detects.
4553
*
4654
* If the analysis is successful, the API returns details of the testing process, optionally including how long each operation took.
@@ -72,7 +80,9 @@ import { Duration } from '@_types/Time'
7280
* You must ensure this load does not affect other users of these systems.
7381
* Analyses respect the repository settings `max_snapshot_bytes_per_sec` and `max_restore_bytes_per_sec` if available and the cluster setting `indices.recovery.max_bytes_per_sec` which you can use to limit the bandwidth they consume.
7482
*
75-
* NOTE: This API is intended for exploratory use by humans. You should expect the request parameters and the response format to vary in future versions.
83+
* NOTE: This API is intended for exploratory use by humans.
84+
* You should expect the request parameters and the response format to vary in future versions.
85+
* The response exposes immplementation details of the analysis which may change from version to version.
7686
*
7787
* NOTE: Different versions of Elasticsearch may perform different checks for repository compatibility, with newer versions typically being stricter than older ones.
7888
* A storage system that passes repository analysis with one version of Elasticsearch may fail with a different version.
@@ -83,7 +93,8 @@ import { Duration } from '@_types/Time'
8393
*
8494
* *Implementation details*
8595
*
86-
* NOTE: This section of documentation describes how the repository analysis API works in this version of Elasticsearch, but you should expect the implementation to vary between versions. The request parameters and response format depend on details of the implementation so may also be different in newer versions.
96+
* NOTE: This section of documentation describes how the repository analysis API works in this version of Elasticsearch, but you should expect the implementation to vary between versions.
97+
* The request parameters and response format depend on details of the implementation so may also be different in newer versions.
8798
*
8899
* The analysis comprises a number of blob-level tasks, as set by the `blob_count` parameter and a number of compare-and-exchange operations on linearizable registers, as set by the `register_operation_count` parameter.
89100
* These tasks are distributed over the data and master-eligible nodes in the cluster for execution.

0 commit comments

Comments
 (0)