diff --git a/docs/reference.asciidoc b/docs/reference.asciidoc index f5079f9..41c56bd 100644 --- a/docs/reference.asciidoc +++ b/docs/reference.asciidoc @@ -386,7 +386,7 @@ client.fieldCaps({ ... }) ** *`index` (Optional, string | string[])*: List of data streams, indices, and aliases used to limit the request. Supports wildcards (*). To target all data streams and indices, omit this parameter or use * or _all. ** *`fields` (Optional, string | string[])*: List of fields to retrieve capabilities for. Wildcard (`*`) expressions are supported. ** *`index_filter` (Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, knn, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, rule, script, script_score, semantic, shape, simple_query_string, span_containing, span_field_masking, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, sparse_vector, term, terms, terms_set, text_expansion, weighted_tokens, wildcard, wrapper, type })*: Allows to filter indices if the provided query rewrites to match_none on every shard. -** *`runtime_mappings` (Optional, Record)*: Defines ad-hoc runtime fields in the request similar to the way it is done in search requests. +** *`runtime_mappings` (Optional, Record)*: Defines ad-hoc runtime fields in the request similar to the way it is done in search requests. These fields exist only as part of the query and take precedence over fields defined with the same name in the index mappings. ** *`allow_no_indices` (Optional, boolean)*: If false, the request returns an error if any wildcard expression, index alias, or `_all` value targets only missing or closed indices. This behavior applies even if the request targets other open indices. For example, a request @@ -659,6 +659,7 @@ client.openPointInTime({ index, keep_alive }) * *Request (object):* ** *`index` (string | string[])*: A list of index names to open point in time; use `_all` or empty string to perform the operation on all indices ** *`keep_alive` (string | -1 | 0)*: Extends the time to live of the corresponding point in time. +** *`index_filter` (Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, knn, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, rule, script, script_score, semantic, shape, simple_query_string, span_containing, span_field_masking, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, sparse_vector, term, terms, terms_set, text_expansion, weighted_tokens, wildcard, wrapper, type })*: Allows to filter indices if the provided query rewrites to `match_none` on every shard. ** *`ignore_unavailable` (Optional, boolean)*: If `false`, the request returns an error if it targets a missing or closed index. ** *`preference` (Optional, string)*: Specifies the node or shard the operation should be performed on. Random by default. @@ -833,7 +834,7 @@ client.search({ ... }) ** *`index` (Optional, string | string[])*: List of data streams, indices, and aliases to search. Supports wildcards (`*`). To search all data streams and indices, omit this parameter or use `*` or `_all`. -** *`aggregations` (Optional, Record)*: Defines the aggregations that are run as part of the search request. +** *`aggregations` (Optional, Record)*: Defines the aggregations that are run as part of the search request. ** *`collapse` (Optional, { field, inner_hits, max_concurrent_group_searches, collapse })*: Collapses search results the values of the specified field. ** *`explain` (Optional, boolean)*: If true, returns detailed information about score computation as part of a hit. ** *`ext` (Optional, Record)*: Configuration of search extensions defined by Elasticsearch plugins. @@ -849,7 +850,6 @@ If `false`, the response does not include the total number of hits matching the ** *`docvalue_fields` (Optional, { field, format, include_unmapped }[])*: Array of wildcard (`*`) patterns. The request returns doc values for field names matching these patterns in the `hits.fields` property of the response. ** *`knn` (Optional, { field, query_vector, query_vector_builder, k, num_candidates, boost, filter, similarity, inner_hits } | { field, query_vector, query_vector_builder, k, num_candidates, boost, filter, similarity, inner_hits }[])*: Defines the approximate kNN search to run. -** *`rank` (Optional, { rrf })*: Defines the Reciprocal Rank Fusion (RRF) to use. ** *`min_score` (Optional, number)*: Minimum `_score` for matching documents. Documents with a lower `_score` are not included in the search results. ** *`post_filter` (Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, knn, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, rule, script, script_score, semantic, shape, simple_query_string, span_containing, span_field_masking, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, sparse_vector, term, terms, terms_set, text_expansion, weighted_tokens, wildcard, wrapper, type })*: Use the `post_filter` parameter to filter search results. @@ -892,7 +892,7 @@ If this field is specified, the `_source` parameter defaults to `false`. You can pass `_source: true` to return both source fields and stored fields in the search response. ** *`pit` (Optional, { id, keep_alive })*: Limits the search to a point in time (PIT). If you provide a PIT, you cannot specify an `` in the request path. -** *`runtime_mappings` (Optional, Record)*: Defines one or more runtime fields in the search request. +** *`runtime_mappings` (Optional, Record)*: Defines one or more runtime fields in the search request. These fields take precedence over mapped fields with the same name. ** *`stats` (Optional, string[])*: Stats groups to associate with the search. Each group maintains a statistics aggregation for its associated searches. @@ -988,7 +988,7 @@ client.searchMvt({ index, field, zoom, x, y }) ** *`zoom` (number)*: Zoom level for the vector tile to search ** *`x` (number)*: X coordinate for the vector tile to search ** *`y` (number)*: Y coordinate for the vector tile to search -** *`aggs` (Optional, Record)*: Sub-aggregations for the geotile_grid. +** *`aggs` (Optional, Record)*: Sub-aggregations for the geotile_grid. Supports the following aggregation types: - avg @@ -1016,7 +1016,7 @@ each feature represents a geotile_grid cell. If 'grid' each feature is a Polygon of the cells bounding box. If 'point' each feature is a Point that is the centroid of the cell. ** *`query` (Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, knn, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, rule, script, script_score, semantic, shape, simple_query_string, span_containing, span_field_masking, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, sparse_vector, term, terms, terms_set, text_expansion, weighted_tokens, wildcard, wrapper, type })*: Query DSL used to filter documents for the search. -** *`runtime_mappings` (Optional, Record)*: Defines one or more runtime fields in the search request. These fields take +** *`runtime_mappings` (Optional, Record)*: Defines one or more runtime fields in the search request. These fields take precedence over mapped fields with the same name. ** *`size` (Optional, number)*: Maximum number of features to return in the hits layer. Accepts 0-10000. If 0, results don’t include the hits layer. @@ -1324,7 +1324,7 @@ client.asyncSearch.submit({ ... }) * *Request (object):* ** *`index` (Optional, string | string[])*: A list of index names to search; use `_all` or empty string to perform the operation on all indices -** *`aggregations` (Optional, Record)* +** *`aggregations` (Optional, Record)* ** *`collapse` (Optional, { field, inner_hits, max_concurrent_group_searches, collapse })* ** *`explain` (Optional, boolean)*: If true, returns detailed information about score computation as part of a hit. ** *`ext` (Optional, Record)*: Configuration of search extensions defined by Elasticsearch plugins. @@ -1374,7 +1374,7 @@ parameter defaults to false. You can pass _source: true to return both source fi and stored fields in the search response. ** *`pit` (Optional, { id, keep_alive })*: Limits the search to a point in time (PIT). If you provide a PIT, you cannot specify an in the request path. -** *`runtime_mappings` (Optional, Record)*: Defines one or more runtime fields in the search request. These fields take +** *`runtime_mappings` (Optional, Record)*: Defines one or more runtime fields in the search request. These fields take precedence over mapped fields with the same name. ** *`stats` (Optional, string[])*: Stats groups to associate with the search. Each group maintains a statistics aggregation for its associated searches. You can retrieve these stats using @@ -2418,7 +2418,7 @@ client.eql.search({ index, query }) ** *`size` (Optional, number)*: For basic queries, the maximum number of matching events to return. Defaults to 10 ** *`fields` (Optional, { field, format, include_unmapped } | { field, format, include_unmapped }[])*: Array of wildcard (*) patterns. The response returns values for field names matching these patterns in the fields property of each hit. ** *`result_position` (Optional, Enum("tail" | "head"))* -** *`runtime_mappings` (Optional, Record)* +** *`runtime_mappings` (Optional, Record)* ** *`allow_no_indices` (Optional, boolean)* ** *`expand_wildcards` (Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[])* ** *`ignore_unavailable` (Optional, boolean)*: If true, missing or closed indices are not included in the response. @@ -2450,7 +2450,7 @@ and its format can change at any time but it can give some insight into the perf of each part of the query. ** *`tables` (Optional, Record>)*: Tables to use with the LOOKUP operation. The top level key is the table name and the next level key is the column name. -** *`format` (Optional, string)*: A short version of the Accept header, e.g. json, yaml. +** *`format` (Optional, Enum("csv" | "json" | "tsv" | "txt" | "yaml" | "cbor" | "smile" | "arrow"))*: A short version of the Accept header, e.g. json, yaml. ** *`delimiter` (Optional, string)*: The character to use between values within a CSV row. Only valid for the CSV format. ** *`drop_null_columns` (Optional, boolean)*: Should columns that are entirely `null` be removed from the `columns` and `values` portion of the results? Defaults to `false`. If `true` then the response will include an extra section under the name `all_columns` which has the name of all columns. @@ -2535,7 +2535,7 @@ If specified, the `analyzer` parameter overrides this value. ** *`normalizer` (Optional, string)*: Normalizer to use to convert text into a single token. ** *`text` (Optional, string | string[])*: Text to analyze. If an array of strings is provided, it is analyzed as a multi-value field. -** *`tokenizer` (Optional, string | { type, tokenize_on_chars, max_token_length } | { type, custom_token_chars, max_gram, min_gram, token_chars } | { type, buffer_size } | { type } | { type } | { type, custom_token_chars, max_gram, min_gram, token_chars } | { type, decompound_mode, discard_punctuation, user_dictionary, user_dictionary_rules } | { type, buffer_size, delimiter, replacement, reverse, skip } | { type, max_token_length } | { type, max_token_length } | { type, max_token_length } | { type, discard_punctuation, mode, nbest_cost, nbest_examples, user_dictionary, user_dictionary_rules, discard_compound_token } | { type, flags, group, pattern } | { type, rule_files })*: Tokenizer to use to convert text into tokens. +** *`tokenizer` (Optional, string | { type, tokenize_on_chars, max_token_length } | { type, max_token_length } | { type, custom_token_chars, max_gram, min_gram, token_chars } | { type, buffer_size } | { type } | { type } | { type, custom_token_chars, max_gram, min_gram, token_chars } | { type, buffer_size, delimiter, replacement, reverse, skip } | { type, flags, group, pattern } | { type, pattern } | { type, pattern } | { type, max_token_length } | { type } | { type, max_token_length } | { type, max_token_length } | { type, rule_files } | { type, discard_punctuation, mode, nbest_cost, nbest_examples, user_dictionary, user_dictionary_rules, discard_compound_token } | { type, decompound_mode, discard_punctuation, user_dictionary, user_dictionary_rules })*: Tokenizer to use to convert text into tokens. [discrete] ==== create @@ -2923,6 +2923,7 @@ Wildcard (`*`) expressions are supported. If omitted, all data streams are retur Supports a list of values, such as `open,hidden`. ** *`include_defaults` (Optional, boolean)*: If true, returns all relevant default configurations for the index template. ** *`master_timeout` (Optional, string | -1 | 0)*: Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. +** *`verbose` (Optional, boolean)*: Whether the maximum timestamp for each data stream should be calculated and returned. [discrete] ==== get_index_template @@ -3209,7 +3210,7 @@ application-specific metadata. - Mapping parameters ** *`_routing` (Optional, { required })*: Enable making a routing value required on indexed documents. ** *`_source` (Optional, { compress, compress_threshold, enabled, excludes, includes, mode })*: Control whether the _source field is enabled on the index. -** *`runtime` (Optional, Record)*: Mapping of runtime fields for the index. +** *`runtime` (Optional, Record)*: Mapping of runtime fields for the index. ** *`allow_no_indices` (Optional, boolean)*: If `false`, the request returns an error if any wildcard expression, index alias, or `_all` value targets only missing or closed indices. This behavior applies even if the request targets other open indices. ** *`expand_wildcards` (Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[])*: Type of index that wildcard patterns can match. @@ -3342,6 +3343,10 @@ Resources on remote clusters can be specified using the ``:`` syn If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. Supports a list of values, such as `open,hidden`. Valid values are: `all`, `open`, `closed`, `hidden`, `none`. +** *`ignore_unavailable` (Optional, boolean)*: If `false`, the request returns an error if it targets a missing or closed index. +** *`allow_no_indices` (Optional, boolean)*: If `false`, the request returns an error if any wildcard expression, index alias, or `_all` value targets only missing or closed indices. +This behavior applies even if the request targets other open indices. +For example, a request targeting `foo*,bar*` returns an error if an index starts with `foo` but no index starts with `bar`. [discrete] ==== rollover @@ -3659,9 +3664,11 @@ client.ingest.putPipeline({ id }) ** *`id` (string)*: ID of the ingest pipeline to create or update. ** *`_meta` (Optional, Record)*: Optional metadata about the ingest pipeline. May have any contents. This map is not automatically generated by Elasticsearch. ** *`description` (Optional, string)*: Description of the ingest pipeline. -** *`on_failure` (Optional, { append, attachment, bytes, circle, convert, csv, date, date_index_name, dissect, dot_expander, drop, enrich, fail, foreach, geoip, grok, gsub, inference, join, json, kv, lowercase, pipeline, remove, rename, reroute, script, set, set_security_user, sort, split, trim, uppercase, urldecode, user_agent }[])*: Processors to run immediately after a processor failure. Each processor supports a processor-level `on_failure` value. If a processor without an `on_failure` value fails, Elasticsearch uses this pipeline-level parameter as a fallback. The processors in this parameter run sequentially in the order specified. Elasticsearch will not attempt to run the pipeline's remaining processors. -** *`processors` (Optional, { append, attachment, bytes, circle, convert, csv, date, date_index_name, dissect, dot_expander, drop, enrich, fail, foreach, geoip, grok, gsub, inference, join, json, kv, lowercase, pipeline, remove, rename, reroute, script, set, set_security_user, sort, split, trim, uppercase, urldecode, user_agent }[])*: Processors used to perform transformations on documents before indexing. Processors run sequentially in the order specified. +** *`on_failure` (Optional, { append, attachment, bytes, circle, convert, csv, date, date_index_name, dissect, dot_expander, drop, enrich, fail, foreach, geo_grid, geoip, grok, gsub, html_strip, inference, join, json, kv, lowercase, pipeline, redact, remove, rename, reroute, script, set, set_security_user, sort, split, trim, uppercase, urldecode, uri_parts, user_agent }[])*: Processors to run immediately after a processor failure. Each processor supports a processor-level `on_failure` value. If a processor without an `on_failure` value fails, Elasticsearch uses this pipeline-level parameter as a fallback. The processors in this parameter run sequentially in the order specified. Elasticsearch will not attempt to run the pipeline's remaining processors. +** *`processors` (Optional, { append, attachment, bytes, circle, convert, csv, date, date_index_name, dissect, dot_expander, drop, enrich, fail, foreach, geo_grid, geoip, grok, gsub, html_strip, inference, join, json, kv, lowercase, pipeline, redact, remove, rename, reroute, script, set, set_security_user, sort, split, trim, uppercase, urldecode, uri_parts, user_agent }[])*: Processors used to perform transformations on documents before indexing. Processors run sequentially in the order specified. ** *`version` (Optional, number)*: Version number used by external systems to track ingest pipelines. This parameter is intended for external systems only. Elasticsearch does not use or validate pipeline version numbers. +** *`deprecated` (Optional, boolean)*: Marks this ingest pipeline as deprecated. +When a deprecated ingest pipeline is referenced as the default or final pipeline when creating or updating a non-deprecated index template, Elasticsearch will emit a deprecation warning. ** *`master_timeout` (Optional, string | -1 | 0)*: Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. ** *`timeout` (Optional, string | -1 | 0)*: Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error. ** *`if_version` (Optional, number)*: Required version for optimistic concurrency control for pipeline updates @@ -3673,17 +3680,17 @@ Executes an ingest pipeline against a set of provided documents. {ref}/simulate-pipeline-api.html[Endpoint documentation] [source,ts] ---- -client.ingest.simulate({ ... }) +client.ingest.simulate({ docs }) ---- [discrete] ==== Arguments * *Request (object):* +** *`docs` ({ _id, _index, _source }[])*: Sample documents to test in the pipeline. ** *`id` (Optional, string)*: Pipeline to test. If you don’t specify a `pipeline` in the request body, this parameter is required. -** *`docs` (Optional, { _id, _index, _source }[])*: Sample documents to test in the pipeline. -** *`pipeline` (Optional, { description, on_failure, processors, version, _meta })*: Pipeline to test. +** *`pipeline` (Optional, { description, on_failure, processors, version, deprecated, _meta })*: Pipeline to test. If you don’t specify the `pipeline` request path parameter, this parameter is required. If you specify both this and the request path parameter, the API only uses the request path parameter. ** *`verbose` (Optional, boolean)*: If `true`, the response includes output data for each processor in the executed pipeline. @@ -3759,7 +3766,7 @@ client.logstash.putPipeline({ id }) * *Request (object):* ** *`id` (string)*: Identifier for the pipeline. -** *`pipeline` (Optional, { description, on_failure, processors, version, _meta })* +** *`pipeline` (Optional, { description, on_failure, processors, version, deprecated, _meta })* [discrete] === ml @@ -4672,7 +4679,7 @@ client.ml.putDatafeed({ datafeed_id }) ** *`datafeed_id` (string)*: A numerical character string that uniquely identifies the datafeed. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -** *`aggregations` (Optional, Record)*: If set, the datafeed performs aggregation searches. +** *`aggregations` (Optional, Record)*: If set, the datafeed performs aggregation searches. Support for aggregations is limited and should be used only with low cardinality data. ** *`chunking_config` (Optional, { mode, time_span })*: Datafeeds might be required to search over long time periods, for several months or years. This search is split into time chunks in order to ensure the load on Elasticsearch is managed. @@ -4703,7 +4710,7 @@ object is passed verbatim to Elasticsearch. not be searchable in Elasticsearch until 10:06 a.m., set this property to 120 seconds. The default value is randomly selected between `60s` and `120s`. This randomness improves the query performance when there are multiple jobs running on the same node. -** *`runtime_mappings` (Optional, Record)*: Specifies runtime fields for the datafeed search. +** *`runtime_mappings` (Optional, Record)*: Specifies runtime fields for the datafeed search. ** *`script_fields` (Optional, Record)*: Specifies scripts that evaluate custom expressions and returns script fields to the datafeed. The detector configuration objects in a job can contain functions that use these script fields. ** *`scroll_size` (Optional, number)*: The size parameter that is used in Elasticsearch searches when the datafeed does not use aggregations. @@ -5149,7 +5156,7 @@ client.ml.updateDatafeed({ datafeed_id }) ** *`datafeed_id` (string)*: A numerical character string that uniquely identifies the datafeed. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -** *`aggregations` (Optional, Record)*: If set, the datafeed performs aggregation searches. Support for aggregations is limited and should be used only +** *`aggregations` (Optional, Record)*: If set, the datafeed performs aggregation searches. Support for aggregations is limited and should be used only with low cardinality data. ** *`chunking_config` (Optional, { mode, time_span })*: Datafeeds might search over long time periods, for several months or years. This search is split into time chunks in order to ensure the load on Elasticsearch is managed. Chunking configuration controls how the size of @@ -5183,7 +5190,7 @@ when you are satisfied with the results of the job. not be searchable in Elasticsearch until 10:06 a.m., set this property to 120 seconds. The default value is randomly selected between `60s` and `120s`. This randomness improves the query performance when there are multiple jobs running on the same node. -** *`runtime_mappings` (Optional, Record)*: Specifies runtime fields for the datafeed search. +** *`runtime_mappings` (Optional, Record)*: Specifies runtime fields for the datafeed search. ** *`script_fields` (Optional, Record)*: Specifies scripts that evaluate custom expressions and returns script fields to the datafeed. The detector configuration objects in a job can contain functions that use these script fields. ** *`scroll_size` (Optional, number)*: The size parameter that is used in Elasticsearch searches when the datafeed does not use aggregations. @@ -5868,7 +5875,7 @@ It ignores other request body parameters. ** *`page_timeout` (Optional, string | -1 | 0)*: The timeout before a pagination request fails. ** *`time_zone` (Optional, string)*: ISO-8601 time zone ID for the search. ** *`field_multi_value_leniency` (Optional, boolean)*: Throw an exception when encountering multiple values for a field (default) or be lenient and return the first value from the list (without any guarantees of what that will be - typically the first in natural ascending order). -** *`runtime_mappings` (Optional, Record)*: Defines one or more runtime fields in the search request. These fields take +** *`runtime_mappings` (Optional, Record)*: Defines one or more runtime fields in the search request. These fields take precedence over mapped fields with the same name. ** *`wait_for_completion_timeout` (Optional, string | -1 | 0)*: Period to wait for complete results. Defaults to no timeout, meaning the request waits for complete search results. If the search doesn’t finish within this period, the search becomes async. ** *`params` (Optional, Record)*: Values for parameters in the query. diff --git a/src/api/api/ingest.ts b/src/api/api/ingest.ts index b26c024..f98bc1c 100644 --- a/src/api/api/ingest.ts +++ b/src/api/api/ingest.ts @@ -155,7 +155,7 @@ export default class Ingest { async putPipeline (this: That, params: T.IngestPutPipelineRequest | TB.IngestPutPipelineRequest, options?: TransportRequestOptions): Promise async putPipeline (this: That, params: T.IngestPutPipelineRequest | TB.IngestPutPipelineRequest, options?: TransportRequestOptions): Promise { const acceptedPath: string[] = ['id'] - const acceptedBody: string[] = ['_meta', 'description', 'on_failure', 'processors', 'version'] + const acceptedBody: string[] = ['_meta', 'description', 'on_failure', 'processors', 'version', 'deprecated'] const querystring: Record = {} // @ts-expect-error const userBody: any = params?.body @@ -194,10 +194,10 @@ export default class Ingest { * Executes an ingest pipeline against a set of provided documents. * @see {@link https://www.elastic.co/guide/en/elasticsearch/reference/master/simulate-pipeline-api.html | Elasticsearch API documentation} */ - async simulate (this: That, params?: T.IngestSimulateRequest | TB.IngestSimulateRequest, options?: TransportRequestOptionsWithOutMeta): Promise - async simulate (this: That, params?: T.IngestSimulateRequest | TB.IngestSimulateRequest, options?: TransportRequestOptionsWithMeta): Promise> - async simulate (this: That, params?: T.IngestSimulateRequest | TB.IngestSimulateRequest, options?: TransportRequestOptions): Promise - async simulate (this: That, params?: T.IngestSimulateRequest | TB.IngestSimulateRequest, options?: TransportRequestOptions): Promise { + async simulate (this: That, params: T.IngestSimulateRequest | TB.IngestSimulateRequest, options?: TransportRequestOptionsWithOutMeta): Promise + async simulate (this: That, params: T.IngestSimulateRequest | TB.IngestSimulateRequest, options?: TransportRequestOptionsWithMeta): Promise> + async simulate (this: That, params: T.IngestSimulateRequest | TB.IngestSimulateRequest, options?: TransportRequestOptions): Promise + async simulate (this: That, params: T.IngestSimulateRequest | TB.IngestSimulateRequest, options?: TransportRequestOptions): Promise { const acceptedPath: string[] = ['id'] const acceptedBody: string[] = ['docs', 'pipeline'] const querystring: Record = {} @@ -210,7 +210,6 @@ export default class Ingest { body = userBody != null ? { ...userBody } : undefined } - params = params ?? {} for (const key in params) { if (acceptedBody.includes(key)) { body = body ?? {} diff --git a/src/api/api/open_point_in_time.ts b/src/api/api/open_point_in_time.ts index f0fdd68..69aabd1 100644 --- a/src/api/api/open_point_in_time.ts +++ b/src/api/api/open_point_in_time.ts @@ -47,11 +47,23 @@ export default async function OpenPointInTimeApi (this: That, params: T.OpenPoin export default async function OpenPointInTimeApi (this: That, params: T.OpenPointInTimeRequest | TB.OpenPointInTimeRequest, options?: TransportRequestOptions): Promise export default async function OpenPointInTimeApi (this: That, params: T.OpenPointInTimeRequest | TB.OpenPointInTimeRequest, options?: TransportRequestOptions): Promise { const acceptedPath: string[] = ['index'] + const acceptedBody: string[] = ['index_filter'] const querystring: Record = {} - const body = undefined + // @ts-expect-error + const userBody: any = params?.body + let body: Record | string + if (typeof userBody === 'string') { + body = userBody + } else { + body = userBody != null ? { ...userBody } : undefined + } for (const key in params) { - if (acceptedPath.includes(key)) { + if (acceptedBody.includes(key)) { + body = body ?? {} + // @ts-expect-error + body[key] = params[key] + } else if (acceptedPath.includes(key)) { continue } else if (key !== 'body') { // @ts-expect-error diff --git a/src/api/api/search.ts b/src/api/api/search.ts index 65dbb6c..1827183 100644 --- a/src/api/api/search.ts +++ b/src/api/api/search.ts @@ -47,7 +47,7 @@ export default async function SearchApi> (this: That, params?: T.SearchRequest | TB.SearchRequest, options?: TransportRequestOptions): Promise> export default async function SearchApi> (this: That, params?: T.SearchRequest | TB.SearchRequest, options?: TransportRequestOptions): Promise { const acceptedPath: string[] = ['index'] - const acceptedBody: string[] = ['aggregations', 'aggs', 'collapse', 'explain', 'ext', 'from', 'highlight', 'track_total_hits', 'indices_boost', 'docvalue_fields', 'knn', 'rank', 'min_score', 'post_filter', 'profile', 'query', 'rescore', 'retriever', 'script_fields', 'search_after', 'size', 'slice', 'sort', '_source', 'fields', 'suggest', 'terminate_after', 'timeout', 'track_scores', 'version', 'seq_no_primary_term', 'stored_fields', 'pit', 'runtime_mappings', 'stats'] + const acceptedBody: string[] = ['aggregations', 'aggs', 'collapse', 'explain', 'ext', 'from', 'highlight', 'track_total_hits', 'indices_boost', 'docvalue_fields', 'knn', 'min_score', 'post_filter', 'profile', 'query', 'rescore', 'retriever', 'script_fields', 'search_after', 'size', 'slice', 'sort', '_source', 'fields', 'suggest', 'terminate_after', 'timeout', 'track_scores', 'version', 'seq_no_primary_term', 'stored_fields', 'pit', 'runtime_mappings', 'stats'] const querystring: Record = {} // @ts-expect-error const userBody: any = params?.body diff --git a/src/api/types.ts b/src/api/types.ts index 6b777ef..e5e79a3 100644 --- a/src/api/types.ts +++ b/src/api/types.ts @@ -882,9 +882,11 @@ export interface OpenPointInTimeRequest extends RequestBase { preference?: string routing?: Routing expand_wildcards?: ExpandWildcards + index_filter?: QueryDslQueryContainer } export interface OpenPointInTimeResponse { + _shards: ShardStatistics id: Id } @@ -1180,7 +1182,6 @@ export interface SearchRequest extends RequestBase { indices_boost?: Record[] docvalue_fields?: (QueryDslFieldAndFormat | Field)[] knn?: KnnSearch | KnnSearch[] - rank?: RankContainer min_score?: double post_filter?: QueryDslQueryContainer profile?: boolean @@ -1748,11 +1749,23 @@ export interface SearchShardsRequest extends RequestBase { } export interface SearchShardsResponse { - nodes: Record + nodes: Record shards: NodeShard[][] indices: Record } +export interface SearchShardsSearchShardsNodeAttributes { + name: NodeName + ephemeral_id: Id + transport_address: TransportAddress + external_id: string + attributes: Record + roles: NodeRoles + version: VersionString + min_index_version: integer + max_index_version: integer +} + export interface SearchShardsShardStoreIndex { aliases?: Name[] filter?: QueryDslQueryContainer @@ -2240,6 +2253,8 @@ export interface GetStats { total: long } +export type GrokPattern = string + export type HealthStatus = 'green' | 'GREEN' | 'yellow' | 'YELLOW' | 'red' | 'RED' export type Host = string @@ -2394,8 +2409,6 @@ export interface NodeAttributes { id?: NodeId name: NodeName transport_address: TransportAddress - roles?: NodeRoles - external_id?: string } export type NodeId = string @@ -2845,7 +2858,7 @@ export interface AggregationsAdjacencyMatrixBucketKeys extends AggregationsMulti export type AggregationsAdjacencyMatrixBucket = AggregationsAdjacencyMatrixBucketKeys & { [property: string]: AggregationsAggregate | string | long } -export type AggregationsAggregate = AggregationsCardinalityAggregate | AggregationsHdrPercentilesAggregate | AggregationsHdrPercentileRanksAggregate | AggregationsTDigestPercentilesAggregate | AggregationsTDigestPercentileRanksAggregate | AggregationsPercentilesBucketAggregate | AggregationsMedianAbsoluteDeviationAggregate | AggregationsMinAggregate | AggregationsMaxAggregate | AggregationsSumAggregate | AggregationsAvgAggregate | AggregationsWeightedAvgAggregate | AggregationsValueCountAggregate | AggregationsSimpleValueAggregate | AggregationsDerivativeAggregate | AggregationsBucketMetricValueAggregate | AggregationsStatsAggregate | AggregationsStatsBucketAggregate | AggregationsExtendedStatsAggregate | AggregationsExtendedStatsBucketAggregate | AggregationsGeoBoundsAggregate | AggregationsGeoCentroidAggregate | AggregationsHistogramAggregate | AggregationsDateHistogramAggregate | AggregationsAutoDateHistogramAggregate | AggregationsVariableWidthHistogramAggregate | AggregationsStringTermsAggregate | AggregationsLongTermsAggregate | AggregationsDoubleTermsAggregate | AggregationsUnmappedTermsAggregate | AggregationsLongRareTermsAggregate | AggregationsStringRareTermsAggregate | AggregationsUnmappedRareTermsAggregate | AggregationsMultiTermsAggregate | AggregationsMissingAggregate | AggregationsNestedAggregate | AggregationsReverseNestedAggregate | AggregationsGlobalAggregate | AggregationsFilterAggregate | AggregationsChildrenAggregate | AggregationsParentAggregate | AggregationsSamplerAggregate | AggregationsUnmappedSamplerAggregate | AggregationsGeoHashGridAggregate | AggregationsGeoTileGridAggregate | AggregationsGeoHexGridAggregate | AggregationsRangeAggregate | AggregationsDateRangeAggregate | AggregationsGeoDistanceAggregate | AggregationsIpRangeAggregate | AggregationsIpPrefixAggregate | AggregationsFiltersAggregate | AggregationsAdjacencyMatrixAggregate | AggregationsSignificantLongTermsAggregate | AggregationsSignificantStringTermsAggregate | AggregationsUnmappedSignificantTermsAggregate | AggregationsCompositeAggregate | AggregationsFrequentItemSetsAggregate | AggregationsScriptedMetricAggregate | AggregationsTopHitsAggregate | AggregationsInferenceAggregate | AggregationsStringStatsAggregate | AggregationsBoxPlotAggregate | AggregationsTopMetricsAggregate | AggregationsTTestAggregate | AggregationsRateAggregate | AggregationsCumulativeCardinalityAggregate | AggregationsMatrixStatsAggregate | AggregationsGeoLineAggregate +export type AggregationsAggregate = AggregationsCardinalityAggregate | AggregationsHdrPercentilesAggregate | AggregationsHdrPercentileRanksAggregate | AggregationsTDigestPercentilesAggregate | AggregationsTDigestPercentileRanksAggregate | AggregationsPercentilesBucketAggregate | AggregationsMedianAbsoluteDeviationAggregate | AggregationsMinAggregate | AggregationsMaxAggregate | AggregationsSumAggregate | AggregationsAvgAggregate | AggregationsWeightedAvgAggregate | AggregationsValueCountAggregate | AggregationsSimpleValueAggregate | AggregationsDerivativeAggregate | AggregationsBucketMetricValueAggregate | AggregationsStatsAggregate | AggregationsStatsBucketAggregate | AggregationsExtendedStatsAggregate | AggregationsExtendedStatsBucketAggregate | AggregationsGeoBoundsAggregate | AggregationsGeoCentroidAggregate | AggregationsHistogramAggregate | AggregationsDateHistogramAggregate | AggregationsAutoDateHistogramAggregate | AggregationsVariableWidthHistogramAggregate | AggregationsStringTermsAggregate | AggregationsLongTermsAggregate | AggregationsDoubleTermsAggregate | AggregationsUnmappedTermsAggregate | AggregationsLongRareTermsAggregate | AggregationsStringRareTermsAggregate | AggregationsUnmappedRareTermsAggregate | AggregationsMultiTermsAggregate | AggregationsMissingAggregate | AggregationsNestedAggregate | AggregationsReverseNestedAggregate | AggregationsGlobalAggregate | AggregationsFilterAggregate | AggregationsChildrenAggregate | AggregationsParentAggregate | AggregationsSamplerAggregate | AggregationsUnmappedSamplerAggregate | AggregationsGeoHashGridAggregate | AggregationsGeoTileGridAggregate | AggregationsGeoHexGridAggregate | AggregationsRangeAggregate | AggregationsDateRangeAggregate | AggregationsGeoDistanceAggregate | AggregationsIpRangeAggregate | AggregationsIpPrefixAggregate | AggregationsFiltersAggregate | AggregationsAdjacencyMatrixAggregate | AggregationsSignificantLongTermsAggregate | AggregationsSignificantStringTermsAggregate | AggregationsUnmappedSignificantTermsAggregate | AggregationsCompositeAggregate | AggregationsFrequentItemSetsAggregate | AggregationsTimeSeriesAggregate | AggregationsScriptedMetricAggregate | AggregationsTopHitsAggregate | AggregationsInferenceAggregate | AggregationsStringStatsAggregate | AggregationsBoxPlotAggregate | AggregationsTopMetricsAggregate | AggregationsTTestAggregate | AggregationsRateAggregate | AggregationsCumulativeCardinalityAggregate | AggregationsMatrixStatsAggregate | AggregationsGeoLineAggregate export interface AggregationsAggregateBase { meta?: Metadata @@ -2930,6 +2943,7 @@ export interface AggregationsAggregationContainer { sum?: AggregationsSumAggregation sum_bucket?: AggregationsSumBucketAggregation terms?: AggregationsTermsAggregation + time_series?: AggregationsTimeSeriesAggregation top_hits?: AggregationsTopHitsAggregation t_test?: AggregationsTTestAggregation top_metrics?: AggregationsTopMetricsAggregation @@ -2939,9 +2953,9 @@ export interface AggregationsAggregationContainer { } export interface AggregationsAggregationRange { - from?: double + from?: double | null key?: string - to?: double + to?: double | null } export interface AggregationsArrayPercentilesItem { @@ -4126,6 +4140,20 @@ export interface AggregationsTestPopulation { filter?: QueryDslQueryContainer } +export interface AggregationsTimeSeriesAggregate extends AggregationsMultiBucketAggregateBase { +} + +export interface AggregationsTimeSeriesAggregation extends AggregationsBucketAggregationBase { + size?: integer + keyed?: boolean +} + +export interface AggregationsTimeSeriesBucketKeys extends AggregationsMultiBucketBase { + key: Record +} +export type AggregationsTimeSeriesBucket = AggregationsTimeSeriesBucketKeys +& { [property: string]: AggregationsAggregate | Record | long } + export interface AggregationsTopHitsAggregate extends AggregationsAggregateBase { hits: SearchHitsMetadata } @@ -4306,6 +4334,11 @@ export interface AnalysisCjkAnalyzer { stopwords_path?: string } +export interface AnalysisClassicTokenizer extends AnalysisTokenizerBase { + type: 'classic' + max_token_length?: integer +} + export interface AnalysisCommonGramsTokenFilter extends AnalysisTokenFilterBase { type: 'common_grams' common_words?: string[] @@ -4392,7 +4425,7 @@ export interface AnalysisEdgeNGramTokenizer extends AnalysisTokenizerBase { custom_token_chars?: string max_gram: integer min_gram: integer - token_chars: AnalysisTokenChar[] + token_chars?: AnalysisTokenChar[] } export interface AnalysisElisionTokenFilter extends AnalysisTokenFilterBase { @@ -4735,7 +4768,7 @@ export interface AnalysisNGramTokenizer extends AnalysisTokenizerBase { custom_token_chars?: string max_gram: integer min_gram: integer - token_chars: AnalysisTokenChar[] + token_chars?: AnalysisTokenChar[] } export interface AnalysisNoriAnalyzer { @@ -4833,11 +4866,11 @@ export type AnalysisPhoneticRuleType = 'approx' | 'exact' export interface AnalysisPhoneticTokenFilter extends AnalysisTokenFilterBase { type: 'phonetic' encoder: AnalysisPhoneticEncoder - languageset: AnalysisPhoneticLanguage | AnalysisPhoneticLanguage[] + languageset?: AnalysisPhoneticLanguage | AnalysisPhoneticLanguage[] max_code_len?: integer - name_type: AnalysisPhoneticNameType + name_type?: AnalysisPhoneticNameType replace?: boolean - rule_type: AnalysisPhoneticRuleType + rule_type?: AnalysisPhoneticRuleType } export interface AnalysisPorterStemTokenFilter extends AnalysisTokenFilterBase { @@ -4900,6 +4933,16 @@ export interface AnalysisSimpleAnalyzer { version?: VersionString } +export interface AnalysisSimplePatternSplitTokenizer extends AnalysisTokenizerBase { + type: 'simple_pattern_split' + pattern?: string +} + +export interface AnalysisSimplePatternTokenizer extends AnalysisTokenizerBase { + type: 'simple_pattern' + pattern?: string +} + export interface AnalysisSnowballAnalyzer { type: 'snowball' version?: VersionString @@ -5007,6 +5050,10 @@ export interface AnalysisThaiAnalyzer { stopwords_path?: string } +export interface AnalysisThaiTokenizer extends AnalysisTokenizerBase { + type: 'thai' +} + export type AnalysisTokenChar = 'letter' | 'digit' | 'whitespace' | 'punctuation' | 'symbol' | 'custom' export type AnalysisTokenFilter = string | AnalysisTokenFilterDefinition @@ -5023,7 +5070,7 @@ export interface AnalysisTokenizerBase { version?: VersionString } -export type AnalysisTokenizerDefinition = AnalysisCharGroupTokenizer | AnalysisEdgeNGramTokenizer | AnalysisKeywordTokenizer | AnalysisLetterTokenizer | AnalysisLowercaseTokenizer | AnalysisNGramTokenizer | AnalysisNoriTokenizer | AnalysisPathHierarchyTokenizer | AnalysisStandardTokenizer | AnalysisUaxEmailUrlTokenizer | AnalysisWhitespaceTokenizer | AnalysisKuromojiTokenizer | AnalysisPatternTokenizer | AnalysisIcuTokenizer +export type AnalysisTokenizerDefinition = AnalysisCharGroupTokenizer | AnalysisClassicTokenizer | AnalysisEdgeNGramTokenizer | AnalysisKeywordTokenizer | AnalysisLetterTokenizer | AnalysisLowercaseTokenizer | AnalysisNGramTokenizer | AnalysisPathHierarchyTokenizer | AnalysisPatternTokenizer | AnalysisSimplePatternTokenizer | AnalysisSimplePatternSplitTokenizer | AnalysisStandardTokenizer | AnalysisThaiTokenizer | AnalysisUaxEmailUrlTokenizer | AnalysisWhitespaceTokenizer | AnalysisIcuTokenizer | AnalysisKuromojiTokenizer | AnalysisNoriTokenizer export interface AnalysisTrimTokenFilter extends AnalysisTokenFilterBase { type: 'trim' @@ -5148,6 +5195,10 @@ export interface MappingCompletionProperty extends MappingDocValuesPropertyBase type: 'completion' } +export interface MappingCompositeSubField { + type: MappingRuntimeFieldType +} + export interface MappingConstantKeywordProperty extends MappingPropertyBase { value?: any type: 'constant_keyword' @@ -5491,6 +5542,7 @@ export interface MappingRoutingField { } export interface MappingRuntimeField { + fields?: Record fetch_fields?: (MappingRuntimeFieldFetchFields | Field)[] format?: string input_field?: Field @@ -6681,39 +6733,39 @@ export interface CatAliasesRequest extends CatCatRequestBase { export type CatAliasesResponse = CatAliasesAliasesRecord[] export interface CatAllocationAllocationRecord { - shards: string - s: string - 'shards.undesired': string | null - 'write_load.forecast': double | null - wlf: double | null - writeLoadForecast: double | null - 'disk.indices.forecast': ByteSize | null - dif: ByteSize | null - diskIndicesForecast: ByteSize | null - 'disk.indices': ByteSize | null - di: ByteSize | null - diskIndices: ByteSize | null - 'disk.used': ByteSize | null - du: ByteSize | null - diskUsed: ByteSize | null - 'disk.avail': ByteSize | null - da: ByteSize | null - diskAvail: ByteSize | null - 'disk.total': ByteSize | null - dt: ByteSize | null - diskTotal: ByteSize | null - 'disk.percent': Percentage | null - dp: Percentage | null - diskPercent: Percentage | null - host: Host | null - h: Host | null - ip: Ip | null - node: string - n: string - 'node.role': string | null - r: string | null - role: string | null - nodeRole: string | null + shards?: string + s?: string + 'shards.undesired'?: string | null + 'write_load.forecast'?: SpecUtilsStringified | null + wlf?: SpecUtilsStringified | null + writeLoadForecast?: SpecUtilsStringified | null + 'disk.indices.forecast'?: ByteSize | null + dif?: ByteSize | null + diskIndicesForecast?: ByteSize | null + 'disk.indices'?: ByteSize | null + di?: ByteSize | null + diskIndices?: ByteSize | null + 'disk.used'?: ByteSize | null + du?: ByteSize | null + diskUsed?: ByteSize | null + 'disk.avail'?: ByteSize | null + da?: ByteSize | null + diskAvail?: ByteSize | null + 'disk.total'?: ByteSize | null + dt?: ByteSize | null + diskTotal?: ByteSize | null + 'disk.percent'?: Percentage | null + dp?: Percentage | null + diskPercent?: Percentage | null + host?: Host | null + h?: Host | null + ip?: Ip | null + node?: string + n?: string + 'node.role'?: string | null + r?: string | null + role?: string | null + nodeRole?: string | null } export interface CatAllocationRequest extends CatCatRequestBase { @@ -6812,6 +6864,10 @@ export interface CatHealthHealthRecord { i?: string 'shards.initializing'?: string shardsInitializing?: string + 'unassign.pri'?: string + up?: string + 'shards.unassigned.primary'?: string + shardsUnassignedPrimary?: string unassign?: string u?: string 'shards.unassigned'?: string @@ -6875,6 +6931,7 @@ export interface CatIndicesIndicesRecord { ss?: string | null storeSize?: string | null 'pri.store.size'?: string | null + 'dataset.size'?: string | null 'completion.size'?: string cs?: string completionSize?: string @@ -7992,6 +8049,7 @@ export interface CatShardsShardsRecord { dc?: string | null store?: string | null sto?: string | null + dataset?: string | null ip?: string | null id?: string node?: string | null @@ -8714,6 +8772,7 @@ export interface ClusterAllocationExplainClusterInfo { export interface ClusterAllocationExplainCurrentNode { id: Id name: Name + roles: NodeRoles attributes: Record transport_address: TransportAddress weight_ranking: integer @@ -8736,6 +8795,7 @@ export interface ClusterAllocationExplainNodeAllocationExplanation { node_decision: ClusterAllocationExplainDecision node_id: Id node_name: Name + roles: NodeRoles store?: ClusterAllocationExplainAllocationStore transport_address: TransportAddress weight_ranking: integer @@ -8867,6 +8927,7 @@ export interface ClusterHealthHealthResponseBody { task_max_waiting_in_queue?: Duration task_max_waiting_in_queue_millis: DurationValue timed_out: boolean + unassigned_primary_shards: integer unassigned_shards: integer } @@ -8880,6 +8941,7 @@ export interface ClusterHealthIndexHealthStats { shards?: Record status: HealthStatus unassigned_shards: integer + unassigned_primary_shards: integer } export interface ClusterHealthRequest extends RequestBase { @@ -8906,6 +8968,7 @@ export interface ClusterHealthShardHealthStats { relocating_shards: integer status: HealthStatus unassigned_shards: integer + unassigned_primary_shards: integer } export interface ClusterInfoRequest extends RequestBase { @@ -9913,8 +9976,11 @@ export interface EnrichStatsCacheStats { node_id: Id count: integer hits: integer + hits_time_in_millis: DurationValue misses: integer + misses_time_in_millis: DurationValue evictions: integer + size_in_bytes: long } export interface EnrichStatsCoordinatorStats { @@ -10034,8 +10100,10 @@ export type EsqlTableValuesLongDouble = double | double[] export type EsqlTableValuesLongValue = long | long[] +export type EsqlQueryEsqlFormat = 'csv' | 'json' | 'tsv' | 'txt' | 'yaml' | 'cbor' | 'smile' | 'arrow' + export interface EsqlQueryRequest extends RequestBase { - format?: string + format?: EsqlQueryEsqlFormat delimiter?: string drop_null_columns?: boolean columnar?: boolean @@ -11443,6 +11511,7 @@ export interface IndicesGetDataStreamRequest extends RequestBase { expand_wildcards?: ExpandWildcards include_defaults?: boolean master_timeout?: Duration + verbose?: boolean } export interface IndicesGetDataStreamResponse { @@ -11827,6 +11896,8 @@ export type IndicesResolveClusterResponse = Record - patterns: string[] + patterns: GrokPattern[] trace_match?: boolean } @@ -12520,6 +12614,12 @@ export interface IngestGsubProcessor extends IngestProcessorBase { target_field?: Field } +export interface IngestHtmlStripProcessor extends IngestProcessorBase { + field: Field + ignore_missing?: boolean + target_field?: Field +} + export interface IngestInferenceConfig { regression?: IngestInferenceConfigRegression classification?: IngestInferenceConfigClassification @@ -12581,11 +12681,16 @@ export interface IngestLowercaseProcessor extends IngestProcessorBase { target_field?: Field } +export interface IngestMaxmind { + account_id: Id +} + export interface IngestPipeline { description?: string on_failure?: IngestProcessorContainer[] processors?: IngestProcessorContainer[] version?: VersionNumber + deprecated?: boolean _meta?: Metadata } @@ -12623,15 +12728,18 @@ export interface IngestProcessorContainer { enrich?: IngestEnrichProcessor fail?: IngestFailProcessor foreach?: IngestForeachProcessor + geo_grid?: IngestGeoGridProcessor geoip?: IngestGeoIpProcessor grok?: IngestGrokProcessor gsub?: IngestGsubProcessor + html_strip?: IngestHtmlStripProcessor inference?: IngestInferenceProcessor join?: IngestJoinProcessor json?: IngestJsonProcessor kv?: IngestKeyValueProcessor lowercase?: IngestLowercaseProcessor pipeline?: IngestPipelineProcessor + redact?: IngestRedactProcessor remove?: IngestRemoveProcessor rename?: IngestRenameProcessor reroute?: IngestRerouteProcessor @@ -12643,9 +12751,20 @@ export interface IngestProcessorContainer { trim?: IngestTrimProcessor uppercase?: IngestUppercaseProcessor urldecode?: IngestUrlDecodeProcessor + uri_parts?: IngestUriPartsProcessor user_agent?: IngestUserAgentProcessor } +export interface IngestRedactProcessor extends IngestProcessorBase { + field: Field + patterns: GrokPattern[] + pattern_definitions?: Record + prefix?: string + suffix?: string + ignore_missing?: boolean + skip_if_unlicensed?: boolean +} + export interface IngestRemoveProcessor extends IngestProcessorBase { field: Fields keep?: Fields @@ -12713,6 +12832,14 @@ export interface IngestUppercaseProcessor extends IngestProcessorBase { target_field?: Field } +export interface IngestUriPartsProcessor extends IngestProcessorBase { + field: Field + ignore_missing?: boolean + keep_original?: boolean + remove_if_successful?: boolean + target_field?: Field +} + export interface IngestUrlDecodeProcessor extends IngestProcessorBase { field: Field ignore_missing?: boolean @@ -12722,12 +12849,21 @@ export interface IngestUrlDecodeProcessor extends IngestProcessorBase { export interface IngestUserAgentProcessor extends IngestProcessorBase { field: Field ignore_missing?: boolean - options?: IngestUserAgentProperty[] regex_file?: string target_field?: Field + properties?: IngestUserAgentProperty[] + extract_device_type?: boolean } -export type IngestUserAgentProperty = 'NAME' | 'MAJOR' | 'MINOR' | 'PATCH' | 'OS' | 'OS_NAME' | 'OS_MAJOR' | 'OS_MINOR' | 'DEVICE' | 'BUILD' +export type IngestUserAgentProperty = 'name' | 'os' | 'device' | 'original' | 'version' + +export interface IngestDeleteGeoipDatabaseRequest extends RequestBase { + id: Ids + master_timeout?: Duration + timeout?: Duration +} + +export type IngestDeleteGeoipDatabaseResponse = AcknowledgedResponseBase export interface IngestDeletePipelineRequest extends RequestBase { id: Id @@ -12741,8 +12877,9 @@ export interface IngestGeoIpStatsGeoIpDownloadStatistics { successful_downloads: integer failed_downloads: integer total_download_time: DurationValue - database_count: integer + databases_count: integer skipped_updates: integer + expired_databases: integer } export interface IngestGeoIpStatsGeoIpNodeDatabaseName { @@ -12762,6 +12899,22 @@ export interface IngestGeoIpStatsResponse { nodes: Record } +export interface IngestGetGeoipDatabaseDatabaseConfigurationMetadata { + id: Id + version: long + modified_date_millis: EpochTime + database: IngestDatabaseConfiguration +} + +export interface IngestGetGeoipDatabaseRequest extends RequestBase { + id?: Ids + master_timeout?: Duration +} + +export interface IngestGetGeoipDatabaseResponse { + databases: IngestGetGeoipDatabaseDatabaseConfigurationMetadata[] +} + export interface IngestGetPipelineRequest extends RequestBase { id?: Id master_timeout?: Duration @@ -12777,6 +12930,16 @@ export interface IngestProcessorGrokResponse { patterns: Record } +export interface IngestPutGeoipDatabaseRequest extends RequestBase { + id: Id + master_timeout?: Duration + timeout?: Duration + name: Name + maxmind: IngestMaxmind +} + +export type IngestPutGeoipDatabaseResponse = AcknowledgedResponseBase + export interface IngestPutPipelineRequest extends RequestBase { id: Id master_timeout?: Duration @@ -12787,6 +12950,7 @@ export interface IngestPutPipelineRequest extends RequestBase { on_failure?: IngestProcessorContainer[] processors?: IngestProcessorContainer[] version?: VersionNumber + deprecated?: boolean } export type IngestPutPipelineResponse = AcknowledgedResponseBase @@ -12816,21 +12980,29 @@ export interface IngestSimulateIngest { export interface IngestSimulatePipelineSimulation { doc?: IngestSimulateDocumentSimulation - processor_results?: IngestSimulatePipelineSimulation[] tag?: string processor_type?: string status?: WatcherActionStatusOptions + description?: string + ignored_error?: ErrorCause + error?: ErrorCause } export interface IngestSimulateRequest extends RequestBase { id?: Id verbose?: boolean - docs?: IngestSimulateDocument[] + docs: IngestSimulateDocument[] pipeline?: IngestPipeline } export interface IngestSimulateResponse { - docs: IngestSimulatePipelineSimulation[] + docs: IngestSimulateSimulateDocumentResult[] +} + +export interface IngestSimulateSimulateDocumentResult { + doc?: IngestSimulateDocumentSimulation + error?: ErrorCause + processor_results?: IngestSimulatePipelineSimulation[] } export interface LicenseLicense { @@ -13182,7 +13354,7 @@ export type MlCategorizationStatus = 'ok' | 'warn' export interface MlCategory { category_id: ulong examples: string[] - grok_pattern?: string + grok_pattern?: GrokPattern job_id: Id max_matching_length: ulong partition_field_name?: string @@ -15661,6 +15833,24 @@ export interface NodesHttp { clients?: NodesClient[] } +export interface NodesHttpRoute { + requests: NodesHttpRouteRequests + responses: NodesHttpRouteResponses +} + +export interface NodesHttpRouteRequests { + count: long + total_size_in_bytes: long + size_histogram: NodesSizeHttpHistogram[] +} + +export interface NodesHttpRouteResponses { + count: long + total_size_in_bytes: long + handling_time_histogram: NodesTimeHttpHistogram[] + size_histogram: NodesSizeHttpHistogram[] +} + export interface NodesIndexingPressure { memory?: NodesIndexingPressureMemory } @@ -15673,16 +15863,25 @@ export interface NodesIndexingPressureMemory { } export interface NodesIngest { - pipelines?: Record + pipelines?: Record total?: NodesIngestTotal } +export interface NodesIngestStats { + count: long + current: long + failed: long + processors: Record[] + time_in_millis: DurationValue + ingested_as_first_pipeline_in_bytes: long + produced_as_first_pipeline_in_bytes: long +} + export interface NodesIngestTotal { - count?: long - current?: long - failed?: long - processors?: Record[] - time_in_millis?: DurationValue + count: long + current: long + failed: long + time_in_millis: DurationValue } export interface NodesIoStatDevice { @@ -15887,6 +16086,12 @@ export interface NodesSerializedClusterStateDetail { compressed_size_in_bytes?: long } +export interface NodesSizeHttpHistogram { + count: long + ge_bytes?: long + lt_bytes?: long +} + export interface NodesStats { adaptive_selection?: Record breakers?: Record @@ -15921,6 +16126,12 @@ export interface NodesThreadCount { threads?: long } +export interface NodesTimeHttpHistogram { + count: long + ge_millis?: long + lt_millis?: long +} + export interface NodesTransport { inbound_handling_time_histogram?: NodesTransportHistogram[] outbound_handling_time_histogram?: NodesTransportHistogram[] @@ -18673,7 +18884,7 @@ export interface TextStructureFindStructureRequest { ecs_compatibility?: string explain?: boolean format?: string - grok_pattern?: string + grok_pattern?: GrokPattern has_header_row?: boolean line_merge_size_limit?: uint lines_to_sample?: uint @@ -18700,7 +18911,7 @@ export interface TextStructureFindStructureResponse { num_lines_analyzed: integer column_names?: string[] explanation?: string[] - grok_pattern?: string + grok_pattern?: GrokPattern multiline_start_pattern?: string exclude_lines_pattern?: string java_timestamp_formats?: string[] @@ -18728,7 +18939,7 @@ export interface TextStructureTestGrokPatternMatchedText { export interface TextStructureTestGrokPatternRequest extends RequestBase { ecs_compatibility?: string - grok_pattern: string + grok_pattern: GrokPattern text: string[] } diff --git a/src/api/typesWithBodyKey.ts b/src/api/typesWithBodyKey.ts index cf01dc9..1b09d73 100644 --- a/src/api/typesWithBodyKey.ts +++ b/src/api/typesWithBodyKey.ts @@ -915,9 +915,14 @@ export interface OpenPointInTimeRequest extends RequestBase { preference?: string routing?: Routing expand_wildcards?: ExpandWildcards + /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ + body?: { + index_filter?: QueryDslQueryContainer + } } export interface OpenPointInTimeResponse { + _shards: ShardStatistics id: Id } @@ -1234,7 +1239,6 @@ export interface SearchRequest extends RequestBase { indices_boost?: Record[] docvalue_fields?: (QueryDslFieldAndFormat | Field)[] knn?: KnnSearch | KnnSearch[] - rank?: RankContainer min_score?: double post_filter?: QueryDslQueryContainer profile?: boolean @@ -1806,11 +1810,23 @@ export interface SearchShardsRequest extends RequestBase { } export interface SearchShardsResponse { - nodes: Record + nodes: Record shards: NodeShard[][] indices: Record } +export interface SearchShardsSearchShardsNodeAttributes { + name: NodeName + ephemeral_id: Id + transport_address: TransportAddress + external_id: string + attributes: Record + roles: NodeRoles + version: VersionString + min_index_version: integer + max_index_version: integer +} + export interface SearchShardsShardStoreIndex { aliases?: Name[] filter?: QueryDslQueryContainer @@ -2313,6 +2329,8 @@ export interface GetStats { total: long } +export type GrokPattern = string + export type HealthStatus = 'green' | 'GREEN' | 'yellow' | 'YELLOW' | 'red' | 'RED' export type Host = string @@ -2467,8 +2485,6 @@ export interface NodeAttributes { id?: NodeId name: NodeName transport_address: TransportAddress - roles?: NodeRoles - external_id?: string } export type NodeId = string @@ -2918,7 +2934,7 @@ export interface AggregationsAdjacencyMatrixBucketKeys extends AggregationsMulti export type AggregationsAdjacencyMatrixBucket = AggregationsAdjacencyMatrixBucketKeys & { [property: string]: AggregationsAggregate | string | long } -export type AggregationsAggregate = AggregationsCardinalityAggregate | AggregationsHdrPercentilesAggregate | AggregationsHdrPercentileRanksAggregate | AggregationsTDigestPercentilesAggregate | AggregationsTDigestPercentileRanksAggregate | AggregationsPercentilesBucketAggregate | AggregationsMedianAbsoluteDeviationAggregate | AggregationsMinAggregate | AggregationsMaxAggregate | AggregationsSumAggregate | AggregationsAvgAggregate | AggregationsWeightedAvgAggregate | AggregationsValueCountAggregate | AggregationsSimpleValueAggregate | AggregationsDerivativeAggregate | AggregationsBucketMetricValueAggregate | AggregationsStatsAggregate | AggregationsStatsBucketAggregate | AggregationsExtendedStatsAggregate | AggregationsExtendedStatsBucketAggregate | AggregationsGeoBoundsAggregate | AggregationsGeoCentroidAggregate | AggregationsHistogramAggregate | AggregationsDateHistogramAggregate | AggregationsAutoDateHistogramAggregate | AggregationsVariableWidthHistogramAggregate | AggregationsStringTermsAggregate | AggregationsLongTermsAggregate | AggregationsDoubleTermsAggregate | AggregationsUnmappedTermsAggregate | AggregationsLongRareTermsAggregate | AggregationsStringRareTermsAggregate | AggregationsUnmappedRareTermsAggregate | AggregationsMultiTermsAggregate | AggregationsMissingAggregate | AggregationsNestedAggregate | AggregationsReverseNestedAggregate | AggregationsGlobalAggregate | AggregationsFilterAggregate | AggregationsChildrenAggregate | AggregationsParentAggregate | AggregationsSamplerAggregate | AggregationsUnmappedSamplerAggregate | AggregationsGeoHashGridAggregate | AggregationsGeoTileGridAggregate | AggregationsGeoHexGridAggregate | AggregationsRangeAggregate | AggregationsDateRangeAggregate | AggregationsGeoDistanceAggregate | AggregationsIpRangeAggregate | AggregationsIpPrefixAggregate | AggregationsFiltersAggregate | AggregationsAdjacencyMatrixAggregate | AggregationsSignificantLongTermsAggregate | AggregationsSignificantStringTermsAggregate | AggregationsUnmappedSignificantTermsAggregate | AggregationsCompositeAggregate | AggregationsFrequentItemSetsAggregate | AggregationsScriptedMetricAggregate | AggregationsTopHitsAggregate | AggregationsInferenceAggregate | AggregationsStringStatsAggregate | AggregationsBoxPlotAggregate | AggregationsTopMetricsAggregate | AggregationsTTestAggregate | AggregationsRateAggregate | AggregationsCumulativeCardinalityAggregate | AggregationsMatrixStatsAggregate | AggregationsGeoLineAggregate +export type AggregationsAggregate = AggregationsCardinalityAggregate | AggregationsHdrPercentilesAggregate | AggregationsHdrPercentileRanksAggregate | AggregationsTDigestPercentilesAggregate | AggregationsTDigestPercentileRanksAggregate | AggregationsPercentilesBucketAggregate | AggregationsMedianAbsoluteDeviationAggregate | AggregationsMinAggregate | AggregationsMaxAggregate | AggregationsSumAggregate | AggregationsAvgAggregate | AggregationsWeightedAvgAggregate | AggregationsValueCountAggregate | AggregationsSimpleValueAggregate | AggregationsDerivativeAggregate | AggregationsBucketMetricValueAggregate | AggregationsStatsAggregate | AggregationsStatsBucketAggregate | AggregationsExtendedStatsAggregate | AggregationsExtendedStatsBucketAggregate | AggregationsGeoBoundsAggregate | AggregationsGeoCentroidAggregate | AggregationsHistogramAggregate | AggregationsDateHistogramAggregate | AggregationsAutoDateHistogramAggregate | AggregationsVariableWidthHistogramAggregate | AggregationsStringTermsAggregate | AggregationsLongTermsAggregate | AggregationsDoubleTermsAggregate | AggregationsUnmappedTermsAggregate | AggregationsLongRareTermsAggregate | AggregationsStringRareTermsAggregate | AggregationsUnmappedRareTermsAggregate | AggregationsMultiTermsAggregate | AggregationsMissingAggregate | AggregationsNestedAggregate | AggregationsReverseNestedAggregate | AggregationsGlobalAggregate | AggregationsFilterAggregate | AggregationsChildrenAggregate | AggregationsParentAggregate | AggregationsSamplerAggregate | AggregationsUnmappedSamplerAggregate | AggregationsGeoHashGridAggregate | AggregationsGeoTileGridAggregate | AggregationsGeoHexGridAggregate | AggregationsRangeAggregate | AggregationsDateRangeAggregate | AggregationsGeoDistanceAggregate | AggregationsIpRangeAggregate | AggregationsIpPrefixAggregate | AggregationsFiltersAggregate | AggregationsAdjacencyMatrixAggregate | AggregationsSignificantLongTermsAggregate | AggregationsSignificantStringTermsAggregate | AggregationsUnmappedSignificantTermsAggregate | AggregationsCompositeAggregate | AggregationsFrequentItemSetsAggregate | AggregationsTimeSeriesAggregate | AggregationsScriptedMetricAggregate | AggregationsTopHitsAggregate | AggregationsInferenceAggregate | AggregationsStringStatsAggregate | AggregationsBoxPlotAggregate | AggregationsTopMetricsAggregate | AggregationsTTestAggregate | AggregationsRateAggregate | AggregationsCumulativeCardinalityAggregate | AggregationsMatrixStatsAggregate | AggregationsGeoLineAggregate export interface AggregationsAggregateBase { meta?: Metadata @@ -3003,6 +3019,7 @@ export interface AggregationsAggregationContainer { sum?: AggregationsSumAggregation sum_bucket?: AggregationsSumBucketAggregation terms?: AggregationsTermsAggregation + time_series?: AggregationsTimeSeriesAggregation top_hits?: AggregationsTopHitsAggregation t_test?: AggregationsTTestAggregation top_metrics?: AggregationsTopMetricsAggregation @@ -3012,9 +3029,9 @@ export interface AggregationsAggregationContainer { } export interface AggregationsAggregationRange { - from?: double + from?: double | null key?: string - to?: double + to?: double | null } export interface AggregationsArrayPercentilesItem { @@ -4199,6 +4216,20 @@ export interface AggregationsTestPopulation { filter?: QueryDslQueryContainer } +export interface AggregationsTimeSeriesAggregate extends AggregationsMultiBucketAggregateBase { +} + +export interface AggregationsTimeSeriesAggregation extends AggregationsBucketAggregationBase { + size?: integer + keyed?: boolean +} + +export interface AggregationsTimeSeriesBucketKeys extends AggregationsMultiBucketBase { + key: Record +} +export type AggregationsTimeSeriesBucket = AggregationsTimeSeriesBucketKeys +& { [property: string]: AggregationsAggregate | Record | long } + export interface AggregationsTopHitsAggregate extends AggregationsAggregateBase { hits: SearchHitsMetadata } @@ -4379,6 +4410,11 @@ export interface AnalysisCjkAnalyzer { stopwords_path?: string } +export interface AnalysisClassicTokenizer extends AnalysisTokenizerBase { + type: 'classic' + max_token_length?: integer +} + export interface AnalysisCommonGramsTokenFilter extends AnalysisTokenFilterBase { type: 'common_grams' common_words?: string[] @@ -4465,7 +4501,7 @@ export interface AnalysisEdgeNGramTokenizer extends AnalysisTokenizerBase { custom_token_chars?: string max_gram: integer min_gram: integer - token_chars: AnalysisTokenChar[] + token_chars?: AnalysisTokenChar[] } export interface AnalysisElisionTokenFilter extends AnalysisTokenFilterBase { @@ -4808,7 +4844,7 @@ export interface AnalysisNGramTokenizer extends AnalysisTokenizerBase { custom_token_chars?: string max_gram: integer min_gram: integer - token_chars: AnalysisTokenChar[] + token_chars?: AnalysisTokenChar[] } export interface AnalysisNoriAnalyzer { @@ -4906,11 +4942,11 @@ export type AnalysisPhoneticRuleType = 'approx' | 'exact' export interface AnalysisPhoneticTokenFilter extends AnalysisTokenFilterBase { type: 'phonetic' encoder: AnalysisPhoneticEncoder - languageset: AnalysisPhoneticLanguage | AnalysisPhoneticLanguage[] + languageset?: AnalysisPhoneticLanguage | AnalysisPhoneticLanguage[] max_code_len?: integer - name_type: AnalysisPhoneticNameType + name_type?: AnalysisPhoneticNameType replace?: boolean - rule_type: AnalysisPhoneticRuleType + rule_type?: AnalysisPhoneticRuleType } export interface AnalysisPorterStemTokenFilter extends AnalysisTokenFilterBase { @@ -4973,6 +5009,16 @@ export interface AnalysisSimpleAnalyzer { version?: VersionString } +export interface AnalysisSimplePatternSplitTokenizer extends AnalysisTokenizerBase { + type: 'simple_pattern_split' + pattern?: string +} + +export interface AnalysisSimplePatternTokenizer extends AnalysisTokenizerBase { + type: 'simple_pattern' + pattern?: string +} + export interface AnalysisSnowballAnalyzer { type: 'snowball' version?: VersionString @@ -5080,6 +5126,10 @@ export interface AnalysisThaiAnalyzer { stopwords_path?: string } +export interface AnalysisThaiTokenizer extends AnalysisTokenizerBase { + type: 'thai' +} + export type AnalysisTokenChar = 'letter' | 'digit' | 'whitespace' | 'punctuation' | 'symbol' | 'custom' export type AnalysisTokenFilter = string | AnalysisTokenFilterDefinition @@ -5096,7 +5146,7 @@ export interface AnalysisTokenizerBase { version?: VersionString } -export type AnalysisTokenizerDefinition = AnalysisCharGroupTokenizer | AnalysisEdgeNGramTokenizer | AnalysisKeywordTokenizer | AnalysisLetterTokenizer | AnalysisLowercaseTokenizer | AnalysisNGramTokenizer | AnalysisNoriTokenizer | AnalysisPathHierarchyTokenizer | AnalysisStandardTokenizer | AnalysisUaxEmailUrlTokenizer | AnalysisWhitespaceTokenizer | AnalysisKuromojiTokenizer | AnalysisPatternTokenizer | AnalysisIcuTokenizer +export type AnalysisTokenizerDefinition = AnalysisCharGroupTokenizer | AnalysisClassicTokenizer | AnalysisEdgeNGramTokenizer | AnalysisKeywordTokenizer | AnalysisLetterTokenizer | AnalysisLowercaseTokenizer | AnalysisNGramTokenizer | AnalysisPathHierarchyTokenizer | AnalysisPatternTokenizer | AnalysisSimplePatternTokenizer | AnalysisSimplePatternSplitTokenizer | AnalysisStandardTokenizer | AnalysisThaiTokenizer | AnalysisUaxEmailUrlTokenizer | AnalysisWhitespaceTokenizer | AnalysisIcuTokenizer | AnalysisKuromojiTokenizer | AnalysisNoriTokenizer export interface AnalysisTrimTokenFilter extends AnalysisTokenFilterBase { type: 'trim' @@ -5221,6 +5271,10 @@ export interface MappingCompletionProperty extends MappingDocValuesPropertyBase type: 'completion' } +export interface MappingCompositeSubField { + type: MappingRuntimeFieldType +} + export interface MappingConstantKeywordProperty extends MappingPropertyBase { value?: any type: 'constant_keyword' @@ -5564,6 +5618,7 @@ export interface MappingRoutingField { } export interface MappingRuntimeField { + fields?: Record fetch_fields?: (MappingRuntimeFieldFetchFields | Field)[] format?: string input_field?: Field @@ -6758,39 +6813,39 @@ export interface CatAliasesRequest extends CatCatRequestBase { export type CatAliasesResponse = CatAliasesAliasesRecord[] export interface CatAllocationAllocationRecord { - shards: string - s: string - 'shards.undesired': string | null - 'write_load.forecast': double | null - wlf: double | null - writeLoadForecast: double | null - 'disk.indices.forecast': ByteSize | null - dif: ByteSize | null - diskIndicesForecast: ByteSize | null - 'disk.indices': ByteSize | null - di: ByteSize | null - diskIndices: ByteSize | null - 'disk.used': ByteSize | null - du: ByteSize | null - diskUsed: ByteSize | null - 'disk.avail': ByteSize | null - da: ByteSize | null - diskAvail: ByteSize | null - 'disk.total': ByteSize | null - dt: ByteSize | null - diskTotal: ByteSize | null - 'disk.percent': Percentage | null - dp: Percentage | null - diskPercent: Percentage | null - host: Host | null - h: Host | null - ip: Ip | null - node: string - n: string - 'node.role': string | null - r: string | null - role: string | null - nodeRole: string | null + shards?: string + s?: string + 'shards.undesired'?: string | null + 'write_load.forecast'?: SpecUtilsStringified | null + wlf?: SpecUtilsStringified | null + writeLoadForecast?: SpecUtilsStringified | null + 'disk.indices.forecast'?: ByteSize | null + dif?: ByteSize | null + diskIndicesForecast?: ByteSize | null + 'disk.indices'?: ByteSize | null + di?: ByteSize | null + diskIndices?: ByteSize | null + 'disk.used'?: ByteSize | null + du?: ByteSize | null + diskUsed?: ByteSize | null + 'disk.avail'?: ByteSize | null + da?: ByteSize | null + diskAvail?: ByteSize | null + 'disk.total'?: ByteSize | null + dt?: ByteSize | null + diskTotal?: ByteSize | null + 'disk.percent'?: Percentage | null + dp?: Percentage | null + diskPercent?: Percentage | null + host?: Host | null + h?: Host | null + ip?: Ip | null + node?: string + n?: string + 'node.role'?: string | null + r?: string | null + role?: string | null + nodeRole?: string | null } export interface CatAllocationRequest extends CatCatRequestBase { @@ -6889,6 +6944,10 @@ export interface CatHealthHealthRecord { i?: string 'shards.initializing'?: string shardsInitializing?: string + 'unassign.pri'?: string + up?: string + 'shards.unassigned.primary'?: string + shardsUnassignedPrimary?: string unassign?: string u?: string 'shards.unassigned'?: string @@ -6952,6 +7011,7 @@ export interface CatIndicesIndicesRecord { ss?: string | null storeSize?: string | null 'pri.store.size'?: string | null + 'dataset.size'?: string | null 'completion.size'?: string cs?: string completionSize?: string @@ -8069,6 +8129,7 @@ export interface CatShardsShardsRecord { dc?: string | null store?: string | null sto?: string | null + dataset?: string | null ip?: string | null id?: string node?: string | null @@ -8803,6 +8864,7 @@ export interface ClusterAllocationExplainClusterInfo { export interface ClusterAllocationExplainCurrentNode { id: Id name: Name + roles: NodeRoles attributes: Record transport_address: TransportAddress weight_ranking: integer @@ -8825,6 +8887,7 @@ export interface ClusterAllocationExplainNodeAllocationExplanation { node_decision: ClusterAllocationExplainDecision node_id: Id node_name: Name + roles: NodeRoles store?: ClusterAllocationExplainAllocationStore transport_address: TransportAddress weight_ranking: integer @@ -8959,6 +9022,7 @@ export interface ClusterHealthHealthResponseBody { task_max_waiting_in_queue?: Duration task_max_waiting_in_queue_millis: DurationValue timed_out: boolean + unassigned_primary_shards: integer unassigned_shards: integer } @@ -8972,6 +9036,7 @@ export interface ClusterHealthIndexHealthStats { shards?: Record status: HealthStatus unassigned_shards: integer + unassigned_primary_shards: integer } export interface ClusterHealthRequest extends RequestBase { @@ -8998,6 +9063,7 @@ export interface ClusterHealthShardHealthStats { relocating_shards: integer status: HealthStatus unassigned_shards: integer + unassigned_primary_shards: integer } export interface ClusterInfoRequest extends RequestBase { @@ -10065,8 +10131,11 @@ export interface EnrichStatsCacheStats { node_id: Id count: integer hits: integer + hits_time_in_millis: DurationValue misses: integer + misses_time_in_millis: DurationValue evictions: integer + size_in_bytes: long } export interface EnrichStatsCoordinatorStats { @@ -10189,8 +10258,10 @@ export type EsqlTableValuesLongDouble = double | double[] export type EsqlTableValuesLongValue = long | long[] +export type EsqlQueryEsqlFormat = 'csv' | 'json' | 'tsv' | 'txt' | 'yaml' | 'cbor' | 'smile' | 'arrow' + export interface EsqlQueryRequest extends RequestBase { - format?: string + format?: EsqlQueryEsqlFormat delimiter?: string drop_null_columns?: boolean /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ @@ -11627,6 +11698,7 @@ export interface IndicesGetDataStreamRequest extends RequestBase { expand_wildcards?: ExpandWildcards include_defaults?: boolean master_timeout?: Duration + verbose?: boolean } export interface IndicesGetDataStreamResponse { @@ -12030,6 +12102,8 @@ export type IndicesResolveClusterResponse = Record - patterns: string[] + patterns: GrokPattern[] trace_match?: boolean } @@ -12745,6 +12842,12 @@ export interface IngestGsubProcessor extends IngestProcessorBase { target_field?: Field } +export interface IngestHtmlStripProcessor extends IngestProcessorBase { + field: Field + ignore_missing?: boolean + target_field?: Field +} + export interface IngestInferenceConfig { regression?: IngestInferenceConfigRegression classification?: IngestInferenceConfigClassification @@ -12806,11 +12909,16 @@ export interface IngestLowercaseProcessor extends IngestProcessorBase { target_field?: Field } +export interface IngestMaxmind { + account_id: Id +} + export interface IngestPipeline { description?: string on_failure?: IngestProcessorContainer[] processors?: IngestProcessorContainer[] version?: VersionNumber + deprecated?: boolean _meta?: Metadata } @@ -12848,15 +12956,18 @@ export interface IngestProcessorContainer { enrich?: IngestEnrichProcessor fail?: IngestFailProcessor foreach?: IngestForeachProcessor + geo_grid?: IngestGeoGridProcessor geoip?: IngestGeoIpProcessor grok?: IngestGrokProcessor gsub?: IngestGsubProcessor + html_strip?: IngestHtmlStripProcessor inference?: IngestInferenceProcessor join?: IngestJoinProcessor json?: IngestJsonProcessor kv?: IngestKeyValueProcessor lowercase?: IngestLowercaseProcessor pipeline?: IngestPipelineProcessor + redact?: IngestRedactProcessor remove?: IngestRemoveProcessor rename?: IngestRenameProcessor reroute?: IngestRerouteProcessor @@ -12868,9 +12979,20 @@ export interface IngestProcessorContainer { trim?: IngestTrimProcessor uppercase?: IngestUppercaseProcessor urldecode?: IngestUrlDecodeProcessor + uri_parts?: IngestUriPartsProcessor user_agent?: IngestUserAgentProcessor } +export interface IngestRedactProcessor extends IngestProcessorBase { + field: Field + patterns: GrokPattern[] + pattern_definitions?: Record + prefix?: string + suffix?: string + ignore_missing?: boolean + skip_if_unlicensed?: boolean +} + export interface IngestRemoveProcessor extends IngestProcessorBase { field: Fields keep?: Fields @@ -12938,6 +13060,14 @@ export interface IngestUppercaseProcessor extends IngestProcessorBase { target_field?: Field } +export interface IngestUriPartsProcessor extends IngestProcessorBase { + field: Field + ignore_missing?: boolean + keep_original?: boolean + remove_if_successful?: boolean + target_field?: Field +} + export interface IngestUrlDecodeProcessor extends IngestProcessorBase { field: Field ignore_missing?: boolean @@ -12947,12 +13077,21 @@ export interface IngestUrlDecodeProcessor extends IngestProcessorBase { export interface IngestUserAgentProcessor extends IngestProcessorBase { field: Field ignore_missing?: boolean - options?: IngestUserAgentProperty[] regex_file?: string target_field?: Field + properties?: IngestUserAgentProperty[] + extract_device_type?: boolean } -export type IngestUserAgentProperty = 'NAME' | 'MAJOR' | 'MINOR' | 'PATCH' | 'OS' | 'OS_NAME' | 'OS_MAJOR' | 'OS_MINOR' | 'DEVICE' | 'BUILD' +export type IngestUserAgentProperty = 'name' | 'os' | 'device' | 'original' | 'version' + +export interface IngestDeleteGeoipDatabaseRequest extends RequestBase { + id: Ids + master_timeout?: Duration + timeout?: Duration +} + +export type IngestDeleteGeoipDatabaseResponse = AcknowledgedResponseBase export interface IngestDeletePipelineRequest extends RequestBase { id: Id @@ -12966,8 +13105,9 @@ export interface IngestGeoIpStatsGeoIpDownloadStatistics { successful_downloads: integer failed_downloads: integer total_download_time: DurationValue - database_count: integer + databases_count: integer skipped_updates: integer + expired_databases: integer } export interface IngestGeoIpStatsGeoIpNodeDatabaseName { @@ -12987,6 +13127,22 @@ export interface IngestGeoIpStatsResponse { nodes: Record } +export interface IngestGetGeoipDatabaseDatabaseConfigurationMetadata { + id: Id + version: long + modified_date_millis: EpochTime + database: IngestDatabaseConfiguration +} + +export interface IngestGetGeoipDatabaseRequest extends RequestBase { + id?: Ids + master_timeout?: Duration +} + +export interface IngestGetGeoipDatabaseResponse { + databases: IngestGetGeoipDatabaseDatabaseConfigurationMetadata[] +} + export interface IngestGetPipelineRequest extends RequestBase { id?: Id master_timeout?: Duration @@ -13002,6 +13158,19 @@ export interface IngestProcessorGrokResponse { patterns: Record } +export interface IngestPutGeoipDatabaseRequest extends RequestBase { + id: Id + master_timeout?: Duration + timeout?: Duration + /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ + body?: { + name: Name + maxmind: IngestMaxmind + } +} + +export type IngestPutGeoipDatabaseResponse = AcknowledgedResponseBase + export interface IngestPutPipelineRequest extends RequestBase { id: Id master_timeout?: Duration @@ -13014,6 +13183,7 @@ export interface IngestPutPipelineRequest extends RequestBase { on_failure?: IngestProcessorContainer[] processors?: IngestProcessorContainer[] version?: VersionNumber + deprecated?: boolean } } @@ -13044,10 +13214,12 @@ export interface IngestSimulateIngest { export interface IngestSimulatePipelineSimulation { doc?: IngestSimulateDocumentSimulation - processor_results?: IngestSimulatePipelineSimulation[] tag?: string processor_type?: string status?: WatcherActionStatusOptions + description?: string + ignored_error?: ErrorCause + error?: ErrorCause } export interface IngestSimulateRequest extends RequestBase { @@ -13055,13 +13227,19 @@ export interface IngestSimulateRequest extends RequestBase { verbose?: boolean /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { - docs?: IngestSimulateDocument[] + docs: IngestSimulateDocument[] pipeline?: IngestPipeline } } export interface IngestSimulateResponse { - docs: IngestSimulatePipelineSimulation[] + docs: IngestSimulateSimulateDocumentResult[] +} + +export interface IngestSimulateSimulateDocumentResult { + doc?: IngestSimulateDocumentSimulation + error?: ErrorCause + processor_results?: IngestSimulatePipelineSimulation[] } export interface LicenseLicense { @@ -13417,7 +13595,7 @@ export type MlCategorizationStatus = 'ok' | 'warn' export interface MlCategory { category_id: ulong examples: string[] - grok_pattern?: string + grok_pattern?: GrokPattern job_id: Id max_matching_length: ulong partition_field_name?: string @@ -16010,6 +16188,24 @@ export interface NodesHttp { clients?: NodesClient[] } +export interface NodesHttpRoute { + requests: NodesHttpRouteRequests + responses: NodesHttpRouteResponses +} + +export interface NodesHttpRouteRequests { + count: long + total_size_in_bytes: long + size_histogram: NodesSizeHttpHistogram[] +} + +export interface NodesHttpRouteResponses { + count: long + total_size_in_bytes: long + handling_time_histogram: NodesTimeHttpHistogram[] + size_histogram: NodesSizeHttpHistogram[] +} + export interface NodesIndexingPressure { memory?: NodesIndexingPressureMemory } @@ -16022,16 +16218,25 @@ export interface NodesIndexingPressureMemory { } export interface NodesIngest { - pipelines?: Record + pipelines?: Record total?: NodesIngestTotal } +export interface NodesIngestStats { + count: long + current: long + failed: long + processors: Record[] + time_in_millis: DurationValue + ingested_as_first_pipeline_in_bytes: long + produced_as_first_pipeline_in_bytes: long +} + export interface NodesIngestTotal { - count?: long - current?: long - failed?: long - processors?: Record[] - time_in_millis?: DurationValue + count: long + current: long + failed: long + time_in_millis: DurationValue } export interface NodesIoStatDevice { @@ -16236,6 +16441,12 @@ export interface NodesSerializedClusterStateDetail { compressed_size_in_bytes?: long } +export interface NodesSizeHttpHistogram { + count: long + ge_bytes?: long + lt_bytes?: long +} + export interface NodesStats { adaptive_selection?: Record breakers?: Record @@ -16270,6 +16481,12 @@ export interface NodesThreadCount { threads?: long } +export interface NodesTimeHttpHistogram { + count: long + ge_millis?: long + lt_millis?: long +} + export interface NodesTransport { inbound_handling_time_histogram?: NodesTransportHistogram[] outbound_handling_time_histogram?: NodesTransportHistogram[] @@ -19152,7 +19369,7 @@ export interface TextStructureFindStructureRequest { ecs_compatibility?: string explain?: boolean format?: string - grok_pattern?: string + grok_pattern?: GrokPattern has_header_row?: boolean line_merge_size_limit?: uint lines_to_sample?: uint @@ -19180,7 +19397,7 @@ export interface TextStructureFindStructureResponse { num_lines_analyzed: integer column_names?: string[] explanation?: string[] - grok_pattern?: string + grok_pattern?: GrokPattern multiline_start_pattern?: string exclude_lines_pattern?: string java_timestamp_formats?: string[] @@ -19210,7 +19427,7 @@ export interface TextStructureTestGrokPatternRequest extends RequestBase { ecs_compatibility?: string /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { - grok_pattern: string + grok_pattern: GrokPattern text: string[] } }