From ef43205879667b8dbc02af0229368bbf5b61d748 Mon Sep 17 00:00:00 2001 From: Elastic Machine Date: Mon, 7 Jul 2025 19:29:45 +0000 Subject: [PATCH] Auto-generated API code --- docs/reference.asciidoc | 73 +++++-- src/api/api/inference.ts | 80 +++++++- src/api/api/streams.ts | 133 ++++++++++++ src/api/index.ts | 8 + src/api/types.ts | 397 +++++++++++++++++++++++++++++------ src/api/typesWithBodyKey.ts | 400 ++++++++++++++++++++++++++++++------ 6 files changed, 936 insertions(+), 155 deletions(-) create mode 100644 src/api/api/streams.ts diff --git a/docs/reference.asciidoc b/docs/reference.asciidoc index bbd1179ba..eab5e41af 100644 --- a/docs/reference.asciidoc +++ b/docs/reference.asciidoc @@ -862,7 +862,6 @@ client.getSource({ id, index }) ** *`_source` (Optional, boolean | string | string[])*: Indicates whether to return the `_source` field (`true` or `false`) or lists the fields to return. ** *`_source_excludes` (Optional, string | string[])*: A list of source fields to exclude in the response. ** *`_source_includes` (Optional, string | string[])*: A list of source fields to include in the response. -** *`stored_fields` (Optional, string | string[])*: A list of stored fields to return as part of a hit. ** *`version` (Optional, number)*: The version number for concurrency control. It must match the current version of the document for the request to succeed. ** *`version_type` (Optional, Enum("internal" | "external" | "external_gte" | "force"))*: The version type. @@ -1048,6 +1047,7 @@ client.index({ index }) ** *`version_type` (Optional, Enum("internal" | "external" | "external_gte" | "force"))*: The version type. ** *`wait_for_active_shards` (Optional, number | Enum("all" | "index-setting"))*: The number of shard copies that must be active before proceeding with the operation. You can set it to `all` or any positive integer up to the total number of shards in the index (`number_of_replicas+1`). The default value of `1` means it waits for each primary shard to be active. ** *`require_alias` (Optional, boolean)*: If `true`, the destination must be an index alias. +** *`require_data_stream` (Optional, boolean)*: If `true`, the request's actions must target a data stream (existing or to be created). [discrete] === info @@ -3131,11 +3131,12 @@ Supports wildcards (`*`). To target all data streams and indices, omit this para ** *`active_only` (Optional, boolean)*: If `true`, the response only includes ongoing shard recoveries. ** *`bytes` (Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb"))*: The unit used to display byte values. ** *`detailed` (Optional, boolean)*: If `true`, the response includes detailed information about shard recoveries. -** *`h` (Optional, string | string[])*: List of columns to appear in the response. Supports simple wildcards. -** *`s` (Optional, string | string[])*: List of columns that determine how the table should be sorted. +** *`h` (Optional, Enum("index" | "shard" | "time" | "type" | "stage" | "source_host" | "source_node" | "target_host" | "target_node" | "repository" | "snapshot" | "files" | "files_recovered" | "files_percent" | "files_total" | "bytes" | "bytes_recovered" | "bytes_percent" | "bytes_total" | "translog_ops" | "translog_ops_recovered" | "translog_ops_percent" | "start_time" | "start_time_millis" | "stop_time" | "stop_time_millis") | Enum("index" | "shard" | "time" | "type" | "stage" | "source_host" | "source_node" | "target_host" | "target_node" | "repository" | "snapshot" | "files" | "files_recovered" | "files_percent" | "files_total" | "bytes" | "bytes_recovered" | "bytes_percent" | "bytes_total" | "translog_ops" | "translog_ops_recovered" | "translog_ops_percent" | "start_time" | "start_time_millis" | "stop_time" | "stop_time_millis")[])*: A list of columns names to display. +It supports simple wildcards. +** *`s` (Optional, string | string[])*: A list of column names or aliases that determines the sort order. Sorting defaults to ascending and can be changed by setting `:asc` or `:desc` as a suffix to the column name. -** *`time` (Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d"))*: Unit used to display time values. +** *`time` (Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d"))*: The unit used to display time values. [discrete] ==== repositories @@ -3186,8 +3187,9 @@ client.cat.segments({ ... }) Supports wildcards (`*`). To target all data streams and indices, omit this parameter or use `*` or `_all`. ** *`bytes` (Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb"))*: The unit used to display byte values. -** *`h` (Optional, string | string[])*: List of columns to appear in the response. Supports simple wildcards. -** *`s` (Optional, string | string[])*: List of columns that determine how the table should be sorted. +** *`h` (Optional, Enum("index" | "shard" | "prirep" | "ip" | "segment" | "generation" | "docs.count" | "docs.deleted" | "size" | "size.memory" | "committed" | "searchable" | "version" | "compound" | "id") | Enum("index" | "shard" | "prirep" | "ip" | "segment" | "generation" | "docs.count" | "docs.deleted" | "size" | "size.memory" | "committed" | "searchable" | "version" | "compound" | "id")[])*: A list of columns names to display. +It supports simple wildcards. +** *`s` (Optional, string | string[])*: A list of column names or aliases that determines the sort order. Sorting defaults to ascending and can be changed by setting `:asc` or `:desc` as a suffix to the column name. ** *`local` (Optional, boolean)*: If `true`, the request computes the list of selected nodes from the @@ -3218,12 +3220,12 @@ client.cat.shards({ ... }) Supports wildcards (`*`). To target all data streams and indices, omit this parameter or use `*` or `_all`. ** *`bytes` (Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb"))*: The unit used to display byte values. -** *`h` (Optional, string | string[])*: List of columns to appear in the response. Supports simple wildcards. -** *`s` (Optional, string | string[])*: List of columns that determine how the table should be sorted. +** *`h` (Optional, Enum("completion.size" | "dataset.size" | "dense_vector.value_count" | "docs" | "fielddata.evictions" | "fielddata.memory_size" | "flush.total" | "flush.total_time" | "get.current" | "get.exists_time" | "get.exists_total" | "get.missing_time" | "get.missing_total" | "get.time" | "get.total" | "id" | "index" | "indexing.delete_current" | "indexing.delete_time" | "indexing.delete_total" | "indexing.index_current" | "indexing.index_failed_due_to_version_conflict" | "indexing.index_failed" | "indexing.index_time" | "indexing.index_total" | "ip" | "merges.current" | "merges.current_docs" | "merges.current_size" | "merges.total" | "merges.total_docs" | "merges.total_size" | "merges.total_time" | "node" | "prirep" | "query_cache.evictions" | "query_cache.memory_size" | "recoverysource.type" | "refresh.time" | "refresh.total" | "search.fetch_current" | "search.fetch_time" | "search.fetch_total" | "search.open_contexts" | "search.query_current" | "search.query_time" | "search.query_total" | "search.scroll_current" | "search.scroll_time" | "search.scroll_total" | "segments.count" | "segments.fixed_bitset_memory" | "segments.index_writer_memory" | "segments.memory" | "segments.version_map_memory" | "seq_no.global_checkpoint" | "seq_no.local_checkpoint" | "seq_no.max" | "shard" | "dsparse_vector.value_count" | "state" | "store" | "suggest.current" | "suggest.time" | "suggest.total" | "sync_id" | "unassigned.at" | "unassigned.details" | "unassigned.for" | "unassigned.reason") | Enum("completion.size" | "dataset.size" | "dense_vector.value_count" | "docs" | "fielddata.evictions" | "fielddata.memory_size" | "flush.total" | "flush.total_time" | "get.current" | "get.exists_time" | "get.exists_total" | "get.missing_time" | "get.missing_total" | "get.time" | "get.total" | "id" | "index" | "indexing.delete_current" | "indexing.delete_time" | "indexing.delete_total" | "indexing.index_current" | "indexing.index_failed_due_to_version_conflict" | "indexing.index_failed" | "indexing.index_time" | "indexing.index_total" | "ip" | "merges.current" | "merges.current_docs" | "merges.current_size" | "merges.total" | "merges.total_docs" | "merges.total_size" | "merges.total_time" | "node" | "prirep" | "query_cache.evictions" | "query_cache.memory_size" | "recoverysource.type" | "refresh.time" | "refresh.total" | "search.fetch_current" | "search.fetch_time" | "search.fetch_total" | "search.open_contexts" | "search.query_current" | "search.query_time" | "search.query_total" | "search.scroll_current" | "search.scroll_time" | "search.scroll_total" | "segments.count" | "segments.fixed_bitset_memory" | "segments.index_writer_memory" | "segments.memory" | "segments.version_map_memory" | "seq_no.global_checkpoint" | "seq_no.local_checkpoint" | "seq_no.max" | "shard" | "dsparse_vector.value_count" | "state" | "store" | "suggest.current" | "suggest.time" | "suggest.total" | "sync_id" | "unassigned.at" | "unassigned.details" | "unassigned.for" | "unassigned.reason")[])*: List of columns to appear in the response. Supports simple wildcards. +** *`s` (Optional, string | string[])*: A list of column names or aliases that determines the sort order. Sorting defaults to ascending and can be changed by setting `:asc` or `:desc` as a suffix to the column name. -** *`master_timeout` (Optional, string | -1 | 0)*: Period to wait for a connection to the master node. -** *`time` (Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d"))*: Unit used to display time values. +** *`master_timeout` (Optional, string | -1 | 0)*: The period to wait for a connection to the master node. +** *`time` (Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d"))*: The unit used to display time values. [discrete] ==== snapshots @@ -3248,7 +3250,8 @@ Accepts wildcard expressions. `_all` returns all repositories. If any repository fails during the request, Elasticsearch returns an error. ** *`ignore_unavailable` (Optional, boolean)*: If `true`, the response does not include information from unavailable snapshots. -** *`h` (Optional, string | string[])*: List of columns to appear in the response. Supports simple wildcards. +** *`h` (Optional, Enum("id" | "repository" | "status" | "start_epoch" | "start_time" | "end_epoch" | "end_time" | "duration" | "indices" | "successful_shards" | "failed_shards" | "total_shards" | "reason") | Enum("build" | "completion.size" | "cpu" | "disk.avail" | "disk.total" | "disk.used" | "disk.used_percent" | "fielddata.evictions" | "fielddata.memory_size" | "file_desc.current" | "file_desc.max" | "file_desc.percent" | "flush.total" | "flush.total_time" | "get.current" | "get.exists_time" | "get.exists_total" | "get.missing_time" | "get.missing_total" | "get.time" | "get.total" | "heap.current" | "heap.max" | "heap.percent" | "http_address" | "id" | "indexing.delete_current" | "indexing.delete_time" | "indexing.delete_total" | "indexing.index_current" | "indexing.index_failed" | "indexing.index_failed_due_to_version_conflict" | "indexing.index_time" | "indexing.index_total" | "ip" | "jdk" | "load_1m" | "load_5m" | "load_15m" | "mappings.total_count" | "mappings.total_estimated_overhead_in_bytes" | "master" | "merges.current" | "merges.current_docs" | "merges.current_size" | "merges.total" | "merges.total_docs" | "merges.total_size" | "merges.total_time" | "name" | "node.role" | "pid" | "port" | "query_cache.memory_size" | "query_cache.evictions" | "query_cache.hit_count" | "query_cache.miss_count" | "ram.current" | "ram.max" | "ram.percent" | "refresh.total" | "refresh.time" | "request_cache.memory_size" | "request_cache.evictions" | "request_cache.hit_count" | "request_cache.miss_count" | "script.compilations" | "script.cache_evictions" | "search.fetch_current" | "search.fetch_time" | "search.fetch_total" | "search.open_contexts" | "search.query_current" | "search.query_time" | "search.query_total" | "search.scroll_current" | "search.scroll_time" | "search.scroll_total" | "segments.count" | "segments.fixed_bitset_memory" | "segments.index_writer_memory" | "segments.memory" | "segments.version_map_memory" | "shard_stats.total_count" | "suggest.current" | "suggest.time" | "suggest.total" | "uptime" | "version")[])*: A list of columns names to display. +It supports simple wildcards. ** *`s` (Optional, string | string[])*: List of columns that determine how the table should be sorted. Sorting defaults to ascending and can be changed by setting `:asc` or `:desc` as a suffix to the column name. @@ -3335,8 +3338,8 @@ client.cat.threadPool({ ... }) * *Request (object):* ** *`thread_pool_patterns` (Optional, string | string[])*: A list of thread pool names used to limit the request. Accepts wildcard expressions. -** *`h` (Optional, string | string[])*: List of columns to appear in the response. Supports simple wildcards. -** *`s` (Optional, string | string[])*: List of columns that determine how the table should be sorted. +** *`h` (Optional, Enum("active" | "completed" | "core" | "ephemeral_id" | "host" | "ip" | "keep_alive" | "largest" | "max" | "name" | "node_id" | "node_name" | "pid" | "pool_size" | "port" | "queue" | "queue_size" | "rejected" | "size" | "type") | Enum("active" | "completed" | "core" | "ephemeral_id" | "host" | "ip" | "keep_alive" | "largest" | "max" | "name" | "node_id" | "node_name" | "pid" | "pool_size" | "port" | "queue" | "queue_size" | "rejected" | "size" | "type")[])*: List of columns to appear in the response. Supports simple wildcards. +** *`s` (Optional, string | string[])*: A list of column names or aliases that determines the sort order. Sorting defaults to ascending and can be changed by setting `:asc` or `:desc` as a suffix to the column name. ** *`time` (Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d"))*: The unit used to display time values. @@ -3344,7 +3347,7 @@ or `:desc` as a suffix to the column name. local cluster state. If `false` the list of selected nodes are computed from the cluster state of the master node. In both cases the coordinating node will send requests for further information to each selected node. -** *`master_timeout` (Optional, string | -1 | 0)*: Period to wait for a connection to the master node. +** *`master_timeout` (Optional, string | -1 | 0)*: The period to wait for a connection to the master node. [discrete] ==== transforms @@ -4038,6 +4041,7 @@ To unset `_meta`, replace the template without specifying this information. ** *`deprecated` (Optional, boolean)*: Marks this index template as deprecated. When creating or updating a non-deprecated index template that uses deprecated components, Elasticsearch will emit a deprecation warning. ** *`create` (Optional, boolean)*: If `true`, this request cannot replace or update existing component templates. +** *`cause` (Optional, string)*: User defined reason for create the component template. ** *`master_timeout` (Optional, string | -1 | 0)*: Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. @@ -5154,6 +5158,7 @@ A query ID is provided in the ES|QL async query API response for a query that do A query ID is also provided when the request was submitted with the `keep_on_completion` parameter set to `true`. ** *`drop_null_columns` (Optional, boolean)*: Indicates whether columns that are entirely `null` will be removed from the `columns` and `values` portion of the results. If `true`, the response will include an extra section under the name `all_columns` which has the name of all the columns. +** *`format` (Optional, Enum("csv" | "json" | "tsv" | "txt" | "yaml" | "cbor" | "smile" | "arrow"))*: A short version of the Accept header, for example `json` or `yaml`. ** *`keep_alive` (Optional, string | -1 | 0)*: The period for which the query and its results are stored in the cluster. When this period expires, the query and its results are deleted, even if the query is still ongoing. ** *`wait_for_completion_timeout` (Optional, string | -1 | 0)*: The period to wait for the request to finish. @@ -5791,7 +5796,7 @@ This could be a built-in analyzer, or an analyzer that’s been configured in th ** *`field` (Optional, string)*: Field used to derive the analyzer. To use this parameter, you must specify an index. If specified, the `analyzer` parameter overrides this value. -** *`filter` (Optional, string | { type } | { type } | { type, preserve_original } | { type, ignored_scripts, output_unigrams } | { type } | { type } | { type, common_words, common_words_path, ignore_case, query_mode } | { type, filter, script } | { type } | { type, delimiter, encoding } | { type, max_gram, min_gram, side, preserve_original } | { type, articles, articles_path, articles_case } | { type, max_output_size, separator } | { type } | { type } | { type } | { type, dedup, dictionary, locale, longest_only } | { type, hyphenation_patterns_path, no_sub_matches, no_overlapping_matches } | { type } | { type, mode, types } | { type, keep_words, keep_words_case, keep_words_path } | { type, ignore_case, keywords, keywords_path, keywords_pattern } | { type } | { type } | { type, max, min } | { type, consume_all_tokens, max_token_count } | { type, language } | { type, bucket_count, hash_count, hash_set_size, with_rotation } | { type, filters, preserve_original } | { type, max_gram, min_gram, preserve_original } | { type, stoptags } | { type, patterns, preserve_original } | { type, all, pattern, replacement } | { type } | { type } | { type, script } | { type } | { type } | { type } | { type } | { type } | { type, filler_token, max_shingle_size, min_shingle_size, output_unigrams, output_unigrams_if_no_shingles, token_separator } | { type, language } | { type } | { type, rules, rules_path } | { type, language } | { type, ignore_case, remove_trailing, stopwords, stopwords_path } | { type } | { type } | { type } | { type, length } | { type, only_on_same_position } | { type } | { type, adjust_offsets, ignore_keywords } | { type } | { type, stopwords } | { type, minimum_length } | { type, use_romaji } | { type, stoptags } | { type, alternate, case_first, case_level, country, decomposition, hiragana_quaternary_mode, language, numeric, rules, strength, variable_top, variant } | { type, unicode_set_filter } | { type, name } | { type, dir, id } | { type, encoder, languageset, max_code_len, name_type, replace, rule_type } | { type }[])*: Array of token filters used to apply after the tokenizer. +** *`filter` (Optional, string | { type } | { type } | { type } | { type, preserve_original } | { type } | { type } | { type, ignored_scripts, output_unigrams } | { type } | { type } | { type, common_words, common_words_path, ignore_case, query_mode } | { type, filter, script } | { type } | { type } | { type, delimiter, encoding } | { type } | { type, max_gram, min_gram, side, preserve_original } | { type, articles, articles_path, articles_case } | { type, max_output_size, separator } | { type } | { type } | { type } | { type } | { type } | { type, dedup, dictionary, locale, longest_only } | { type, hyphenation_patterns_path, no_sub_matches, no_overlapping_matches } | { type } | { type, mode, types } | { type, keep_words, keep_words_case, keep_words_path } | { type, ignore_case, keywords, keywords_path, keywords_pattern } | { type } | { type } | { type, max, min } | { type, consume_all_tokens, max_token_count } | { type, language } | { type, bucket_count, hash_count, hash_set_size, with_rotation } | { type, filters, preserve_original } | { type, max_gram, min_gram, preserve_original } | { type, stoptags } | { type, patterns, preserve_original } | { type, all, flags, pattern, replacement } | { type } | { type } | { type } | { type, script } | { type } | { type } | { type } | { type } | { type } | { type } | { type, filler_token, max_shingle_size, min_shingle_size, output_unigrams, output_unigrams_if_no_shingles, token_separator } | { type, language } | { type } | { type, rules, rules_path } | { type, language } | { type, ignore_case, remove_trailing, stopwords, stopwords_path } | { type } | { type } | { type } | { type, length } | { type, only_on_same_position } | { type } | { type, adjust_offsets, ignore_keywords } | { type } | { type, stopwords } | { type, minimum_length } | { type, use_romaji } | { type, stoptags } | { type, alternate, caseFirst, caseLevel, country, decomposition, hiraganaQuaternaryMode, language, numeric, rules, strength, variableTop, variant } | { type, unicode_set_filter } | { type, name } | { type, dir, id } | { type, encoder, languageset, max_code_len, name_type, replace, rule_type } | { type }[])*: Array of token filters used to apply after the tokenizer. ** *`normalizer` (Optional, string)*: Normalizer to use to convert text into a single token. ** *`text` (Optional, string | string[])*: Text to analyze. If an array of strings is provided, it is analyzed as a multi-value field. @@ -7214,7 +7219,7 @@ a new date field is added instead of string. not used at all by Elasticsearch, but can be used to store application-specific metadata. ** *`numeric_detection` (Optional, boolean)*: Automatically map strings into numeric data types for all fields. -** *`properties` (Optional, Record)*: Mapping for a field. For new fields, this mapping can include: +** *`properties` (Optional, Record)*: Mapping for a field. For new fields, this mapping can include: - Field name - Field data type @@ -8183,6 +8188,7 @@ The following integrations are available through the inference API. You can find * Azure AI Studio (`completion`, `text_embedding`) * Azure OpenAI (`completion`, `text_embedding`) * Cohere (`completion`, `rerank`, `text_embedding`) +* DeepSeek (`completion`, `chat_completion`) * Elasticsearch (`rerank`, `sparse_embedding`, `text_embedding` - this service is for built-in models and models uploaded through Eland) * ELSER (`sparse_embedding`) * Google AI Studio (`completion`, `text_embedding`) @@ -8259,6 +8265,15 @@ client.inference.putAmazonbedrock({ task_type, amazonbedrock_inference_id, servi ** *`task_settings` (Optional, { max_new_tokens, temperature, top_k, top_p })*: Settings to configure the inference task. These settings are specific to the task type you specified. +[discrete] +==== put_amazonsagemaker +Configure a Amazon SageMaker inference endpoint +[source,ts] +---- +client.inference.putAmazonsagemaker() +---- + + [discrete] ==== put_anthropic Create an Anthropic inference endpoint. @@ -8365,6 +8380,27 @@ These settings are specific to the `cohere` service. ** *`task_settings` (Optional, { input_type, return_documents, top_n, truncate })*: Settings to configure the inference task. These settings are specific to the task type you specified. +[discrete] +==== put_deepseek +Create a DeepSeek inference endpoint. + +Create an inference endpoint to perform an inference task with the `deepseek` service. +[source,ts] +---- +client.inference.putDeepseek({ task_type, deepseek_inference_id, service, service_settings }) +---- + +[discrete] +==== Arguments + +* *Request (object):* +** *`task_type` (Enum("completion" | "chat_completion"))*: The type of the inference task that the model will perform. +** *`deepseek_inference_id` (string)*: The unique identifier of the inference endpoint. +** *`service` (Enum("deepseek"))*: The type of service supported for the specified task type. In this case, `deepseek`. +** *`service_settings` ({ api_key, model_id, url })*: Settings used to install the inference model. +These settings are specific to the `deepseek` service. +** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. + [discrete] ==== put_elasticsearch Create an Elasticsearch inference endpoint. @@ -8900,9 +8936,6 @@ client.ingest.getIpLocationDatabase({ ... }) ** *`id` (Optional, string | string[])*: List of database configuration IDs to retrieve. Wildcard (`*`) expressions are supported. To get all database configurations, omit this parameter or use `*`. -** *`master_timeout` (Optional, string | -1 | 0)*: The period to wait for a connection to the master node. -If no response is received before the timeout expires, the request fails and returns an error. -A value of `-1` indicates that the request should never time out. [discrete] ==== get_pipeline @@ -9186,7 +9219,7 @@ client.license.postStartTrial({ ... }) * *Request (object):* ** *`acknowledge` (Optional, boolean)*: whether the user has acknowledged acknowledge messages (default: false) -** *`type_query_string` (Optional, string)* +** *`type` (Optional, string)*: The type of trial license to generate (default: "trial") ** *`master_timeout` (Optional, string | -1 | 0)*: Period to wait for a connection to the master node. [discrete] diff --git a/src/api/api/inference.ts b/src/api/api/inference.ts index 85ebc695a..06e30dd75 100644 --- a/src/api/api/inference.ts +++ b/src/api/api/inference.ts @@ -262,7 +262,7 @@ export default class Inference { } /** - * Create an inference endpoint. IMPORTANT: The inference APIs enable you to use certain services, such as built-in machine learning models (ELSER, E5), models uploaded through Eland, Cohere, OpenAI, Mistral, Azure OpenAI, Google AI Studio, Google Vertex AI, Anthropic, Watsonx.ai, or Hugging Face. For built-in models and models uploaded through Eland, the inference APIs offer an alternative way to use and manage trained models. However, if you do not plan to use the inference APIs to use these models or if you want to use non-NLP models, use the machine learning trained model APIs. The following integrations are available through the inference API. You can find the available task types next to the integration name: * AlibabaCloud AI Search (`completion`, `rerank`, `sparse_embedding`, `text_embedding`) * Amazon Bedrock (`completion`, `text_embedding`) * Anthropic (`completion`) * Azure AI Studio (`completion`, `text_embedding`) * Azure OpenAI (`completion`, `text_embedding`) * Cohere (`completion`, `rerank`, `text_embedding`) * Elasticsearch (`rerank`, `sparse_embedding`, `text_embedding` - this service is for built-in models and models uploaded through Eland) * ELSER (`sparse_embedding`) * Google AI Studio (`completion`, `text_embedding`) * Google Vertex AI (`rerank`, `text_embedding`) * Hugging Face (`chat_completion`, `completion`, `rerank`, `text_embedding`) * Mistral (`chat_completion`, `completion`, `text_embedding`) * OpenAI (`chat_completion`, `completion`, `text_embedding`) * VoyageAI (`text_embedding`, `rerank`) * Watsonx inference integration (`text_embedding`) * JinaAI (`text_embedding`, `rerank`) + * Create an inference endpoint. IMPORTANT: The inference APIs enable you to use certain services, such as built-in machine learning models (ELSER, E5), models uploaded through Eland, Cohere, OpenAI, Mistral, Azure OpenAI, Google AI Studio, Google Vertex AI, Anthropic, Watsonx.ai, or Hugging Face. For built-in models and models uploaded through Eland, the inference APIs offer an alternative way to use and manage trained models. However, if you do not plan to use the inference APIs to use these models or if you want to use non-NLP models, use the machine learning trained model APIs. The following integrations are available through the inference API. You can find the available task types next to the integration name: * AlibabaCloud AI Search (`completion`, `rerank`, `sparse_embedding`, `text_embedding`) * Amazon Bedrock (`completion`, `text_embedding`) * Anthropic (`completion`) * Azure AI Studio (`completion`, `text_embedding`) * Azure OpenAI (`completion`, `text_embedding`) * Cohere (`completion`, `rerank`, `text_embedding`) * DeepSeek (`completion`, `chat_completion`) * Elasticsearch (`rerank`, `sparse_embedding`, `text_embedding` - this service is for built-in models and models uploaded through Eland) * ELSER (`sparse_embedding`) * Google AI Studio (`completion`, `text_embedding`) * Google Vertex AI (`rerank`, `text_embedding`) * Hugging Face (`chat_completion`, `completion`, `rerank`, `text_embedding`) * Mistral (`chat_completion`, `completion`, `text_embedding`) * OpenAI (`chat_completion`, `completion`, `text_embedding`) * VoyageAI (`text_embedding`, `rerank`) * Watsonx inference integration (`text_embedding`) * JinaAI (`text_embedding`, `rerank`) * @see {@link https://www.elastic.co/guide/en/elasticsearch/reference/8.19/put-inference-api.html | Elasticsearch API documentation} */ async put (this: That, params: T.InferencePutRequest | TB.InferencePutRequest, options?: TransportRequestOptionsWithOutMeta): Promise @@ -396,6 +396,39 @@ export default class Inference { return await this.transport.request({ path, method, querystring, body, meta }, options) } + /** + * Configure a Amazon SageMaker inference endpoint + * @see {@link https://www.elastic.co/guide/en/elasticsearch/reference/8.19/infer-service-amazon-sagemaker.html | Elasticsearch API documentation} + */ + async putAmazonsagemaker (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptionsWithOutMeta): Promise + async putAmazonsagemaker (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptionsWithMeta): Promise> + async putAmazonsagemaker (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptions): Promise + async putAmazonsagemaker (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptions): Promise { + const acceptedPath: string[] = ['task_type', 'amazonsagemaker_inference_id'] + const querystring: Record = {} + const body = undefined + + params = params ?? {} + for (const key in params) { + if (acceptedPath.includes(key)) { + continue + } else if (key !== 'body') { + querystring[key] = params[key] + } + } + + const method = 'PUT' + const path = `/_inference/${encodeURIComponent(params.task_type.toString())}/${encodeURIComponent(params.amazonsagemaker_inference_id.toString())}` + const meta: TransportRequestMetadata = { + name: 'inference.put_amazonsagemaker', + pathParts: { + task_type: params.task_type, + amazonsagemaker_inference_id: params.amazonsagemaker_inference_id + } + } + return await this.transport.request({ path, method, querystring, body, meta }, options) + } + /** * Create an Anthropic inference endpoint. Create an inference endpoint to perform an inference task with the `anthropic` service. * @see {@link https://www.elastic.co/guide/en/elasticsearch/reference/8.19/infer-service-anthropic.html | Elasticsearch API documentation} @@ -576,6 +609,51 @@ export default class Inference { return await this.transport.request({ path, method, querystring, body, meta }, options) } + /** + * Create a DeepSeek inference endpoint. Create an inference endpoint to perform an inference task with the `deepseek` service. + * @see {@link https://www.elastic.co/guide/en/elasticsearch/reference/8.19/infer-service-deepseek.html | Elasticsearch API documentation} + */ + async putDeepseek (this: That, params: T.InferencePutDeepseekRequest | TB.InferencePutDeepseekRequest, options?: TransportRequestOptionsWithOutMeta): Promise + async putDeepseek (this: That, params: T.InferencePutDeepseekRequest | TB.InferencePutDeepseekRequest, options?: TransportRequestOptionsWithMeta): Promise> + async putDeepseek (this: That, params: T.InferencePutDeepseekRequest | TB.InferencePutDeepseekRequest, options?: TransportRequestOptions): Promise + async putDeepseek (this: That, params: T.InferencePutDeepseekRequest | TB.InferencePutDeepseekRequest, options?: TransportRequestOptions): Promise { + const acceptedPath: string[] = ['task_type', 'deepseek_inference_id'] + const acceptedBody: string[] = ['chunking_settings', 'service', 'service_settings'] + const querystring: Record = {} + // @ts-expect-error + const userBody: any = params?.body + let body: Record | string + if (typeof userBody === 'string') { + body = userBody + } else { + body = userBody != null ? { ...userBody } : undefined + } + + for (const key in params) { + if (acceptedBody.includes(key)) { + body = body ?? {} + // @ts-expect-error + body[key] = params[key] + } else if (acceptedPath.includes(key)) { + continue + } else if (key !== 'body') { + // @ts-expect-error + querystring[key] = params[key] + } + } + + const method = 'PUT' + const path = `/_inference/${encodeURIComponent(params.task_type.toString())}/${encodeURIComponent(params.deepseek_inference_id.toString())}` + const meta: TransportRequestMetadata = { + name: 'inference.put_deepseek', + pathParts: { + task_type: params.task_type, + deepseek_inference_id: params.deepseek_inference_id + } + } + return await this.transport.request({ path, method, querystring, body, meta }, options) + } + /** * Create an Elasticsearch inference endpoint. Create an inference endpoint to perform an inference task with the `elasticsearch` service. > info > Your Elasticsearch deployment contains preconfigured ELSER and E5 inference endpoints, you only need to create the enpoints using the API if you want to customize the settings. If you use the ELSER or the E5 model through the `elasticsearch` service, the API request will automatically download and deploy the model if it isn't downloaded yet. > info > You might see a 502 bad gateway error in the response when using the Kibana Console. This error usually just reflects a timeout, while the model downloads in the background. You can check the download progress in the Machine Learning UI. If using the Python client, you can set the timeout parameter to a higher value. After creating the endpoint, wait for the model deployment to complete before using it. To verify the deployment status, use the get trained model statistics API. Look for `"state": "fully_allocated"` in the response and ensure that the `"allocation_count"` matches the `"target_allocation_count"`. Avoid creating multiple endpoints for the same model unless required, as each endpoint consumes significant resources. * @see {@link https://www.elastic.co/guide/en/elasticsearch/reference/8.19/infer-service-elasticsearch.html | Elasticsearch API documentation} diff --git a/src/api/api/streams.ts b/src/api/api/streams.ts new file mode 100644 index 000000000..0762d9fa1 --- /dev/null +++ b/src/api/api/streams.ts @@ -0,0 +1,133 @@ +/* + * Licensed to Elasticsearch B.V. under one or more contributor + * license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright + * ownership. Elasticsearch B.V. licenses this file to you under + * the Apache License, Version 2.0 (the "License"); you may + * not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +/* eslint-disable import/export */ +/* eslint-disable @typescript-eslint/no-misused-new */ +/* eslint-disable @typescript-eslint/no-extraneous-class */ +/* eslint-disable @typescript-eslint/no-unused-vars */ + +// This file was automatically generated by elastic/elastic-client-generator-js +// DO NOT MODIFY IT BY HAND. Instead, modify the source open api file, +// and elastic/elastic-client-generator-js to regenerate this file again. + +import { + Transport, + TransportRequestMetadata, + TransportRequestOptions, + TransportRequestOptionsWithMeta, + TransportRequestOptionsWithOutMeta, + TransportResult +} from '@elastic/transport' +import * as T from '../types' +import * as TB from '../typesWithBodyKey' +interface That { transport: Transport } + +export default class Streams { + transport: Transport + constructor (transport: Transport) { + this.transport = transport + } + + /** + * Disable the Logs Streams feature for this cluster + * @see {@link https://www.elastic.co/guide/en/elasticsearch/reference/8.19/streams-logs-disable.html | Elasticsearch API documentation} + */ + async logsDisable (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptionsWithOutMeta): Promise + async logsDisable (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptionsWithMeta): Promise> + async logsDisable (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptions): Promise + async logsDisable (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptions): Promise { + const acceptedPath: string[] = [] + const querystring: Record = {} + const body = undefined + + params = params ?? {} + for (const key in params) { + if (acceptedPath.includes(key)) { + continue + } else if (key !== 'body') { + querystring[key] = params[key] + } + } + + const method = 'POST' + const path = '/_streams/logs/_disable' + const meta: TransportRequestMetadata = { + name: 'streams.logs_disable' + } + return await this.transport.request({ path, method, querystring, body, meta }, options) + } + + /** + * Enable the Logs Streams feature for this cluster + * @see {@link https://www.elastic.co/guide/en/elasticsearch/reference/8.19/streams-logs-enable.html | Elasticsearch API documentation} + */ + async logsEnable (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptionsWithOutMeta): Promise + async logsEnable (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptionsWithMeta): Promise> + async logsEnable (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptions): Promise + async logsEnable (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptions): Promise { + const acceptedPath: string[] = [] + const querystring: Record = {} + const body = undefined + + params = params ?? {} + for (const key in params) { + if (acceptedPath.includes(key)) { + continue + } else if (key !== 'body') { + querystring[key] = params[key] + } + } + + const method = 'POST' + const path = '/_streams/logs/_enable' + const meta: TransportRequestMetadata = { + name: 'streams.logs_enable' + } + return await this.transport.request({ path, method, querystring, body, meta }, options) + } + + /** + * Return the current status of the streams feature for each streams type + * @see {@link https://www.elastic.co/guide/en/elasticsearch/reference/8.19/streams-status.html | Elasticsearch API documentation} + */ + async status (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptionsWithOutMeta): Promise + async status (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptionsWithMeta): Promise> + async status (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptions): Promise + async status (this: That, params?: T.TODO | TB.TODO, options?: TransportRequestOptions): Promise { + const acceptedPath: string[] = [] + const querystring: Record = {} + const body = undefined + + params = params ?? {} + for (const key in params) { + if (acceptedPath.includes(key)) { + continue + } else if (key !== 'body') { + querystring[key] = params[key] + } + } + + const method = 'GET' + const path = '/_streams/status' + const meta: TransportRequestMetadata = { + name: 'streams.status' + } + return await this.transport.request({ path, method, querystring, body, meta }, options) + } +} diff --git a/src/api/index.ts b/src/api/index.ts index f69eb473d..e6c74a674 100644 --- a/src/api/index.ts +++ b/src/api/index.ts @@ -101,6 +101,7 @@ import SlmApi from './api/slm' import SnapshotApi from './api/snapshot' import SqlApi from './api/sql' import SslApi from './api/ssl' +import StreamsApi from './api/streams' import SynonymsApi from './api/synonyms' import TasksApi from './api/tasks' import termsEnumApi from './api/terms_enum' @@ -190,6 +191,7 @@ export default interface API { snapshot: SnapshotApi sql: SqlApi ssl: SslApi + streams: StreamsApi synonyms: SynonymsApi tasks: TasksApi termsEnum: typeof termsEnumApi @@ -238,6 +240,7 @@ const kSlm = Symbol('Slm') const kSnapshot = Symbol('Snapshot') const kSql = Symbol('Sql') const kSsl = Symbol('Ssl') +const kStreams = Symbol('Streams') const kSynonyms = Symbol('Synonyms') const kTasks = Symbol('Tasks') const kTextStructure = Symbol('TextStructure') @@ -281,6 +284,7 @@ export default class API { [kSnapshot]: symbol | null [kSql]: symbol | null [kSsl]: symbol | null + [kStreams]: symbol | null [kSynonyms]: symbol | null [kTasks]: symbol | null [kTextStructure]: symbol | null @@ -323,6 +327,7 @@ export default class API { this[kSnapshot] = null this[kSql] = null this[kSsl] = null + this[kStreams] = null this[kSynonyms] = null this[kTasks] = null this[kTextStructure] = null @@ -484,6 +489,9 @@ Object.defineProperties(API.prototype, { ssl: { get () { return this[kSsl] === null ? (this[kSsl] = new SslApi(this.transport)) : this[kSsl] } }, + streams: { + get () { return this[kStreams] === null ? (this[kStreams] = new StreamsApi(this.transport)) : this[kStreams] } + }, synonyms: { get () { return this[kSynonyms] === null ? (this[kSynonyms] = new SynonymsApi(this.transport)) : this[kSynonyms] } }, diff --git a/src/api/types.ts b/src/api/types.ts index 27bbbf9cc..628466584 100644 --- a/src/api/types.ts +++ b/src/api/types.ts @@ -458,7 +458,6 @@ export interface GetSourceRequest extends RequestBase { _source?: SearchSourceConfigParam _source_excludes?: Fields _source_includes?: Fields - stored_fields?: Fields version?: VersionNumber version_type?: VersionType } @@ -659,6 +658,7 @@ export interface IndexRequest extends RequestBase { version_type?: VersionType wait_for_active_shards?: WaitForActiveShards require_alias?: boolean + require_data_stream?: boolean document?: TDocument } @@ -2258,8 +2258,6 @@ export interface ErrorResponseBase { status: integer } -export type EsqlResult = ArrayBuffer - export type ExpandWildcard = 'all' | 'open' | 'closed' | 'hidden' | 'none' export type ExpandWildcards = ExpandWildcard | ExpandWildcard[] @@ -2480,7 +2478,7 @@ export type LifecycleOperationMode = 'RUNNING' | 'STOPPING' | 'STOPPED' export interface LinearRetriever extends RetrieverBase { retrievers?: InnerRetriever[] - rank_window_size: integer + rank_window_size?: integer } export type MapboxVectorTiles = ArrayBuffer @@ -2575,7 +2573,7 @@ export interface PinnedRetriever extends RetrieverBase { retriever: RetrieverContainer ids?: string[] docs?: SpecifiedDocument[] - rank_window_size: integer + rank_window_size?: integer } export type PipelineName = string @@ -2910,8 +2908,8 @@ export interface TextSimilarityReranker extends RetrieverBase { retriever: RetrieverContainer rank_window_size?: integer inference_id?: string - inference_text?: string - field?: string + inference_text: string + field: string } export type ThreadType = 'cpu' | 'wait' | 'block' | 'gpu' | 'mem' @@ -2922,6 +2920,12 @@ export type TimeUnit = 'nanos' | 'micros' | 'ms' | 's' | 'm' | 'h' | 'd' export type TimeZone = string +export interface TokenPruningConfig { + tokens_freq_ratio_threshold?: integer + tokens_weight_threshold?: float + only_score_pruned_tokens?: boolean +} + export interface TopLeftBottomRightGeoBounds { top_left: GeoLocation bottom_right: GeoLocation @@ -4450,6 +4454,10 @@ export interface AnalysisArabicNormalizationTokenFilter extends AnalysisTokenFil type: 'arabic_normalization' } +export interface AnalysisArabicStemTokenFilter extends AnalysisTokenFilterBase { + type: 'arabic_stem' +} + export interface AnalysisArmenianAnalyzer { type: 'armenian' stopwords?: AnalysisStopWords @@ -4476,12 +4484,20 @@ export interface AnalysisBengaliAnalyzer { stem_exclusion?: string[] } +export interface AnalysisBengaliNormalizationTokenFilter extends AnalysisTokenFilterBase { + type: 'bengali_normalization' +} + export interface AnalysisBrazilianAnalyzer { type: 'brazilian' stopwords?: AnalysisStopWords stopwords_path?: string } +export interface AnalysisBrazilianStemTokenFilter extends AnalysisTokenFilterBase { + type: 'brazilian_stem' +} + export interface AnalysisBulgarianAnalyzer { type: 'bulgarian' stopwords?: AnalysisStopWords @@ -4588,6 +4604,10 @@ export interface AnalysisCzechAnalyzer { stem_exclusion?: string[] } +export interface AnalysisCzechStemTokenFilter extends AnalysisTokenFilterBase { + type: 'czech_stem' +} + export interface AnalysisDanishAnalyzer { type: 'danish' stopwords?: AnalysisStopWords @@ -4617,6 +4637,10 @@ export interface AnalysisDutchAnalyzer { stem_exclusion?: string[] } +export interface AnalysisDutchStemTokenFilter extends AnalysisTokenFilterBase { + type: 'dutch_stem' +} + export type AnalysisEdgeNGramSide = 'front' | 'back' export interface AnalysisEdgeNGramTokenFilter extends AnalysisTokenFilterBase { @@ -4632,7 +4656,7 @@ export interface AnalysisEdgeNGramTokenizer extends AnalysisTokenizerBase { custom_token_chars?: string max_gram?: integer min_gram?: integer - token_chars?: AnalysisTokenChar[] + token_chars?: string | AnalysisTokenChar[] } export interface AnalysisElisionTokenFilter extends AnalysisTokenFilterBase { @@ -4689,6 +4713,10 @@ export interface AnalysisFrenchAnalyzer { stem_exclusion?: string[] } +export interface AnalysisFrenchStemTokenFilter extends AnalysisTokenFilterBase { + type: 'french_stem' +} + export interface AnalysisGalicianAnalyzer { type: 'galician' stopwords?: AnalysisStopWords @@ -4707,6 +4735,10 @@ export interface AnalysisGermanNormalizationTokenFilter extends AnalysisTokenFil type: 'german_normalization' } +export interface AnalysisGermanStemTokenFilter extends AnalysisTokenFilterBase { + type: 'german_stem' +} + export interface AnalysisGreekAnalyzer { type: 'greek' stopwords?: AnalysisStopWords @@ -4770,16 +4802,16 @@ export type AnalysisIcuCollationStrength = 'primary' | 'secondary' | 'tertiary' export interface AnalysisIcuCollationTokenFilter extends AnalysisTokenFilterBase { type: 'icu_collation' alternate?: AnalysisIcuCollationAlternate - case_first?: AnalysisIcuCollationCaseFirst - case_level?: boolean + caseFirst?: AnalysisIcuCollationCaseFirst + caseLevel?: boolean country?: string decomposition?: AnalysisIcuCollationDecomposition - hiragana_quaternary_mode?: boolean + hiraganaQuaternaryMode?: boolean language?: string numeric?: boolean rules?: string strength?: AnalysisIcuCollationStrength - variable_top?: string + variableTop?: string variant?: string } @@ -5074,6 +5106,7 @@ export interface AnalysisPatternReplaceCharFilter extends AnalysisCharFilterBase export interface AnalysisPatternReplaceTokenFilter extends AnalysisTokenFilterBase { type: 'pattern_replace' all?: boolean + flags?: string pattern: string replacement?: string } @@ -5095,6 +5128,10 @@ export interface AnalysisPersianNormalizationTokenFilter extends AnalysisTokenFi type: 'persian_normalization' } +export interface AnalysisPersianStemTokenFilter extends AnalysisTokenFilterBase { + type: 'persian_stem' +} + export type AnalysisPhoneticEncoder = 'metaphone' | 'double_metaphone' | 'soundex' | 'refined_soundex' | 'caverphone1' | 'caverphone2' | 'cologne' | 'nysiis' | 'koelnerphonetik' | 'haasephonetik' | 'beider_morse' | 'daitch_mokotoff' export type AnalysisPhoneticLanguage = 'any' | 'common' | 'cyrillic' | 'english' | 'french' | 'german' | 'hebrew' | 'hungarian' | 'polish' | 'romanian' | 'russian' | 'spanish' @@ -5151,6 +5188,10 @@ export interface AnalysisRussianAnalyzer { stem_exclusion?: string[] } +export interface AnalysisRussianStemTokenFilter extends AnalysisTokenFilterBase { + type: 'russian_stem' +} + export interface AnalysisScandinavianFoldingTokenFilter extends AnalysisTokenFilterBase { type: 'scandinavian_folding' } @@ -5315,7 +5356,7 @@ export interface AnalysisTokenFilterBase { version?: VersionString } -export type AnalysisTokenFilterDefinition = AnalysisApostropheTokenFilter | AnalysisArabicNormalizationTokenFilter | AnalysisAsciiFoldingTokenFilter | AnalysisCjkBigramTokenFilter | AnalysisCjkWidthTokenFilter | AnalysisClassicTokenFilter | AnalysisCommonGramsTokenFilter | AnalysisConditionTokenFilter | AnalysisDecimalDigitTokenFilter | AnalysisDelimitedPayloadTokenFilter | AnalysisEdgeNGramTokenFilter | AnalysisElisionTokenFilter | AnalysisFingerprintTokenFilter | AnalysisFlattenGraphTokenFilter | AnalysisGermanNormalizationTokenFilter | AnalysisHindiNormalizationTokenFilter | AnalysisHunspellTokenFilter | AnalysisHyphenationDecompounderTokenFilter | AnalysisIndicNormalizationTokenFilter | AnalysisKeepTypesTokenFilter | AnalysisKeepWordsTokenFilter | AnalysisKeywordMarkerTokenFilter | AnalysisKeywordRepeatTokenFilter | AnalysisKStemTokenFilter | AnalysisLengthTokenFilter | AnalysisLimitTokenCountTokenFilter | AnalysisLowercaseTokenFilter | AnalysisMinHashTokenFilter | AnalysisMultiplexerTokenFilter | AnalysisNGramTokenFilter | AnalysisNoriPartOfSpeechTokenFilter | AnalysisPatternCaptureTokenFilter | AnalysisPatternReplaceTokenFilter | AnalysisPersianNormalizationTokenFilter | AnalysisPorterStemTokenFilter | AnalysisPredicateTokenFilter | AnalysisRemoveDuplicatesTokenFilter | AnalysisReverseTokenFilter | AnalysisScandinavianFoldingTokenFilter | AnalysisScandinavianNormalizationTokenFilter | AnalysisSerbianNormalizationTokenFilter | AnalysisShingleTokenFilter | AnalysisSnowballTokenFilter | AnalysisSoraniNormalizationTokenFilter | AnalysisStemmerOverrideTokenFilter | AnalysisStemmerTokenFilter | AnalysisStopTokenFilter | AnalysisSynonymGraphTokenFilter | AnalysisSynonymTokenFilter | AnalysisTrimTokenFilter | AnalysisTruncateTokenFilter | AnalysisUniqueTokenFilter | AnalysisUppercaseTokenFilter | AnalysisWordDelimiterGraphTokenFilter | AnalysisWordDelimiterTokenFilter | AnalysisJaStopTokenFilter | AnalysisKuromojiStemmerTokenFilter | AnalysisKuromojiReadingFormTokenFilter | AnalysisKuromojiPartOfSpeechTokenFilter | AnalysisIcuCollationTokenFilter | AnalysisIcuFoldingTokenFilter | AnalysisIcuNormalizationTokenFilter | AnalysisIcuTransformTokenFilter | AnalysisPhoneticTokenFilter | AnalysisDictionaryDecompounderTokenFilter +export type AnalysisTokenFilterDefinition = AnalysisApostropheTokenFilter | AnalysisArabicStemTokenFilter | AnalysisArabicNormalizationTokenFilter | AnalysisAsciiFoldingTokenFilter | AnalysisBengaliNormalizationTokenFilter | AnalysisBrazilianStemTokenFilter | AnalysisCjkBigramTokenFilter | AnalysisCjkWidthTokenFilter | AnalysisClassicTokenFilter | AnalysisCommonGramsTokenFilter | AnalysisConditionTokenFilter | AnalysisCzechStemTokenFilter | AnalysisDecimalDigitTokenFilter | AnalysisDelimitedPayloadTokenFilter | AnalysisDutchStemTokenFilter | AnalysisEdgeNGramTokenFilter | AnalysisElisionTokenFilter | AnalysisFingerprintTokenFilter | AnalysisFlattenGraphTokenFilter | AnalysisFrenchStemTokenFilter | AnalysisGermanNormalizationTokenFilter | AnalysisGermanStemTokenFilter | AnalysisHindiNormalizationTokenFilter | AnalysisHunspellTokenFilter | AnalysisHyphenationDecompounderTokenFilter | AnalysisIndicNormalizationTokenFilter | AnalysisKeepTypesTokenFilter | AnalysisKeepWordsTokenFilter | AnalysisKeywordMarkerTokenFilter | AnalysisKeywordRepeatTokenFilter | AnalysisKStemTokenFilter | AnalysisLengthTokenFilter | AnalysisLimitTokenCountTokenFilter | AnalysisLowercaseTokenFilter | AnalysisMinHashTokenFilter | AnalysisMultiplexerTokenFilter | AnalysisNGramTokenFilter | AnalysisNoriPartOfSpeechTokenFilter | AnalysisPatternCaptureTokenFilter | AnalysisPatternReplaceTokenFilter | AnalysisPersianNormalizationTokenFilter | AnalysisPersianStemTokenFilter | AnalysisPorterStemTokenFilter | AnalysisPredicateTokenFilter | AnalysisRemoveDuplicatesTokenFilter | AnalysisReverseTokenFilter | AnalysisRussianStemTokenFilter | AnalysisScandinavianFoldingTokenFilter | AnalysisScandinavianNormalizationTokenFilter | AnalysisSerbianNormalizationTokenFilter | AnalysisShingleTokenFilter | AnalysisSnowballTokenFilter | AnalysisSoraniNormalizationTokenFilter | AnalysisStemmerOverrideTokenFilter | AnalysisStemmerTokenFilter | AnalysisStopTokenFilter | AnalysisSynonymGraphTokenFilter | AnalysisSynonymTokenFilter | AnalysisTrimTokenFilter | AnalysisTruncateTokenFilter | AnalysisUniqueTokenFilter | AnalysisUppercaseTokenFilter | AnalysisWordDelimiterGraphTokenFilter | AnalysisWordDelimiterTokenFilter | AnalysisJaStopTokenFilter | AnalysisKuromojiStemmerTokenFilter | AnalysisKuromojiReadingFormTokenFilter | AnalysisKuromojiPartOfSpeechTokenFilter | AnalysisIcuCollationTokenFilter | AnalysisIcuFoldingTokenFilter | AnalysisIcuNormalizationTokenFilter | AnalysisIcuTransformTokenFilter | AnalysisPhoneticTokenFilter | AnalysisDictionaryDecompounderTokenFilter export type AnalysisTokenizer = string | AnalysisTokenizerDefinition @@ -5394,6 +5435,7 @@ export interface AnalysisWordDelimiterTokenFilterBase extends AnalysisTokenFilte export interface MappingAggregateMetricDoubleProperty extends MappingPropertyBase { type: 'aggregate_metric_double' default_metric: string + ignore_malformed?: boolean metrics: string[] time_series_metric?: MappingTimeSeriesMetricType } @@ -5604,6 +5646,7 @@ export interface MappingFlattenedProperty extends MappingPropertyBase { null_value?: string similarity?: string split_queries_on_whitespace?: boolean + time_series_dimensions?: string[] type: 'flattened' } @@ -5618,6 +5661,8 @@ export interface MappingFloatRangeProperty extends MappingRangePropertyBase { export type MappingGeoOrientation = 'right' | 'RIGHT' | 'counterclockwise' | 'ccw' | 'left' | 'LEFT' | 'clockwise' | 'cw' +export type MappingGeoPointMetricType = 'gauge' | 'counter' | 'position' + export interface MappingGeoPointProperty extends MappingDocValuesPropertyBase { ignore_malformed?: boolean ignore_z_value?: boolean @@ -5626,6 +5671,7 @@ export interface MappingGeoPointProperty extends MappingDocValuesPropertyBase { on_script_error?: MappingOnScriptError script?: Script | string type: 'geo_point' + time_series_metric?: MappingGeoPointMetricType } export interface MappingGeoShapeProperty extends MappingDocValuesPropertyBase { @@ -5788,7 +5834,7 @@ export interface MappingPointProperty extends MappingDocValuesPropertyBase { type: 'point' } -export type MappingProperty = MappingBinaryProperty | MappingBooleanProperty | MappingDynamicProperty | MappingJoinProperty | MappingKeywordProperty | MappingMatchOnlyTextProperty | MappingPercolatorProperty | MappingRankFeatureProperty | MappingRankFeaturesProperty | MappingSearchAsYouTypeProperty | MappingTextProperty | MappingVersionProperty | MappingWildcardProperty | MappingDateNanosProperty | MappingDateProperty | MappingAggregateMetricDoubleProperty | MappingDenseVectorProperty | MappingFlattenedProperty | MappingNestedProperty | MappingObjectProperty | MappingPassthroughObjectProperty | MappingSemanticTextProperty | MappingSparseVectorProperty | MappingCompletionProperty | MappingConstantKeywordProperty | MappingCountedKeywordProperty | MappingFieldAliasProperty | MappingHistogramProperty | MappingIpProperty | MappingMurmur3HashProperty | MappingTokenCountProperty | MappingGeoPointProperty | MappingGeoShapeProperty | MappingPointProperty | MappingShapeProperty | MappingByteNumberProperty | MappingDoubleNumberProperty | MappingFloatNumberProperty | MappingHalfFloatNumberProperty | MappingIntegerNumberProperty | MappingLongNumberProperty | MappingScaledFloatNumberProperty | MappingShortNumberProperty | MappingUnsignedLongNumberProperty | MappingDateRangeProperty | MappingDoubleRangeProperty | MappingFloatRangeProperty | MappingIntegerRangeProperty | MappingIpRangeProperty | MappingLongRangeProperty | MappingIcuCollationProperty +export type MappingProperty = MappingBinaryProperty | MappingBooleanProperty | MappingDynamicProperty | MappingJoinProperty | MappingKeywordProperty | MappingMatchOnlyTextProperty | MappingPercolatorProperty | MappingRankFeatureProperty | MappingRankFeaturesProperty | MappingSearchAsYouTypeProperty | MappingTextProperty | MappingVersionProperty | MappingWildcardProperty | MappingDateNanosProperty | MappingDateProperty | MappingAggregateMetricDoubleProperty | MappingDenseVectorProperty | MappingFlattenedProperty | MappingNestedProperty | MappingObjectProperty | MappingPassthroughObjectProperty | MappingRankVectorProperty | MappingSemanticTextProperty | MappingSparseVectorProperty | MappingCompletionProperty | MappingConstantKeywordProperty | MappingCountedKeywordProperty | MappingFieldAliasProperty | MappingHistogramProperty | MappingIpProperty | MappingMurmur3HashProperty | MappingTokenCountProperty | MappingGeoPointProperty | MappingGeoShapeProperty | MappingPointProperty | MappingShapeProperty | MappingByteNumberProperty | MappingDoubleNumberProperty | MappingFloatNumberProperty | MappingHalfFloatNumberProperty | MappingIntegerNumberProperty | MappingLongNumberProperty | MappingScaledFloatNumberProperty | MappingShortNumberProperty | MappingUnsignedLongNumberProperty | MappingDateRangeProperty | MappingDoubleRangeProperty | MappingFloatRangeProperty | MappingIntegerRangeProperty | MappingIpRangeProperty | MappingLongRangeProperty | MappingIcuCollationProperty export interface MappingPropertyBase { meta?: Record @@ -5815,6 +5861,14 @@ export interface MappingRankFeaturesProperty extends MappingPropertyBase { type: 'rank_features' } +export type MappingRankVectorElementType = 'byte' | 'float' | 'bit' + +export interface MappingRankVectorProperty extends MappingPropertyBase { + type: 'rank_vectors' + element_type?: MappingRankVectorElementType + dims?: integer +} + export interface MappingRoutingField { required: boolean } @@ -5858,11 +5912,16 @@ export interface MappingSearchAsYouTypeProperty extends MappingCorePropertyBase type: 'search_as_you_type' } +export interface MappingSemanticTextIndexOptions { + dense_vector?: MappingDenseVectorIndexOptions +} + export interface MappingSemanticTextProperty { type: 'semantic_text' meta?: Record inference_id?: Id search_inference_id?: Id + index_options?: MappingSemanticTextIndexOptions chunking_settings?: MappingChunkingSettings } @@ -5894,8 +5953,15 @@ export interface MappingSourceField { export type MappingSourceFieldMode = 'disabled' | 'stored' | 'synthetic' +export interface MappingSparseVectorIndexOptions { + prune?: boolean + pruning_config?: TokenPruningConfig +} + export interface MappingSparseVectorProperty extends MappingPropertyBase { + store?: boolean type: 'sparse_vector' + index_options?: MappingSparseVectorIndexOptions } export type MappingSubobjects = boolean | 'true' | 'false' | 'auto' @@ -6725,7 +6791,7 @@ export interface QueryDslSparseVectorQuery extends QueryDslQueryBase { inference_id?: Id query?: string prune?: boolean - pruning_config?: QueryDslTokenPruningConfig + pruning_config?: TokenPruningConfig } export interface QueryDslTermQuery extends QueryDslQueryBase { @@ -6760,17 +6826,11 @@ export interface QueryDslTermsSetQuery extends QueryDslQueryBase { export interface QueryDslTextExpansionQuery extends QueryDslQueryBase { model_id: string model_text: string - pruning_config?: QueryDslTokenPruningConfig + pruning_config?: TokenPruningConfig } export type QueryDslTextQueryType = 'best_fields' | 'most_fields' | 'cross_fields' | 'phrase' | 'phrase_prefix' | 'bool_prefix' -export interface QueryDslTokenPruningConfig { - tokens_freq_ratio_threshold?: integer - tokens_weight_threshold?: float - only_score_pruned_tokens?: boolean -} - export interface QueryDslTypeQuery extends QueryDslQueryBase { value: string } @@ -6790,7 +6850,7 @@ export interface QueryDslUntypedRangeQuery extends QueryDslRangeQueryBase { export interface QueryDslWeightedTokensQuery extends QueryDslQueryBase { tokens: Record - pruning_config?: QueryDslTokenPruningConfig + pruning_config?: TokenPruningConfig } export interface QueryDslWildcardQuery extends QueryDslQueryBase { @@ -7017,9 +7077,29 @@ export type CatCatNodeColumn = 'build' | 'b' | 'completion.size' | 'cs' | 'compl export type CatCatNodeColumns = CatCatNodeColumn | CatCatNodeColumn[] +export type CatCatRecoveryColumn = 'index' | 'i' | 'idx' | 'shard' | 's' | 'sh' | 'time' | 't' | 'ti' | 'primaryOrReplica' | 'type' | 'stage' | 'st' | 'source_host' | 'shost' | 'source_node' | 'snode' | 'target_host' | 'thost' | 'target_node' | 'tnode' | 'repository' | 'tnode' | 'snapshot' | 'snap' | 'files' | 'f' | 'files_recovered' | 'fr' | 'files_percent' | 'fp' | 'files_total' | 'tf' | 'bytes' | 'b' | 'bytes_recovered' | 'br' | 'bytes_percent' | 'bp' | 'bytes_total' | 'tb' | 'translog_ops' | 'to' | 'translog_ops_recovered' | 'tor' | 'translog_ops_percent' | 'top' | 'start_time' | 'start' | 'start_time_millis' | 'start_millis' | 'stop_time' | 'stop' | 'stop_time_millis' | 'stop_millis' | string + +export type CatCatRecoveryColumns = CatCatRecoveryColumn | CatCatRecoveryColumn[] + export interface CatCatRequestBase extends RequestBase, SpecUtilsCommonCatQueryParameters { } +export type CatCatSegmentsColumn = 'index' | 'i' | 'idx' | 'shard' | 's' | 'sh' | 'prirep' | 'p' | 'pr' | 'primaryOrReplica' | 'ip' | 'segment' | 'generation' | 'docs.count' | 'docs.deleted' | 'size' | 'size.memory' | 'committed' | 'searchable' | 'version' | 'compound' | 'id' | string + +export type CatCatSegmentsColumns = CatCatSegmentsColumn | CatCatSegmentsColumn[] + +export type CatCatShardColumn = 'completion.size' | 'cs' | 'completionSize' | 'dataset.size' | 'dense_vector.value_count' | 'dvc' | 'denseVectorCount' | 'docs' | 'd' | 'dc' | 'fielddata.evictions' | 'fe' | 'fielddataEvictions' | 'fielddata.memory_size' | 'fm' | 'fielddataMemory' | 'flush.total' | 'ft' | 'flushTotal' | 'flush.total_time' | 'ftt' | 'flushTotalTime' | 'get.current' | 'gc' | 'getCurrent' | 'get.exists_time' | 'geti' | 'getExistsTime' | 'get.exists_total' | 'geto' | 'getExistsTotal' | 'get.missing_time' | 'gmti' | 'getMissingTime' | 'get.missing_total' | 'gmto' | 'getMissingTotal' | 'get.time' | 'gti' | 'getTime' | 'get.total' | 'gto' | 'getTotal' | 'id' | 'index' | 'i' | 'idx' | 'indexing.delete_current' | 'idc' | 'indexingDeleteCurrent' | 'indexing.delete_time' | 'idti' | 'indexingDeleteTime' | 'indexing.delete_total' | 'idto' | 'indexingDeleteTotal' | 'indexing.index_current' | 'iic' | 'indexingIndexCurrent' | 'indexing.index_failed_due_to_version_conflict' | 'iifvc' | 'indexingIndexFailedDueToVersionConflict' | 'indexing.index_failed' | 'iif' | 'indexingIndexFailed' | 'indexing.index_time' | 'iiti' | 'indexingIndexTime' | 'indexing.index_total' | 'iito' | 'indexingIndexTotal' | 'ip' | 'merges.current' | 'mc' | 'mergesCurrent' | 'merges.current_docs' | 'mcd' | 'mergesCurrentDocs' | 'merges.current_size' | 'mcs' | 'mergesCurrentSize' | 'merges.total' | 'mt' | 'mergesTotal' | 'merges.total_docs' | 'mtd' | 'mergesTotalDocs' | 'merges.total_size' | 'mts' | 'mergesTotalSize' | 'merges.total_time' | 'mtt' | 'mergesTotalTime' | 'node' | 'n' | 'prirep' | 'p' | 'pr' | 'primaryOrReplica' | 'query_cache.evictions' | 'qce' | 'queryCacheEvictions' | 'query_cache.memory_size' | 'qcm' | 'queryCacheMemory' | 'recoverysource.type' | 'rs' | 'refresh.time' | 'rti' | 'refreshTime' | 'refresh.total' | 'rto' | 'refreshTotal' | 'search.fetch_current' | 'sfc' | 'searchFetchCurrent' | 'search.fetch_time' | 'sfti' | 'searchFetchTime' | 'search.fetch_total' | 'sfto' | 'searchFetchTotal' | 'search.open_contexts' | 'so' | 'searchOpenContexts' | 'search.query_current' | 'sqc' | 'searchQueryCurrent' | 'search.query_time' | 'sqti' | 'searchQueryTime' | 'search.query_total' | 'sqto' | 'searchQueryTotal' | 'search.scroll_current' | 'scc' | 'searchScrollCurrent' | 'search.scroll_time' | 'scti' | 'searchScrollTime' | 'search.scroll_total' | 'scto' | 'searchScrollTotal' | 'segments.count' | 'sc' | 'segmentsCount' | 'segments.fixed_bitset_memory' | 'sfbm' | 'fixedBitsetMemory' | 'segments.index_writer_memory' | 'siwm' | 'segmentsIndexWriterMemory' | 'segments.memory' | 'sm' | 'segmentsMemory' | 'segments.version_map_memory' | 'svmm' | 'segmentsVersionMapMemory' | 'seq_no.global_checkpoint' | 'sqg' | 'globalCheckpoint' | 'seq_no.local_checkpoint' | 'sql' | 'localCheckpoint' | 'seq_no.max' | 'sqm' | 'maxSeqNo' | 'shard' | 's' | 'sh' | 'dsparse_vector.value_count' | 'svc' | 'sparseVectorCount' | 'state' | 'st' | 'store' | 'sto' | 'suggest.current' | 'suc' | 'suggestCurrent' | 'suggest.time' | 'suti' | 'suggestTime' | 'suggest.total' | 'suto' | 'suggestTotal' | 'sync_id' | 'unassigned.at' | 'ua' | 'unassigned.details' | 'ud' | 'unassigned.for' | 'uf' | 'unassigned.reason' | 'ur' | string + +export type CatCatShardColumns = CatCatShardColumn | CatCatShardColumn[] + +export type CatCatSnapshotsColumn = 'id' | 'snapshot' | 'repository' | 're' | 'repo' | 'status' | 's' | 'start_epoch' | 'ste' | 'startEpoch' | 'start_time' | 'sti' | 'startTime' | 'end_epoch' | 'ete' | 'endEpoch' | 'end_time' | 'eti' | 'endTime' | 'duration' | 'dur' | 'indices' | 'i' | 'successful_shards' | 'ss' | 'failed_shards' | 'fs' | 'total_shards' | 'ts' | 'reason' | 'r' | string + +export type CatCatSnapshotsColumns = CatCatSnapshotsColumn | CatCatNodeColumn[] + +export type CatCatThreadPoolColumn = 'active' | 'a' | 'completed' | 'c' | 'core' | 'cr' | 'ephemeral_id' | 'eid' | 'host' | 'h' | 'ip' | 'i' | 'keep_alive' | 'k' | 'largest' | 'l' | 'max' | 'mx' | 'name' | 'node_id' | 'id' | 'node_name' | 'pid' | 'p' | 'pool_size' | 'psz' | 'port' | 'po' | 'queue' | 'q' | 'queue_size' | 'qs' | 'rejected' | 'r' | 'size' | 'sz' | 'type' | 't' | string + +export type CatCatThreadPoolColumns = CatCatThreadPoolColumn | CatCatThreadPoolColumn[] + export type CatCatTrainedModelsColumn = 'create_time' | 'ct' | 'created_by' | 'c' | 'createdBy' | 'data_frame_analytics_id' | 'df' | 'dataFrameAnalytics' | 'dfid' | 'description' | 'd' | 'heap_size' | 'hs' | 'modelHeapSize' | 'id' | 'ingest.count' | 'ic' | 'ingestCount' | 'ingest.current' | 'icurr' | 'ingestCurrent' | 'ingest.failed' | 'if' | 'ingestFailed' | 'ingest.pipelines' | 'ip' | 'ingestPipelines' | 'ingest.time' | 'it' | 'ingestTime' | 'license' | 'l' | 'operations' | 'o' | 'modelOperations' | 'version' | 'v' export type CatCatTrainedModelsColumns = CatCatTrainedModelsColumn | CatCatTrainedModelsColumn[] @@ -8323,7 +8403,7 @@ export interface CatRecoveryRequest extends CatCatRequestBase { active_only?: boolean bytes?: Bytes detailed?: boolean - h?: Names + h?: CatCatRecoveryColumns s?: Names time?: TimeUnit } @@ -8349,7 +8429,7 @@ export type CatRepositoriesResponse = CatRepositoriesRepositoriesRecord[] export interface CatSegmentsRequest extends CatCatRequestBase { index?: Indices bytes?: Bytes - h?: Names + h?: CatCatSegmentsColumns s?: Names local?: boolean master_timeout?: Duration @@ -8402,7 +8482,7 @@ export interface CatSegmentsSegmentsRecord { export interface CatShardsRequest extends CatCatRequestBase { index?: Indices bytes?: Bytes - h?: Names + h?: CatCatShardColumns s?: Names master_timeout?: Duration time?: TimeUnit @@ -8628,7 +8708,7 @@ export interface CatShardsShardsRecord { export interface CatSnapshotsRequest extends CatCatRequestBase { repository?: Names ignore_unavailable?: boolean - h?: Names + h?: CatCatSnapshotsColumns s?: Names master_timeout?: Duration time?: TimeUnit @@ -8745,7 +8825,7 @@ export interface CatTemplatesTemplatesRecord { export interface CatThreadPoolRequest extends CatCatRequestBase { thread_pool_patterns?: Names - h?: Names + h?: CatCatThreadPoolColumns s?: Names time?: TimeUnit local?: boolean @@ -9431,6 +9511,7 @@ export type ClusterPostVotingConfigExclusionsResponse = boolean export interface ClusterPutComponentTemplateRequest extends RequestBase { name: Name create?: boolean + cause?: string master_timeout?: Duration template: IndicesIndexState version?: VersionNumber @@ -10613,6 +10694,60 @@ export type EqlSearchResponse = EqlEqlSearchResponseBase +} + +export interface EsqlColumnInfo { + name: string + type: string +} + +export interface EsqlEsqlClusterDetails { + status: EsqlEsqlClusterStatus + indices: string + took?: DurationValue + _shards?: EsqlEsqlShardInfo +} + +export type EsqlEsqlClusterStatus = 'running' | 'successful' | 'partial' | 'skipped' | 'failed' + +export interface EsqlEsqlResult { + took?: DurationValue + is_partial?: boolean + all_columns?: EsqlColumnInfo[] + columns: EsqlColumnInfo[] + values: FieldValue[][] + _clusters?: EsqlClusterInfo + profile?: any +} + +export interface EsqlEsqlShardFailure { + shard: Id + index: IndexName + node?: NodeId + reason: ErrorCause +} + +export interface EsqlEsqlShardInfo { + total: integer + successful?: integer + skipped?: integer + failed?: integer + failures?: EsqlEsqlShardFailure[] +} + export interface EsqlTableValuesContainer { integer?: EsqlTableValuesIntegerValue[] keyword?: EsqlTableValuesKeywordValue[] @@ -10646,7 +10781,7 @@ export interface EsqlAsyncQueryRequest extends RequestBase { keep_on_completion?: boolean } -export type EsqlAsyncQueryResponse = EsqlResult +export type EsqlAsyncQueryResponse = EsqlEsqlResult export interface EsqlAsyncQueryDeleteRequest extends RequestBase { id: Id @@ -10657,18 +10792,19 @@ export type EsqlAsyncQueryDeleteResponse = AcknowledgedResponseBase export interface EsqlAsyncQueryGetRequest extends RequestBase { id: Id drop_null_columns?: boolean + format?: EsqlQueryEsqlFormat keep_alive?: Duration wait_for_completion_timeout?: Duration } -export type EsqlAsyncQueryGetResponse = EsqlResult +export type EsqlAsyncQueryGetResponse = EsqlAsyncEsqlResult export interface EsqlAsyncQueryStopRequest extends RequestBase { id: Id drop_null_columns?: boolean } -export type EsqlAsyncQueryStopResponse = EsqlResult +export type EsqlAsyncQueryStopResponse = EsqlEsqlResult export type EsqlQueryEsqlFormat = 'csv' | 'json' | 'tsv' | 'txt' | 'yaml' | 'cbor' | 'smile' | 'arrow' @@ -10687,7 +10823,7 @@ export interface EsqlQueryRequest extends RequestBase { include_ccs_metadata?: boolean } -export type EsqlQueryResponse = EsqlResult +export type EsqlQueryResponse = EsqlEsqlResult export interface FeaturesFeature { name: string @@ -11613,6 +11749,7 @@ export type IndicesSourceMode = 'disabled' | 'stored' | 'synthetic' export interface IndicesStorage { type: IndicesStorageType allow_mmap?: boolean + stats_refresh_interval?: Duration } export type IndicesStorageType = 'fs' | 'niofs' | 'mmapfs' | 'hybridfs' | string @@ -12087,10 +12224,6 @@ export interface IndicesGetRequest extends RequestBase { export type IndicesGetResponse = Record -export interface IndicesGetAliasIndexAliases { - aliases: Record -} - export interface IndicesGetAliasRequest extends RequestBase { name?: Names index?: Indices @@ -12102,6 +12235,17 @@ export interface IndicesGetAliasRequest extends RequestBase { export type IndicesGetAliasResponse = Record +export interface IndicesGetAliasIndexAliases { + aliases: Record +} + +export interface IndicesGetAliasNotFoundAliasesKeys { + error: string + status: number +} +export type IndicesGetAliasNotFoundAliases = IndicesGetAliasNotFoundAliasesKeys +& { [property: string]: IndicesGetAliasIndexAliases | string | number } + export interface IndicesGetDataLifecycleDataStreamWithLifecycle { name: DataStreamName lifecycle?: IndicesDataStreamLifecycleWithRollover @@ -13210,6 +13354,14 @@ export interface InferenceContentObject { type: string } +export interface InferenceDeepSeekServiceSettings { + api_key: string + model_id: string + url?: string +} + +export type InferenceDeepSeekServiceType = 'deepseek' + export interface InferenceDeleteInferenceEndpointResult extends AcknowledgedResponseBase { pipelines: string[] } @@ -13311,11 +13463,86 @@ export interface InferenceInferenceEndpointInfoAlibabaCloudAI extends InferenceI task_type: InferenceTaskTypeAlibabaCloudAI } +export interface InferenceInferenceEndpointInfoAmazonBedrock extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeAmazonBedrock +} + +export interface InferenceInferenceEndpointInfoAnthropic extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeAnthropic +} + +export interface InferenceInferenceEndpointInfoAzureAIStudio extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeAzureAIStudio +} + +export interface InferenceInferenceEndpointInfoAzureOpenAI extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeAzureOpenAI +} + +export interface InferenceInferenceEndpointInfoCohere extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeCohere +} + +export interface InferenceInferenceEndpointInfoDeepSeek extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeDeepSeek +} + +export interface InferenceInferenceEndpointInfoELSER extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeELSER +} + +export interface InferenceInferenceEndpointInfoElasticsearch extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeElasticsearch +} + +export interface InferenceInferenceEndpointInfoGoogleAIStudio extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeGoogleAIStudio +} + +export interface InferenceInferenceEndpointInfoGoogleVertexAI extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeGoogleVertexAI +} + +export interface InferenceInferenceEndpointInfoHuggingFace extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeHuggingFace +} + export interface InferenceInferenceEndpointInfoJinaAi extends InferenceInferenceEndpoint { inference_id: string task_type: InferenceTaskTypeJinaAi } +export interface InferenceInferenceEndpointInfoMistral extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeMistral +} + +export interface InferenceInferenceEndpointInfoOpenAI extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeOpenAI +} + +export interface InferenceInferenceEndpointInfoVoyageAI extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeVoyageAI +} + +export interface InferenceInferenceEndpointInfoWatsonx extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeWatsonx +} + export interface InferenceInferenceResult { text_embedding_bytes?: InferenceTextEmbeddingByteResult[] text_embedding_bits?: InferenceTextEmbeddingByteResult[] @@ -13426,8 +13653,38 @@ export type InferenceTaskType = 'sparse_embedding' | 'text_embedding' | 'rerank' export type InferenceTaskTypeAlibabaCloudAI = 'text_embedding' | 'rerank' | 'completion' | 'sparse_embedding' +export type InferenceTaskTypeAmazonBedrock = 'text_embedding' | 'completion' + +export type InferenceTaskTypeAnthropic = 'completion' + +export type InferenceTaskTypeAzureAIStudio = 'text_embedding' | 'completion' + +export type InferenceTaskTypeAzureOpenAI = 'text_embedding' | 'completion' + +export type InferenceTaskTypeCohere = 'text_embedding' | 'rerank' | 'completion' + +export type InferenceTaskTypeDeepSeek = 'completion' | 'chat_completion' + +export type InferenceTaskTypeELSER = 'sparse_embedding' + +export type InferenceTaskTypeElasticsearch = 'sparse_embedding' | 'text_embedding' | 'rerank' + +export type InferenceTaskTypeGoogleAIStudio = 'text_embedding' | 'completion' + +export type InferenceTaskTypeGoogleVertexAI = 'text_embedding' | 'rerank' + +export type InferenceTaskTypeHuggingFace = 'text_embedding' + export type InferenceTaskTypeJinaAi = 'text_embedding' | 'rerank' +export type InferenceTaskTypeMistral = 'text_embedding' + +export type InferenceTaskTypeOpenAI = 'text_embedding' | 'chat_completion' | 'completion' + +export type InferenceTaskTypeVoyageAI = 'text_embedding' | 'rerank' + +export type InferenceTaskTypeWatsonx = 'text_embedding' + export interface InferenceTextEmbeddingByteResult { embedding: InferenceDenseByteVector } @@ -13558,7 +13815,7 @@ export interface InferencePutAmazonbedrockRequest extends RequestBase { task_settings?: InferenceAmazonBedrockTaskSettings } -export type InferencePutAmazonbedrockResponse = InferenceInferenceEndpointInfo +export type InferencePutAmazonbedrockResponse = InferenceInferenceEndpointInfoAmazonBedrock export interface InferencePutAnthropicRequest extends RequestBase { task_type: InferenceAnthropicTaskType @@ -13569,7 +13826,7 @@ export interface InferencePutAnthropicRequest extends RequestBase { task_settings?: InferenceAnthropicTaskSettings } -export type InferencePutAnthropicResponse = InferenceInferenceEndpointInfo +export type InferencePutAnthropicResponse = InferenceInferenceEndpointInfoAnthropic export interface InferencePutAzureaistudioRequest extends RequestBase { task_type: InferenceAzureAiStudioTaskType @@ -13580,7 +13837,7 @@ export interface InferencePutAzureaistudioRequest extends RequestBase { task_settings?: InferenceAzureAiStudioTaskSettings } -export type InferencePutAzureaistudioResponse = InferenceInferenceEndpointInfo +export type InferencePutAzureaistudioResponse = InferenceInferenceEndpointInfoAzureAIStudio export interface InferencePutAzureopenaiRequest extends RequestBase { task_type: InferenceAzureOpenAITaskType @@ -13591,7 +13848,7 @@ export interface InferencePutAzureopenaiRequest extends RequestBase { task_settings?: InferenceAzureOpenAITaskSettings } -export type InferencePutAzureopenaiResponse = InferenceInferenceEndpointInfo +export type InferencePutAzureopenaiResponse = InferenceInferenceEndpointInfoAzureOpenAI export interface InferencePutCohereRequest extends RequestBase { task_type: InferenceCohereTaskType @@ -13602,7 +13859,17 @@ export interface InferencePutCohereRequest extends RequestBase { task_settings?: InferenceCohereTaskSettings } -export type InferencePutCohereResponse = InferenceInferenceEndpointInfo +export type InferencePutCohereResponse = InferenceInferenceEndpointInfoCohere + +export interface InferencePutDeepseekRequest extends RequestBase { + task_type: InferenceTaskTypeDeepSeek + deepseek_inference_id: Id + chunking_settings?: InferenceInferenceChunkingSettings + service: InferenceDeepSeekServiceType + service_settings: InferenceDeepSeekServiceSettings +} + +export type InferencePutDeepseekResponse = InferenceInferenceEndpointInfoDeepSeek export interface InferencePutElasticsearchRequest extends RequestBase { task_type: InferenceElasticsearchTaskType @@ -13613,7 +13880,7 @@ export interface InferencePutElasticsearchRequest extends RequestBase { task_settings?: InferenceElasticsearchTaskSettings } -export type InferencePutElasticsearchResponse = InferenceInferenceEndpointInfo +export type InferencePutElasticsearchResponse = InferenceInferenceEndpointInfoElasticsearch export interface InferencePutElserRequest extends RequestBase { task_type: InferenceElserTaskType @@ -13623,7 +13890,7 @@ export interface InferencePutElserRequest extends RequestBase { service_settings: InferenceElserServiceSettings } -export type InferencePutElserResponse = InferenceInferenceEndpointInfo +export type InferencePutElserResponse = InferenceInferenceEndpointInfoELSER export interface InferencePutGoogleaistudioRequest extends RequestBase { task_type: InferenceGoogleAiStudioTaskType @@ -13633,7 +13900,7 @@ export interface InferencePutGoogleaistudioRequest extends RequestBase { service_settings: InferenceGoogleAiStudioServiceSettings } -export type InferencePutGoogleaistudioResponse = InferenceInferenceEndpointInfo +export type InferencePutGoogleaistudioResponse = InferenceInferenceEndpointInfoGoogleAIStudio export interface InferencePutGooglevertexaiRequest extends RequestBase { task_type: InferenceGoogleVertexAITaskType @@ -13644,7 +13911,7 @@ export interface InferencePutGooglevertexaiRequest extends RequestBase { task_settings?: InferenceGoogleVertexAITaskSettings } -export type InferencePutGooglevertexaiResponse = InferenceInferenceEndpointInfo +export type InferencePutGooglevertexaiResponse = InferenceInferenceEndpointInfoGoogleVertexAI export interface InferencePutHuggingFaceRequest extends RequestBase { task_type: InferenceHuggingFaceTaskType @@ -13655,7 +13922,7 @@ export interface InferencePutHuggingFaceRequest extends RequestBase { task_settings?: InferenceHuggingFaceTaskSettings } -export type InferencePutHuggingFaceResponse = InferenceInferenceEndpointInfo +export type InferencePutHuggingFaceResponse = InferenceInferenceEndpointInfoHuggingFace export interface InferencePutJinaaiRequest extends RequestBase { task_type: InferenceJinaAITaskType @@ -13676,7 +13943,7 @@ export interface InferencePutMistralRequest extends RequestBase { service_settings: InferenceMistralServiceSettings } -export type InferencePutMistralResponse = InferenceInferenceEndpointInfo +export type InferencePutMistralResponse = InferenceInferenceEndpointInfoMistral export interface InferencePutOpenaiRequest extends RequestBase { task_type: InferenceOpenAITaskType @@ -13687,7 +13954,7 @@ export interface InferencePutOpenaiRequest extends RequestBase { task_settings?: InferenceOpenAITaskSettings } -export type InferencePutOpenaiResponse = InferenceInferenceEndpointInfo +export type InferencePutOpenaiResponse = InferenceInferenceEndpointInfoOpenAI export interface InferencePutVoyageaiRequest extends RequestBase { task_type: InferenceVoyageAITaskType @@ -13698,7 +13965,7 @@ export interface InferencePutVoyageaiRequest extends RequestBase { task_settings?: InferenceVoyageAITaskSettings } -export type InferencePutVoyageaiResponse = InferenceInferenceEndpointInfo +export type InferencePutVoyageaiResponse = InferenceInferenceEndpointInfoVoyageAI export interface InferencePutWatsonxRequest extends RequestBase { task_type: InferenceWatsonxTaskType @@ -13707,7 +13974,7 @@ export interface InferencePutWatsonxRequest extends RequestBase { service_settings: InferenceWatsonxServiceSettings } -export type InferencePutWatsonxResponse = InferenceInferenceEndpointInfo +export type InferencePutWatsonxResponse = InferenceInferenceEndpointInfoWatsonx export interface InferenceRerankRequest extends RequestBase { inference_id: Id @@ -14355,7 +14622,6 @@ export interface IngestGetIpLocationDatabaseDatabaseConfigurationMetadata { export interface IngestGetIpLocationDatabaseRequest extends RequestBase { id?: Ids - master_timeout?: Duration } export interface IngestGetIpLocationDatabaseResponse { @@ -14519,7 +14785,7 @@ export interface LicensePostStartBasicResponse { export interface LicensePostStartTrialRequest extends RequestBase { acknowledge?: boolean - type_query_string?: string + type?: string master_timeout?: Duration } @@ -17741,12 +18007,13 @@ export interface NodesInfoNodeInfo { build_flavor: string build_hash: string build_type: string + component_versions: Record host: Host http?: NodesInfoNodeInfoHttp + index_version: VersionNumber ip: Ip jvm?: NodesInfoNodeJvmInfo name: Name - network?: NodesInfoNodeInfoNetwork os?: NodesInfoNodeOperatingSystemInfo plugins?: PluginStats[] process?: NodesInfoNodeProcessInfo @@ -17757,10 +18024,12 @@ export interface NodesInfoNodeInfo { total_indexing_buffer_in_bytes?: ByteSize transport?: NodesInfoNodeInfoTransport transport_address: TransportAddress + transport_version: VersionNumber version: VersionString modules?: PluginStats[] ingest?: NodesInfoNodeInfoIngest aggregations?: Record + remote_cluster_server?: NodesInfoRemoveClusterServer } export interface NodesInfoNodeInfoAction { @@ -17780,7 +18049,7 @@ export interface NodesInfoNodeInfoClient { } export interface NodesInfoNodeInfoDiscoverKeys { - seed_hosts?: string[] + seed_hosts?: string[] | string type?: string seed_providers?: string[] } @@ -17828,17 +18097,6 @@ export interface NodesInfoNodeInfoMemory { total_in_bytes: long } -export interface NodesInfoNodeInfoNetwork { - primary_interface: NodesInfoNodeInfoNetworkInterface - refresh_interval: integer -} - -export interface NodesInfoNodeInfoNetworkInterface { - address: string - mac_address: string - name: Name -} - export interface NodesInfoNodeInfoOSCPU { cache_size: string cache_size_in_bytes: integer @@ -17900,7 +18158,7 @@ export interface NodesInfoNodeInfoSettingsCluster { name: Name routing?: IndicesIndexRouting election: NodesInfoNodeInfoSettingsClusterElection - initial_master_nodes?: string[] + initial_master_nodes?: string[] | string deprecation_indexing?: NodesInfoDeprecationIndexing } @@ -17970,6 +18228,7 @@ export interface NodesInfoNodeInfoSettingsTransport { type: NodesInfoNodeInfoSettingsTransportType | string 'type.default'?: string features?: NodesInfoNodeInfoSettingsTransportFeatures + ignore_deserialization_errors?: SpecUtilsStringified } export interface NodesInfoNodeInfoSettingsTransportFeatures { @@ -18047,7 +18306,6 @@ export interface NodesInfoNodeJvmInfo { vm_vendor: string vm_version: VersionString using_bundled_jdk: boolean - bundled_jdk: boolean using_compressed_ordinary_object_pointers?: boolean | string input_arguments: string[] } @@ -18080,6 +18338,11 @@ export interface NodesInfoNodeThreadPoolInfo { type: string } +export interface NodesInfoRemoveClusterServer { + bound_address: TransportAddress[] + publish_address: TransportAddress +} + export interface NodesInfoRequest extends RequestBase { node_id?: NodeIds metric?: Metrics diff --git a/src/api/typesWithBodyKey.ts b/src/api/typesWithBodyKey.ts index 11cf3e0e5..3b1ab9c97 100644 --- a/src/api/typesWithBodyKey.ts +++ b/src/api/typesWithBodyKey.ts @@ -479,7 +479,6 @@ export interface GetSourceRequest extends RequestBase { _source?: SearchSourceConfigParam _source_excludes?: Fields _source_includes?: Fields - stored_fields?: Fields version?: VersionNumber version_type?: VersionType } @@ -680,6 +679,7 @@ export interface IndexRequest extends RequestBase { version_type?: VersionType wait_for_active_shards?: WaitForActiveShards require_alias?: boolean + require_data_stream?: boolean /** @deprecated The use of the 'body' key has been deprecated, use 'document' instead. */ body?: TDocument } @@ -2335,8 +2335,6 @@ export interface ErrorResponseBase { status: integer } -export type EsqlResult = ArrayBuffer - export type ExpandWildcard = 'all' | 'open' | 'closed' | 'hidden' | 'none' export type ExpandWildcards = ExpandWildcard | ExpandWildcard[] @@ -2557,7 +2555,7 @@ export type LifecycleOperationMode = 'RUNNING' | 'STOPPING' | 'STOPPED' export interface LinearRetriever extends RetrieverBase { retrievers?: InnerRetriever[] - rank_window_size: integer + rank_window_size?: integer } export type MapboxVectorTiles = ArrayBuffer @@ -2652,7 +2650,7 @@ export interface PinnedRetriever extends RetrieverBase { retriever: RetrieverContainer ids?: string[] docs?: SpecifiedDocument[] - rank_window_size: integer + rank_window_size?: integer } export type PipelineName = string @@ -2987,8 +2985,8 @@ export interface TextSimilarityReranker extends RetrieverBase { retriever: RetrieverContainer rank_window_size?: integer inference_id?: string - inference_text?: string - field?: string + inference_text: string + field: string } export type ThreadType = 'cpu' | 'wait' | 'block' | 'gpu' | 'mem' @@ -2999,6 +2997,12 @@ export type TimeUnit = 'nanos' | 'micros' | 'ms' | 's' | 'm' | 'h' | 'd' export type TimeZone = string +export interface TokenPruningConfig { + tokens_freq_ratio_threshold?: integer + tokens_weight_threshold?: float + only_score_pruned_tokens?: boolean +} + export interface TopLeftBottomRightGeoBounds { top_left: GeoLocation bottom_right: GeoLocation @@ -4527,6 +4531,10 @@ export interface AnalysisArabicNormalizationTokenFilter extends AnalysisTokenFil type: 'arabic_normalization' } +export interface AnalysisArabicStemTokenFilter extends AnalysisTokenFilterBase { + type: 'arabic_stem' +} + export interface AnalysisArmenianAnalyzer { type: 'armenian' stopwords?: AnalysisStopWords @@ -4553,12 +4561,20 @@ export interface AnalysisBengaliAnalyzer { stem_exclusion?: string[] } +export interface AnalysisBengaliNormalizationTokenFilter extends AnalysisTokenFilterBase { + type: 'bengali_normalization' +} + export interface AnalysisBrazilianAnalyzer { type: 'brazilian' stopwords?: AnalysisStopWords stopwords_path?: string } +export interface AnalysisBrazilianStemTokenFilter extends AnalysisTokenFilterBase { + type: 'brazilian_stem' +} + export interface AnalysisBulgarianAnalyzer { type: 'bulgarian' stopwords?: AnalysisStopWords @@ -4665,6 +4681,10 @@ export interface AnalysisCzechAnalyzer { stem_exclusion?: string[] } +export interface AnalysisCzechStemTokenFilter extends AnalysisTokenFilterBase { + type: 'czech_stem' +} + export interface AnalysisDanishAnalyzer { type: 'danish' stopwords?: AnalysisStopWords @@ -4694,6 +4714,10 @@ export interface AnalysisDutchAnalyzer { stem_exclusion?: string[] } +export interface AnalysisDutchStemTokenFilter extends AnalysisTokenFilterBase { + type: 'dutch_stem' +} + export type AnalysisEdgeNGramSide = 'front' | 'back' export interface AnalysisEdgeNGramTokenFilter extends AnalysisTokenFilterBase { @@ -4709,7 +4733,7 @@ export interface AnalysisEdgeNGramTokenizer extends AnalysisTokenizerBase { custom_token_chars?: string max_gram?: integer min_gram?: integer - token_chars?: AnalysisTokenChar[] + token_chars?: string | AnalysisTokenChar[] } export interface AnalysisElisionTokenFilter extends AnalysisTokenFilterBase { @@ -4766,6 +4790,10 @@ export interface AnalysisFrenchAnalyzer { stem_exclusion?: string[] } +export interface AnalysisFrenchStemTokenFilter extends AnalysisTokenFilterBase { + type: 'french_stem' +} + export interface AnalysisGalicianAnalyzer { type: 'galician' stopwords?: AnalysisStopWords @@ -4784,6 +4812,10 @@ export interface AnalysisGermanNormalizationTokenFilter extends AnalysisTokenFil type: 'german_normalization' } +export interface AnalysisGermanStemTokenFilter extends AnalysisTokenFilterBase { + type: 'german_stem' +} + export interface AnalysisGreekAnalyzer { type: 'greek' stopwords?: AnalysisStopWords @@ -4847,16 +4879,16 @@ export type AnalysisIcuCollationStrength = 'primary' | 'secondary' | 'tertiary' export interface AnalysisIcuCollationTokenFilter extends AnalysisTokenFilterBase { type: 'icu_collation' alternate?: AnalysisIcuCollationAlternate - case_first?: AnalysisIcuCollationCaseFirst - case_level?: boolean + caseFirst?: AnalysisIcuCollationCaseFirst + caseLevel?: boolean country?: string decomposition?: AnalysisIcuCollationDecomposition - hiragana_quaternary_mode?: boolean + hiraganaQuaternaryMode?: boolean language?: string numeric?: boolean rules?: string strength?: AnalysisIcuCollationStrength - variable_top?: string + variableTop?: string variant?: string } @@ -5151,6 +5183,7 @@ export interface AnalysisPatternReplaceCharFilter extends AnalysisCharFilterBase export interface AnalysisPatternReplaceTokenFilter extends AnalysisTokenFilterBase { type: 'pattern_replace' all?: boolean + flags?: string pattern: string replacement?: string } @@ -5172,6 +5205,10 @@ export interface AnalysisPersianNormalizationTokenFilter extends AnalysisTokenFi type: 'persian_normalization' } +export interface AnalysisPersianStemTokenFilter extends AnalysisTokenFilterBase { + type: 'persian_stem' +} + export type AnalysisPhoneticEncoder = 'metaphone' | 'double_metaphone' | 'soundex' | 'refined_soundex' | 'caverphone1' | 'caverphone2' | 'cologne' | 'nysiis' | 'koelnerphonetik' | 'haasephonetik' | 'beider_morse' | 'daitch_mokotoff' export type AnalysisPhoneticLanguage = 'any' | 'common' | 'cyrillic' | 'english' | 'french' | 'german' | 'hebrew' | 'hungarian' | 'polish' | 'romanian' | 'russian' | 'spanish' @@ -5228,6 +5265,10 @@ export interface AnalysisRussianAnalyzer { stem_exclusion?: string[] } +export interface AnalysisRussianStemTokenFilter extends AnalysisTokenFilterBase { + type: 'russian_stem' +} + export interface AnalysisScandinavianFoldingTokenFilter extends AnalysisTokenFilterBase { type: 'scandinavian_folding' } @@ -5392,7 +5433,7 @@ export interface AnalysisTokenFilterBase { version?: VersionString } -export type AnalysisTokenFilterDefinition = AnalysisApostropheTokenFilter | AnalysisArabicNormalizationTokenFilter | AnalysisAsciiFoldingTokenFilter | AnalysisCjkBigramTokenFilter | AnalysisCjkWidthTokenFilter | AnalysisClassicTokenFilter | AnalysisCommonGramsTokenFilter | AnalysisConditionTokenFilter | AnalysisDecimalDigitTokenFilter | AnalysisDelimitedPayloadTokenFilter | AnalysisEdgeNGramTokenFilter | AnalysisElisionTokenFilter | AnalysisFingerprintTokenFilter | AnalysisFlattenGraphTokenFilter | AnalysisGermanNormalizationTokenFilter | AnalysisHindiNormalizationTokenFilter | AnalysisHunspellTokenFilter | AnalysisHyphenationDecompounderTokenFilter | AnalysisIndicNormalizationTokenFilter | AnalysisKeepTypesTokenFilter | AnalysisKeepWordsTokenFilter | AnalysisKeywordMarkerTokenFilter | AnalysisKeywordRepeatTokenFilter | AnalysisKStemTokenFilter | AnalysisLengthTokenFilter | AnalysisLimitTokenCountTokenFilter | AnalysisLowercaseTokenFilter | AnalysisMinHashTokenFilter | AnalysisMultiplexerTokenFilter | AnalysisNGramTokenFilter | AnalysisNoriPartOfSpeechTokenFilter | AnalysisPatternCaptureTokenFilter | AnalysisPatternReplaceTokenFilter | AnalysisPersianNormalizationTokenFilter | AnalysisPorterStemTokenFilter | AnalysisPredicateTokenFilter | AnalysisRemoveDuplicatesTokenFilter | AnalysisReverseTokenFilter | AnalysisScandinavianFoldingTokenFilter | AnalysisScandinavianNormalizationTokenFilter | AnalysisSerbianNormalizationTokenFilter | AnalysisShingleTokenFilter | AnalysisSnowballTokenFilter | AnalysisSoraniNormalizationTokenFilter | AnalysisStemmerOverrideTokenFilter | AnalysisStemmerTokenFilter | AnalysisStopTokenFilter | AnalysisSynonymGraphTokenFilter | AnalysisSynonymTokenFilter | AnalysisTrimTokenFilter | AnalysisTruncateTokenFilter | AnalysisUniqueTokenFilter | AnalysisUppercaseTokenFilter | AnalysisWordDelimiterGraphTokenFilter | AnalysisWordDelimiterTokenFilter | AnalysisJaStopTokenFilter | AnalysisKuromojiStemmerTokenFilter | AnalysisKuromojiReadingFormTokenFilter | AnalysisKuromojiPartOfSpeechTokenFilter | AnalysisIcuCollationTokenFilter | AnalysisIcuFoldingTokenFilter | AnalysisIcuNormalizationTokenFilter | AnalysisIcuTransformTokenFilter | AnalysisPhoneticTokenFilter | AnalysisDictionaryDecompounderTokenFilter +export type AnalysisTokenFilterDefinition = AnalysisApostropheTokenFilter | AnalysisArabicStemTokenFilter | AnalysisArabicNormalizationTokenFilter | AnalysisAsciiFoldingTokenFilter | AnalysisBengaliNormalizationTokenFilter | AnalysisBrazilianStemTokenFilter | AnalysisCjkBigramTokenFilter | AnalysisCjkWidthTokenFilter | AnalysisClassicTokenFilter | AnalysisCommonGramsTokenFilter | AnalysisConditionTokenFilter | AnalysisCzechStemTokenFilter | AnalysisDecimalDigitTokenFilter | AnalysisDelimitedPayloadTokenFilter | AnalysisDutchStemTokenFilter | AnalysisEdgeNGramTokenFilter | AnalysisElisionTokenFilter | AnalysisFingerprintTokenFilter | AnalysisFlattenGraphTokenFilter | AnalysisFrenchStemTokenFilter | AnalysisGermanNormalizationTokenFilter | AnalysisGermanStemTokenFilter | AnalysisHindiNormalizationTokenFilter | AnalysisHunspellTokenFilter | AnalysisHyphenationDecompounderTokenFilter | AnalysisIndicNormalizationTokenFilter | AnalysisKeepTypesTokenFilter | AnalysisKeepWordsTokenFilter | AnalysisKeywordMarkerTokenFilter | AnalysisKeywordRepeatTokenFilter | AnalysisKStemTokenFilter | AnalysisLengthTokenFilter | AnalysisLimitTokenCountTokenFilter | AnalysisLowercaseTokenFilter | AnalysisMinHashTokenFilter | AnalysisMultiplexerTokenFilter | AnalysisNGramTokenFilter | AnalysisNoriPartOfSpeechTokenFilter | AnalysisPatternCaptureTokenFilter | AnalysisPatternReplaceTokenFilter | AnalysisPersianNormalizationTokenFilter | AnalysisPersianStemTokenFilter | AnalysisPorterStemTokenFilter | AnalysisPredicateTokenFilter | AnalysisRemoveDuplicatesTokenFilter | AnalysisReverseTokenFilter | AnalysisRussianStemTokenFilter | AnalysisScandinavianFoldingTokenFilter | AnalysisScandinavianNormalizationTokenFilter | AnalysisSerbianNormalizationTokenFilter | AnalysisShingleTokenFilter | AnalysisSnowballTokenFilter | AnalysisSoraniNormalizationTokenFilter | AnalysisStemmerOverrideTokenFilter | AnalysisStemmerTokenFilter | AnalysisStopTokenFilter | AnalysisSynonymGraphTokenFilter | AnalysisSynonymTokenFilter | AnalysisTrimTokenFilter | AnalysisTruncateTokenFilter | AnalysisUniqueTokenFilter | AnalysisUppercaseTokenFilter | AnalysisWordDelimiterGraphTokenFilter | AnalysisWordDelimiterTokenFilter | AnalysisJaStopTokenFilter | AnalysisKuromojiStemmerTokenFilter | AnalysisKuromojiReadingFormTokenFilter | AnalysisKuromojiPartOfSpeechTokenFilter | AnalysisIcuCollationTokenFilter | AnalysisIcuFoldingTokenFilter | AnalysisIcuNormalizationTokenFilter | AnalysisIcuTransformTokenFilter | AnalysisPhoneticTokenFilter | AnalysisDictionaryDecompounderTokenFilter export type AnalysisTokenizer = string | AnalysisTokenizerDefinition @@ -5471,6 +5512,7 @@ export interface AnalysisWordDelimiterTokenFilterBase extends AnalysisTokenFilte export interface MappingAggregateMetricDoubleProperty extends MappingPropertyBase { type: 'aggregate_metric_double' default_metric: string + ignore_malformed?: boolean metrics: string[] time_series_metric?: MappingTimeSeriesMetricType } @@ -5681,6 +5723,7 @@ export interface MappingFlattenedProperty extends MappingPropertyBase { null_value?: string similarity?: string split_queries_on_whitespace?: boolean + time_series_dimensions?: string[] type: 'flattened' } @@ -5695,6 +5738,8 @@ export interface MappingFloatRangeProperty extends MappingRangePropertyBase { export type MappingGeoOrientation = 'right' | 'RIGHT' | 'counterclockwise' | 'ccw' | 'left' | 'LEFT' | 'clockwise' | 'cw' +export type MappingGeoPointMetricType = 'gauge' | 'counter' | 'position' + export interface MappingGeoPointProperty extends MappingDocValuesPropertyBase { ignore_malformed?: boolean ignore_z_value?: boolean @@ -5703,6 +5748,7 @@ export interface MappingGeoPointProperty extends MappingDocValuesPropertyBase { on_script_error?: MappingOnScriptError script?: Script | string type: 'geo_point' + time_series_metric?: MappingGeoPointMetricType } export interface MappingGeoShapeProperty extends MappingDocValuesPropertyBase { @@ -5865,7 +5911,7 @@ export interface MappingPointProperty extends MappingDocValuesPropertyBase { type: 'point' } -export type MappingProperty = MappingBinaryProperty | MappingBooleanProperty | MappingDynamicProperty | MappingJoinProperty | MappingKeywordProperty | MappingMatchOnlyTextProperty | MappingPercolatorProperty | MappingRankFeatureProperty | MappingRankFeaturesProperty | MappingSearchAsYouTypeProperty | MappingTextProperty | MappingVersionProperty | MappingWildcardProperty | MappingDateNanosProperty | MappingDateProperty | MappingAggregateMetricDoubleProperty | MappingDenseVectorProperty | MappingFlattenedProperty | MappingNestedProperty | MappingObjectProperty | MappingPassthroughObjectProperty | MappingSemanticTextProperty | MappingSparseVectorProperty | MappingCompletionProperty | MappingConstantKeywordProperty | MappingCountedKeywordProperty | MappingFieldAliasProperty | MappingHistogramProperty | MappingIpProperty | MappingMurmur3HashProperty | MappingTokenCountProperty | MappingGeoPointProperty | MappingGeoShapeProperty | MappingPointProperty | MappingShapeProperty | MappingByteNumberProperty | MappingDoubleNumberProperty | MappingFloatNumberProperty | MappingHalfFloatNumberProperty | MappingIntegerNumberProperty | MappingLongNumberProperty | MappingScaledFloatNumberProperty | MappingShortNumberProperty | MappingUnsignedLongNumberProperty | MappingDateRangeProperty | MappingDoubleRangeProperty | MappingFloatRangeProperty | MappingIntegerRangeProperty | MappingIpRangeProperty | MappingLongRangeProperty | MappingIcuCollationProperty +export type MappingProperty = MappingBinaryProperty | MappingBooleanProperty | MappingDynamicProperty | MappingJoinProperty | MappingKeywordProperty | MappingMatchOnlyTextProperty | MappingPercolatorProperty | MappingRankFeatureProperty | MappingRankFeaturesProperty | MappingSearchAsYouTypeProperty | MappingTextProperty | MappingVersionProperty | MappingWildcardProperty | MappingDateNanosProperty | MappingDateProperty | MappingAggregateMetricDoubleProperty | MappingDenseVectorProperty | MappingFlattenedProperty | MappingNestedProperty | MappingObjectProperty | MappingPassthroughObjectProperty | MappingRankVectorProperty | MappingSemanticTextProperty | MappingSparseVectorProperty | MappingCompletionProperty | MappingConstantKeywordProperty | MappingCountedKeywordProperty | MappingFieldAliasProperty | MappingHistogramProperty | MappingIpProperty | MappingMurmur3HashProperty | MappingTokenCountProperty | MappingGeoPointProperty | MappingGeoShapeProperty | MappingPointProperty | MappingShapeProperty | MappingByteNumberProperty | MappingDoubleNumberProperty | MappingFloatNumberProperty | MappingHalfFloatNumberProperty | MappingIntegerNumberProperty | MappingLongNumberProperty | MappingScaledFloatNumberProperty | MappingShortNumberProperty | MappingUnsignedLongNumberProperty | MappingDateRangeProperty | MappingDoubleRangeProperty | MappingFloatRangeProperty | MappingIntegerRangeProperty | MappingIpRangeProperty | MappingLongRangeProperty | MappingIcuCollationProperty export interface MappingPropertyBase { meta?: Record @@ -5892,6 +5938,14 @@ export interface MappingRankFeaturesProperty extends MappingPropertyBase { type: 'rank_features' } +export type MappingRankVectorElementType = 'byte' | 'float' | 'bit' + +export interface MappingRankVectorProperty extends MappingPropertyBase { + type: 'rank_vectors' + element_type?: MappingRankVectorElementType + dims?: integer +} + export interface MappingRoutingField { required: boolean } @@ -5935,11 +5989,16 @@ export interface MappingSearchAsYouTypeProperty extends MappingCorePropertyBase type: 'search_as_you_type' } +export interface MappingSemanticTextIndexOptions { + dense_vector?: MappingDenseVectorIndexOptions +} + export interface MappingSemanticTextProperty { type: 'semantic_text' meta?: Record inference_id?: Id search_inference_id?: Id + index_options?: MappingSemanticTextIndexOptions chunking_settings?: MappingChunkingSettings } @@ -5971,8 +6030,15 @@ export interface MappingSourceField { export type MappingSourceFieldMode = 'disabled' | 'stored' | 'synthetic' +export interface MappingSparseVectorIndexOptions { + prune?: boolean + pruning_config?: TokenPruningConfig +} + export interface MappingSparseVectorProperty extends MappingPropertyBase { + store?: boolean type: 'sparse_vector' + index_options?: MappingSparseVectorIndexOptions } export type MappingSubobjects = boolean | 'true' | 'false' | 'auto' @@ -6802,7 +6868,7 @@ export interface QueryDslSparseVectorQuery extends QueryDslQueryBase { inference_id?: Id query?: string prune?: boolean - pruning_config?: QueryDslTokenPruningConfig + pruning_config?: TokenPruningConfig } export interface QueryDslTermQuery extends QueryDslQueryBase { @@ -6837,17 +6903,11 @@ export interface QueryDslTermsSetQuery extends QueryDslQueryBase { export interface QueryDslTextExpansionQuery extends QueryDslQueryBase { model_id: string model_text: string - pruning_config?: QueryDslTokenPruningConfig + pruning_config?: TokenPruningConfig } export type QueryDslTextQueryType = 'best_fields' | 'most_fields' | 'cross_fields' | 'phrase' | 'phrase_prefix' | 'bool_prefix' -export interface QueryDslTokenPruningConfig { - tokens_freq_ratio_threshold?: integer - tokens_weight_threshold?: float - only_score_pruned_tokens?: boolean -} - export interface QueryDslTypeQuery extends QueryDslQueryBase { value: string } @@ -6867,7 +6927,7 @@ export interface QueryDslUntypedRangeQuery extends QueryDslRangeQueryBase { export interface QueryDslWeightedTokensQuery extends QueryDslQueryBase { tokens: Record - pruning_config?: QueryDslTokenPruningConfig + pruning_config?: TokenPruningConfig } export interface QueryDslWildcardQuery extends QueryDslQueryBase { @@ -7098,9 +7158,29 @@ export type CatCatNodeColumn = 'build' | 'b' | 'completion.size' | 'cs' | 'compl export type CatCatNodeColumns = CatCatNodeColumn | CatCatNodeColumn[] +export type CatCatRecoveryColumn = 'index' | 'i' | 'idx' | 'shard' | 's' | 'sh' | 'time' | 't' | 'ti' | 'primaryOrReplica' | 'type' | 'stage' | 'st' | 'source_host' | 'shost' | 'source_node' | 'snode' | 'target_host' | 'thost' | 'target_node' | 'tnode' | 'repository' | 'tnode' | 'snapshot' | 'snap' | 'files' | 'f' | 'files_recovered' | 'fr' | 'files_percent' | 'fp' | 'files_total' | 'tf' | 'bytes' | 'b' | 'bytes_recovered' | 'br' | 'bytes_percent' | 'bp' | 'bytes_total' | 'tb' | 'translog_ops' | 'to' | 'translog_ops_recovered' | 'tor' | 'translog_ops_percent' | 'top' | 'start_time' | 'start' | 'start_time_millis' | 'start_millis' | 'stop_time' | 'stop' | 'stop_time_millis' | 'stop_millis' | string + +export type CatCatRecoveryColumns = CatCatRecoveryColumn | CatCatRecoveryColumn[] + export interface CatCatRequestBase extends RequestBase, SpecUtilsCommonCatQueryParameters { } +export type CatCatSegmentsColumn = 'index' | 'i' | 'idx' | 'shard' | 's' | 'sh' | 'prirep' | 'p' | 'pr' | 'primaryOrReplica' | 'ip' | 'segment' | 'generation' | 'docs.count' | 'docs.deleted' | 'size' | 'size.memory' | 'committed' | 'searchable' | 'version' | 'compound' | 'id' | string + +export type CatCatSegmentsColumns = CatCatSegmentsColumn | CatCatSegmentsColumn[] + +export type CatCatShardColumn = 'completion.size' | 'cs' | 'completionSize' | 'dataset.size' | 'dense_vector.value_count' | 'dvc' | 'denseVectorCount' | 'docs' | 'd' | 'dc' | 'fielddata.evictions' | 'fe' | 'fielddataEvictions' | 'fielddata.memory_size' | 'fm' | 'fielddataMemory' | 'flush.total' | 'ft' | 'flushTotal' | 'flush.total_time' | 'ftt' | 'flushTotalTime' | 'get.current' | 'gc' | 'getCurrent' | 'get.exists_time' | 'geti' | 'getExistsTime' | 'get.exists_total' | 'geto' | 'getExistsTotal' | 'get.missing_time' | 'gmti' | 'getMissingTime' | 'get.missing_total' | 'gmto' | 'getMissingTotal' | 'get.time' | 'gti' | 'getTime' | 'get.total' | 'gto' | 'getTotal' | 'id' | 'index' | 'i' | 'idx' | 'indexing.delete_current' | 'idc' | 'indexingDeleteCurrent' | 'indexing.delete_time' | 'idti' | 'indexingDeleteTime' | 'indexing.delete_total' | 'idto' | 'indexingDeleteTotal' | 'indexing.index_current' | 'iic' | 'indexingIndexCurrent' | 'indexing.index_failed_due_to_version_conflict' | 'iifvc' | 'indexingIndexFailedDueToVersionConflict' | 'indexing.index_failed' | 'iif' | 'indexingIndexFailed' | 'indexing.index_time' | 'iiti' | 'indexingIndexTime' | 'indexing.index_total' | 'iito' | 'indexingIndexTotal' | 'ip' | 'merges.current' | 'mc' | 'mergesCurrent' | 'merges.current_docs' | 'mcd' | 'mergesCurrentDocs' | 'merges.current_size' | 'mcs' | 'mergesCurrentSize' | 'merges.total' | 'mt' | 'mergesTotal' | 'merges.total_docs' | 'mtd' | 'mergesTotalDocs' | 'merges.total_size' | 'mts' | 'mergesTotalSize' | 'merges.total_time' | 'mtt' | 'mergesTotalTime' | 'node' | 'n' | 'prirep' | 'p' | 'pr' | 'primaryOrReplica' | 'query_cache.evictions' | 'qce' | 'queryCacheEvictions' | 'query_cache.memory_size' | 'qcm' | 'queryCacheMemory' | 'recoverysource.type' | 'rs' | 'refresh.time' | 'rti' | 'refreshTime' | 'refresh.total' | 'rto' | 'refreshTotal' | 'search.fetch_current' | 'sfc' | 'searchFetchCurrent' | 'search.fetch_time' | 'sfti' | 'searchFetchTime' | 'search.fetch_total' | 'sfto' | 'searchFetchTotal' | 'search.open_contexts' | 'so' | 'searchOpenContexts' | 'search.query_current' | 'sqc' | 'searchQueryCurrent' | 'search.query_time' | 'sqti' | 'searchQueryTime' | 'search.query_total' | 'sqto' | 'searchQueryTotal' | 'search.scroll_current' | 'scc' | 'searchScrollCurrent' | 'search.scroll_time' | 'scti' | 'searchScrollTime' | 'search.scroll_total' | 'scto' | 'searchScrollTotal' | 'segments.count' | 'sc' | 'segmentsCount' | 'segments.fixed_bitset_memory' | 'sfbm' | 'fixedBitsetMemory' | 'segments.index_writer_memory' | 'siwm' | 'segmentsIndexWriterMemory' | 'segments.memory' | 'sm' | 'segmentsMemory' | 'segments.version_map_memory' | 'svmm' | 'segmentsVersionMapMemory' | 'seq_no.global_checkpoint' | 'sqg' | 'globalCheckpoint' | 'seq_no.local_checkpoint' | 'sql' | 'localCheckpoint' | 'seq_no.max' | 'sqm' | 'maxSeqNo' | 'shard' | 's' | 'sh' | 'dsparse_vector.value_count' | 'svc' | 'sparseVectorCount' | 'state' | 'st' | 'store' | 'sto' | 'suggest.current' | 'suc' | 'suggestCurrent' | 'suggest.time' | 'suti' | 'suggestTime' | 'suggest.total' | 'suto' | 'suggestTotal' | 'sync_id' | 'unassigned.at' | 'ua' | 'unassigned.details' | 'ud' | 'unassigned.for' | 'uf' | 'unassigned.reason' | 'ur' | string + +export type CatCatShardColumns = CatCatShardColumn | CatCatShardColumn[] + +export type CatCatSnapshotsColumn = 'id' | 'snapshot' | 'repository' | 're' | 'repo' | 'status' | 's' | 'start_epoch' | 'ste' | 'startEpoch' | 'start_time' | 'sti' | 'startTime' | 'end_epoch' | 'ete' | 'endEpoch' | 'end_time' | 'eti' | 'endTime' | 'duration' | 'dur' | 'indices' | 'i' | 'successful_shards' | 'ss' | 'failed_shards' | 'fs' | 'total_shards' | 'ts' | 'reason' | 'r' | string + +export type CatCatSnapshotsColumns = CatCatSnapshotsColumn | CatCatNodeColumn[] + +export type CatCatThreadPoolColumn = 'active' | 'a' | 'completed' | 'c' | 'core' | 'cr' | 'ephemeral_id' | 'eid' | 'host' | 'h' | 'ip' | 'i' | 'keep_alive' | 'k' | 'largest' | 'l' | 'max' | 'mx' | 'name' | 'node_id' | 'id' | 'node_name' | 'pid' | 'p' | 'pool_size' | 'psz' | 'port' | 'po' | 'queue' | 'q' | 'queue_size' | 'qs' | 'rejected' | 'r' | 'size' | 'sz' | 'type' | 't' | string + +export type CatCatThreadPoolColumns = CatCatThreadPoolColumn | CatCatThreadPoolColumn[] + export type CatCatTrainedModelsColumn = 'create_time' | 'ct' | 'created_by' | 'c' | 'createdBy' | 'data_frame_analytics_id' | 'df' | 'dataFrameAnalytics' | 'dfid' | 'description' | 'd' | 'heap_size' | 'hs' | 'modelHeapSize' | 'id' | 'ingest.count' | 'ic' | 'ingestCount' | 'ingest.current' | 'icurr' | 'ingestCurrent' | 'ingest.failed' | 'if' | 'ingestFailed' | 'ingest.pipelines' | 'ip' | 'ingestPipelines' | 'ingest.time' | 'it' | 'ingestTime' | 'license' | 'l' | 'operations' | 'o' | 'modelOperations' | 'version' | 'v' export type CatCatTrainedModelsColumns = CatCatTrainedModelsColumn | CatCatTrainedModelsColumn[] @@ -8404,7 +8484,7 @@ export interface CatRecoveryRequest extends CatCatRequestBase { active_only?: boolean bytes?: Bytes detailed?: boolean - h?: Names + h?: CatCatRecoveryColumns s?: Names time?: TimeUnit } @@ -8430,7 +8510,7 @@ export type CatRepositoriesResponse = CatRepositoriesRepositoriesRecord[] export interface CatSegmentsRequest extends CatCatRequestBase { index?: Indices bytes?: Bytes - h?: Names + h?: CatCatSegmentsColumns s?: Names local?: boolean master_timeout?: Duration @@ -8483,7 +8563,7 @@ export interface CatSegmentsSegmentsRecord { export interface CatShardsRequest extends CatCatRequestBase { index?: Indices bytes?: Bytes - h?: Names + h?: CatCatShardColumns s?: Names master_timeout?: Duration time?: TimeUnit @@ -8709,7 +8789,7 @@ export interface CatShardsShardsRecord { export interface CatSnapshotsRequest extends CatCatRequestBase { repository?: Names ignore_unavailable?: boolean - h?: Names + h?: CatCatSnapshotsColumns s?: Names master_timeout?: Duration time?: TimeUnit @@ -8826,7 +8906,7 @@ export interface CatTemplatesTemplatesRecord { export interface CatThreadPoolRequest extends CatCatRequestBase { thread_pool_patterns?: Names - h?: Names + h?: CatCatThreadPoolColumns s?: Names time?: TimeUnit local?: boolean @@ -9527,6 +9607,7 @@ export type ClusterPostVotingConfigExclusionsResponse = boolean export interface ClusterPutComponentTemplateRequest extends RequestBase { name: Name create?: boolean + cause?: string master_timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { @@ -10784,6 +10865,60 @@ export type EqlSearchResponse = EqlEqlSearchResponseBase +} + +export interface EsqlColumnInfo { + name: string + type: string +} + +export interface EsqlEsqlClusterDetails { + status: EsqlEsqlClusterStatus + indices: string + took?: DurationValue + _shards?: EsqlEsqlShardInfo +} + +export type EsqlEsqlClusterStatus = 'running' | 'successful' | 'partial' | 'skipped' | 'failed' + +export interface EsqlEsqlResult { + took?: DurationValue + is_partial?: boolean + all_columns?: EsqlColumnInfo[] + columns: EsqlColumnInfo[] + values: FieldValue[][] + _clusters?: EsqlClusterInfo + profile?: any +} + +export interface EsqlEsqlShardFailure { + shard: Id + index: IndexName + node?: NodeId + reason: ErrorCause +} + +export interface EsqlEsqlShardInfo { + total: integer + successful?: integer + skipped?: integer + failed?: integer + failures?: EsqlEsqlShardFailure[] +} + export interface EsqlTableValuesContainer { integer?: EsqlTableValuesIntegerValue[] keyword?: EsqlTableValuesKeywordValue[] @@ -10820,7 +10955,7 @@ export interface EsqlAsyncQueryRequest extends RequestBase { } } -export type EsqlAsyncQueryResponse = EsqlResult +export type EsqlAsyncQueryResponse = EsqlEsqlResult export interface EsqlAsyncQueryDeleteRequest extends RequestBase { id: Id @@ -10831,18 +10966,19 @@ export type EsqlAsyncQueryDeleteResponse = AcknowledgedResponseBase export interface EsqlAsyncQueryGetRequest extends RequestBase { id: Id drop_null_columns?: boolean + format?: EsqlQueryEsqlFormat keep_alive?: Duration wait_for_completion_timeout?: Duration } -export type EsqlAsyncQueryGetResponse = EsqlResult +export type EsqlAsyncQueryGetResponse = EsqlAsyncEsqlResult export interface EsqlAsyncQueryStopRequest extends RequestBase { id: Id drop_null_columns?: boolean } -export type EsqlAsyncQueryStopResponse = EsqlResult +export type EsqlAsyncQueryStopResponse = EsqlEsqlResult export type EsqlQueryEsqlFormat = 'csv' | 'json' | 'tsv' | 'txt' | 'yaml' | 'cbor' | 'smile' | 'arrow' @@ -10864,7 +11000,7 @@ export interface EsqlQueryRequest extends RequestBase { } } -export type EsqlQueryResponse = EsqlResult +export type EsqlQueryResponse = EsqlEsqlResult export interface FeaturesFeature { name: string @@ -11806,6 +11942,7 @@ export type IndicesSourceMode = 'disabled' | 'stored' | 'synthetic' export interface IndicesStorage { type: IndicesStorageType allow_mmap?: boolean + stats_refresh_interval?: Duration } export type IndicesStorageType = 'fs' | 'niofs' | 'mmapfs' | 'hybridfs' | string @@ -12291,10 +12428,6 @@ export interface IndicesGetRequest extends RequestBase { export type IndicesGetResponse = Record -export interface IndicesGetAliasIndexAliases { - aliases: Record -} - export interface IndicesGetAliasRequest extends RequestBase { name?: Names index?: Indices @@ -12306,6 +12439,17 @@ export interface IndicesGetAliasRequest extends RequestBase { export type IndicesGetAliasResponse = Record +export interface IndicesGetAliasIndexAliases { + aliases: Record +} + +export interface IndicesGetAliasNotFoundAliasesKeys { + error: string + status: number +} +export type IndicesGetAliasNotFoundAliases = IndicesGetAliasNotFoundAliasesKeys +& { [property: string]: IndicesGetAliasIndexAliases | string | number } + export interface IndicesGetDataLifecycleDataStreamWithLifecycle { name: DataStreamName lifecycle?: IndicesDataStreamLifecycleWithRollover @@ -13452,6 +13596,14 @@ export interface InferenceContentObject { type: string } +export interface InferenceDeepSeekServiceSettings { + api_key: string + model_id: string + url?: string +} + +export type InferenceDeepSeekServiceType = 'deepseek' + export interface InferenceDeleteInferenceEndpointResult extends AcknowledgedResponseBase { pipelines: string[] } @@ -13553,11 +13705,86 @@ export interface InferenceInferenceEndpointInfoAlibabaCloudAI extends InferenceI task_type: InferenceTaskTypeAlibabaCloudAI } +export interface InferenceInferenceEndpointInfoAmazonBedrock extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeAmazonBedrock +} + +export interface InferenceInferenceEndpointInfoAnthropic extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeAnthropic +} + +export interface InferenceInferenceEndpointInfoAzureAIStudio extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeAzureAIStudio +} + +export interface InferenceInferenceEndpointInfoAzureOpenAI extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeAzureOpenAI +} + +export interface InferenceInferenceEndpointInfoCohere extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeCohere +} + +export interface InferenceInferenceEndpointInfoDeepSeek extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeDeepSeek +} + +export interface InferenceInferenceEndpointInfoELSER extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeELSER +} + +export interface InferenceInferenceEndpointInfoElasticsearch extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeElasticsearch +} + +export interface InferenceInferenceEndpointInfoGoogleAIStudio extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeGoogleAIStudio +} + +export interface InferenceInferenceEndpointInfoGoogleVertexAI extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeGoogleVertexAI +} + +export interface InferenceInferenceEndpointInfoHuggingFace extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeHuggingFace +} + export interface InferenceInferenceEndpointInfoJinaAi extends InferenceInferenceEndpoint { inference_id: string task_type: InferenceTaskTypeJinaAi } +export interface InferenceInferenceEndpointInfoMistral extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeMistral +} + +export interface InferenceInferenceEndpointInfoOpenAI extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeOpenAI +} + +export interface InferenceInferenceEndpointInfoVoyageAI extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeVoyageAI +} + +export interface InferenceInferenceEndpointInfoWatsonx extends InferenceInferenceEndpoint { + inference_id: string + task_type: InferenceTaskTypeWatsonx +} + export interface InferenceInferenceResult { text_embedding_bytes?: InferenceTextEmbeddingByteResult[] text_embedding_bits?: InferenceTextEmbeddingByteResult[] @@ -13668,8 +13895,38 @@ export type InferenceTaskType = 'sparse_embedding' | 'text_embedding' | 'rerank' export type InferenceTaskTypeAlibabaCloudAI = 'text_embedding' | 'rerank' | 'completion' | 'sparse_embedding' +export type InferenceTaskTypeAmazonBedrock = 'text_embedding' | 'completion' + +export type InferenceTaskTypeAnthropic = 'completion' + +export type InferenceTaskTypeAzureAIStudio = 'text_embedding' | 'completion' + +export type InferenceTaskTypeAzureOpenAI = 'text_embedding' | 'completion' + +export type InferenceTaskTypeCohere = 'text_embedding' | 'rerank' | 'completion' + +export type InferenceTaskTypeDeepSeek = 'completion' | 'chat_completion' + +export type InferenceTaskTypeELSER = 'sparse_embedding' + +export type InferenceTaskTypeElasticsearch = 'sparse_embedding' | 'text_embedding' | 'rerank' + +export type InferenceTaskTypeGoogleAIStudio = 'text_embedding' | 'completion' + +export type InferenceTaskTypeGoogleVertexAI = 'text_embedding' | 'rerank' + +export type InferenceTaskTypeHuggingFace = 'text_embedding' + export type InferenceTaskTypeJinaAi = 'text_embedding' | 'rerank' +export type InferenceTaskTypeMistral = 'text_embedding' + +export type InferenceTaskTypeOpenAI = 'text_embedding' | 'chat_completion' | 'completion' + +export type InferenceTaskTypeVoyageAI = 'text_embedding' | 'rerank' + +export type InferenceTaskTypeWatsonx = 'text_embedding' + export interface InferenceTextEmbeddingByteResult { embedding: InferenceDenseByteVector } @@ -13814,7 +14071,7 @@ export interface InferencePutAmazonbedrockRequest extends RequestBase { } } -export type InferencePutAmazonbedrockResponse = InferenceInferenceEndpointInfo +export type InferencePutAmazonbedrockResponse = InferenceInferenceEndpointInfoAmazonBedrock export interface InferencePutAnthropicRequest extends RequestBase { task_type: InferenceAnthropicTaskType @@ -13828,7 +14085,7 @@ export interface InferencePutAnthropicRequest extends RequestBase { } } -export type InferencePutAnthropicResponse = InferenceInferenceEndpointInfo +export type InferencePutAnthropicResponse = InferenceInferenceEndpointInfoAnthropic export interface InferencePutAzureaistudioRequest extends RequestBase { task_type: InferenceAzureAiStudioTaskType @@ -13842,7 +14099,7 @@ export interface InferencePutAzureaistudioRequest extends RequestBase { } } -export type InferencePutAzureaistudioResponse = InferenceInferenceEndpointInfo +export type InferencePutAzureaistudioResponse = InferenceInferenceEndpointInfoAzureAIStudio export interface InferencePutAzureopenaiRequest extends RequestBase { task_type: InferenceAzureOpenAITaskType @@ -13856,7 +14113,7 @@ export interface InferencePutAzureopenaiRequest extends RequestBase { } } -export type InferencePutAzureopenaiResponse = InferenceInferenceEndpointInfo +export type InferencePutAzureopenaiResponse = InferenceInferenceEndpointInfoAzureOpenAI export interface InferencePutCohereRequest extends RequestBase { task_type: InferenceCohereTaskType @@ -13870,7 +14127,20 @@ export interface InferencePutCohereRequest extends RequestBase { } } -export type InferencePutCohereResponse = InferenceInferenceEndpointInfo +export type InferencePutCohereResponse = InferenceInferenceEndpointInfoCohere + +export interface InferencePutDeepseekRequest extends RequestBase { + task_type: InferenceTaskTypeDeepSeek + deepseek_inference_id: Id + /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ + body?: { + chunking_settings?: InferenceInferenceChunkingSettings + service: InferenceDeepSeekServiceType + service_settings: InferenceDeepSeekServiceSettings + } +} + +export type InferencePutDeepseekResponse = InferenceInferenceEndpointInfoDeepSeek export interface InferencePutElasticsearchRequest extends RequestBase { task_type: InferenceElasticsearchTaskType @@ -13884,7 +14154,7 @@ export interface InferencePutElasticsearchRequest extends RequestBase { } } -export type InferencePutElasticsearchResponse = InferenceInferenceEndpointInfo +export type InferencePutElasticsearchResponse = InferenceInferenceEndpointInfoElasticsearch export interface InferencePutElserRequest extends RequestBase { task_type: InferenceElserTaskType @@ -13897,7 +14167,7 @@ export interface InferencePutElserRequest extends RequestBase { } } -export type InferencePutElserResponse = InferenceInferenceEndpointInfo +export type InferencePutElserResponse = InferenceInferenceEndpointInfoELSER export interface InferencePutGoogleaistudioRequest extends RequestBase { task_type: InferenceGoogleAiStudioTaskType @@ -13910,7 +14180,7 @@ export interface InferencePutGoogleaistudioRequest extends RequestBase { } } -export type InferencePutGoogleaistudioResponse = InferenceInferenceEndpointInfo +export type InferencePutGoogleaistudioResponse = InferenceInferenceEndpointInfoGoogleAIStudio export interface InferencePutGooglevertexaiRequest extends RequestBase { task_type: InferenceGoogleVertexAITaskType @@ -13924,7 +14194,7 @@ export interface InferencePutGooglevertexaiRequest extends RequestBase { } } -export type InferencePutGooglevertexaiResponse = InferenceInferenceEndpointInfo +export type InferencePutGooglevertexaiResponse = InferenceInferenceEndpointInfoGoogleVertexAI export interface InferencePutHuggingFaceRequest extends RequestBase { task_type: InferenceHuggingFaceTaskType @@ -13938,7 +14208,7 @@ export interface InferencePutHuggingFaceRequest extends RequestBase { } } -export type InferencePutHuggingFaceResponse = InferenceInferenceEndpointInfo +export type InferencePutHuggingFaceResponse = InferenceInferenceEndpointInfoHuggingFace export interface InferencePutJinaaiRequest extends RequestBase { task_type: InferenceJinaAITaskType @@ -13965,7 +14235,7 @@ export interface InferencePutMistralRequest extends RequestBase { } } -export type InferencePutMistralResponse = InferenceInferenceEndpointInfo +export type InferencePutMistralResponse = InferenceInferenceEndpointInfoMistral export interface InferencePutOpenaiRequest extends RequestBase { task_type: InferenceOpenAITaskType @@ -13979,7 +14249,7 @@ export interface InferencePutOpenaiRequest extends RequestBase { } } -export type InferencePutOpenaiResponse = InferenceInferenceEndpointInfo +export type InferencePutOpenaiResponse = InferenceInferenceEndpointInfoOpenAI export interface InferencePutVoyageaiRequest extends RequestBase { task_type: InferenceVoyageAITaskType @@ -13993,7 +14263,7 @@ export interface InferencePutVoyageaiRequest extends RequestBase { } } -export type InferencePutVoyageaiResponse = InferenceInferenceEndpointInfo +export type InferencePutVoyageaiResponse = InferenceInferenceEndpointInfoVoyageAI export interface InferencePutWatsonxRequest extends RequestBase { task_type: InferenceWatsonxTaskType @@ -14005,7 +14275,7 @@ export interface InferencePutWatsonxRequest extends RequestBase { } } -export type InferencePutWatsonxResponse = InferenceInferenceEndpointInfo +export type InferencePutWatsonxResponse = InferenceInferenceEndpointInfoWatsonx export interface InferenceRerankRequest extends RequestBase { inference_id: Id @@ -14666,7 +14936,6 @@ export interface IngestGetIpLocationDatabaseDatabaseConfigurationMetadata { export interface IngestGetIpLocationDatabaseRequest extends RequestBase { id?: Ids - master_timeout?: Duration } export interface IngestGetIpLocationDatabaseResponse { @@ -14843,7 +15112,7 @@ export interface LicensePostStartBasicResponse { export interface LicensePostStartTrialRequest extends RequestBase { acknowledge?: boolean - type_query_string?: string + type?: string master_timeout?: Duration } @@ -18184,12 +18453,13 @@ export interface NodesInfoNodeInfo { build_flavor: string build_hash: string build_type: string + component_versions: Record host: Host http?: NodesInfoNodeInfoHttp + index_version: VersionNumber ip: Ip jvm?: NodesInfoNodeJvmInfo name: Name - network?: NodesInfoNodeInfoNetwork os?: NodesInfoNodeOperatingSystemInfo plugins?: PluginStats[] process?: NodesInfoNodeProcessInfo @@ -18200,10 +18470,12 @@ export interface NodesInfoNodeInfo { total_indexing_buffer_in_bytes?: ByteSize transport?: NodesInfoNodeInfoTransport transport_address: TransportAddress + transport_version: VersionNumber version: VersionString modules?: PluginStats[] ingest?: NodesInfoNodeInfoIngest aggregations?: Record + remote_cluster_server?: NodesInfoRemoveClusterServer } export interface NodesInfoNodeInfoAction { @@ -18223,7 +18495,7 @@ export interface NodesInfoNodeInfoClient { } export interface NodesInfoNodeInfoDiscoverKeys { - seed_hosts?: string[] + seed_hosts?: string[] | string type?: string seed_providers?: string[] } @@ -18271,17 +18543,6 @@ export interface NodesInfoNodeInfoMemory { total_in_bytes: long } -export interface NodesInfoNodeInfoNetwork { - primary_interface: NodesInfoNodeInfoNetworkInterface - refresh_interval: integer -} - -export interface NodesInfoNodeInfoNetworkInterface { - address: string - mac_address: string - name: Name -} - export interface NodesInfoNodeInfoOSCPU { cache_size: string cache_size_in_bytes: integer @@ -18343,7 +18604,7 @@ export interface NodesInfoNodeInfoSettingsCluster { name: Name routing?: IndicesIndexRouting election: NodesInfoNodeInfoSettingsClusterElection - initial_master_nodes?: string[] + initial_master_nodes?: string[] | string deprecation_indexing?: NodesInfoDeprecationIndexing } @@ -18413,6 +18674,7 @@ export interface NodesInfoNodeInfoSettingsTransport { type: NodesInfoNodeInfoSettingsTransportType | string 'type.default'?: string features?: NodesInfoNodeInfoSettingsTransportFeatures + ignore_deserialization_errors?: SpecUtilsStringified } export interface NodesInfoNodeInfoSettingsTransportFeatures { @@ -18490,7 +18752,6 @@ export interface NodesInfoNodeJvmInfo { vm_vendor: string vm_version: VersionString using_bundled_jdk: boolean - bundled_jdk: boolean using_compressed_ordinary_object_pointers?: boolean | string input_arguments: string[] } @@ -18523,6 +18784,11 @@ export interface NodesInfoNodeThreadPoolInfo { type: string } +export interface NodesInfoRemoveClusterServer { + bound_address: TransportAddress[] + publish_address: TransportAddress +} + export interface NodesInfoRequest extends RequestBase { node_id?: NodeIds metric?: Metrics