From 17ce636b377d43dc5ad0eec5a3f7676100aec8d2 Mon Sep 17 00:00:00 2001 From: David Kyle Date: Mon, 14 Jul 2025 09:43:16 +0100 Subject: [PATCH 1/3] Add timeout params for PUT inference and POST inference/_stream --- output/openapi/elasticsearch-openapi.json | 32 +++++++++++++++++ .../elasticsearch-serverless-openapi.json | 16 +++++++++ output/schema/schema-serverless.json | 18 ++++++++-- output/schema/schema.json | 36 ++++++++++++++++--- output/typescript/types.ts | 2 ++ specification/inference/put/PutRequest.ts | 8 +++++ .../StreamInferenceRequest.ts | 8 +++++ 7 files changed, 114 insertions(+), 6 deletions(-) diff --git a/output/openapi/elasticsearch-openapi.json b/output/openapi/elasticsearch-openapi.json index 6ec158e75c..f3b2c4365a 100644 --- a/output/openapi/elasticsearch-openapi.json +++ b/output/openapi/elasticsearch-openapi.json @@ -16311,6 +16311,9 @@ "parameters": [ { "$ref": "#/components/parameters/inference.put#inference_id" + }, + { + "$ref": "#/components/parameters/inference.put#timeout" } ], "requestBody": { @@ -16408,6 +16411,9 @@ }, { "$ref": "#/components/parameters/inference.put#inference_id" + }, + { + "$ref": "#/components/parameters/inference.put#timeout" } ], "requestBody": { @@ -16502,6 +16508,9 @@ "parameters": [ { "$ref": "#/components/parameters/inference.stream_inference#inference_id" + }, + { + "$ref": "#/components/parameters/inference.stream_inference#timeout" } ], "requestBody": { @@ -16529,6 +16538,9 @@ }, { "$ref": "#/components/parameters/inference.stream_inference#inference_id" + }, + { + "$ref": "#/components/parameters/inference.stream_inference#timeout" } ], "requestBody": { @@ -103479,6 +103491,16 @@ }, "style": "simple" }, + "inference.put#timeout": { + "in": "query", + "name": "timeout", + "description": "Specifies the amount of time to wait for the inference endpoint to be created.", + "deprecated": false, + "schema": { + "$ref": "#/components/schemas/_types:Duration" + }, + "style": "form" + }, "inference.stream_inference#inference_id": { "in": "path", "name": "inference_id", @@ -103501,6 +103523,16 @@ }, "style": "simple" }, + "inference.stream_inference#timeout": { + "in": "query", + "name": "timeout", + "description": "The amount of time to wait for the inference request to complete.", + "deprecated": false, + "schema": { + "$ref": "#/components/schemas/_types:Duration" + }, + "style": "form" + }, "inference.unified_inference#task_type": { "in": "path", "name": "task_type", diff --git a/output/openapi/elasticsearch-serverless-openapi.json b/output/openapi/elasticsearch-serverless-openapi.json index 5761bc6349..1db7a559c9 100644 --- a/output/openapi/elasticsearch-serverless-openapi.json +++ b/output/openapi/elasticsearch-serverless-openapi.json @@ -8796,6 +8796,9 @@ "parameters": [ { "$ref": "#/components/parameters/inference.put#inference_id" + }, + { + "$ref": "#/components/parameters/inference.put#timeout" } ], "requestBody": { @@ -8893,6 +8896,9 @@ }, { "$ref": "#/components/parameters/inference.put#inference_id" + }, + { + "$ref": "#/components/parameters/inference.put#timeout" } ], "requestBody": { @@ -61097,6 +61103,16 @@ }, "style": "simple" }, + "inference.put#timeout": { + "in": "query", + "name": "timeout", + "description": "Specifies the amount of time to wait for the inference endpoint to be created.", + "deprecated": false, + "schema": { + "$ref": "#/components/schemas/_types:Duration" + }, + "style": "form" + }, "inference.unified_inference#task_type": { "in": "path", "name": "task_type", diff --git a/output/schema/schema-serverless.json b/output/schema/schema-serverless.json index dd4fafcc1f..b475345d16 100644 --- a/output/schema/schema-serverless.json +++ b/output/schema/schema-serverless.json @@ -25595,8 +25595,22 @@ } } ], - "query": [], - "specLocation": "inference/put/PutRequest.ts#L25-L65" + "query": [ + { + "description": "Specifies the amount of time to wait for the inference endpoint to be created.", + "name": "timeout", + "required": false, + "serverDefault": "30s", + "type": { + "kind": "instance_of", + "type": { + "name": "Duration", + "namespace": "_types" + } + } + } + ], + "specLocation": "inference/put/PutRequest.ts#L26-L73" }, { "body": { diff --git a/output/schema/schema.json b/output/schema/schema.json index ad9f981b8c..9cd7617d7d 100644 --- a/output/schema/schema.json +++ b/output/schema/schema.json @@ -146800,8 +146800,22 @@ } } ], - "query": [], - "specLocation": "inference/put/PutRequest.ts#L25-L65" + "query": [ + { + "description": "Specifies the amount of time to wait for the inference endpoint to be created.", + "name": "timeout", + "required": false, + "serverDefault": "30s", + "type": { + "kind": "instance_of", + "type": { + "name": "Duration", + "namespace": "_types" + } + } + } + ], + "specLocation": "inference/put/PutRequest.ts#L26-L73" }, { "kind": "response", @@ -146895,8 +146909,22 @@ } } ], - "query": [], - "specLocation": "inference/stream_inference/StreamInferenceRequest.ts#L24-L67" + "query": [ + { + "description": "The amount of time to wait for the inference request to complete.", + "name": "timeout", + "required": false, + "serverDefault": "30s", + "type": { + "kind": "instance_of", + "type": { + "name": "Duration", + "namespace": "_types" + } + } + } + ], + "specLocation": "inference/stream_inference/StreamInferenceRequest.ts#L25-L75" }, { "kind": "response", diff --git a/output/typescript/types.ts b/output/typescript/types.ts index 5db4dbcfad..0c479b4a2b 100644 --- a/output/typescript/types.ts +++ b/output/typescript/types.ts @@ -13081,6 +13081,7 @@ export type InferenceInferenceResponse = InferenceInferenceResult export interface InferencePutRequest extends RequestBase { task_type?: InferenceTaskType inference_id: Id + timeout?: Duration body?: InferenceInferenceEndpoint } @@ -13089,6 +13090,7 @@ export type InferencePutResponse = InferenceInferenceEndpointInfo export interface InferenceStreamInferenceRequest extends RequestBase { inference_id: Id task_type?: InferenceTaskType + timeout?: Duration body?: { input: string | string[] } diff --git a/specification/inference/put/PutRequest.ts b/specification/inference/put/PutRequest.ts index 0706189d1b..5559432f7e 100644 --- a/specification/inference/put/PutRequest.ts +++ b/specification/inference/put/PutRequest.ts @@ -21,6 +21,7 @@ import { InferenceEndpoint } from '@inference/_types/Services' import { TaskType } from '@inference/_types/TaskType' import { RequestBase } from '@_types/Base' import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an inference endpoint. @@ -60,6 +61,13 @@ export interface Request extends RequestBase { */ inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } /** @codegen_name inference_config */ body: InferenceEndpoint } diff --git a/specification/inference/stream_inference/StreamInferenceRequest.ts b/specification/inference/stream_inference/StreamInferenceRequest.ts index 0bb675c5fb..3605eb07a7 100644 --- a/specification/inference/stream_inference/StreamInferenceRequest.ts +++ b/specification/inference/stream_inference/StreamInferenceRequest.ts @@ -20,6 +20,7 @@ import { TaskType } from '@inference/_types/TaskType' import { RequestBase } from '@_types/Base' import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Perform streaming inference. @@ -55,6 +56,13 @@ export interface Request extends RequestBase { */ task_type?: TaskType } + query_parameters: { + /** + * The amount of time to wait for the inference request to complete. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The text on which you want to perform the inference task. From 1feef1bcd04102969157b9a684e8d4c7b3ae4a44 Mon Sep 17 00:00:00 2001 From: David Kyle Date: Mon, 14 Jul 2025 17:21:43 +0100 Subject: [PATCH 2/3] timeouts per service --- specification/inference/put/PutRequest.ts | 5 +++-- .../put_alibabacloud/PutAlibabaCloudRequest.ts | 12 ++++++++++-- .../put_amazonbedrock/PutAmazonBedrockRequest.ts | 12 ++++++++++-- .../inference/put_anthropic/PutAnthropicRequest.ts | 12 ++++++++++-- .../put_azureaistudio/PutAzureAiStudioRequest.ts | 12 ++++++++++-- .../put_azureopenai/PutAzureOpenAiRequest.ts | 12 ++++++++++-- .../inference/put_cohere/PutCohereRequest.ts | 12 ++++++++++-- .../inference/put_deepseek/PutDeepSeekRequest.ts | 12 ++++++++++-- .../put_elasticsearch/PutElasticsearchRequest.ts | 12 ++++++++++-- specification/inference/put_elser/PutElserRequest.ts | 12 ++++++++++-- .../put_googleaistudio/PutGoogleAiStudioRequest.ts | 12 ++++++++++-- .../put_googlevertexai/PutGoogleVertexAiRequest.ts | 12 ++++++++++-- .../put_hugging_face/PutHuggingFaceRequest.ts | 12 ++++++++++-- .../inference/put_jinaai/PutJinaAiRequest.ts | 12 ++++++++++-- .../inference/put_mistral/PutMistralRequest.ts | 12 ++++++++++-- .../inference/put_openai/PutOpenAiRequest.ts | 12 ++++++++++-- .../inference/put_voyageai/PutVoyageAIRequest.ts | 12 ++++++++++-- .../inference/put_watsonx/PutWatsonxRequest.ts | 12 ++++++++++-- .../stream_completion/StreamInferenceRequest.ts | 3 ++- 19 files changed, 175 insertions(+), 37 deletions(-) diff --git a/specification/inference/put/PutRequest.ts b/specification/inference/put/PutRequest.ts index 9957a031d7..9227273850 100644 --- a/specification/inference/put/PutRequest.ts +++ b/specification/inference/put/PutRequest.ts @@ -17,10 +17,11 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { InferenceEndpoint } from '@inference/_types/Services' import { TaskType } from '@inference/_types/TaskType' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an inference endpoint. diff --git a/specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts b/specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts index f390250bbd..07a558043a 100644 --- a/specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts +++ b/specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { AlibabaCloudServiceSettings, AlibabaCloudServiceType, @@ -26,6 +24,9 @@ import { AlibabaCloudTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an AlibabaCloud AI Search inference endpoint. @@ -54,6 +55,13 @@ export interface Request extends RequestBase { */ alibabacloud_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts b/specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts index 0420c26263..b55338a9b8 100644 --- a/specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts +++ b/specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { AmazonBedrockServiceSettings, AmazonBedrockServiceType, @@ -26,6 +24,9 @@ import { AmazonBedrockTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an Amazon Bedrock inference endpoint. @@ -57,6 +58,13 @@ export interface Request extends RequestBase { */ amazonbedrock_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_anthropic/PutAnthropicRequest.ts b/specification/inference/put_anthropic/PutAnthropicRequest.ts index d0e0b87ed3..f982944825 100644 --- a/specification/inference/put_anthropic/PutAnthropicRequest.ts +++ b/specification/inference/put_anthropic/PutAnthropicRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { AnthropicServiceSettings, AnthropicServiceType, @@ -26,6 +24,9 @@ import { AnthropicTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an Anthropic inference endpoint. @@ -55,6 +56,13 @@ export interface Request extends RequestBase { */ anthropic_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts b/specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts index d09f31b75c..deb5b55e83 100644 --- a/specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts +++ b/specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { AzureAiStudioServiceSettings, AzureAiStudioServiceType, @@ -26,6 +24,9 @@ import { AzureAiStudioTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an Azure AI studio inference endpoint. @@ -54,6 +55,13 @@ export interface Request extends RequestBase { */ azureaistudio_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts b/specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts index 63f0c42a8b..629c738274 100644 --- a/specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts +++ b/specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { AzureOpenAIServiceSettings, AzureOpenAIServiceType, @@ -26,6 +24,9 @@ import { AzureOpenAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an Azure OpenAI inference endpoint. @@ -62,6 +63,13 @@ export interface Request extends RequestBase { */ azureopenai_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_cohere/PutCohereRequest.ts b/specification/inference/put_cohere/PutCohereRequest.ts index 52ddd382e7..6f42331dbd 100644 --- a/specification/inference/put_cohere/PutCohereRequest.ts +++ b/specification/inference/put_cohere/PutCohereRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { CohereServiceSettings, CohereServiceType, @@ -26,6 +24,9 @@ import { CohereTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create a Cohere inference endpoint. @@ -54,6 +55,13 @@ export interface Request extends RequestBase { */ cohere_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_deepseek/PutDeepSeekRequest.ts b/specification/inference/put_deepseek/PutDeepSeekRequest.ts index 8abd332dcb..3d70b2df8a 100644 --- a/specification/inference/put_deepseek/PutDeepSeekRequest.ts +++ b/specification/inference/put_deepseek/PutDeepSeekRequest.ts @@ -17,14 +17,15 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { DeepSeekServiceSettings, DeepSeekServiceType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' import { TaskTypeDeepSeek } from '@inference/_types/TaskType' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create a DeepSeek inference endpoint. @@ -53,6 +54,13 @@ export interface Request extends RequestBase { */ deepseek_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_elasticsearch/PutElasticsearchRequest.ts b/specification/inference/put_elasticsearch/PutElasticsearchRequest.ts index bdf2a8d991..28f715eddd 100644 --- a/specification/inference/put_elasticsearch/PutElasticsearchRequest.ts +++ b/specification/inference/put_elasticsearch/PutElasticsearchRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { ElasticsearchServiceSettings, ElasticsearchServiceType, @@ -26,6 +24,9 @@ import { ElasticsearchTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an Elasticsearch inference endpoint. @@ -68,6 +69,13 @@ export interface Request extends RequestBase { */ elasticsearch_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_elser/PutElserRequest.ts b/specification/inference/put_elser/PutElserRequest.ts index d9a4812243..6119c5acd2 100644 --- a/specification/inference/put_elser/PutElserRequest.ts +++ b/specification/inference/put_elser/PutElserRequest.ts @@ -17,14 +17,15 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { ElserServiceSettings, ElserServiceType, ElserTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an ELSER inference endpoint. @@ -68,6 +69,13 @@ export interface Request extends RequestBase { */ elser_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_googleaistudio/PutGoogleAiStudioRequest.ts b/specification/inference/put_googleaistudio/PutGoogleAiStudioRequest.ts index 691710a32e..7f264e08a9 100644 --- a/specification/inference/put_googleaistudio/PutGoogleAiStudioRequest.ts +++ b/specification/inference/put_googleaistudio/PutGoogleAiStudioRequest.ts @@ -17,14 +17,15 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { GoogleAiServiceType, GoogleAiStudioServiceSettings, GoogleAiStudioTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an Google AI Studio inference endpoint. @@ -53,6 +54,13 @@ export interface Request extends RequestBase { */ googleaistudio_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_googlevertexai/PutGoogleVertexAiRequest.ts b/specification/inference/put_googlevertexai/PutGoogleVertexAiRequest.ts index 40a65dbb56..008cc6caf0 100644 --- a/specification/inference/put_googlevertexai/PutGoogleVertexAiRequest.ts +++ b/specification/inference/put_googlevertexai/PutGoogleVertexAiRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { GoogleVertexAIServiceSettings, GoogleVertexAIServiceType, @@ -26,6 +24,9 @@ import { GoogleVertexAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create a Google Vertex AI inference endpoint. @@ -54,6 +55,13 @@ export interface Request extends RequestBase { */ googlevertexai_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_hugging_face/PutHuggingFaceRequest.ts b/specification/inference/put_hugging_face/PutHuggingFaceRequest.ts index 9181c42149..5de54d462b 100644 --- a/specification/inference/put_hugging_face/PutHuggingFaceRequest.ts +++ b/specification/inference/put_hugging_face/PutHuggingFaceRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { HuggingFaceServiceSettings, HuggingFaceServiceType, @@ -26,6 +24,9 @@ import { HuggingFaceTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create a Hugging Face inference endpoint. @@ -90,6 +91,13 @@ export interface Request extends RequestBase { */ huggingface_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_jinaai/PutJinaAiRequest.ts b/specification/inference/put_jinaai/PutJinaAiRequest.ts index c34b80d4a4..bb64af96bf 100644 --- a/specification/inference/put_jinaai/PutJinaAiRequest.ts +++ b/specification/inference/put_jinaai/PutJinaAiRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { JinaAIServiceSettings, JinaAIServiceType, @@ -26,6 +24,9 @@ import { JinaAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an JinaAI inference endpoint. @@ -57,6 +58,13 @@ export interface Request extends RequestBase { */ jinaai_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_mistral/PutMistralRequest.ts b/specification/inference/put_mistral/PutMistralRequest.ts index 17508e1505..e46be97654 100644 --- a/specification/inference/put_mistral/PutMistralRequest.ts +++ b/specification/inference/put_mistral/PutMistralRequest.ts @@ -17,14 +17,15 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { MistralServiceSettings, MistralServiceType, MistralTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create a Mistral inference endpoint. @@ -53,6 +54,13 @@ export interface Request extends RequestBase { */ mistral_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_openai/PutOpenAiRequest.ts b/specification/inference/put_openai/PutOpenAiRequest.ts index d322b89c36..d5425e2ab6 100644 --- a/specification/inference/put_openai/PutOpenAiRequest.ts +++ b/specification/inference/put_openai/PutOpenAiRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { OpenAIServiceSettings, OpenAIServiceType, @@ -26,6 +24,9 @@ import { OpenAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create an OpenAI inference endpoint. @@ -55,6 +56,13 @@ export interface Request extends RequestBase { */ openai_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_voyageai/PutVoyageAIRequest.ts b/specification/inference/put_voyageai/PutVoyageAIRequest.ts index 9d1c2e0006..389428cd4e 100644 --- a/specification/inference/put_voyageai/PutVoyageAIRequest.ts +++ b/specification/inference/put_voyageai/PutVoyageAIRequest.ts @@ -17,8 +17,6 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { VoyageAIServiceSettings, VoyageAIServiceType, @@ -26,6 +24,9 @@ import { VoyageAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create a VoyageAI inference endpoint. @@ -56,6 +57,13 @@ export interface Request extends RequestBase { */ voyageai_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The chunking configuration object. diff --git a/specification/inference/put_watsonx/PutWatsonxRequest.ts b/specification/inference/put_watsonx/PutWatsonxRequest.ts index f582fe0790..ffa095a93a 100644 --- a/specification/inference/put_watsonx/PutWatsonxRequest.ts +++ b/specification/inference/put_watsonx/PutWatsonxRequest.ts @@ -17,13 +17,14 @@ * under the License. */ -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' import { WatsonxServiceSettings, WatsonxServiceType, WatsonxTaskType } from '@inference/_types/CommonTypes' +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' /** * Create a Watsonx inference endpoint. @@ -54,6 +55,13 @@ export interface Request extends RequestBase { */ watsonx_inference_id: Id } + query_parameters: { + /** + * Specifies the amount of time to wait for the inference endpoint to be created. + * @server_default 30s + */ + timeout?: Duration + } body: { /** * The type of service supported for the specified task type. In this case, `watsonxai`. diff --git a/specification/inference/stream_completion/StreamInferenceRequest.ts b/specification/inference/stream_completion/StreamInferenceRequest.ts index cff1e6a62a..48f814cb59 100644 --- a/specification/inference/stream_completion/StreamInferenceRequest.ts +++ b/specification/inference/stream_completion/StreamInferenceRequest.ts @@ -17,9 +17,10 @@ * under the License. */ +import { TaskSettings } from '@inference/_types/Services' import { RequestBase } from '@_types/Base' import { Id } from '@_types/common' -import { TaskSettings } from '@inference/_types/Services' +import { Duration } from '@_types/Time' /** * Perform streaming inference. From b85672d06531b815584e68dba20a6db09cdfe6c3 Mon Sep 17 00:00:00 2001 From: David Kyle Date: Tue, 15 Jul 2025 10:56:03 +0100 Subject: [PATCH 3/3] code style --- specification/inference/put/PutRequest.ts | 4 ++-- .../inference/put_alibabacloud/PutAlibabaCloudRequest.ts | 6 +++--- .../inference/put_amazonbedrock/PutAmazonBedrockRequest.ts | 6 +++--- .../inference/put_anthropic/PutAnthropicRequest.ts | 6 +++--- .../inference/put_azureaistudio/PutAzureAiStudioRequest.ts | 6 +++--- .../inference/put_azureopenai/PutAzureOpenAiRequest.ts | 6 +++--- specification/inference/put_cohere/PutCohereRequest.ts | 6 +++--- specification/inference/put_deepseek/PutDeepSeekRequest.ts | 6 +++--- .../inference/put_elasticsearch/PutElasticsearchRequest.ts | 6 +++--- specification/inference/put_elser/PutElserRequest.ts | 6 +++--- .../put_googleaistudio/PutGoogleAiStudioRequest.ts | 6 +++--- .../put_googlevertexai/PutGoogleVertexAiRequest.ts | 6 +++--- .../inference/put_hugging_face/PutHuggingFaceRequest.ts | 6 +++--- specification/inference/put_jinaai/PutJinaAiRequest.ts | 6 +++--- specification/inference/put_mistral/PutMistralRequest.ts | 6 +++--- specification/inference/put_openai/PutOpenAiRequest.ts | 6 +++--- specification/inference/put_voyageai/PutVoyageAIRequest.ts | 6 +++--- specification/inference/put_watsonx/PutWatsonxRequest.ts | 6 +++--- .../inference/stream_completion/StreamInferenceRequest.ts | 2 +- 19 files changed, 54 insertions(+), 54 deletions(-) diff --git a/specification/inference/put/PutRequest.ts b/specification/inference/put/PutRequest.ts index 9227273850..4554574e32 100644 --- a/specification/inference/put/PutRequest.ts +++ b/specification/inference/put/PutRequest.ts @@ -17,11 +17,11 @@ * under the License. */ -import { InferenceEndpoint } from '@inference/_types/Services' -import { TaskType } from '@inference/_types/TaskType' import { RequestBase } from '@_types/Base' import { Id } from '@_types/common' import { Duration } from '@_types/Time' +import { InferenceEndpoint } from '@inference/_types/Services' +import { TaskType } from '@inference/_types/TaskType' /** * Create an inference endpoint. diff --git a/specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts b/specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts index 07a558043a..c725397056 100644 --- a/specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts +++ b/specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { AlibabaCloudServiceSettings, AlibabaCloudServiceType, @@ -24,9 +27,6 @@ import { AlibabaCloudTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an AlibabaCloud AI Search inference endpoint. diff --git a/specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts b/specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts index b55338a9b8..61927ce3bf 100644 --- a/specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts +++ b/specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { AmazonBedrockServiceSettings, AmazonBedrockServiceType, @@ -24,9 +27,6 @@ import { AmazonBedrockTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an Amazon Bedrock inference endpoint. diff --git a/specification/inference/put_anthropic/PutAnthropicRequest.ts b/specification/inference/put_anthropic/PutAnthropicRequest.ts index f982944825..d7942495c3 100644 --- a/specification/inference/put_anthropic/PutAnthropicRequest.ts +++ b/specification/inference/put_anthropic/PutAnthropicRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { AnthropicServiceSettings, AnthropicServiceType, @@ -24,9 +27,6 @@ import { AnthropicTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an Anthropic inference endpoint. diff --git a/specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts b/specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts index deb5b55e83..6ab0d8b029 100644 --- a/specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts +++ b/specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { AzureAiStudioServiceSettings, AzureAiStudioServiceType, @@ -24,9 +27,6 @@ import { AzureAiStudioTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an Azure AI studio inference endpoint. diff --git a/specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts b/specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts index 629c738274..af881b68ef 100644 --- a/specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts +++ b/specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { AzureOpenAIServiceSettings, AzureOpenAIServiceType, @@ -24,9 +27,6 @@ import { AzureOpenAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an Azure OpenAI inference endpoint. diff --git a/specification/inference/put_cohere/PutCohereRequest.ts b/specification/inference/put_cohere/PutCohereRequest.ts index 6f42331dbd..6ebfc47292 100644 --- a/specification/inference/put_cohere/PutCohereRequest.ts +++ b/specification/inference/put_cohere/PutCohereRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { CohereServiceSettings, CohereServiceType, @@ -24,9 +27,6 @@ import { CohereTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create a Cohere inference endpoint. diff --git a/specification/inference/put_deepseek/PutDeepSeekRequest.ts b/specification/inference/put_deepseek/PutDeepSeekRequest.ts index 3d70b2df8a..c188ffd857 100644 --- a/specification/inference/put_deepseek/PutDeepSeekRequest.ts +++ b/specification/inference/put_deepseek/PutDeepSeekRequest.ts @@ -17,15 +17,15 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { DeepSeekServiceSettings, DeepSeekServiceType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' import { TaskTypeDeepSeek } from '@inference/_types/TaskType' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create a DeepSeek inference endpoint. diff --git a/specification/inference/put_elasticsearch/PutElasticsearchRequest.ts b/specification/inference/put_elasticsearch/PutElasticsearchRequest.ts index 28f715eddd..b57b062f67 100644 --- a/specification/inference/put_elasticsearch/PutElasticsearchRequest.ts +++ b/specification/inference/put_elasticsearch/PutElasticsearchRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { ElasticsearchServiceSettings, ElasticsearchServiceType, @@ -24,9 +27,6 @@ import { ElasticsearchTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an Elasticsearch inference endpoint. diff --git a/specification/inference/put_elser/PutElserRequest.ts b/specification/inference/put_elser/PutElserRequest.ts index 6119c5acd2..179bb1e379 100644 --- a/specification/inference/put_elser/PutElserRequest.ts +++ b/specification/inference/put_elser/PutElserRequest.ts @@ -17,15 +17,15 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { ElserServiceSettings, ElserServiceType, ElserTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an ELSER inference endpoint. diff --git a/specification/inference/put_googleaistudio/PutGoogleAiStudioRequest.ts b/specification/inference/put_googleaistudio/PutGoogleAiStudioRequest.ts index 7f264e08a9..6871ceb750 100644 --- a/specification/inference/put_googleaistudio/PutGoogleAiStudioRequest.ts +++ b/specification/inference/put_googleaistudio/PutGoogleAiStudioRequest.ts @@ -17,15 +17,15 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { GoogleAiServiceType, GoogleAiStudioServiceSettings, GoogleAiStudioTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an Google AI Studio inference endpoint. diff --git a/specification/inference/put_googlevertexai/PutGoogleVertexAiRequest.ts b/specification/inference/put_googlevertexai/PutGoogleVertexAiRequest.ts index 008cc6caf0..80cf04059e 100644 --- a/specification/inference/put_googlevertexai/PutGoogleVertexAiRequest.ts +++ b/specification/inference/put_googlevertexai/PutGoogleVertexAiRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { GoogleVertexAIServiceSettings, GoogleVertexAIServiceType, @@ -24,9 +27,6 @@ import { GoogleVertexAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create a Google Vertex AI inference endpoint. diff --git a/specification/inference/put_hugging_face/PutHuggingFaceRequest.ts b/specification/inference/put_hugging_face/PutHuggingFaceRequest.ts index 5de54d462b..d4c16ffb84 100644 --- a/specification/inference/put_hugging_face/PutHuggingFaceRequest.ts +++ b/specification/inference/put_hugging_face/PutHuggingFaceRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { HuggingFaceServiceSettings, HuggingFaceServiceType, @@ -24,9 +27,6 @@ import { HuggingFaceTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create a Hugging Face inference endpoint. diff --git a/specification/inference/put_jinaai/PutJinaAiRequest.ts b/specification/inference/put_jinaai/PutJinaAiRequest.ts index bb64af96bf..6685c2c874 100644 --- a/specification/inference/put_jinaai/PutJinaAiRequest.ts +++ b/specification/inference/put_jinaai/PutJinaAiRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { JinaAIServiceSettings, JinaAIServiceType, @@ -24,9 +27,6 @@ import { JinaAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an JinaAI inference endpoint. diff --git a/specification/inference/put_mistral/PutMistralRequest.ts b/specification/inference/put_mistral/PutMistralRequest.ts index e46be97654..3a7b5eaace 100644 --- a/specification/inference/put_mistral/PutMistralRequest.ts +++ b/specification/inference/put_mistral/PutMistralRequest.ts @@ -17,15 +17,15 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { MistralServiceSettings, MistralServiceType, MistralTaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create a Mistral inference endpoint. diff --git a/specification/inference/put_openai/PutOpenAiRequest.ts b/specification/inference/put_openai/PutOpenAiRequest.ts index d5425e2ab6..b6bb675f46 100644 --- a/specification/inference/put_openai/PutOpenAiRequest.ts +++ b/specification/inference/put_openai/PutOpenAiRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { OpenAIServiceSettings, OpenAIServiceType, @@ -24,9 +27,6 @@ import { OpenAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create an OpenAI inference endpoint. diff --git a/specification/inference/put_voyageai/PutVoyageAIRequest.ts b/specification/inference/put_voyageai/PutVoyageAIRequest.ts index 389428cd4e..2bb4b7ef46 100644 --- a/specification/inference/put_voyageai/PutVoyageAIRequest.ts +++ b/specification/inference/put_voyageai/PutVoyageAIRequest.ts @@ -17,6 +17,9 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { VoyageAIServiceSettings, VoyageAIServiceType, @@ -24,9 +27,6 @@ import { VoyageAITaskType } from '@inference/_types/CommonTypes' import { InferenceChunkingSettings } from '@inference/_types/Services' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create a VoyageAI inference endpoint. diff --git a/specification/inference/put_watsonx/PutWatsonxRequest.ts b/specification/inference/put_watsonx/PutWatsonxRequest.ts index ffa095a93a..b48fb3aa4c 100644 --- a/specification/inference/put_watsonx/PutWatsonxRequest.ts +++ b/specification/inference/put_watsonx/PutWatsonxRequest.ts @@ -17,14 +17,14 @@ * under the License. */ +import { RequestBase } from '@_types/Base' +import { Id } from '@_types/common' +import { Duration } from '@_types/Time' import { WatsonxServiceSettings, WatsonxServiceType, WatsonxTaskType } from '@inference/_types/CommonTypes' -import { RequestBase } from '@_types/Base' -import { Id } from '@_types/common' -import { Duration } from '@_types/Time' /** * Create a Watsonx inference endpoint. diff --git a/specification/inference/stream_completion/StreamInferenceRequest.ts b/specification/inference/stream_completion/StreamInferenceRequest.ts index 48f814cb59..0e08af6a6f 100644 --- a/specification/inference/stream_completion/StreamInferenceRequest.ts +++ b/specification/inference/stream_completion/StreamInferenceRequest.ts @@ -17,10 +17,10 @@ * under the License. */ -import { TaskSettings } from '@inference/_types/Services' import { RequestBase } from '@_types/Base' import { Id } from '@_types/common' import { Duration } from '@_types/Time' +import { TaskSettings } from '@inference/_types/Services' /** * Perform streaming inference.