File tree Expand file tree Collapse file tree 19 files changed +175
-37
lines changed Expand file tree Collapse file tree 19 files changed +175
-37
lines changed Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import { InferenceEndpoint } from '@inference/_types/Services'
2321import { TaskType } from '@inference/_types/TaskType'
22+ import { RequestBase } from '@_types/Base'
23+ import { Id } from '@_types/common'
24+ import { Duration } from '@_types/Time'
2425
2526/**
2627 * Create an inference endpoint.
Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import {
2321 AlibabaCloudServiceSettings ,
2422 AlibabaCloudServiceType ,
2523 AlibabaCloudTaskSettings ,
2624 AlibabaCloudTaskType
2725} from '@inference/_types/CommonTypes'
2826import { InferenceChunkingSettings } from '@inference/_types/Services'
27+ import { RequestBase } from '@_types/Base'
28+ import { Id } from '@_types/common'
29+ import { Duration } from '@_types/Time'
2930
3031/**
3132 * Create an AlibabaCloud AI Search inference endpoint.
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
5455 */
5556 alibabacloud_inference_id : Id
5657 }
58+ query_parameters : {
59+ /**
60+ * Specifies the amount of time to wait for the inference endpoint to be created.
61+ * @server_default 30s
62+ */
63+ timeout ?: Duration
64+ }
5765 body : {
5866 /**
5967 * The chunking configuration object.
Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import {
2321 AmazonBedrockServiceSettings ,
2422 AmazonBedrockServiceType ,
2523 AmazonBedrockTaskSettings ,
2624 AmazonBedrockTaskType
2725} from '@inference/_types/CommonTypes'
2826import { InferenceChunkingSettings } from '@inference/_types/Services'
27+ import { RequestBase } from '@_types/Base'
28+ import { Id } from '@_types/common'
29+ import { Duration } from '@_types/Time'
2930
3031/**
3132 * Create an Amazon Bedrock inference endpoint.
@@ -57,6 +58,13 @@ export interface Request extends RequestBase {
5758 */
5859 amazonbedrock_inference_id : Id
5960 }
61+ query_parameters : {
62+ /**
63+ * Specifies the amount of time to wait for the inference endpoint to be created.
64+ * @server_default 30s
65+ */
66+ timeout ?: Duration
67+ }
6068 body : {
6169 /**
6270 * The chunking configuration object.
Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import {
2321 AnthropicServiceSettings ,
2422 AnthropicServiceType ,
2523 AnthropicTaskSettings ,
2624 AnthropicTaskType
2725} from '@inference/_types/CommonTypes'
2826import { InferenceChunkingSettings } from '@inference/_types/Services'
27+ import { RequestBase } from '@_types/Base'
28+ import { Id } from '@_types/common'
29+ import { Duration } from '@_types/Time'
2930
3031/**
3132 * Create an Anthropic inference endpoint.
@@ -55,6 +56,13 @@ export interface Request extends RequestBase {
5556 */
5657 anthropic_inference_id : Id
5758 }
59+ query_parameters : {
60+ /**
61+ * Specifies the amount of time to wait for the inference endpoint to be created.
62+ * @server_default 30s
63+ */
64+ timeout ?: Duration
65+ }
5866 body : {
5967 /**
6068 * The chunking configuration object.
Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import {
2321 AzureAiStudioServiceSettings ,
2422 AzureAiStudioServiceType ,
2523 AzureAiStudioTaskSettings ,
2624 AzureAiStudioTaskType
2725} from '@inference/_types/CommonTypes'
2826import { InferenceChunkingSettings } from '@inference/_types/Services'
27+ import { RequestBase } from '@_types/Base'
28+ import { Id } from '@_types/common'
29+ import { Duration } from '@_types/Time'
2930
3031/**
3132 * Create an Azure AI studio inference endpoint.
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
5455 */
5556 azureaistudio_inference_id : Id
5657 }
58+ query_parameters : {
59+ /**
60+ * Specifies the amount of time to wait for the inference endpoint to be created.
61+ * @server_default 30s
62+ */
63+ timeout ?: Duration
64+ }
5765 body : {
5866 /**
5967 * The chunking configuration object.
Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import {
2321 AzureOpenAIServiceSettings ,
2422 AzureOpenAIServiceType ,
2523 AzureOpenAITaskSettings ,
2624 AzureOpenAITaskType
2725} from '@inference/_types/CommonTypes'
2826import { InferenceChunkingSettings } from '@inference/_types/Services'
27+ import { RequestBase } from '@_types/Base'
28+ import { Id } from '@_types/common'
29+ import { Duration } from '@_types/Time'
2930
3031/**
3132 * Create an Azure OpenAI inference endpoint.
@@ -62,6 +63,13 @@ export interface Request extends RequestBase {
6263 */
6364 azureopenai_inference_id : Id
6465 }
66+ query_parameters : {
67+ /**
68+ * Specifies the amount of time to wait for the inference endpoint to be created.
69+ * @server_default 30s
70+ */
71+ timeout ?: Duration
72+ }
6573 body : {
6674 /**
6775 * The chunking configuration object.
Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import {
2321 CohereServiceSettings ,
2422 CohereServiceType ,
2523 CohereTaskSettings ,
2624 CohereTaskType
2725} from '@inference/_types/CommonTypes'
2826import { InferenceChunkingSettings } from '@inference/_types/Services'
27+ import { RequestBase } from '@_types/Base'
28+ import { Id } from '@_types/common'
29+ import { Duration } from '@_types/Time'
2930
3031/**
3132 * Create a Cohere inference endpoint.
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
5455 */
5556 cohere_inference_id : Id
5657 }
58+ query_parameters : {
59+ /**
60+ * Specifies the amount of time to wait for the inference endpoint to be created.
61+ * @server_default 30s
62+ */
63+ timeout ?: Duration
64+ }
5765 body : {
5866 /**
5967 * The chunking configuration object.
Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import {
2321 DeepSeekServiceSettings ,
2422 DeepSeekServiceType
2523} from '@inference/_types/CommonTypes'
2624import { InferenceChunkingSettings } from '@inference/_types/Services'
2725import { TaskTypeDeepSeek } from '@inference/_types/TaskType'
26+ import { RequestBase } from '@_types/Base'
27+ import { Id } from '@_types/common'
28+ import { Duration } from '@_types/Time'
2829
2930/**
3031 * Create a DeepSeek inference endpoint.
@@ -53,6 +54,13 @@ export interface Request extends RequestBase {
5354 */
5455 deepseek_inference_id : Id
5556 }
57+ query_parameters : {
58+ /**
59+ * Specifies the amount of time to wait for the inference endpoint to be created.
60+ * @server_default 30s
61+ */
62+ timeout ?: Duration
63+ }
5664 body : {
5765 /**
5866 * The chunking configuration object.
Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import {
2321 ElasticsearchServiceSettings ,
2422 ElasticsearchServiceType ,
2523 ElasticsearchTaskSettings ,
2624 ElasticsearchTaskType
2725} from '@inference/_types/CommonTypes'
2826import { InferenceChunkingSettings } from '@inference/_types/Services'
27+ import { RequestBase } from '@_types/Base'
28+ import { Id } from '@_types/common'
29+ import { Duration } from '@_types/Time'
2930
3031/**
3132 * Create an Elasticsearch inference endpoint.
@@ -68,6 +69,13 @@ export interface Request extends RequestBase {
6869 */
6970 elasticsearch_inference_id : Id
7071 }
72+ query_parameters : {
73+ /**
74+ * Specifies the amount of time to wait for the inference endpoint to be created.
75+ * @server_default 30s
76+ */
77+ timeout ?: Duration
78+ }
7179 body : {
7280 /**
7381 * The chunking configuration object.
Original file line number Diff line number Diff line change 1717 * under the License.
1818 */
1919
20- import { RequestBase } from '@_types/Base'
21- import { Id } from '@_types/common'
2220import {
2321 ElserServiceSettings ,
2422 ElserServiceType ,
2523 ElserTaskType
2624} from '@inference/_types/CommonTypes'
2725import { InferenceChunkingSettings } from '@inference/_types/Services'
26+ import { RequestBase } from '@_types/Base'
27+ import { Id } from '@_types/common'
28+ import { Duration } from '@_types/Time'
2829
2930/**
3031 * Create an ELSER inference endpoint.
@@ -68,6 +69,13 @@ export interface Request extends RequestBase {
6869 */
6970 elser_inference_id : Id
7071 }
72+ query_parameters : {
73+ /**
74+ * Specifies the amount of time to wait for the inference endpoint to be created.
75+ * @server_default 30s
76+ */
77+ timeout ?: Duration
78+ }
7179 body : {
7280 /**
7381 * The chunking configuration object.
You can’t perform that action at this time.
0 commit comments