Skip to content

Commit 1feef1b

Browse files
committed
timeouts per service
1 parent 8af6a83 commit 1feef1b

File tree

19 files changed

+175
-37
lines changed

19 files changed

+175
-37
lines changed

specification/inference/put/PutRequest.ts

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,10 +17,11 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import { InferenceEndpoint } from '@inference/_types/Services'
2321
import { TaskType } from '@inference/_types/TaskType'
22+
import { RequestBase } from '@_types/Base'
23+
import { Id } from '@_types/common'
24+
import { Duration } from '@_types/Time'
2425

2526
/**
2627
* Create an inference endpoint.

specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,15 +17,16 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import {
2321
AlibabaCloudServiceSettings,
2422
AlibabaCloudServiceType,
2523
AlibabaCloudTaskSettings,
2624
AlibabaCloudTaskType
2725
} from '@inference/_types/CommonTypes'
2826
import { InferenceChunkingSettings } from '@inference/_types/Services'
27+
import { RequestBase } from '@_types/Base'
28+
import { Id } from '@_types/common'
29+
import { Duration } from '@_types/Time'
2930

3031
/**
3132
* Create an AlibabaCloud AI Search inference endpoint.
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
5455
*/
5556
alibabacloud_inference_id: Id
5657
}
58+
query_parameters: {
59+
/**
60+
* Specifies the amount of time to wait for the inference endpoint to be created.
61+
* @server_default 30s
62+
*/
63+
timeout?: Duration
64+
}
5765
body: {
5866
/**
5967
* The chunking configuration object.

specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,15 +17,16 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import {
2321
AmazonBedrockServiceSettings,
2422
AmazonBedrockServiceType,
2523
AmazonBedrockTaskSettings,
2624
AmazonBedrockTaskType
2725
} from '@inference/_types/CommonTypes'
2826
import { InferenceChunkingSettings } from '@inference/_types/Services'
27+
import { RequestBase } from '@_types/Base'
28+
import { Id } from '@_types/common'
29+
import { Duration } from '@_types/Time'
2930

3031
/**
3132
* Create an Amazon Bedrock inference endpoint.
@@ -57,6 +58,13 @@ export interface Request extends RequestBase {
5758
*/
5859
amazonbedrock_inference_id: Id
5960
}
61+
query_parameters: {
62+
/**
63+
* Specifies the amount of time to wait for the inference endpoint to be created.
64+
* @server_default 30s
65+
*/
66+
timeout?: Duration
67+
}
6068
body: {
6169
/**
6270
* The chunking configuration object.

specification/inference/put_anthropic/PutAnthropicRequest.ts

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,15 +17,16 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import {
2321
AnthropicServiceSettings,
2422
AnthropicServiceType,
2523
AnthropicTaskSettings,
2624
AnthropicTaskType
2725
} from '@inference/_types/CommonTypes'
2826
import { InferenceChunkingSettings } from '@inference/_types/Services'
27+
import { RequestBase } from '@_types/Base'
28+
import { Id } from '@_types/common'
29+
import { Duration } from '@_types/Time'
2930

3031
/**
3132
* Create an Anthropic inference endpoint.
@@ -55,6 +56,13 @@ export interface Request extends RequestBase {
5556
*/
5657
anthropic_inference_id: Id
5758
}
59+
query_parameters: {
60+
/**
61+
* Specifies the amount of time to wait for the inference endpoint to be created.
62+
* @server_default 30s
63+
*/
64+
timeout?: Duration
65+
}
5866
body: {
5967
/**
6068
* The chunking configuration object.

specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,15 +17,16 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import {
2321
AzureAiStudioServiceSettings,
2422
AzureAiStudioServiceType,
2523
AzureAiStudioTaskSettings,
2624
AzureAiStudioTaskType
2725
} from '@inference/_types/CommonTypes'
2826
import { InferenceChunkingSettings } from '@inference/_types/Services'
27+
import { RequestBase } from '@_types/Base'
28+
import { Id } from '@_types/common'
29+
import { Duration } from '@_types/Time'
2930

3031
/**
3132
* Create an Azure AI studio inference endpoint.
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
5455
*/
5556
azureaistudio_inference_id: Id
5657
}
58+
query_parameters: {
59+
/**
60+
* Specifies the amount of time to wait for the inference endpoint to be created.
61+
* @server_default 30s
62+
*/
63+
timeout?: Duration
64+
}
5765
body: {
5866
/**
5967
* The chunking configuration object.

specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,15 +17,16 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import {
2321
AzureOpenAIServiceSettings,
2422
AzureOpenAIServiceType,
2523
AzureOpenAITaskSettings,
2624
AzureOpenAITaskType
2725
} from '@inference/_types/CommonTypes'
2826
import { InferenceChunkingSettings } from '@inference/_types/Services'
27+
import { RequestBase } from '@_types/Base'
28+
import { Id } from '@_types/common'
29+
import { Duration } from '@_types/Time'
2930

3031
/**
3132
* Create an Azure OpenAI inference endpoint.
@@ -62,6 +63,13 @@ export interface Request extends RequestBase {
6263
*/
6364
azureopenai_inference_id: Id
6465
}
66+
query_parameters: {
67+
/**
68+
* Specifies the amount of time to wait for the inference endpoint to be created.
69+
* @server_default 30s
70+
*/
71+
timeout?: Duration
72+
}
6573
body: {
6674
/**
6775
* The chunking configuration object.

specification/inference/put_cohere/PutCohereRequest.ts

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,15 +17,16 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import {
2321
CohereServiceSettings,
2422
CohereServiceType,
2523
CohereTaskSettings,
2624
CohereTaskType
2725
} from '@inference/_types/CommonTypes'
2826
import { InferenceChunkingSettings } from '@inference/_types/Services'
27+
import { RequestBase } from '@_types/Base'
28+
import { Id } from '@_types/common'
29+
import { Duration } from '@_types/Time'
2930

3031
/**
3132
* Create a Cohere inference endpoint.
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
5455
*/
5556
cohere_inference_id: Id
5657
}
58+
query_parameters: {
59+
/**
60+
* Specifies the amount of time to wait for the inference endpoint to be created.
61+
* @server_default 30s
62+
*/
63+
timeout?: Duration
64+
}
5765
body: {
5866
/**
5967
* The chunking configuration object.

specification/inference/put_deepseek/PutDeepSeekRequest.ts

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,14 +17,15 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import {
2321
DeepSeekServiceSettings,
2422
DeepSeekServiceType
2523
} from '@inference/_types/CommonTypes'
2624
import { InferenceChunkingSettings } from '@inference/_types/Services'
2725
import { TaskTypeDeepSeek } from '@inference/_types/TaskType'
26+
import { RequestBase } from '@_types/Base'
27+
import { Id } from '@_types/common'
28+
import { Duration } from '@_types/Time'
2829

2930
/**
3031
* Create a DeepSeek inference endpoint.
@@ -53,6 +54,13 @@ export interface Request extends RequestBase {
5354
*/
5455
deepseek_inference_id: Id
5556
}
57+
query_parameters: {
58+
/**
59+
* Specifies the amount of time to wait for the inference endpoint to be created.
60+
* @server_default 30s
61+
*/
62+
timeout?: Duration
63+
}
5664
body: {
5765
/**
5866
* The chunking configuration object.

specification/inference/put_elasticsearch/PutElasticsearchRequest.ts

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,15 +17,16 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import {
2321
ElasticsearchServiceSettings,
2422
ElasticsearchServiceType,
2523
ElasticsearchTaskSettings,
2624
ElasticsearchTaskType
2725
} from '@inference/_types/CommonTypes'
2826
import { InferenceChunkingSettings } from '@inference/_types/Services'
27+
import { RequestBase } from '@_types/Base'
28+
import { Id } from '@_types/common'
29+
import { Duration } from '@_types/Time'
2930

3031
/**
3132
* Create an Elasticsearch inference endpoint.
@@ -68,6 +69,13 @@ export interface Request extends RequestBase {
6869
*/
6970
elasticsearch_inference_id: Id
7071
}
72+
query_parameters: {
73+
/**
74+
* Specifies the amount of time to wait for the inference endpoint to be created.
75+
* @server_default 30s
76+
*/
77+
timeout?: Duration
78+
}
7179
body: {
7280
/**
7381
* The chunking configuration object.

specification/inference/put_elser/PutElserRequest.ts

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,14 +17,15 @@
1717
* under the License.
1818
*/
1919

20-
import { RequestBase } from '@_types/Base'
21-
import { Id } from '@_types/common'
2220
import {
2321
ElserServiceSettings,
2422
ElserServiceType,
2523
ElserTaskType
2624
} from '@inference/_types/CommonTypes'
2725
import { InferenceChunkingSettings } from '@inference/_types/Services'
26+
import { RequestBase } from '@_types/Base'
27+
import { Id } from '@_types/common'
28+
import { Duration } from '@_types/Time'
2829

2930
/**
3031
* Create an ELSER inference endpoint.
@@ -68,6 +69,13 @@ export interface Request extends RequestBase {
6869
*/
6970
elser_inference_id: Id
7071
}
72+
query_parameters: {
73+
/**
74+
* Specifies the amount of time to wait for the inference endpoint to be created.
75+
* @server_default 30s
76+
*/
77+
timeout?: Duration
78+
}
7179
body: {
7280
/**
7381
* The chunking configuration object.

0 commit comments

Comments
 (0)