Skip to content

Commit 07a4922

Browse files
Add timeout params for PUT inference and POST inference/_stream (#4895) (#4921)
(cherry picked from commit 7fbfa61) Co-authored-by: David Kyle <[email protected]>
1 parent 97f7f1d commit 07a4922

File tree

20 files changed

+153
-0
lines changed

20 files changed

+153
-0
lines changed

output/typescript/types.ts

Lines changed: 1 addition & 0 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

specification/inference/put/PutRequest.ts

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919

2020
import { RequestBase } from '@_types/Base'
2121
import { Id } from '@_types/common'
22+
import { Duration } from '@_types/Time'
2223
import { InferenceEndpoint } from '@inference/_types/Services'
2324
import { TaskType } from '@inference/_types/TaskType'
2425

@@ -74,6 +75,13 @@ export interface Request extends RequestBase {
7475
*/
7576
inference_id: Id
7677
}
78+
query_parameters: {
79+
/**
80+
* Specifies the amount of time to wait for the inference endpoint to be created.
81+
* @server_default 30s
82+
*/
83+
timeout?: Duration
84+
}
7785
/** @codegen_name inference_config */
7886
body: InferenceEndpoint
7987
}

specification/inference/put_alibabacloud/PutAlibabaCloudRequest.ts

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919

2020
import { RequestBase } from '@_types/Base'
2121
import { Id } from '@_types/common'
22+
import { Duration } from '@_types/Time'
2223
import {
2324
AlibabaCloudServiceSettings,
2425
AlibabaCloudServiceType,
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
5455
*/
5556
alibabacloud_inference_id: Id
5657
}
58+
query_parameters: {
59+
/**
60+
* Specifies the amount of time to wait for the inference endpoint to be created.
61+
* @server_default 30s
62+
*/
63+
timeout?: Duration
64+
}
5765
body: {
5866
/**
5967
* The chunking configuration object.

specification/inference/put_amazonbedrock/PutAmazonBedrockRequest.ts

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919

2020
import { RequestBase } from '@_types/Base'
2121
import { Id } from '@_types/common'
22+
import { Duration } from '@_types/Time'
2223
import {
2324
AmazonBedrockServiceSettings,
2425
AmazonBedrockServiceType,
@@ -57,6 +58,13 @@ export interface Request extends RequestBase {
5758
*/
5859
amazonbedrock_inference_id: Id
5960
}
61+
query_parameters: {
62+
/**
63+
* Specifies the amount of time to wait for the inference endpoint to be created.
64+
* @server_default 30s
65+
*/
66+
timeout?: Duration
67+
}
6068
body: {
6169
/**
6270
* The chunking configuration object.

specification/inference/put_anthropic/PutAnthropicRequest.ts

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919

2020
import { RequestBase } from '@_types/Base'
2121
import { Id } from '@_types/common'
22+
import { Duration } from '@_types/Time'
2223
import {
2324
AnthropicServiceSettings,
2425
AnthropicServiceType,
@@ -55,6 +56,13 @@ export interface Request extends RequestBase {
5556
*/
5657
anthropic_inference_id: Id
5758
}
59+
query_parameters: {
60+
/**
61+
* Specifies the amount of time to wait for the inference endpoint to be created.
62+
* @server_default 30s
63+
*/
64+
timeout?: Duration
65+
}
5866
body: {
5967
/**
6068
* The chunking configuration object.

specification/inference/put_azureaistudio/PutAzureAiStudioRequest.ts

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919

2020
import { RequestBase } from '@_types/Base'
2121
import { Id } from '@_types/common'
22+
import { Duration } from '@_types/Time'
2223
import {
2324
AzureAiStudioServiceSettings,
2425
AzureAiStudioServiceType,
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
5455
*/
5556
azureaistudio_inference_id: Id
5657
}
58+
query_parameters: {
59+
/**
60+
* Specifies the amount of time to wait for the inference endpoint to be created.
61+
* @server_default 30s
62+
*/
63+
timeout?: Duration
64+
}
5765
body: {
5866
/**
5967
* The chunking configuration object.

specification/inference/put_azureopenai/PutAzureOpenAiRequest.ts

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919

2020
import { RequestBase } from '@_types/Base'
2121
import { Id } from '@_types/common'
22+
import { Duration } from '@_types/Time'
2223
import {
2324
AzureOpenAIServiceSettings,
2425
AzureOpenAIServiceType,
@@ -62,6 +63,13 @@ export interface Request extends RequestBase {
6263
*/
6364
azureopenai_inference_id: Id
6465
}
66+
query_parameters: {
67+
/**
68+
* Specifies the amount of time to wait for the inference endpoint to be created.
69+
* @server_default 30s
70+
*/
71+
timeout?: Duration
72+
}
6573
body: {
6674
/**
6775
* The chunking configuration object.

specification/inference/put_cohere/PutCohereRequest.ts

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919

2020
import { RequestBase } from '@_types/Base'
2121
import { Id } from '@_types/common'
22+
import { Duration } from '@_types/Time'
2223
import {
2324
CohereServiceSettings,
2425
CohereServiceType,
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
5455
*/
5556
cohere_inference_id: Id
5657
}
58+
query_parameters: {
59+
/**
60+
* Specifies the amount of time to wait for the inference endpoint to be created.
61+
* @server_default 30s
62+
*/
63+
timeout?: Duration
64+
}
5765
body: {
5866
/**
5967
* The chunking configuration object.

specification/inference/put_deepseek/PutDeepSeekRequest.ts

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919

2020
import { RequestBase } from '@_types/Base'
2121
import { Id } from '@_types/common'
22+
import { Duration } from '@_types/Time'
2223
import {
2324
DeepSeekServiceSettings,
2425
DeepSeekServiceType
@@ -53,6 +54,13 @@ export interface Request extends RequestBase {
5354
*/
5455
deepseek_inference_id: Id
5556
}
57+
query_parameters: {
58+
/**
59+
* Specifies the amount of time to wait for the inference endpoint to be created.
60+
* @server_default 30s
61+
*/
62+
timeout?: Duration
63+
}
5664
body: {
5765
/**
5866
* The chunking configuration object.

specification/inference/put_elasticsearch/PutElasticsearchRequest.ts

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919

2020
import { RequestBase } from '@_types/Base'
2121
import { Id } from '@_types/common'
22+
import { Duration } from '@_types/Time'
2223
import {
2324
ElasticsearchServiceSettings,
2425
ElasticsearchServiceType,
@@ -68,6 +69,13 @@ export interface Request extends RequestBase {
6869
*/
6970
elasticsearch_inference_id: Id
7071
}
72+
query_parameters: {
73+
/**
74+
* Specifies the amount of time to wait for the inference endpoint to be created.
75+
* @server_default 30s
76+
*/
77+
timeout?: Duration
78+
}
7179
body: {
7280
/**
7381
* The chunking configuration object.

0 commit comments

Comments
 (0)