File tree Expand file tree Collapse file tree 19 files changed +175
-37
lines changed Expand file tree Collapse file tree 19 files changed +175
-37
lines changed Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import { InferenceEndpoint } from '@inference/_types/Services'
23
21
import { TaskType } from '@inference/_types/TaskType'
22
+ import { RequestBase } from '@_types/Base'
23
+ import { Id } from '@_types/common'
24
+ import { Duration } from '@_types/Time'
24
25
25
26
/**
26
27
* Create an inference endpoint.
Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import {
23
21
AlibabaCloudServiceSettings ,
24
22
AlibabaCloudServiceType ,
25
23
AlibabaCloudTaskSettings ,
26
24
AlibabaCloudTaskType
27
25
} from '@inference/_types/CommonTypes'
28
26
import { InferenceChunkingSettings } from '@inference/_types/Services'
27
+ import { RequestBase } from '@_types/Base'
28
+ import { Id } from '@_types/common'
29
+ import { Duration } from '@_types/Time'
29
30
30
31
/**
31
32
* Create an AlibabaCloud AI Search inference endpoint.
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
54
55
*/
55
56
alibabacloud_inference_id : Id
56
57
}
58
+ query_parameters : {
59
+ /**
60
+ * Specifies the amount of time to wait for the inference endpoint to be created.
61
+ * @server_default 30s
62
+ */
63
+ timeout ?: Duration
64
+ }
57
65
body : {
58
66
/**
59
67
* The chunking configuration object.
Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import {
23
21
AmazonBedrockServiceSettings ,
24
22
AmazonBedrockServiceType ,
25
23
AmazonBedrockTaskSettings ,
26
24
AmazonBedrockTaskType
27
25
} from '@inference/_types/CommonTypes'
28
26
import { InferenceChunkingSettings } from '@inference/_types/Services'
27
+ import { RequestBase } from '@_types/Base'
28
+ import { Id } from '@_types/common'
29
+ import { Duration } from '@_types/Time'
29
30
30
31
/**
31
32
* Create an Amazon Bedrock inference endpoint.
@@ -57,6 +58,13 @@ export interface Request extends RequestBase {
57
58
*/
58
59
amazonbedrock_inference_id : Id
59
60
}
61
+ query_parameters : {
62
+ /**
63
+ * Specifies the amount of time to wait for the inference endpoint to be created.
64
+ * @server_default 30s
65
+ */
66
+ timeout ?: Duration
67
+ }
60
68
body : {
61
69
/**
62
70
* The chunking configuration object.
Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import {
23
21
AnthropicServiceSettings ,
24
22
AnthropicServiceType ,
25
23
AnthropicTaskSettings ,
26
24
AnthropicTaskType
27
25
} from '@inference/_types/CommonTypes'
28
26
import { InferenceChunkingSettings } from '@inference/_types/Services'
27
+ import { RequestBase } from '@_types/Base'
28
+ import { Id } from '@_types/common'
29
+ import { Duration } from '@_types/Time'
29
30
30
31
/**
31
32
* Create an Anthropic inference endpoint.
@@ -55,6 +56,13 @@ export interface Request extends RequestBase {
55
56
*/
56
57
anthropic_inference_id : Id
57
58
}
59
+ query_parameters : {
60
+ /**
61
+ * Specifies the amount of time to wait for the inference endpoint to be created.
62
+ * @server_default 30s
63
+ */
64
+ timeout ?: Duration
65
+ }
58
66
body : {
59
67
/**
60
68
* The chunking configuration object.
Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import {
23
21
AzureAiStudioServiceSettings ,
24
22
AzureAiStudioServiceType ,
25
23
AzureAiStudioTaskSettings ,
26
24
AzureAiStudioTaskType
27
25
} from '@inference/_types/CommonTypes'
28
26
import { InferenceChunkingSettings } from '@inference/_types/Services'
27
+ import { RequestBase } from '@_types/Base'
28
+ import { Id } from '@_types/common'
29
+ import { Duration } from '@_types/Time'
29
30
30
31
/**
31
32
* Create an Azure AI studio inference endpoint.
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
54
55
*/
55
56
azureaistudio_inference_id : Id
56
57
}
58
+ query_parameters : {
59
+ /**
60
+ * Specifies the amount of time to wait for the inference endpoint to be created.
61
+ * @server_default 30s
62
+ */
63
+ timeout ?: Duration
64
+ }
57
65
body : {
58
66
/**
59
67
* The chunking configuration object.
Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import {
23
21
AzureOpenAIServiceSettings ,
24
22
AzureOpenAIServiceType ,
25
23
AzureOpenAITaskSettings ,
26
24
AzureOpenAITaskType
27
25
} from '@inference/_types/CommonTypes'
28
26
import { InferenceChunkingSettings } from '@inference/_types/Services'
27
+ import { RequestBase } from '@_types/Base'
28
+ import { Id } from '@_types/common'
29
+ import { Duration } from '@_types/Time'
29
30
30
31
/**
31
32
* Create an Azure OpenAI inference endpoint.
@@ -62,6 +63,13 @@ export interface Request extends RequestBase {
62
63
*/
63
64
azureopenai_inference_id : Id
64
65
}
66
+ query_parameters : {
67
+ /**
68
+ * Specifies the amount of time to wait for the inference endpoint to be created.
69
+ * @server_default 30s
70
+ */
71
+ timeout ?: Duration
72
+ }
65
73
body : {
66
74
/**
67
75
* The chunking configuration object.
Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import {
23
21
CohereServiceSettings ,
24
22
CohereServiceType ,
25
23
CohereTaskSettings ,
26
24
CohereTaskType
27
25
} from '@inference/_types/CommonTypes'
28
26
import { InferenceChunkingSettings } from '@inference/_types/Services'
27
+ import { RequestBase } from '@_types/Base'
28
+ import { Id } from '@_types/common'
29
+ import { Duration } from '@_types/Time'
29
30
30
31
/**
31
32
* Create a Cohere inference endpoint.
@@ -54,6 +55,13 @@ export interface Request extends RequestBase {
54
55
*/
55
56
cohere_inference_id : Id
56
57
}
58
+ query_parameters : {
59
+ /**
60
+ * Specifies the amount of time to wait for the inference endpoint to be created.
61
+ * @server_default 30s
62
+ */
63
+ timeout ?: Duration
64
+ }
57
65
body : {
58
66
/**
59
67
* The chunking configuration object.
Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import {
23
21
DeepSeekServiceSettings ,
24
22
DeepSeekServiceType
25
23
} from '@inference/_types/CommonTypes'
26
24
import { InferenceChunkingSettings } from '@inference/_types/Services'
27
25
import { TaskTypeDeepSeek } from '@inference/_types/TaskType'
26
+ import { RequestBase } from '@_types/Base'
27
+ import { Id } from '@_types/common'
28
+ import { Duration } from '@_types/Time'
28
29
29
30
/**
30
31
* Create a DeepSeek inference endpoint.
@@ -53,6 +54,13 @@ export interface Request extends RequestBase {
53
54
*/
54
55
deepseek_inference_id : Id
55
56
}
57
+ query_parameters : {
58
+ /**
59
+ * Specifies the amount of time to wait for the inference endpoint to be created.
60
+ * @server_default 30s
61
+ */
62
+ timeout ?: Duration
63
+ }
56
64
body : {
57
65
/**
58
66
* The chunking configuration object.
Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import {
23
21
ElasticsearchServiceSettings ,
24
22
ElasticsearchServiceType ,
25
23
ElasticsearchTaskSettings ,
26
24
ElasticsearchTaskType
27
25
} from '@inference/_types/CommonTypes'
28
26
import { InferenceChunkingSettings } from '@inference/_types/Services'
27
+ import { RequestBase } from '@_types/Base'
28
+ import { Id } from '@_types/common'
29
+ import { Duration } from '@_types/Time'
29
30
30
31
/**
31
32
* Create an Elasticsearch inference endpoint.
@@ -68,6 +69,13 @@ export interface Request extends RequestBase {
68
69
*/
69
70
elasticsearch_inference_id : Id
70
71
}
72
+ query_parameters : {
73
+ /**
74
+ * Specifies the amount of time to wait for the inference endpoint to be created.
75
+ * @server_default 30s
76
+ */
77
+ timeout ?: Duration
78
+ }
71
79
body : {
72
80
/**
73
81
* The chunking configuration object.
Original file line number Diff line number Diff line change 17
17
* under the License.
18
18
*/
19
19
20
- import { RequestBase } from '@_types/Base'
21
- import { Id } from '@_types/common'
22
20
import {
23
21
ElserServiceSettings ,
24
22
ElserServiceType ,
25
23
ElserTaskType
26
24
} from '@inference/_types/CommonTypes'
27
25
import { InferenceChunkingSettings } from '@inference/_types/Services'
26
+ import { RequestBase } from '@_types/Base'
27
+ import { Id } from '@_types/common'
28
+ import { Duration } from '@_types/Time'
28
29
29
30
/**
30
31
* Create an ELSER inference endpoint.
@@ -68,6 +69,13 @@ export interface Request extends RequestBase {
68
69
*/
69
70
elser_inference_id : Id
70
71
}
72
+ query_parameters : {
73
+ /**
74
+ * Specifies the amount of time to wait for the inference endpoint to be created.
75
+ * @server_default 30s
76
+ */
77
+ timeout ?: Duration
78
+ }
71
79
body : {
72
80
/**
73
81
* The chunking configuration object.
You can’t perform that action at this time.
0 commit comments