@@ -3217,8 +3217,9 @@ def decorator():
3217
3217
def text_completion_input (self ,
3218
3218
arg_name : str ,
3219
3219
prompt : str ,
3220
- chat_model : Optional [
3221
- OpenAIModels ] = OpenAIModels .DefaultChatModel ,
3220
+ chat_model : Optional
3221
+ [OpenAIModels ]
3222
+ = OpenAIModels .DefaultChatModel ,
3222
3223
temperature : Optional [str ] = "0.5" ,
3223
3224
top_p : Optional [str ] = None ,
3224
3225
max_tokens : Optional [str ] = "100" ,
@@ -3374,8 +3375,9 @@ def decorator():
3374
3375
def assistant_post_input (self , arg_name : str ,
3375
3376
id : str ,
3376
3377
user_message : str ,
3377
- chat_model : Optional [
3378
- OpenAIModels ] = OpenAIModels .DefaultChatModel ,
3378
+ chat_model : Optional
3379
+ [OpenAIModels ]
3380
+ = OpenAIModels .DefaultChatModel ,
3379
3381
chat_storage_connection_setting : Optional [str ] = "AzureWebJobsStorage" , # noqa: E501
3380
3382
collection_name : Optional [str ] = "ChatState" , # noqa: E501
3381
3383
temperature : Optional [str ] = "0.5" ,
@@ -3394,7 +3396,8 @@ def assistant_post_input(self, arg_name: str,
3394
3396
:param id: The ID of the assistant to update.
3395
3397
:param user_message: The user message that user has entered for
3396
3398
assistant to respond to.
3397
- :param chat_model: The deployment name or model name of OpenAI Chat Completion API.
3399
+ :param chat_model: The deployment name or model name of OpenAI Chat
3400
+ Completion API.
3398
3401
:param chat_storage_connection_setting: The configuration section name
3399
3402
for the table settings for assistant chat storage. The default value is
3400
3403
"AzureWebJobsStorage".
@@ -3413,8 +3416,8 @@ def assistant_post_input(self, arg_name: str,
3413
3416
exceed the model's context length. Most models have a context length of
3414
3417
2048 tokens (except for the newest models, which support 4096)
3415
3418
:param is_reasoning_model: Whether the configured chat completion model
3416
- is a reasoning model or not. Properties max_tokens and temperature are not
3417
- supported for reasoning models.
3419
+ is a reasoning model or not. Properties max_tokens and temperature are
3420
+ not supported for reasoning models.
3418
3421
:param data_type: Defines how Functions runtime should treat the
3419
3422
parameter value
3420
3423
:param kwargs: Keyword arguments for specifying additional binding
@@ -3451,8 +3454,9 @@ def embeddings_input(self,
3451
3454
arg_name : str ,
3452
3455
input : str ,
3453
3456
input_type : InputType ,
3454
- embeddings_model : Optional [
3455
- OpenAIModels ] = OpenAIModels .DefaultEmbeddingsModel ,
3457
+ embeddings_model : Optional
3458
+ [OpenAIModels ]
3459
+ = OpenAIModels .DefaultEmbeddingsModel ,
3456
3460
max_chunk_length : Optional [int ] = 8 * 1024 ,
3457
3461
max_overlap : Optional [int ] = 128 ,
3458
3462
data_type : Optional [
@@ -3469,7 +3473,8 @@ def embeddings_input(self,
3469
3473
:param input: The input source containing the data to generate
3470
3474
embeddings for.
3471
3475
:param input_type: The type of the input.
3472
- :param embeddings_model: The deployment name or model name for OpenAI Embeddings.
3476
+ :param embeddings_model: The deployment name or model name for OpenAI
3477
+ Embeddings.
3473
3478
:param max_chunk_length: The maximum number of characters to chunk the
3474
3479
input into. Default value: 8 * 1024
3475
3480
:param max_overlap: The maximum number of characters to overlap
@@ -3507,10 +3512,12 @@ def semantic_search_input(self,
3507
3512
search_connection_name : str ,
3508
3513
collection : str ,
3509
3514
query : Optional [str ] = None ,
3510
- embeddings_model : Optional [
3511
- OpenAIModels ] = OpenAIModels .DefaultEmbeddingsModel ,
3512
- chat_model : Optional [
3513
- OpenAIModels ] = OpenAIModels .DefaultChatModel ,
3515
+ embeddings_model : Optional
3516
+ [OpenAIModels ]
3517
+ = OpenAIModels .DefaultEmbeddingsModel ,
3518
+ chat_model : Optional
3519
+ [OpenAIModels ]
3520
+ = OpenAIModels .DefaultChatModel ,
3514
3521
system_prompt : Optional [str ] = semantic_search_system_prompt , # NoQA
3515
3522
max_knowledge_count : Optional [int ] = 1 ,
3516
3523
temperature : Optional [str ] = "0.5" ,
@@ -3533,14 +3540,15 @@ def semantic_search_input(self,
3533
3540
Ref: https://platform.openai.com/docs/guides/embeddings
3534
3541
3535
3542
:param arg_name: The name of binding parameter in the function code.
3536
- :param search_connection_name: app setting or environment variable which
3537
- contains a vector search connection setting value.
3543
+ :param search_connection_name: app setting or environment variable
3544
+ which contains a vector search connection setting value.
3538
3545
:param collection: The name of the collection or table to search or
3539
3546
store.
3540
3547
:param query: The semantic query text to use for searching.
3541
- :param embeddings_model: The deployment name or model name for OpenAI Embeddings.
3542
- The default value is "text-embedding-ada-002".
3543
- :param chat_model: The deployment name or model name of OpenAI Chat Completion API.
3548
+ :param embeddings_model: The deployment name or model name for OpenAI
3549
+ Embeddings. The default value is "text-embedding-ada-002".
3550
+ :param chat_model: The deployment name or model name of OpenAI Chat
3551
+ Completion API.
3544
3552
:param system_prompt: Optional. The system prompt to use for prompting
3545
3553
the large language model.
3546
3554
:param max_knowledge_count: Optional. The number of knowledge items to
@@ -3556,10 +3564,10 @@ def semantic_search_input(self,
3556
3564
:param max_tokens: The maximum number of tokens to generate in the
3557
3565
completion. The token count of your prompt plus max_tokens cannot
3558
3566
exceed the model's context length. Most models have a context length of
3559
- 2048 tokens (except for the newest models, which support 4096)
3567
+ 2048 tokens (except for the newest models, which support 4096)
3560
3568
:param is_reasoning_model: Whether the configured chat completion model
3561
- is a reasoning model or not. Properties max_tokens and temperature are not
3562
- supported for reasoning models.
3569
+ is a reasoning model or not. Properties max_tokens and temperature are
3570
+ not supported for reasoning models.
3563
3571
:param data_type: Optional. Defines how Functions runtime should treat
3564
3572
the parameter value. Default value: None
3565
3573
:param kwargs: Keyword arguments for specifying additional binding
@@ -3600,8 +3608,9 @@ def embeddings_store_output(self,
3600
3608
input_type : InputType ,
3601
3609
store_connection_name : str ,
3602
3610
collection : str ,
3603
- embeddings_model : Optional [
3604
- OpenAIModels ] = OpenAIModels .DefaultEmbeddingsModel ,
3611
+ embeddings_model : Optional
3612
+ [OpenAIModels ]
3613
+ = OpenAIModels .DefaultEmbeddingsModel ,
3605
3614
max_chunk_length : Optional [int ] = 8 * 1024 ,
3606
3615
max_overlap : Optional [int ] = 128 ,
3607
3616
data_type : Optional [
@@ -3624,7 +3633,8 @@ def embeddings_store_output(self,
3624
3633
:param store_connection_name: The name of an app setting or environment
3625
3634
variable which contains a vectore store connection setting value
3626
3635
:param collection: The collection or table to search.
3627
- :param embeddings_model: The deployment name or model name for OpenAI Embeddings.
3636
+ :param embeddings_model: The deployment name or model name for OpenAI
3637
+ Embeddings.
3628
3638
:param max_chunk_length: The maximum number of characters to chunk the
3629
3639
input into.
3630
3640
:param max_overlap: The maximum number of characters to overlap between
0 commit comments