@@ -3217,7 +3217,8 @@ def decorator():
32173217 def text_completion_input (self ,
32183218 arg_name : str ,
32193219 prompt : str ,
3220- chat_model : Optional [OpenAIModels ] = OpenAIModels .DefaultChatModel , # NoQA
3220+ chat_model : Optional [
3221+ OpenAIModels ] = OpenAIModels .DefaultChatModel ,
32213222 temperature : Optional [str ] = "0.5" ,
32223223 top_p : Optional [str ] = None ,
32233224 max_tokens : Optional [str ] = "100" ,
@@ -3373,7 +3374,8 @@ def decorator():
33733374 def assistant_post_input (self , arg_name : str ,
33743375 id : str ,
33753376 user_message : str ,
3376- chat_model : Optional [str ] = OpenAIModels .DefaultChatModel ,
3377+ chat_model : Optional [
3378+ OpenAIModels ] = OpenAIModels .DefaultChatModel ,
33773379 chat_storage_connection_setting : Optional [str ] = "AzureWebJobsStorage" , # noqa: E501
33783380 collection_name : Optional [str ] = "ChatState" , # noqa: E501
33793381 temperature : Optional [str ] = "0.5" ,
@@ -3409,7 +3411,7 @@ def assistant_post_input(self, arg_name: str,
34093411 :param max_tokens: The maximum number of tokens to generate in the
34103412 completion. The token count of your prompt plus max_tokens cannot
34113413 exceed the model's context length. Most models have a context length of
3412- 2048 tokens (except for the newest models, which support 4096)
3414+ 2048 tokens (except for the newest models, which support 4096)
34133415 :param is_reasoning_model: Whether the configured chat completion model
34143416 is a reasoning model or not. Properties max_tokens and temperature are not
34153417 supported for reasoning models.
@@ -3449,7 +3451,8 @@ def embeddings_input(self,
34493451 arg_name : str ,
34503452 input : str ,
34513453 input_type : InputType ,
3452- embeddings_model : Optional [str ] = OpenAIModels .DefaultEmbeddingsModel ,
3454+ embeddings_model : Optional [
3455+ OpenAIModels ] = OpenAIModels .DefaultEmbeddingsModel ,
34533456 max_chunk_length : Optional [int ] = 8 * 1024 ,
34543457 max_overlap : Optional [int ] = 128 ,
34553458 data_type : Optional [
@@ -3504,8 +3507,10 @@ def semantic_search_input(self,
35043507 search_connection_name : str ,
35053508 collection : str ,
35063509 query : Optional [str ] = None ,
3507- embeddings_model : Optional [OpenAIModels ] = OpenAIModels .DefaultEmbeddingsModel , # NoQA
3508- chat_model : Optional [OpenAIModels ] = OpenAIModels .DefaultChatModel , # NoQA
3510+ embeddings_model : Optional [
3511+ OpenAIModels ] = OpenAIModels .DefaultEmbeddingsModel ,
3512+ chat_model : Optional [
3513+ OpenAIModels ] = OpenAIModels .DefaultChatModel ,
35093514 system_prompt : Optional [str ] = semantic_search_system_prompt , # NoQA
35103515 max_knowledge_count : Optional [int ] = 1 ,
35113516 temperature : Optional [str ] = "0.5" ,
@@ -3595,7 +3600,8 @@ def embeddings_store_output(self,
35953600 input_type : InputType ,
35963601 store_connection_name : str ,
35973602 collection : str ,
3598- embeddings_model : Optional [OpenAIModels ] = OpenAIModels .DefaultEmbeddingsModel , # NoQA
3603+ embeddings_model : Optional [
3604+ OpenAIModels ] = OpenAIModels .DefaultEmbeddingsModel ,
35993605 max_chunk_length : Optional [int ] = 8 * 1024 ,
36003606 max_overlap : Optional [int ] = 128 ,
36013607 data_type : Optional [
0 commit comments