|
@@ -312,39 +312,39 @@ class AzureOpenAILargeLanguageModel(_CommonAzureOpenAI, LargeLanguageModel):
|
|
|
if user:
|
|
|
extra_model_kwargs["user"] = user
|
|
|
|
|
|
- # clear illegal prompt messages
|
|
|
- prompt_messages = self._clear_illegal_prompt_messages(model, prompt_messages)
|
|
|
-
|
|
|
- block_as_stream = False
|
|
|
- if model.startswith("o1"):
|
|
|
- if stream:
|
|
|
- block_as_stream = True
|
|
|
- stream = False
|
|
|
-
|
|
|
- if "stream_options" in extra_model_kwargs:
|
|
|
- del extra_model_kwargs["stream_options"]
|
|
|
-
|
|
|
- if "stop" in extra_model_kwargs:
|
|
|
- del extra_model_kwargs["stop"]
|
|
|
-
|
|
|
- # chat model
|
|
|
- response = client.chat.completions.create(
|
|
|
- messages=[self._convert_prompt_message_to_dict(m) for m in prompt_messages],
|
|
|
- model=model,
|
|
|
- stream=stream,
|
|
|
- **model_parameters,
|
|
|
- **extra_model_kwargs,
|
|
|
- )
|
|
|
+ # clear illegal prompt messages
|
|
|
+ prompt_messages = self._clear_illegal_prompt_messages(model, prompt_messages)
|
|
|
|
|
|
+ block_as_stream = False
|
|
|
+ if model.startswith("o1"):
|
|
|
if stream:
|
|
|
- return self._handle_chat_generate_stream_response(model, credentials, response, prompt_messages, tools)
|
|
|
+ block_as_stream = True
|
|
|
+ stream = False
|
|
|
+
|
|
|
+ if "stream_options" in extra_model_kwargs:
|
|
|
+ del extra_model_kwargs["stream_options"]
|
|
|
+
|
|
|
+ if "stop" in extra_model_kwargs:
|
|
|
+ del extra_model_kwargs["stop"]
|
|
|
+
|
|
|
+ # chat model
|
|
|
+ response = client.chat.completions.create(
|
|
|
+ messages=[self._convert_prompt_message_to_dict(m) for m in prompt_messages],
|
|
|
+ model=model,
|
|
|
+ stream=stream,
|
|
|
+ **model_parameters,
|
|
|
+ **extra_model_kwargs,
|
|
|
+ )
|
|
|
+
|
|
|
+ if stream:
|
|
|
+ return self._handle_chat_generate_stream_response(model, credentials, response, prompt_messages, tools)
|
|
|
|
|
|
- block_result = self._handle_chat_generate_response(model, credentials, response, prompt_messages, tools)
|
|
|
+ block_result = self._handle_chat_generate_response(model, credentials, response, prompt_messages, tools)
|
|
|
|
|
|
- if block_as_stream:
|
|
|
- return self._handle_chat_block_as_stream_response(block_result, prompt_messages, stop)
|
|
|
+ if block_as_stream:
|
|
|
+ return self._handle_chat_block_as_stream_response(block_result, prompt_messages, stop)
|
|
|
|
|
|
- return block_result
|
|
|
+ return block_result
|
|
|
|
|
|
def _handle_chat_block_as_stream_response(
|
|
|
self,
|