|
2 | 2 |
|
3 | 3 | import json |
4 | 4 | import time |
5 | | -from functools import lru_cache, cached_property |
| 5 | +from functools import cached_property, lru_cache |
6 | 6 |
|
7 | 7 | from anthropic import Anthropic |
8 | 8 | from anthropic.types import Message, MessageParam, TextBlockParam |
|
14 | 14 | ChatCompletionToolParam, |
15 | 15 | completion_create_params, |
16 | 16 | ) |
17 | | -from pydantic_ai.messages import ModelMessage, ModelResponse |
18 | | -from pydantic_ai.models import ModelRequestParameters, StreamedResponse, Model |
19 | | -from pydantic_ai.settings import ModelSettings |
20 | | -from pydantic_ai.usage import Usage |
21 | | -from pydantic_ai.models.anthropic import AnthropicModel |
22 | 17 | from openai.types.chat.chat_completion import Choice, CompletionUsage |
23 | 18 | from openai.types.chat.chat_completion_message_tool_call import ( |
24 | 19 | ChatCompletionMessageToolCall, |
25 | 20 | Function, |
26 | 21 | ) |
27 | 22 | from openai.types.completion_usage import CompletionUsage |
28 | | -from typing_extensions import Dict, Iterable, List, Optional, Union, AsyncIterator |
| 23 | +from pydantic_ai.messages import ModelMessage, ModelResponse |
| 24 | +from pydantic_ai.models import Model, ModelRequestParameters, StreamedResponse |
| 25 | +from pydantic_ai.models.anthropic import AnthropicModel |
| 26 | +from pydantic_ai.settings import ModelSettings |
| 27 | +from pydantic_ai.usage import Usage |
| 28 | +from typing_extensions import AsyncIterator, Dict, Iterable, List, Optional, Union |
29 | 29 |
|
30 | 30 | from patchwork.common.client.llm.protocol import NOT_GIVEN, LlmClient, NotGiven |
31 | 31 |
|
@@ -94,19 +94,19 @@ def __get_pydantic_model(self, model_settings: ModelSettings | None) -> Model: |
94 | 94 | return AnthropicModel(model_name, api_key=self.__api_key) |
95 | 95 |
|
96 | 96 | async def request( |
97 | | - self, |
98 | | - messages: list[ModelMessage], |
99 | | - model_settings: ModelSettings | None, |
100 | | - model_request_parameters: ModelRequestParameters, |
| 97 | + self, |
| 98 | + messages: list[ModelMessage], |
| 99 | + model_settings: ModelSettings | None, |
| 100 | + model_request_parameters: ModelRequestParameters, |
101 | 101 | ) -> tuple[ModelResponse, Usage]: |
102 | 102 | model = self.__get_pydantic_model(model_settings) |
103 | 103 | return await model.request(messages, model_settings, model_request_parameters) |
104 | 104 |
|
105 | 105 | async def request_stream( |
106 | | - self, |
107 | | - messages: list[ModelMessage], |
108 | | - model_settings: ModelSettings | None, |
109 | | - model_request_parameters: ModelRequestParameters, |
| 106 | + self, |
| 107 | + messages: list[ModelMessage], |
| 108 | + model_settings: ModelSettings | None, |
| 109 | + model_request_parameters: ModelRequestParameters, |
110 | 110 | ) -> AsyncIterator[StreamedResponse]: |
111 | 111 | model = self.__get_pydantic_model(model_settings) |
112 | 112 | yield model.request_stream(messages, model_settings, model_request_parameters) |
|
0 commit comments