Skip to content

Commit 9746b8b

Browse files
committed
lint and fix
1 parent ee1c514 commit 9746b8b

File tree

10 files changed

+261
-260
lines changed

10 files changed

+261
-260
lines changed

patchwork/common/client/llm/aio.py

Lines changed: 11 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -9,11 +9,12 @@
99
ChatCompletionToolParam,
1010
completion_create_params,
1111
)
12-
from typing_extensions import Dict, Iterable, List, Optional, Union, AsyncIterator
1312
from pydantic_ai.messages import ModelMessage, ModelResponse
14-
from pydantic_ai.models import ModelRequestParameters, StreamedResponse, Model
13+
from pydantic_ai.models import ModelRequestParameters, StreamedResponse
1514
from pydantic_ai.settings import ModelSettings
1615
from pydantic_ai.usage import Usage
16+
from typing_extensions import AsyncIterator, Dict, Iterable, List, Optional, Union
17+
1718
from patchwork.common.client.llm.anthropic import AnthropicLlmClient
1819
from patchwork.common.client.llm.google import GoogleLlmClient
1920
from patchwork.common.client.llm.openai_ import OpenAiLlmClient
@@ -44,10 +45,10 @@ def __get_model(self, model_settings: ModelSettings | None) -> None:
4445
return model_name
4546

4647
async def request(
47-
self,
48-
messages: list[ModelMessage],
49-
model_settings: ModelSettings | None,
50-
model_request_parameters: ModelRequestParameters,
48+
self,
49+
messages: list[ModelMessage],
50+
model_settings: ModelSettings | None,
51+
model_request_parameters: ModelRequestParameters,
5152
) -> tuple[ModelResponse, Usage]:
5253
model = self.__get_model(model_settings)
5354
if model is None:
@@ -64,10 +65,10 @@ async def request(
6465
)
6566

6667
async def request_stream(
67-
self,
68-
messages: list[ModelMessage],
69-
model_settings: ModelSettings | None,
70-
model_request_parameters: ModelRequestParameters,
68+
self,
69+
messages: list[ModelMessage],
70+
model_settings: ModelSettings | None,
71+
model_request_parameters: ModelRequestParameters,
7172
) -> AsyncIterator[StreamedResponse]:
7273
model = self.__get_model(model_settings)
7374
if model is None:

patchwork/common/client/llm/anthropic.py

Lines changed: 15 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
import json
44
import time
5-
from functools import lru_cache, cached_property
5+
from functools import cached_property, lru_cache
66

77
from anthropic import Anthropic
88
from anthropic.types import Message, MessageParam, TextBlockParam
@@ -14,18 +14,18 @@
1414
ChatCompletionToolParam,
1515
completion_create_params,
1616
)
17-
from pydantic_ai.messages import ModelMessage, ModelResponse
18-
from pydantic_ai.models import ModelRequestParameters, StreamedResponse, Model
19-
from pydantic_ai.settings import ModelSettings
20-
from pydantic_ai.usage import Usage
21-
from pydantic_ai.models.anthropic import AnthropicModel
2217
from openai.types.chat.chat_completion import Choice, CompletionUsage
2318
from openai.types.chat.chat_completion_message_tool_call import (
2419
ChatCompletionMessageToolCall,
2520
Function,
2621
)
2722
from openai.types.completion_usage import CompletionUsage
28-
from typing_extensions import Dict, Iterable, List, Optional, Union, AsyncIterator
23+
from pydantic_ai.messages import ModelMessage, ModelResponse
24+
from pydantic_ai.models import Model, ModelRequestParameters, StreamedResponse
25+
from pydantic_ai.models.anthropic import AnthropicModel
26+
from pydantic_ai.settings import ModelSettings
27+
from pydantic_ai.usage import Usage
28+
from typing_extensions import AsyncIterator, Dict, Iterable, List, Optional, Union
2929

3030
from patchwork.common.client.llm.protocol import NOT_GIVEN, LlmClient, NotGiven
3131

@@ -94,19 +94,19 @@ def __get_pydantic_model(self, model_settings: ModelSettings | None) -> Model:
9494
return AnthropicModel(model_name, api_key=self.__api_key)
9595

9696
async def request(
97-
self,
98-
messages: list[ModelMessage],
99-
model_settings: ModelSettings | None,
100-
model_request_parameters: ModelRequestParameters,
97+
self,
98+
messages: list[ModelMessage],
99+
model_settings: ModelSettings | None,
100+
model_request_parameters: ModelRequestParameters,
101101
) -> tuple[ModelResponse, Usage]:
102102
model = self.__get_pydantic_model(model_settings)
103103
return await model.request(messages, model_settings, model_request_parameters)
104104

105105
async def request_stream(
106-
self,
107-
messages: list[ModelMessage],
108-
model_settings: ModelSettings | None,
109-
model_request_parameters: ModelRequestParameters,
106+
self,
107+
messages: list[ModelMessage],
108+
model_settings: ModelSettings | None,
109+
model_request_parameters: ModelRequestParameters,
110110
) -> AsyncIterator[StreamedResponse]:
111111
model = self.__get_pydantic_model(model_settings)
112112
yield model.request_stream(messages, model_settings, model_request_parameters)

patchwork/common/client/llm/google.py

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -20,13 +20,13 @@
2020
ChatCompletionToolParam,
2121
completion_create_params,
2222
)
23+
from openai.types.chat.chat_completion import ChatCompletion, Choice
2324
from pydantic_ai.messages import ModelMessage, ModelResponse
24-
from pydantic_ai.models import ModelRequestParameters, StreamedResponse, Model
25+
from pydantic_ai.models import Model, ModelRequestParameters, StreamedResponse
26+
from pydantic_ai.models.gemini import GeminiModel
2527
from pydantic_ai.settings import ModelSettings
2628
from pydantic_ai.usage import Usage
27-
from pydantic_ai.models.gemini import GeminiModel
28-
from openai.types.chat.chat_completion import ChatCompletion, Choice
29-
from typing_extensions import Any, Dict, Iterable, List, Optional, Union, AsyncIterator
29+
from typing_extensions import Any, AsyncIterator, Dict, Iterable, List, Optional, Union
3030

3131
from patchwork.common.client.llm.protocol import NOT_GIVEN, LlmClient, NotGiven
3232
from patchwork.common.client.llm.utils import json_schema_to_model
@@ -60,19 +60,19 @@ def __get_pydantic_model(self, model_settings: ModelSettings | None) -> Model:
6060
return GeminiModel(model_name, api_key=self.__api_key)
6161

6262
async def request(
63-
self,
64-
messages: list[ModelMessage],
65-
model_settings: ModelSettings | None,
66-
model_request_parameters: ModelRequestParameters,
63+
self,
64+
messages: list[ModelMessage],
65+
model_settings: ModelSettings | None,
66+
model_request_parameters: ModelRequestParameters,
6767
) -> tuple[ModelResponse, Usage]:
6868
model = self.__get_pydantic_model(model_settings)
6969
return await model.request(messages, model_settings, model_request_parameters)
7070

7171
async def request_stream(
72-
self,
73-
messages: list[ModelMessage],
74-
model_settings: ModelSettings | None,
75-
model_request_parameters: ModelRequestParameters,
72+
self,
73+
messages: list[ModelMessage],
74+
model_settings: ModelSettings | None,
75+
model_request_parameters: ModelRequestParameters,
7676
) -> AsyncIterator[StreamedResponse]:
7777
model = self.__get_pydantic_model(model_settings)
7878
yield model.request_stream(messages, model_settings, model_request_parameters)

patchwork/common/client/llm/openai_.py

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -13,11 +13,11 @@
1313
completion_create_params,
1414
)
1515
from pydantic_ai.messages import ModelMessage, ModelResponse
16-
from pydantic_ai.models import ModelRequestParameters, StreamedResponse, Model
16+
from pydantic_ai.models import Model, ModelRequestParameters, StreamedResponse
17+
from pydantic_ai.models.openai import OpenAIModel
1718
from pydantic_ai.settings import ModelSettings
1819
from pydantic_ai.usage import Usage
19-
from pydantic_ai.models.openai import OpenAIModel
20-
from typing_extensions import Dict, Iterable, List, Optional, Union, AsyncIterator
20+
from typing_extensions import AsyncIterator, Dict, Iterable, List, Optional, Union
2121

2222
from patchwork.common.client.llm.protocol import NOT_GIVEN, LlmClient, NotGiven
2323
from patchwork.logger import logger
@@ -65,19 +65,19 @@ def __get_pydantic_model(self, model_settings: ModelSettings | None) -> Model:
6565
return OpenAIModel(model_name, base_url=self.__base_url, api_key=self.__api_key)
6666

6767
async def request(
68-
self,
69-
messages: list[ModelMessage],
70-
model_settings: ModelSettings | None,
71-
model_request_parameters: ModelRequestParameters,
68+
self,
69+
messages: list[ModelMessage],
70+
model_settings: ModelSettings | None,
71+
model_request_parameters: ModelRequestParameters,
7272
) -> tuple[ModelResponse, Usage]:
7373
model = self.__get_pydantic_model(model_settings)
7474
return await model.request(messages, model_settings, model_request_parameters)
7575

7676
async def request_stream(
77-
self,
78-
messages: list[ModelMessage],
79-
model_settings: ModelSettings | None,
80-
model_request_parameters: ModelRequestParameters,
77+
self,
78+
messages: list[ModelMessage],
79+
model_settings: ModelSettings | None,
80+
model_request_parameters: ModelRequestParameters,
8181
) -> AsyncIterator[StreamedResponse]:
8282
model = self.__get_pydantic_model(model_settings)
8383
yield model.request_stream(messages, model_settings, model_request_parameters)

patchwork/common/client/llm/protocol.py

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,7 @@
11
from __future__ import annotations
22

33
from abc import abstractmethod
4-
from typing import Dict, Any, List
5-
6-
from pydantic_ai.models import Model
4+
from typing import Any, Dict, List
75

86
from openai.types.chat import (
97
ChatCompletion,
@@ -12,6 +10,7 @@
1210
ChatCompletionToolParam,
1311
completion_create_params,
1412
)
13+
from pydantic_ai.models import Model
1514
from typing_extensions import Any, Dict, Iterable, List, Optional, Union
1615

1716

@@ -147,4 +146,3 @@ def chat_completion(
147146
top_p: Optional[float] | NotGiven = NOT_GIVEN,
148147
) -> ChatCompletion:
149148
...
150-

patchwork/common/tools/code_edit_tools.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -169,7 +169,6 @@ def execute(
169169
except Exception as e:
170170
return f"Error: {str(e)}"
171171

172-
173172
self.modified_files.update({abs_path})
174173
return result
175174

patchwork/patchflows/LogAnalysis/LogAnalysis.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ def __init__(self, inputs: dict):
2424
validate_steps_with_inputs(
2525
set(final_inputs.keys()).union({""}),
2626
# CallSQL,
27-
AgenticLLM,
27+
AgenticLLMV2,
2828
)
2929

3030
self.inputs = final_inputs

patchwork/steps/AgenticLLMV2/AgenticLLMV2.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,9 @@
11
from pathlib import Path
22

3-
from patchwork.common.multiturn_strategy.agentic_strategy import (
3+
from patchwork.common.multiturn_strategy.agentic_strategy_v2 import (
44
AgentConfig,
5-
AgenticStrategy,
5+
AgenticStrategyV2,
66
)
7-
from patchwork.common.multiturn_strategy.agentic_strategy_v2 import AgenticStrategyV2
87
from patchwork.common.tools import Tool
98
from patchwork.step import Step
109
from patchwork.steps.AgenticLLMV2.typed import AgenticLLMV2Inputs, AgenticLLMV2Outputs

patchwork/steps/SendEmail/typed.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,4 @@
1-
from typing_extensions import Annotated, Any, Dict, TypedDict
2-
3-
from patchwork.common.utils.step_typing import StepTypeConfig
1+
from typing_extensions import Any, TypedDict
42

53

64
class __SendEmailRequiredInputs(TypedDict):

0 commit comments

Comments
 (0)