Skip to content

Commit a2ad5ac

Browse files
authored
1 parent 5ecbb5f commit a2ad5ac

File tree

23 files changed

+54
-53
lines changed

23 files changed

+54
-53
lines changed

libs/langchain/langchain/agents/agent.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1185,7 +1185,7 @@ def _action_agent(self) -> Union[BaseSingleActionAgent, BaseMultiActionAgent]:
11851185
to reflect the changes made in the root_validator.
11861186
"""
11871187
if isinstance(self.agent, Runnable):
1188-
return cast(RunnableAgentType, self.agent)
1188+
return cast("RunnableAgentType", self.agent)
11891189
return self.agent
11901190

11911191
def save(self, file_path: Union[Path, str]) -> None:

libs/langchain/langchain/callbacks/streaming_aiter.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ async def aiter(self) -> AsyncIterator[str]:
7373
other.pop().cancel()
7474

7575
# Extract the value of the first completed task
76-
token_or_done = cast(Union[str, Literal[True]], done.pop().result())
76+
token_or_done = cast("Union[str, Literal[True]]", done.pop().result())
7777

7878
# If the extracted value is the boolean True, the done event was set
7979
if token_or_done is True:

libs/langchain/langchain/chains/base.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -411,7 +411,7 @@ def __call__(
411411

412412
return self.invoke(
413413
inputs,
414-
cast(RunnableConfig, {k: v for k, v in config.items() if v is not None}),
414+
cast("RunnableConfig", {k: v for k, v in config.items() if v is not None}),
415415
return_only_outputs=return_only_outputs,
416416
include_run_info=include_run_info,
417417
)
@@ -461,7 +461,7 @@ async def acall(
461461
}
462462
return await self.ainvoke(
463463
inputs,
464-
cast(RunnableConfig, {k: v for k, v in config.items() if k is not None}),
464+
cast("RunnableConfig", {k: v for k, v in config.items() if k is not None}),
465465
return_only_outputs=return_only_outputs,
466466
include_run_info=include_run_info,
467467
)

libs/langchain/langchain/chains/combine_documents/map_rerank.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -229,7 +229,7 @@ def _process_results(
229229
docs: list[Document],
230230
results: Sequence[Union[str, list[str], dict[str, str]]],
231231
) -> tuple[str, dict]:
232-
typed_results = cast(list[dict], results)
232+
typed_results = cast("list[dict]", results)
233233
sorted_res = sorted(
234234
zip(typed_results, docs),
235235
key=lambda x: -int(x[0][self.rank_key]),

libs/langchain/langchain/chains/llm.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -145,7 +145,7 @@ def generate(
145145
**self.llm_kwargs,
146146
)
147147
results = self.llm.bind(stop=stop, **self.llm_kwargs).batch(
148-
cast(list, prompts),
148+
cast("list", prompts),
149149
{"callbacks": callbacks},
150150
)
151151
generations: list[list[Generation]] = []
@@ -172,7 +172,7 @@ async def agenerate(
172172
**self.llm_kwargs,
173173
)
174174
results = await self.llm.bind(stop=stop, **self.llm_kwargs).abatch(
175-
cast(list, prompts),
175+
cast("list", prompts),
176176
{"callbacks": callbacks},
177177
)
178178
generations: list[list[Generation]] = []

libs/langchain/langchain/chains/openai_functions/qa_with_structure.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -76,11 +76,11 @@ def create_qa_with_structure_chain(
7676
raise ValueError(msg)
7777
if isinstance(schema, type) and is_basemodel_subclass(schema):
7878
if hasattr(schema, "model_json_schema"):
79-
schema_dict = cast(dict, schema.model_json_schema())
79+
schema_dict = cast("dict", schema.model_json_schema())
8080
else:
81-
schema_dict = cast(dict, schema.schema())
81+
schema_dict = cast("dict", schema.schema())
8282
else:
83-
schema_dict = cast(dict, schema)
83+
schema_dict = cast("dict", schema)
8484
function = {
8585
"name": schema_dict["title"],
8686
"description": schema_dict["description"],

libs/langchain/langchain/chains/query_constructor/base.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -91,7 +91,7 @@ def from_components(
9191

9292
def ast_parse(raw_filter: str) -> Optional[FilterDirective]:
9393
filter_directive = cast(
94-
Optional[FilterDirective],
94+
"Optional[FilterDirective]",
9595
get_parser().parse(raw_filter),
9696
)
9797
return fix_filter_directive(
@@ -144,7 +144,7 @@ def fix_filter_directive(
144144
return None
145145
args = [
146146
cast(
147-
FilterDirective,
147+
"FilterDirective",
148148
fix_filter_directive(
149149
arg,
150150
allowed_comparators=allowed_comparators,

libs/langchain/langchain/chains/router/llm_router.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -137,7 +137,7 @@ def _call(
137137

138138
prediction = self.llm_chain.predict(callbacks=callbacks, **inputs)
139139
return cast(
140-
dict[str, Any],
140+
"dict[str, Any]",
141141
self.llm_chain.prompt.output_parser.parse(prediction),
142142
)
143143

@@ -149,7 +149,7 @@ async def _acall(
149149
_run_manager = run_manager or CallbackManagerForChainRun.get_noop_manager()
150150
callbacks = _run_manager.get_child()
151151
return cast(
152-
dict[str, Any],
152+
"dict[str, Any]",
153153
await self.llm_chain.apredict_and_parse(callbacks=callbacks, **inputs),
154154
)
155155

libs/langchain/langchain/chat_models/base.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -322,7 +322,7 @@ class GetPopulation(BaseModel):
322322

323323
if not configurable_fields:
324324
return _init_chat_model_helper(
325-
cast(str, model),
325+
cast("str", model),
326326
model_provider=model_provider,
327327
**kwargs,
328328
)
@@ -632,7 +632,7 @@ def with_config(
632632
**kwargs: Any,
633633
) -> _ConfigurableModel:
634634
"""Bind config to a Runnable, returning a new Runnable."""
635-
config = RunnableConfig(**(config or {}), **cast(RunnableConfig, kwargs))
635+
config = RunnableConfig(**(config or {}), **cast("RunnableConfig", kwargs))
636636
model_params = self._model_params(config)
637637
remaining_config = {k: v for k, v in config.items() if k != "configurable"}
638638
remaining_config["configurable"] = {
@@ -781,7 +781,7 @@ def batch_as_completed(
781781
if config is None or isinstance(config, dict) or len(config) <= 1:
782782
if isinstance(config, list):
783783
config = config[0]
784-
yield from self._model(cast(RunnableConfig, config)).batch_as_completed( # type: ignore[call-overload]
784+
yield from self._model(cast("RunnableConfig", config)).batch_as_completed( # type: ignore[call-overload]
785785
inputs,
786786
config=config,
787787
return_exceptions=return_exceptions,
@@ -811,7 +811,7 @@ async def abatch_as_completed(
811811
if isinstance(config, list):
812812
config = config[0]
813813
async for x in self._model(
814-
cast(RunnableConfig, config),
814+
cast("RunnableConfig", config),
815815
).abatch_as_completed( # type: ignore[call-overload]
816816
inputs,
817817
config=config,

libs/langchain/langchain/document_loaders/blob_loaders/schema.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,9 @@
1-
from typing import TYPE_CHECKING, Any
1+
from typing import Any
22

33
from langchain_core.document_loaders import Blob, BlobLoader
44

55
from langchain._api import create_importer
66

7-
if TYPE_CHECKING:
8-
pass
9-
107
# Create a way to dynamically look up deprecated imports.
118
# Used to consolidate logic for raising deprecation warnings and
129
# handling optional imports.

0 commit comments

Comments
 (0)