diff --git a/backend/Makefile b/backend/Makefile index 72878508..3033c70f 100644 --- a/backend/Makefile +++ b/backend/Makefile @@ -1,7 +1,8 @@ init: @python3 -m venv .venv && \ . .venv/bin/activate && \ - pip install -r requirements.txt + pip install -r requirements.txt && \ + pip install -e . init-dev: init @. .venv/bin/activate && \ @@ -9,9 +10,9 @@ init-dev: init format: @. .venv/bin/activate && \ - ruff format && \ - mypy . + ruff format check: @. .venv/bin/activate && \ - ruff check --fix + mypy . && \ + ruff check diff --git a/backend/__init__.py b/backend/__init__.py deleted file mode 100644 index e69de29b..00000000 diff --git a/backend/pyproject.toml b/backend/pyproject.toml index 50ad191d..9e26a877 100644 --- a/backend/pyproject.toml +++ b/backend/pyproject.toml @@ -20,6 +20,9 @@ classifiers = [ dependencies = { file = ["requirements.txt"] } optional-dependencies = { test = { file = ["requirements-test.txt"] } } +[tool.setuptools.packages.find] +where = ["./"] + [tool.mypy] python_version = "3.12" warn_unused_configs = true diff --git a/backend/src/agents/retriever_graph.py b/backend/src/agents/retriever_graph.py index b22256d4..d90f7c77 100644 --- a/backend/src/agents/retriever_graph.py +++ b/backend/src/agents/retriever_graph.py @@ -104,7 +104,7 @@ def __init__( self.tool_descriptions = "" for tool in self.tools: - text_desc = render_text_description([tool]) # type: ignore + text_desc = render_text_description([tool]) text_desc.replace("(query: str) -> Tuple[str, list[str], list[str]]", " ") self.tool_descriptions += text_desc + "\n\n" @@ -196,7 +196,7 @@ def route(self, state: AgentState) -> list[str]: return ["retrieve_general"] if self.inbuit_tool_calling: - tool_names = [tool["name"] for tool in tools if "name" in tool] # type: ignore + tool_names = [tool["name"] for tool in tools if "name" in tool] # type: ignore return tool_names else: return tools @@ -204,12 +204,12 @@ def route(self, state: AgentState) -> list[str]: def initialize(self) -> None: workflow = StateGraph(AgentState) - commands = ToolNode(self.retriever_tools.retrieve_cmds) # type: ignore - install = ToolNode(self.retriever_tools.retrieve_install) # type: ignore - general = ToolNode(self.retriever_tools.retrieve_general) # type: ignore - klayout_docs = ToolNode(self.retriever_tools.retrieve_klayout_docs) # type: ignore - errinfo = ToolNode(self.retriever_tools.retrieve_errinfo) # type: ignore - yosys_rtdocs = ToolNode(self.retriever_tools.retrieve_yosys_rtdocs) # type: ignore + commands = ToolNode(self.retriever_tools.retrieve_cmds) + install = ToolNode(self.retriever_tools.retrieve_install) + general = ToolNode(self.retriever_tools.retrieve_general) + klayout_docs = ToolNode(self.retriever_tools.retrieve_klayout_docs) + errinfo = ToolNode(self.retriever_tools.retrieve_errinfo) + yosys_rtdocs = ToolNode(self.retriever_tools.retrieve_yosys_rtdocs) workflow.add_node("agent", self.agent) workflow.add_node("generate", self.generate) diff --git a/backend/src/api/routers/chains.py b/backend/src/api/routers/chains.py index fda070ad..b1bbd089 100644 --- a/backend/src/api/routers/chains.py +++ b/backend/src/api/routers/chains.py @@ -55,7 +55,7 @@ elif os.getenv("LLM_MODEL") == "gemini": if os.getenv("GOOGLE_GEMINI") == "1_pro": - llm = ChatGoogleGenerativeAI(model="gemini-pro", temperature=llm_temp) # type: ignore + llm = ChatGoogleGenerativeAI(model="gemini-pro", temperature=llm_temp) elif os.getenv("GOOGLE_GEMINI") == "1.5_flash": llm = ChatVertexAI(model_name="gemini-1.5-flash", temperature=llm_temp) elif os.getenv("GOOGLE_GEMINI") == "1.5_pro": diff --git a/backend/src/api/routers/graphs.py b/backend/src/api/routers/graphs.py index 93e8b13f..e7750dfd 100644 --- a/backend/src/api/routers/graphs.py +++ b/backend/src/api/routers/graphs.py @@ -58,7 +58,7 @@ elif os.getenv("LLM_MODEL") == "gemini": if os.getenv("GOOGLE_GEMINI") == "1_pro": - llm = ChatGoogleGenerativeAI(model="gemini-pro", temperature=llm_temp) # type: ignore + llm = ChatGoogleGenerativeAI(model="gemini-pro", temperature=llm_temp) elif os.getenv("GOOGLE_GEMINI") == "1.5_flash": llm = ChatVertexAI(model_name="gemini-1.5-flash", temperature=llm_temp) elif os.getenv("GOOGLE_GEMINI") == "1.5_pro": diff --git a/backend/src/tools/format_docs.py b/backend/src/tools/format_docs.py index a2376c41..fea704fe 100644 --- a/backend/src/tools/format_docs.py +++ b/backend/src/tools/format_docs.py @@ -23,7 +23,7 @@ def format_docs(docs: list[Document]) -> tuple[str, list[str], list[str]]: if "url" in doc.metadata: doc_urls.append(doc.metadata["url"]) - + doc_output = "\n\n -------------------------- \n\n".join(doc_texts) return doc_output, doc_srcs, doc_urls diff --git a/evaluation/Makefile b/evaluation/Makefile index d0dd015f..f624b0ec 100644 --- a/evaluation/Makefile +++ b/evaluation/Makefile @@ -12,12 +12,12 @@ init-dev: init format: @. .venv/bin/activate && \ - ruff format && \ - mypy . + ruff format check: @. .venv/bin/activate && \ - ruff check --fix + mypy . && \ + ruff check clean: @rm -f llm_tests_output.txt diff --git a/frontend/Makefile b/frontend/Makefile index 72878508..c4423b64 100644 --- a/frontend/Makefile +++ b/frontend/Makefile @@ -9,9 +9,9 @@ init-dev: init format: @. .venv/bin/activate && \ - ruff format && \ - mypy . + ruff format check: @. .venv/bin/activate && \ - ruff check --fix + mypy . && \ + ruff check