From 3c58e3cf20483548740239fd1ce8825e01533ec5 Mon Sep 17 00:00:00 2001 From: Jack Zhang <32371937+jackzhxng@users.noreply.github.com> Date: Tue, 27 May 2025 16:41:47 -0700 Subject: [PATCH] fix: add missing LlmConfig import and parameter in export_llama_lib.py [ghstack-poisoned] --- examples/models/llama/export_llama_lib.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/examples/models/llama/export_llama_lib.py b/examples/models/llama/export_llama_lib.py index 28a2ae3debb..c1dd27dc390 100644 --- a/examples/models/llama/export_llama_lib.py +++ b/examples/models/llama/export_llama_lib.py @@ -29,6 +29,7 @@ from executorch.devtools.etrecord import generate_etrecord as generate_etrecord_func +from executorch.examples.models.llama.config.llm_config import LlmConfig from executorch.examples.models.llama.config.llm_config_utils import ( convert_args_to_llm_config, ) @@ -1198,6 +1199,7 @@ def _load_llama_model( use_qnn: bool = False, export_only: bool = False, args, + llm_config: Optional[LlmConfig] = None, ) -> "LLMEdgeManager": """ A helper util that builds a Llama2 model. It returns a LLMEdgeManager that