Skip to content

Commit 984ece0

Browse files
committed
Merge branch 'jz/move-llm-config-2' into jz/move-modeling
2 parents ab34e18 + 9a8282d commit 984ece0

File tree

2 files changed

+4
-4
lines changed

2 files changed

+4
-4
lines changed

examples/models/llama/config/test_llm_config.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@
88

99
import unittest
1010

11-
from executorch.examples.models.llama.config.llm_config import (
11+
from executorch.extension.llm.export.config.llm_config import (
1212
BackendConfig,
1313
BaseConfig,
1414
CoreMLComputeUnit,

examples/models/llama/eval_llama_lib.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -175,7 +175,7 @@ def gen_eval_wrapper(
175175
"""
176176
# If llm_config is not provided, convert args to llm_config
177177
if llm_config is None:
178-
from executorch.examples.models.llama.config.llm_config import LlmConfig
178+
from executorch.extension.llm.export.config.llm_config import LlmConfig
179179

180180
llm_config = LlmConfig.from_args(args)
181181

@@ -306,7 +306,7 @@ def eval_llama(
306306
args: argparse.ArgumentParser,
307307
) -> None:
308308
# Convert args to LlmConfig
309-
from executorch.examples.models.llama.config.llm_config import LlmConfig
309+
from executorch.extension.llm.export.config.llm_config import LlmConfig
310310

311311
llm_config = LlmConfig.from_args(args)
312312

@@ -340,7 +340,7 @@ def eval_llama_with_attention_sink(model_name: str, args: argparse.ArgumentParse
340340
This is mostly copied from https://github.com/mit-han-lab/streaming-llm/blob/main/examples/eval_long_ppl.py
341341
"""
342342
# Convert args to LlmConfig
343-
from executorch.examples.models.llama.config.llm_config import LlmConfig
343+
from executorch.extension.llm.export.config.llm_config import LlmConfig
344344

345345
llm_config = LlmConfig.from_args(args)
346346

0 commit comments

Comments
 (0)