Skip to content

Commit 1a338e1

Browse files
authored
[None][chore] use cached vila model (#8788)
Signed-off-by: Hui Gao <[email protected]>
1 parent 1d4a186 commit 1a338e1

File tree

1 file changed

+3
-2
lines changed

1 file changed

+3
-2
lines changed

tests/unittest/_torch/modeling/test_modeling_vila.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,7 @@
44

55
import torch
66
from parameterized import parameterized
7+
from utils.llm_data import llm_models_root
78

89
import tensorrt_llm
910
from tensorrt_llm._torch.attention_backend.utils import get_attention_backend
@@ -16,7 +17,7 @@
1617
from tensorrt_llm.mapping import Mapping
1718

1819
VILA_1_5_3B_CONFIG = {
19-
"_name_or_path": "Efficient-Large-Model/VILA1.5-3b",
20+
"_name_or_path": f"{llm_models_root()}/vila/VILA1.5-3b",
2021
"architectures": ["LlavaLlamaModel"],
2122
"drop_path_rate": 0.0,
2223
"hidden_size": 2560,
@@ -263,7 +264,7 @@
263264

264265
NVILA_8B_CONFIG = {
265266
"_attn_implementation_autoset": True,
266-
"_name_or_path": "Efficient-Large-Model/NVILA-8B",
267+
"_name_or_path": f"{llm_models_root()}/vila/NVILA-8B",
267268
"architectures": ["LlavaLlamaModel"],
268269
"chat_template": None,
269270
"drop_path_rate": 0.0,

0 commit comments

Comments
 (0)