File tree Expand file tree Collapse file tree 1 file changed +2
-3
lines changed Expand file tree Collapse file tree 1 file changed +2
-3
lines changed Original file line number Diff line number Diff line change 3
3
import torch
4
4
import torch .nn as nn
5
5
import torchair
6
- import vllm .envs as envs_vllm
7
6
from torchair import patch_for_hcom
8
7
from vllm .attention .layer import Attention
9
8
from vllm .config import (VllmConfig , get_layers_from_vllm_config ,
@@ -607,7 +606,7 @@ def _get_torchair_lazy_compiled_model(self, batch_size: int):
607
606
self .torchair_compiled_model = torch .compile (
608
607
self .model ,
609
608
dynamic = not get_ascend_config ().use_sfa ,
610
- fullgraph = envs_vllm . VLLM_TEST_DYNAMO_FULLGRAPH_CAPTURE ,
609
+ fullgraph = True ,
611
610
backend = npu_backend )
612
611
return self .torchair_compiled_model
613
612
else :
@@ -630,7 +629,7 @@ def _get_torchair_lazy_compiled_model(self, batch_size: int):
630
629
batch_size ] = torchair .inference .cache_compile (
631
630
self .model .__dict__ [forward_proxy_name ],
632
631
dynamic = not get_ascend_config ().use_sfa ,
633
- fullgraph = envs_vllm . VLLM_TEST_DYNAMO_FULLGRAPH_CAPTURE ,
632
+ fullgraph = True ,
634
633
cache_dir = TORCHAIR_CACHE_DIR ,
635
634
config = config ,
636
635
ge_cache = False )
You can’t perform that action at this time.
0 commit comments