File tree Expand file tree Collapse file tree 1 file changed +12
-3
lines changed
vllm/model_executor/layers/fused_moe Expand file tree Collapse file tree 1 file changed +12
-3
lines changed Original file line number Diff line number Diff line change 5
5
import torch
6
6
7
7
import vllm .model_executor .layers .fused_moe .modular_kernel as mk
8
+ from vllm .logger import init_logger
8
9
from vllm .model_executor .layers .fused_moe .topk_weight_and_reduce import (
9
10
TopKWeightAndReduceDelegate )
10
11
from vllm .model_executor .layers .fused_moe .utils import extract_required_args
11
12
from vllm .utils import has_triton_kernels
12
13
14
+ logger = init_logger (__name__ )
15
+
13
16
if has_triton_kernels ():
14
- import triton_kernels .swiglu
15
- from triton_kernels .matmul_ogs import FnSpecs , FusedActivation , matmul_ogs
16
- from triton_kernels .routing import routing
17
+ try :
18
+ import triton_kernels .swiglu
19
+ from triton_kernels .matmul_ogs import (FnSpecs , FusedActivation ,
20
+ matmul_ogs )
21
+ from triton_kernels .routing import routing
22
+ except ModuleNotFoundError :
23
+ logger .error (
24
+ "Failed to import Triton kernels. Please make sure your triton "
25
+ "version is compatible." )
17
26
18
27
if TYPE_CHECKING :
19
28
from triton_kernels .matmul_ogs import PrecisionConfig
You can’t perform that action at this time.
0 commit comments