We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent bc16bc3 commit 8b01d3dCopy full SHA for 8b01d3d
test/benchmark/kernel/benchmark_fused_moe_triton.py
@@ -4,7 +4,6 @@
4
5
import torch
6
import triton
7
-import vllm
8
from transformers import AutoConfig
9
from lightllm.common.fused_moe.topk_select import select_experts
10
from lightllm.common.fused_moe.grouped_fused_moe import fused_experts_impl
@@ -173,8 +172,7 @@ def fused_moe_sglang_api(
173
172
a2_scale=None,
174
block_shape=None,
175
):
176
- from sglang.srt.layers.moe.moe_runner import MoeRunnerConfig
177
- from sglang.srt.layers.moe.topk import TopK, TopKConfig, select_experts
+ from sglang.srt.layers.moe.topk import TopKConfig, select_experts
178
from sglang.srt.layers.moe.fused_moe_triton.fused_moe import (
179
fused_moe as fused_moe_sglang,
180
)
0 commit comments