Skip to content

Commit 1b3d4ff

Browse files
author
wanghao7
committed
back move ffn.
1 parent 428f2b5 commit 1b3d4ff

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

lightllm/models/deepseek2/layer_infer/transformer_layer_infer.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -759,10 +759,6 @@ def overlap_tpsp_context_forward(
759759

760760
_0_overlap_event = Buffer.capture()
761761

762-
# 0 shared expert
763-
if self.n_shared_experts is not None:
764-
_0_shared_output = LlamaTransformerLayerInfer._ffn(self, _0_input1, infer_state, layer_weight)
765-
766762
# 1 attention
767763
_1_input1 = self._att_norm(input_embdings1, infer_state1, layer_weight)
768764
_1_cache_kv = self._pre_cache_kv(infer_state1, layer_weight)
@@ -801,6 +797,10 @@ def overlap_tpsp_context_forward(
801797

802798
_1_overlap_event = Buffer.capture()
803799

800+
# 0 shared expert
801+
if self.n_shared_experts is not None:
802+
_0_shared_output = LlamaTransformerLayerInfer._ffn(self, _0_input1, infer_state, layer_weight)
803+
804804
# 1 shared expert
805805
if self.n_shared_experts is not None:
806806
_1_shared_output = LlamaTransformerLayerInfer._ffn(self, _1_input1, infer_state1, layer_weight)

0 commit comments

Comments
 (0)