Skip to content

Commit ab24acf

Browse files
committed
fix functional tests
Signed-off-by: yaoyu-33 <yaoyu.094@gmail.com>
1 parent b753ab8 commit ab24acf

File tree

4 files changed

+12
-180
lines changed

4 files changed

+12
-180
lines changed

src/megatron/bridge/models/glm/glm45_bridge.py

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,10 @@
1313
# limitations under the License.
1414

1515
import logging
16+
from functools import partial
1617

1718
import torch
19+
from megatron.core.models.gpt.gpt_layer_specs import get_gpt_decoder_block_spec
1820
from megatron.core.models.gpt.gpt_model import GPTModel
1921
from transformers import Glm4MoeForCausalLM
2022

@@ -29,6 +31,14 @@
2931
from megatron.bridge.models.hf_pretrained.causal_lm import PreTrainedCausalLM
3032

3133

34+
try:
35+
import transformer_engine # noqa: F401
36+
37+
HAVE_TE = True
38+
except (ImportError, ModuleNotFoundError):
39+
HAVE_TE = False
40+
41+
3242
logger = logging.getLogger(__name__)
3343

3444

@@ -51,6 +61,8 @@ def provider_bridge(self, hf_pretrained: PreTrainedCausalLM) -> GPTModelProvider
5161
provider = super().provider_bridge(hf_pretrained)
5262
hf_config = hf_pretrained.config
5363

64+
# Use decoder block spec to properly handle moe_layer_freq (mixed dense/MoE layers)
65+
provider.transformer_layer_spec = partial(get_gpt_decoder_block_spec, use_transformer_engine=HAVE_TE)
5466
provider.normalization = "RMSNorm"
5567
provider.gated_linear_unit = True
5668
provider.position_embedding_type = "rope"

src/megatron/bridge/models/kimi/kimi_bridge.py

Lines changed: 0 additions & 80 deletions
This file was deleted.

tests/functional_tests/models/deepseek/test_deepseek_provider_mapping.py

Lines changed: 0 additions & 52 deletions
This file was deleted.

tests/functional_tests/models/qwen/test_qwen3_moe_provider.py

Lines changed: 0 additions & 48 deletions
This file was deleted.

0 commit comments

Comments
 (0)