Skip to content

Commit d8180ef

Browse files
committed
Merge branch 'ko3n1g/chore/cherry-pick-3529' into 'core_r0.13.0'
ADLR/megatron-lm!3526 - Skip fused rope check if te version < 1.4.0 See merge request ADLR/megatron-lm!3528
2 parents 99c1fce + c43e1f1 commit d8180ef

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

tests/unit_tests/transformer/test_attention.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,7 @@
1313
from megatron.core.transformer.attention import SelfAttention
1414
from megatron.core.transformer.enums import AttnMaskType
1515
from megatron.core.transformer.transformer_config import TransformerConfig
16+
from megatron.core.utils import is_te_min_version
1617
from tests.unit_tests.test_utilities import Utils
1718

1819

@@ -74,6 +75,7 @@ def test_gpu_forward(self):
7475
assert output.shape[2] == config.hidden_size
7576
assert bias.shape[0] == config.hidden_size
7677

78+
@pytest.mark.skipif(not is_te_min_version("1.4.0"), reason="Fused RoPE requires TE >= 1.4.0")
7779
def test_fused_rope_gpu_forward(self):
7880
self.parallel_attention.config.apply_rope_fusion = True
7981
config = self.parallel_attention.config

0 commit comments

Comments
 (0)