Skip to content

Commit 363165a

Browse files
shanmugamr1992Shanmugam Ramasamy
andauthored
fix: Set use_flashinfer_fused_rope to False (#1636)
Signed-off-by: Shanmugam Ramasamy <[email protected]> Co-authored-by: Shanmugam Ramasamy <[email protected]>
1 parent 995efaa commit 363165a

File tree

2 files changed

+1
-4
lines changed

2 files changed

+1
-4
lines changed

nemo_rl/models/policy/workers/megatron_policy_worker.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1850,7 +1850,7 @@ def generate(
18501850
"tensor_model_parallel_size"
18511851
],
18521852
use_cuda_graphs_for_non_decode_steps=use_cuda_graphs_for_non_decode_steps,
1853-
use_flashinfer_fused_rope=None,
1853+
use_flashinfer_fused_rope=False,
18541854
unified_memory_level=unified_memory_level,
18551855
max_tokens_override=max_tokens,
18561856
)

tests/unit/models/policy/test_megatron_worker.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -547,9 +547,6 @@ def generation_setup(request, tiny_llama_model_path):
547547
cluster.shutdown()
548548

549549

550-
@pytest.mark.skip(
551-
reason="Skipping mcore generation test to unblock mcore bump. Current issue summarized in https://github.com/NVIDIA-NeMo/RL/issues/1633"
552-
)
553550
@pytest.mark.timeout(240)
554551
@pytest.mark.parametrize(
555552
"generation_setup",

0 commit comments

Comments
 (0)