Skip to content

Commit a82ab33

Browse files
committed
Update on "[rl] Add CI for numerics test against vllm native inference"
Test cases: 1. Integration tests: - single GPU, no compile + cudagraph - multiple GPU (with TP), no compile + cudagraph - multiple GPU, with compile + cudagraph - This test runs on A10G (default CI GPU type) 3. Numerics parity test: vLLM native model vs vLLM + TorchTitan wrapper. - test_weights_match: max_diff <= 1e-5 (exact weight loading) - test_attention_module: atol=1e-5 (TP=1) - test_end_to_end_logits: atol=1e-3 (TP=1) - We would need to run numerics test for only TP=1. This is because we are assuming both torchtitan and vllm will make sure their multi-GPU implementation is on par with single GPU. And we can add more numerics test under parallelism if needed. - This test runs on H100, and runs FA3 kernel for attention. [ghstack-poisoned]
2 parents 72978cf + 8fbb949 commit a82ab33

File tree

1 file changed

+4
-2
lines changed

1 file changed

+4
-2
lines changed

torchtitan/experiments/rl/tests/integration_tests.py

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,8 @@ def build_rl_test_list() -> list[OverrideDefinitions]:
3636
"--config rl_grpo_qwen3_0_6b",
3737
"--trainer.parallelism.tensor_parallel_degree 2",
3838
"--generator.parallelism.tensor_parallel_degree 2",
39-
"--generator.max_model_len 2048",
39+
"--generator.num_samples_per_prompt 2",
40+
"--no_batch_invariant_mode",
4041
"--generator.compile.backend none",
4142
"--generator.compile.cudagraph_mode none",
4243
],
@@ -52,7 +53,8 @@ def build_rl_test_list() -> list[OverrideDefinitions]:
5253
"--config rl_grpo_qwen3_0_6b",
5354
"--trainer.parallelism.tensor_parallel_degree 2",
5455
"--generator.parallelism.tensor_parallel_degree 2",
55-
"--generator.max_model_len 2048",
56+
"--generator.num_samples_per_prompt 2",
57+
"--no_batch_invariant_mode",
5658
],
5759
],
5860
"RL GRPO TP=2 compile",

0 commit comments

Comments
 (0)