Skip to content

Update LLAMA3 70B LoRa Base Configs for GB300, GB200 and H100#2265

Open
rhmukundan wants to merge 2 commits intomainfrom
rmukundan/llama3_lora_baseconfigs_change
Open

Update LLAMA3 70B LoRa Base Configs for GB300, GB200 and H100#2265
rhmukundan wants to merge 2 commits intomainfrom
rmukundan/llama3_lora_baseconfigs_change

Conversation

@rhmukundan
Copy link
Contributor

@rhmukundan rhmukundan commented Feb 6, 2026

Summary by CodeRabbit

  • Chores
    • Updated Llama 3 70B LoRA workload configurations for GB300, GB200, and H100 hardware platforms
    • Adjusted parallelism parameters (tensor, pipeline, and context) across model variants
    • Modified recomputation settings for H100 configurations

@rhmukundan rhmukundan self-assigned this Feb 6, 2026
@copy-pr-bot
Copy link

copy-pr-bot bot commented Feb 6, 2026

This pull request requires additional validation before any workflows can run on NVIDIA's runners.

Pull request vetters can view their responsibilities here.

Contributors can view more details about this message here.

@coderabbitai
Copy link
Contributor

coderabbitai bot commented Feb 6, 2026

📝 Walkthrough

Walkthrough

This change modifies Llama3 70B LoRA configuration settings for GB300, GB200, and H100 hardware platforms. Updates include adjusting tensor/pipeline/context parallelism dimensions and recomputation parameters across base configs and their variants through replace() calls.

Changes

Cohort / File(s) Summary
Llama3 LoRA Performance Config Updates
scripts/performance/configs/llama/llama3_workload_base_configs.py
Modified parallelism dimensions (tensor_model_parallel_size, pipeline_model_parallel_size, context_parallel_size, virtual_pipeline_model_parallel_size) for GB300, GB200, and H100 base configs. Updated public aliases to derive from refactored bases using replace() calls. Adjusted recompute_num_layers for H100_BF16_V1 from 2 to 1.

Estimated code review effort

🎯 3 (Moderate) | ⏱️ ~20 minutes

Suggested labels

Run CICD, performance

Suggested reviewers

  • malay-nagda
  • erhoo82
  • thomasdhc
🚥 Pre-merge checks | ✅ 3 | ❌ 1
❌ Failed checks (1 warning)
Check name Status Explanation Resolution
Test Results For Major Changes ⚠️ Warning PR contains major configuration changes for LLAMA3 70B LoRA model parallelism across multiple platforms but lacks comprehensive testing information, performance metrics, and validation details in the PR description. Add test results validating configurations, before-and-after performance metrics for each platform, batch sizes used, and specify which configurations have been validated.
✅ Passed checks (3 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title accurately and specifically describes the main change: updating LLAMA3 70B LoRa base configurations for three hardware platforms (GB300, GB200, H100).
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing touches
  • 📝 Generate docstrings
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Post copyable unit tests in a comment
  • Commit unit tests in branch rmukundan/llama3_lora_baseconfigs_change

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
@rhmukundan rhmukundan force-pushed the rmukundan/llama3_lora_baseconfigs_change branch from 6b69798 to 31d7bcf Compare February 6, 2026 20:58
@rhmukundan
Copy link
Contributor Author

/ok to test 16aa00e

@rhmukundan rhmukundan enabled auto-merge (squash) February 6, 2026 20:58
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant