-
Notifications
You must be signed in to change notification settings - Fork 209
Open
Labels
community-requestenhancementNew feature or requestNew feature or requestneeds-follow-upIssue needs follow-upIssue needs follow-up
Description
Is your feature request related to a problem? Please describe.
When attempting to train (SFT) a base version of Nemotron Nano v3 30B-A3B, I've encountered this error:
AssertionError: FP8 block scaled GEMM requires Hopper and CUDA >= 12.9.
I assume that might be because nano-v3 branch does not include that specific commit that addresses this issue (as mentioned in the docs). Any chance it could be added there?
Describe the solution you'd like
Adding that commit into nano-v3 branch.
Describe alternatives you've considered
- I've tried using main branch directly, but it looks like the pinned Megatron-LM version is not compatible with Nemotron Nano v3 - it always fails on this line:
num_query_groups (2) must be a multiple of tensor_model_parallel_size (4).
Current latest version of Megatron-LM seems to use a different check, and it should work with nemotron then, but I haven't had time to test it.
coderabbitai
Metadata
Metadata
Assignees
Labels
community-requestenhancementNew feature or requestNew feature or requestneeds-follow-upIssue needs follow-upIssue needs follow-up