-
Notifications
You must be signed in to change notification settings - Fork 31k
fix llama issue #39942
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
fix llama issue #39942
Conversation
Signed-off-by: Yao, Matrix <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nice
|
@bot /style |
|
Style fix runs successfully without any file modified. |
|
You need to propagate the changes to other repos also using make fix-copies. Btw since we are touching llama model, can you share more details to the issue you had ? |
@SunMarc when I enable this multi-adapter lora inference in peft huggingface/peft#2711 on a 2-card env, the issue raised, it's because this line https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py#L66 is not a Module, so not hooked, same issue as before we discussed. |
Signed-off-by: Yao, Matrix <[email protected]>
Head branch was pushed to by a user without write access
it is a nn.Module no ? |
I mean |
|
seems i lost the env to reproduce it, let's put it on hold here, and once i reproduced it, i'll put the log back |
|
[For maintainers] Suggested jobs to run (before merge) run-slow: arcee, aria, bitnet, cohere, cohere2, deepseek_v2, deepseek_v3, diffllama, doge, dots1, ernie4_5, gemma, gemma2, glm, glm4, glm4_moe |
similar like this PR #39646, fix the same issue found while enabling llama lora finetuning across multiple card.
@SunMarc , pls help review, thx very much.