Replies: 1 comment 1 reply
-
For this section : lora_target_modules:
I see you have two v_projs above, is that just a typo? One should be k_proj? Tom |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm attempting to fine-tune Llama-3-70B on my personal data with qlora r: 512 and qlora_alpha: 256. It's currently using 8 H100 80GB SXM5, and I'm wondering if this is normal to consume this much compute for qlora or if I've made a mistake somewhere. Is it possible to optimize the pipeline with the same sequence length and lora r & alpha?
Here is my full config
Beta Was this translation helpful? Give feedback.
All reactions