-
Notifications
You must be signed in to change notification settings - Fork 9
Add initial LoRA finetuning support; vulkan OUT_PROD; vulkan cross-entropy-backward #5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: temp-finetuning
Are you sure you want to change the base?
Conversation
Signed-off-by: vineet <[email protected]>
Signed-off-by: vineet <[email protected]>
Signed-off-by: vineet <[email protected]>
Signed-off-by: vineet <[email protected]>
Signed-off-by: vineet <[email protected]>
Signed-off-by: vineet <[email protected]>
Signed-off-by: vineet <[email protected]>
Signed-off-by: vineet <[email protected]>
…lation Signed-off-by: vineet <[email protected]>
Steps to test llama.cpp inference on Android:
make sure to checkout the
|
For testing I'll reference the updated README: https://github.com/tetherto/qvac-ext-lib-llama.cpp/blob/bc7dd9f9288222394da37eac3d7adf71d409ad83/examples/training/README.md#using-trained-adapters |
command we used for testing |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Changes LGTM in general, just some small comments/nits overall, feel free to ignore the nitpicks :).
ggml/src/ggml-vulkan/ggml-vulkan.cpp
Outdated
case GGML_OP_ADD: | ||
case GGML_OP_SUB: | ||
case GGML_OP_MUL: | ||
case GGML_OP_DIV: | ||
return (op->src[0]->type == GGML_TYPE_F32 || op->src[0]->type == GGML_TYPE_F16) && | ||
return (op->src[0]->type == GGML_TYPE_F32 || op->src[0]->type == GGML_TYPE_F16) && |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: spurious change?
Looks like there are some CI failures also related to these changes - see https://github.com/tetherto/qvac-ext-lib-llama.cpp/actions/runs/17076253696/job/48418341198?pr=5 for example:
|
@JamieBohannaWebDev On our side, I think for a test we ran it with 10-20% of the data. Took much less time. Also we have checkpoint saving resuming integration going on which would allows us to train in bursts by saving a checkpoint and resuming later from the same point. |
@JamieBohannaWebDev Did we test the output with some prompts after the completion of the fine-tuned model? (before vs. after) |
This fixes the vkDeviceLostError on Mali
25dfd75
to
53f2e8e
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Changes LGTM, looks like I can't merge it though so will defer to someone else with perms to do it.
The current CI failures seem unrelated to the changes here, both are failing with:
|
cb9e955
to
ca99485
Compare
ca99485
to
191dd7e
Compare
The PR adds: