Replies: 1 comment
-
DPO is good enough for me, hope someone MLX ninja can make it happen |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello!
I was reading this article here on finetuning llama3 on ORPO (https://mlabonne.github.io/blog/posts/2024-04-19_Fine_tune_Llama_3_with_ORPO.html)
I was wondering how to do the same finetuning process with mlx.
The article is using OrpoTrainer from https://github.com/huggingface/trl
is there a similar functionality in mlx, if not would any pointers on implementing myself in mlx?
Beta Was this translation helpful? Give feedback.
All reactions