Replies: 4 comments 1 reply
-
Its a good idea. I am testing them and also looking at adding support for different embeddings. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Just saw the new accelerate update. I think it will now be possible. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Support for GPTQ models would be awesome |
Beta Was this translation helpful? Give feedback.
0 replies
-
You can use GPTQ now. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
7B HF models take more than 12Gb of memory, so is there a way to use GPTQ ones?
Beta Was this translation helpful? Give feedback.
All reactions