How to load a model in bf16 format #1540
Closed
georgel2004
started this conversation in
General
Replies: 1 comment
-
turbomind doesn't hard code lmdeploy/lmdeploy/turbomind/turbomind.py Line 271 in 7c4e75b |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
In turbomind.py the data_type is hard coded to fp16.
If I change the value to bf16 I could see increase in performance speed.
How could I load a model in bf16 format using the pipeline?
Beta Was this translation helpful? Give feedback.
All reactions