flash attention 2, batch size, etc.? #2126
Replies: 2 comments 1 reply
-
https://github.com/Vaibhavs10/insanely-fast-whisper the author working at huggingface so he should be very well informed |
Beta Was this translation helpful? Give feedback.
-
i checked that. i am not entirely sure what technique can be used to speed up whisper in general. https://github.com/igorcosta/insanely-fast-whisper-cli simply says this: import torch pipe = pipeline("automatic-speech-recognition", outputs = pipe("<FILE_NAME>", outputs["text"] i was looking for examples that call whisper directly versus going through pipeline. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
does anyone have any experience on using flash attention 2 and different batch size or other parameters like that to make whisper as fast as possible?
carl
Beta Was this translation helpful? Give feedback.
All reactions