how to solve low inference speed of my model #1815
Unanswered
Topology2333
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Inquiry about Inference Time and Custom Tokenizer in Whisper-Small
Dear Whisper community,
I hope this message finds you well. As a beginner in the field, I am currently working on leveraging Whisper's robust alignment capabilities to fine-tune the model for low-resource languages. My approach involves training a language-to-IPA model to address specific phonetic requirements. However, I'm encountering challenges, particularly with the support for the fricative sound ɬ.
Issue Description:
After fine-tuning the model with a custom tokenizer (IPA), I noticed a significant increase in the time required for predicting a single sentence—approximately 7-8 minutes. Strangely, when I used the default Whisper tokenizer, the predictions were much faster, taking only a few seconds. However, the results were in multiple languages, which is unexpected.
Details:
Questions:
Additionally, I'm in the process of building a language-to-IPA model due to specific phonetic requirements. This is the context that led to the aforementioned issues. Given Whisper's strengths in alignment, I believe it's a powerful tool for my intended workflow: speech-to-IPA-to-low-resource language text. Any advice or insights, especially considering my beginner status, on optimizing the model for handling phonetic nuances, particularly the fricative sound ɬ, would be immensely valuable.
Thank you for your time and assistance.
Best regards,
Topology2333
Beta Was this translation helpful? Give feedback.
All reactions