Happy to announce easy_whisper #1971
gyllila
started this conversation in
Show and tell
Replies: 4 comments 26 replies
-
What makes it faster? |
Beta Was this translation helpful? Give feedback.
6 replies
-
More correctly: by splitting up in sentences, I reduced the context length to within the sentence, which makes it faster while maintaining quality. I can assure you that already with base.en you can transcribe an English audiobook almost free of errors. I‘m really impressed by the good quality of OpenAI Whisper.Von meinem iPhone gesendetAm 22.01.2024 um 18:42 schrieb Purfview ***@***.***>:
Quick looked at your repo, I don't see what would make it faster, with splitting up files you lose context, that's it.
Use --condition_on_previous_text=False to get similar effect.
—Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you authored the thread.Message ID: ***@***.***>
|
Beta Was this translation helpful? Give feedback.
5 replies
-
It doesn't work for me: Empty txt files. |
Beta Was this translation helpful? Give feedback.
14 replies
-
I am getting this error,i have tried reinstalling the libraries,can you help me? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I’ve written a Python script to make transcription/translation with Whisper easier for me. Because I find it very useful, I also made a Python library out of it. Feel free to use it. :)
The main features are:
Here the GitHub link:
https://github.com/gyllila/easy_whisper
PS. „faster processing“ is possibly inaccurate. Out of curiosity, I just tested transcribing a 13:23 mp3 file with plain vanilla OpenAI Whisper („short“ in easy_whisper), it only took 200s to finish! I‘m really impressed. To be fair, I used the base.en model instead of a large one, but the transcript is nonetheless of high quality, almost free of error, and that on my more than 5y old laptop! These 200s don’t include time for loading the model, because in the GUI, once a model is loaded, it won’t be reloaded until another model is selected.
PSS. The next update, version 1.1.0, is even faster and only took 197s for the above audio. The higher speed results from lower-level access to Whisper, not splitting in sentences, which I thus removed. I‘ll then add the option of API access for whose who prefer using OpenAI‘s online transcription/translation service.
Beta Was this translation helpful? Give feedback.
All reactions