Replies: 2 comments 1 reply
-
i dont have any exp with finetune but here my thoughts suppose the transcript is in the section "Load WhisperTokenizer" of the finetune guide, the transcript must be tokenized into something like
after digging into source files
so i guess u have to find a way with the tokenizer to make this |
Beta Was this translation helpful? Give feedback.
-
@phineas-pta Thanks for your response. What should be the |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi All,
How are you?
I’m trying to finetune Whisper by resuming its pre-training task and adding initial prompts as part of the model’s forward pass. I saw this amazing tutorial, however, it does not contain a section about using prompts as part of the fine-tuning dataset.
My motivation is we witness that Whisper is not acting as expected when transcribing with prompts. Sometimes the output is blank text and on other occasions the output text contains reoccurrence. We want to solve such behaviors by fine-tuning Whisper with prompts.
Any help will be appreciated!
Beta Was this translation helpful? Give feedback.
All reactions