-
I have a small Python script that's translating a batch of WAV files, and recently made some improvements to speed it up. The following has remained unchanged though: model = whisper.load_model('medium.en') # detects the applicable device automatically
text = model.transcribe(
audio=wav,
fp16=has_cuda,
language='en',
condition_on_previous_text=False,
verbose=True
)['text'].strip() I'll get warnings about the SHA256 checksum not matching and it downloads the model over and over again, until eventually crashing with this error:
I'm using CUDA 12.1, CUDANN 12.x, and Whisper v20231117. My old program used CUDA 11 so maybe that has something to do with it? Also if there are better parameters I could give the model or there's a better way to load the model and transcribe text please let me know! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 7 replies
-
Suggest a quick test with a single file using the command line and Some relevant past discussions here: |
Beta Was this translation helpful? Give feedback.
I changed the
download_root
to be a directory one level above the CWD, so it shouldn't be a problem. I've tried doing this:And it's still happening. I've been able to fix it, but in order to do so I downloaded the files manually, then assigned the proper path as the
name
parameter. If you want to download the model one time, first callload_model
effectively like a void statement. Then load…