Why is "FP16 is not supported on CPU; using FP32 instead" on Ampere A1? #978
-
I'm running whisper on a free instance on Oracle Cloud. I'm only performing basic usage and do not yet have a sound understanding
According to https://learncloudnative.com/blog/2022-03-16-running_ai_on_ampere_instance FP16 is actually supported on Amphere A1. How do I forcefully enable FP16? Is it as simple as commenting away the check at https://github.com/openai/whisper/blob/main/whisper/transcribe.py#L79? Here is the CPU information.
|
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 3 replies
-
So, I tried commenting out the check for CPU and got:
Looks like pytorch is not implementing this. |
Beta Was this translation helpful? Give feedback.
-
Hi, You may find this implementation of whisper handy: https://github.com/ggerganov/whisper.cpp
Also you could find Ampere optimized Torch interesting: https://amperecomputing.com/solutions/ampere-ai |
Beta Was this translation helpful? Give feedback.
-
result = whisper.decode(model, mel, options, fp16=False) |
Beta Was this translation helpful? Give feedback.
Hi,
You may find this implementation of whisper handy: https://github.com/ggerganov/whisper.cpp
fp16 can be enabled there on Ampere A1 systems by modifying Makefile:
Also you could find Ampere optimized Torch interesting: https://amperecomputing.com/solutions/ampere-ai