The latest optimum-cli command to download and convert whisper model to OV as given in the README does not create the decoder model with KV cache.
optimum-cli export openvino --trust-remote-code --model openai/whisper-base whisper-base
Since NPU inference pipeline needs the KV cache decoder model, running the sample throws out this error.

The following change to the exporting command can solve this issue as it generates the KV cache decoder also
optimum-cli export openvino --model openai/whisper-base whisper-base --task automatic-speech-recognition-with-past --disable-stateful
NOTE
Tried without the --disable-stateful flag, but did not produce the KV cache decoder.
As mentioned in this #1726 , --trust-remote-code is also not needed.
Environment:
- transformers 4.46.2
- openvino 2025.0.0
- openvino-genai 2025.0.0.0
- openvino-telemetry 2024.1.0
- openvino-tokenizers 2025.0.0.0
- optimum 1.24.0
- optimum-intel 1.23.0.dev0+dd4fe68