Skip to content

[Whisper][NPU] Default optimum export of whisper model does not create KV cache decoder #1728

@soumendukrg

Description

@soumendukrg

The latest optimum-cli command to download and convert whisper model to OV as given in the README does not create the decoder model with KV cache.

optimum-cli export openvino --trust-remote-code --model openai/whisper-base whisper-base

Since NPU inference pipeline needs the KV cache decoder model, running the sample throws out this error.

Image

The following change to the exporting command can solve this issue as it generates the KV cache decoder also

optimum-cli export openvino --model openai/whisper-base whisper-base --task automatic-speech-recognition-with-past --disable-stateful

NOTE
Tried without the --disable-stateful flag, but did not produce the KV cache decoder.

As mentioned in this #1726 , --trust-remote-code is also not needed.

Environment:

  • transformers 4.46.2
  • openvino 2025.0.0
  • openvino-genai 2025.0.0.0
  • openvino-telemetry 2024.1.0
  • openvino-tokenizers 2025.0.0.0
  • optimum 1.24.0
  • optimum-intel 1.23.0.dev0+dd4fe68

Metadata

Metadata

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions