Add support for shallow biasing of Whisper#1889
Add support for shallow biasing of Whisper#1889anthonyrathe wants to merge 7 commits intoOpenNMT:masterfrom
Conversation
Do you have the python wheels for this branch? I would like to use it |
Yes, you can find the links to the wheels via the checks below. |
|
@minhthuc2502 would you mind taking a look at this PR? Would love to get this released! |
There might be a bug in this PR. When I use sequence bias with a model compiled in int8_float16 I get this error ValueError: expected storage to be of type float16, but is of type float32 which I dont get when I use sequence_bias as None. The exact same input |
Sorry I resolved. Does this increase latency? |
It does, unfortunately. Especially if the number of sequences to bias is large. |
Where does this latency increase come from? |
|
would it be possible to allow a list-of-lists of sequences to bias, so each sample in a batch can be biased separately? this would match how the |
Attempting to fix #1789 by @zwycl