We encountered deployment failures while attempting to integrate both vLLM and a speech recognition model from speeches-ai into an OpenAI-API-compatible interface.General Discussion #20280
Unanswered
ZHJ19970917
asked this question in
Help
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Self Checks
Content
I'm trying to integrate a self-hosted speech recognition model into the Dify platform using an OpenAI-compatible API interface, but the configuration consistently fails during model registration.
Details:
First Attempt:
Second Attempt:
# vllm command vllm serve openai/whisper-large-v3-turbo \ --trust-remote-code \ --served-model-name whisper-3 \ --port 8015 \ --api-key Aa7410.. \ --task transcription
The error seems to occur around the following stage(provider-configuration.py):

In both experiments, logs from the backend program confirm that Dify successfully reaches the model's inference endpoint. The model receives the request and returns a valid response.
However, despite this successful interaction, Dify still fails to complete the model configuration process, and the model does not appear in the available model list.
Beta Was this translation helpful? Give feedback.
All reactions