Skip to content

Conversation

yossiovadia
Copy link
Collaborator

Summary

Fixes llm-katan OpenAI API compatibility issues causing OpenWebUI to hang when connecting to llm-katan endpoints.

Changes Made

  • Add missing OpenAI API response fields: system_fingerprint, logprobs, detailed usage object
  • Fix streaming response Content-Type: Changed from text/plain to text/event-stream
  • Add token_usage alias: For better SDK compatibility
  • Apply fixes to both backends: TransformersBackend and VLLMBackend
  • Bump version to 0.1.9: Published to PyPI

Problem Solved

  • Resolves OpenWebUI hanging issue when using llm-katan as backend
  • Improves OpenAI SDK compatibility with proper response format
  • Fixes streaming responses with correct SSE format and final usage chunk

Testing

  • Tested with curl (non-streaming and streaming)
  • Verified all OpenAI compatibility fields present
  • Confirmed OpenWebUI now works correctly
  • Published v0.1.9 to PyPI

Fixes #241

…t#241

- Add missing OpenAI API response fields (system_fingerprint, logprobs, detailed usage)
- Fix streaming response Content-Type from text/plain to text/event-stream
- Ensure both static and streaming responses include all compatibility fields
- Add token_usage alias for better SDK compatibility
- Apply fixes to both TransformersBackend and VLLMBackend

Resolves OpenWebUI hanging issue when connecting to llm-katan endpoints.

Signed-off-by: Yossi Ovadia <[email protected]>
Published llm-katan v0.1.9 to PyPI with OpenAI API compatibility fixes.

Signed-off-by: Yossi Ovadia <[email protected]>
@yossiovadia yossiovadia requested a review from rootfs as a code owner October 6, 2025 20:34
Copy link

netlify bot commented Oct 6, 2025

Deploy Preview for vllm-semantic-router ready!

Name Link
🔨 Latest commit 528bc6d
🔍 Latest deploy log https://app.netlify.com/projects/vllm-semantic-router/deploys/68e4954519cb16000809f343
😎 Deploy Preview https://deploy-preview-354--vllm-semantic-router.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

rootfs
rootfs previously approved these changes Oct 6, 2025
@rootfs
Copy link
Collaborator

rootfs commented Oct 6, 2025

@yossiovadia can you run pre-commit?

Trigger CI re-run to verify if Black formatting issues are resolved.

Signed-off-by: Yossi Ovadia <[email protected]>
Copy link

github-actions bot commented Oct 7, 2025

👥 vLLM Semantic Team Notification

The following members have been identified for the changed files in this PR and have been automatically assigned:

📁 e2e-tests

Owners: @yossiovadia
Files changed:

  • e2e-tests/llm-katan/llm_katan/model.py
  • e2e-tests/llm-katan/llm_katan/server.py
  • e2e-tests/llm-katan/pyproject.toml

vLLM

🎉 Thanks for your contributions!

This comment was automatically generated based on the OWNER files in the repository.

@Xunzhuo Xunzhuo merged commit 88eec24 into vllm-project:main Oct 7, 2025
9 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Enhance llm-katan OpenAI API compatibility with missing response fields
3 participants