Replies: 1 comment 2 replies
-
I really like the idea of using the routes for cost efficiency. Though, could you elaborate a bit on what you mean by "compatible with"? Looking at the example issue you posted, Wilmer's currently style of routing would at least work somewhat closely to what they're imagining:
Having python requests route to say chatgpt 4o while Truescript go to Claude, things like that, would be something they could do now. But anything more dynamic than that would require a little work. Are you thinking in terms of Wilmer dynamically adjusting the routes based on price similar to RouteLLM, or something else entirely? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Sounds weird, but I think Wilmer can be compatible with RouteLLM when it comes to picking some model over others in different times based on computation cost and accuracy https://github.com/lm-sys/RouteLLM
According to LiveBench for SOTA open weight models, here are the usage:
qwen/qwq-32b-preview
for general reasoning and simple mathsqwen/qwen-2.5-coder-32b-instruct
for simple programming tasksdeepseek/deepseek-chat
for data science, heavy duty programming, and simple instructionsmeta-llama/llama-3.3-70b-instruct
for complex Instruction Following (IF)P.S. It would be useful in specific applications as cost-to-accuracy All-Hands-AI/OpenHands#5869
Beta Was this translation helpful? Give feedback.
All reactions