🚀 Feature Launch: Intelligent Model Routing arrives in v0.11.2 #12375
Replies: 3 comments 3 replies
-
|
works quite well, thanks! for the automatic setting, it might be nice to see a model indicator in the UI to know which model produced the response. by knowing this, I could more readily understand the trust I can put into the response from experience with those three models. |
Beta Was this translation helpful? Give feedback.
-
|
Awful future, suggest to disable it for good. It's switching between different models and may ruin your project by removing important code lines. |
Beta Was this translation helpful? Give feedback.
-
|
@abhipatel12 The router should be smart enough to not try to use a model that is out of token when others are not. Plus it's not that clear what is happening, in the /model settings I'm on Auto for 3 yet it complains about being unable to use 2.5 without telling me that maybe I can switch to another model. For a while actually I thought I was all out and it was trying to weakest one as a last recourse, it's only yesterday that I switched to manual 3pro and realized it worked just fine... |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hey everyone!
We're excited to announce that Intelligent Model Routing has graduated from our experimental channel post and is now standard in version
0.11.2!A huge thank you to everyone who tested this in
previewand provided the feedback needed to stabilize this feature.Why Model Routing? 🤔
Until now, Gemini CLI used one primary model (like Gemini 2.5 Pro) for your entire session. While fantastic for complex tasks, using such a heavy-duty model for simpler requests often increased latency and burned through your Pro quota faster than necessary.
Intelligent Model Routing solves this by dynamically directing your requests to the model best suited for the job. Simple queries are handled instantly by faster, lighter models (like Flash), while complex analytical or creative tasks continue to leverage the full power of Pro.
By upgrading, you gain:
How to get it 📦
Routing is now enabled by default in our latest stable release. No configuration required.
Just update to the latest version:
Taking Control: The new
/modelslash command 🛠️While the router is great for 95% of cases, sometimes you need guaranteed raw power, or perhaps you specifically want to iterate quickly on small tasks with a lighter model.
We’ve introduced
/model, allowing you to instantly switch routing behavior mid-session without restarting Gemini CLI.This will open an interactive dialog to help you switch between routing and models mid-session!
(You can also still use the
-mstartup flag when starting withgemini).Here's what it looks like!
Verifying it's active ⚙️
You can verify routing is active by running the standard
/statscommand during a session. You will see an increase in requests to lighter models (likegemini-2.5-flash) alongside your usual Pro requests. You'll also see a new, small stream ofgemini-2.5-flash-litemodel requests that we use to help route your request!Feedback 💭
While this feature is now stable, we are always looking to refine the routing logic. Please continue to share your experiences in this thread, specifically if you encounter cases where the router consistently underproduces on complex tasks.
Have questions? If you're curious about the mechanics of how the router makes decisions, or how this fits into your specific workflows, please ask away in this thread! We're happy to dive into the details.
Thanks for helping us build Gemini CLI! ❤️
Beta Was this translation helpful? Give feedback.
All reactions