Improve Autocomplete Suggestions by Experimenting with models/prompts #906
hassoncs
started this conversation in
1. Feature requests
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Summary
Experiment with multiple OpenRouter models for autocomplete to identify the best trade-off between quality, latency, and cost, then expose a dropdown UI allowing end users to select their preferred model.
Background
Currently using Gemini Flash 2.5 for inline autocomplete with good accuracy but significant latency issues (taking at least 2 seconds) under load. We need to determine if Gemini 2.5 is actually the best choice by comparing it against potentially cheaper/faster models using a consistent prompt template.
Goals
Acceptance Criteria
Tasks
Technical Notes
Beta Was this translation helpful? Give feedback.
All reactions