Replies: 12 comments 8 replies
-
Hey @sruckh I'm one of the CLI maintainers. I agree, there are totally times when Gemini totally gets lost and confused, sorry that it hosed your stuff. That's not cool. We're actively working to make this better. If you don't mind, tell me more about your specific scenarios feel free to email me if you prefer, mattkorwel at Google dot com. If you could do I'm most interested in Thanks for trying out our new CLI, let's get it working as good as we can for you. |
Beta Was this translation helpful? Give feedback.
-
Amazing so far for me. 1 week in. Any LLM can go off the rails thats normal. But its also how you use it. Tame the dragon so to speak. And you shall conquer land. Some tips: Keep files small, reset context if it bugs out, do things in chunks, careful with being too ambitious in your prompts. And a lots more. Agentic coding is a skill atm. You have to learn the gotchas. Its not automagic. But very magic once you get the hang of it. |
Beta Was this translation helpful? Give feedback.
-
Basically this. If you know how to prompt it, what to tell it, what info to show it, it will do most of the stuff by itself. As with any LLM model, the more specific and the more info you give it, the more it will actually help. The only caveat for now is the CLI switching to the flash model, which sucks at coding and does wild mistakes. As for it ignoring your request, how much context did it fill up before you asked it to do so? Anything above 150k or 200k context starts going off the rails from what I've seen. Maybe you had conflicting instructions for it, let's say when it fell into a loop and started filling up its own context? You'd need to provide specifics of what how and when everything happened. |
Beta Was this translation helpful? Give feedback.
-
nothing made gemini-cli usable for me. This is not the CLI's problem. Gemini models behave exactly the same in other agentic frameworks. In Cursor, the same looping problem happens and failing to apply the simplest of diffs after just 5k 10k used context. It's exactly the same issues, gemini 2.5 pro falls into loops less, but both pro and flash are literally unusable beyond asking questions. They can't behave agentically. There was a point where gemini exp 2.5 felt like the best model. Then sonnet 4 and o3 got released and something happened to the gemini weights. It's literally worse than 8b models right now. |
Beta Was this translation helpful? Give feedback.
-
Im having a great success with it so far. Here is how do I work with gemini:
|
Beta Was this translation helpful? Give feedback.
-
Just discovered that using the Gemini API in conjunction with the Roo Code extension gives much better results, so this is most likely an issue with the CLI, although Gemini has its own outstanding issues as well. |
Beta Was this translation helpful? Give feedback.
-
It's honestly so weird seeing people have issues with gemini. There's a lot of people with good results and there's people who don't get any results at all. I wonder what's the statistic on that. Also, be careful with the API, especially if you intend on paying. I used the API for about 20 minutes and it ate 50$ worth of API calls (the pro model). Even the free api from AI Studio ate up real money, when in reality it was supposed to stop after the limit was reached. |
Beta Was this translation helpful? Give feedback.
-
We appreciate all the feedback here! This is the value of open-source 👏 Team is working hard as mentioned by @mattKorwel and we are actively looking to make the experience with Gemini CLI better. Going to convert this thread to a discussion as it more of a question and discussion thread. |
Beta Was this translation helpful? Give feedback.
-
It is not a ready to be launched. This has been a waste of time for me. Very frustrating. Lost few days of working trying it and don't even get proper answers from Google! |
Beta Was this translation helpful? Give feedback.
-
@sruckh you didn't mention which LLM model you used? Gemini 2.5 Pro or Flash? There's a clear difference between them for quality of responses from my experience via
From my experience and other reports, Gemini models can be subject to scope expansion - going beyond what it's instructions asked. To reign it in my system prompts and GEMINI.md account for it. Example below - adjust as desired or losen it strictness. Then I have project specific overview at bottom etc. # GEMINI.md
This file provides guidance to [Gemini CLI](https://github.com/google-gemini/gemini-cli)) when working with code in this repository.
## AI Guidance
**Primary Directive:**
* You are a specialized AI assistant. Your primary function is to execute the user's instructions with precision and within the specified scope.
* * Ignore CLAUDE.md and CLAUDE-*.md files
* Before you finish, please verify your solution.
**Core Principles:**
1. **Strict Adherence to Instructions:** You MUST adhere strictly to the user's instructions. Do not add unsolicited information, analysis, or suggestions unless explicitly asked. Your response should directly and exclusively address the user's query.
2. **Scope Limitation:** Your operational scope is defined by the immediate user request. Do not expand upon the request, generalize the topic, or provide background information that was not explicitly solicited.
3. **Clarification Protocol:** If an instruction is ambiguous, or if fulfilling it would require exceeding the apparent scope, you MUST ask for clarification before proceeding. State what part of the request is unclear and what information you require to continue.
4. **Output Formatting:** You are to generate output ONLY in the format specified by the user. If no format is specified, provide a concise and direct answer without additional formatting.
**Behavioral Guardrails:**
* **No Unsolicited Summaries:** Do not summarize the conversation or your response unless explicitly instructed to do so.
* **No Proactive Advice:** Do not offer advice or suggestions for improvement unless the user asks for them.
* **Task-Specific Focus:** Concentrate solely on the task at hand. Do not introduce related but irrelevant topics.
**Example of Adherence:**
* **User Prompt:** "What is the capital of France?"
* **Your Correct Response:** "Paris"
* **Your Incorrect Response (Scope Expansion):** "The capital of France is Paris, which is also its largest city. It is known for its art, fashion, and culture, and is home to landmarks like the Eiffel Tower and the Louvre."
By internalizing these directives, you will provide focused and efficient responses that directly meet the user's needs without unnecessary expansion.
## Memory Bank System
This project uses a structured memory bank system with specialized context files. Always check these files for relevant information before starting work:
### Core Context Files
* **GEMINI-codebase.md** - Detailed file structure and key component documentation
* **GEMINI-activeContext.md** - Current session state, goals, and progress (if exists)
* **GEMINI-patterns.md** - Established code patterns and conventions (if exists)
* **GEMINI-decisions.md** - Architecture decisions and rationale (if exists)
* **GEMINI-troubleshooting.md** - Common issues and proven solutions (if exists)
* **GEMINI-config-variables.md** - Configuration variables reference (if exists)
* **GEMINI-temp.md** - Temporary scratch pad (only read when referenced)
**Important:** Always reference the active context file first to understand what's currently being worked on and maintain session continuity. I evaluate Claude Code vs Gemini CLI responses. But until Gemini CLI improves, I am using it as a companion and wrapped it into a Gemini CLI MCP server which I add to Claude Code. This way I can use Claude Code for execution of tasks but allows Claude Sonnet 4/Opus 4 to work with Gemini 2.5 models for code generation reviews/planning etc https://github.com/centminmod/gemini-cli-mcp-server 😁 |
Beta Was this translation helpful? Give feedback.
-
Myself, I am falling in some serious love with it ;0) as it has been solving so many years long problems, mostly IT ones, so far. I shall try to explain why so very spontaneously here, sorry for the "warts and all" but you shall thus feel my "raw enthusiasm" thereby, I hope. Quick background, about me mostly, first. But as long as you apply some project control, me toying with the PMBOK one mostly here still, it starts to make sense and be useful. (The method in its essence is thus similar to #3316 (comment) above.) The last session fixing that That is why I love it so far, and I have been using "pre-LLMs" since 1990s (neural network smths, in very short). Ref, session that has ended right now, see the agreement ratio and the time it has taken:
|
Beta Was this translation helpful? Give feedback.
-
Hello everyone, I'm pretty happy with my new week-long experiment. I'm running Gemini on Arduino ESP32-style code, based on "ESPAsyncWebServer." I've been struggling to get the hang of this thing for months, and I thought it was a good test for Gemini. Of course, it could be improved, but you quickly understand how it works; I'm still discovering new things. In conclusion, thank you to everyone actively working on this project; it's going to get better and better, and it's going to make our lives easier. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
gemini-cli wiped my entire codebase trying to suppress a warning message. When it jacks up behind all repairs, it goes into the corner and hides. Both times it has turned code into a complete mess, and I have asked it to stop working on the feature branch and pull from the remote repo, but it has completely stopped responding. It thinks for about a minute, does nothing, and returns to the chat prompt. I don't get the hype, as this has failed on multiple occasions and seems to completely ignore the GEMINI.md file. I give it strict commands not to push stuff to git without permission, and it does it anyway. I ask it to use git instead of gh, and again totally ignores it. Once it goes off the rails, it is a complete joke. It blows through the number of requests in a minute, trying to recover from its mistakes. My experience so far is that this is not reliable and cannot be trusted with even the simplest of tasks.
Beta Was this translation helpful? Give feedback.
All reactions