Skip to content

Commit f88bb75

Browse files
feat(docs): Release v3.18.0 & Enhance Context Condensing Docs (#198)
* feat(docs): Release v3.18.0 & Enhance Context Condensing Docs This commit introduces the release notes for Roo Code v3.18.0 and significantly updates documentation for Intelligent Context Condensing, Custom Modes (YAML), API Cost Control, and AI provider capabilities. **Release Notes & Structure:** * **v3.18.0 Release:** * Adds detailed release notes for v3.18.0 ([`docs/update-notes/v3.18.0.mdx`](docs/update-notes/v3.18.0.mdx:0)). * Creates a combined release notes page for the 3.18 series ([`docs/update-notes/v3.18.mdx`](docs/update-notes/v3.18.mdx:0)), initially populated with v3.18.0 content. * Updates the main update notes index ([`docs/update-notes/index.md`](docs/update-notes/index.md:0)) to include the 3.18 series. * Adds new images for v3.18.0 features ([`static/img/v3.18.0/`](static/img/v3.18.0/)). * **Sidebar Navigation:** Updates [`sidebars.ts`](sidebars.ts:0) to include the new v3.18.0 release notes and correct the path for Intelligent Context Condensing documentation. **Feature Documentation Updates:** * **Intelligent Context Condensing:** * Replaces `intelligent-context-condensation.md` with a new, comprehensive [`docs/features/experimental/intelligent-context-condensing.mdx`](docs/features/experimental/intelligent-context-condensing.mdx:0). * Details enabling the feature, manual/automatic condensing controls, UI indicators, token counting, internationalization, technical implementation, and performance. * Adds new illustrative images ([`static/img/intelligent-context-condensation/`](static/img/intelligent-context-condensation/)). * Adds redirect in [`docusaurus.config.ts`](docusaurus.config.ts:0) for the old path. * Updates links in v3.17.0 and v3.17 release notes to point to the new path. * **Custom Modes (YAML Focus):** * Updates [`docs/features/custom-modes.mdx`](docs/features/custom-modes.mdx:0) to reflect YAML as the primary format, including table headers and link updates. * Adds a title to the embedded overview video. * **API Cost Control:** * Adds a new section "Limiting Auto-Approved Requests" to [`docs/advanced-usage/rate-limits-costs.md`](docs/advanced-usage/rate-limits-costs.md:0), explaining the "Max Requests" setting with images. **Provider Documentation Updates:** * **Gemini & Vertex AI:** Updates [`docs/providers/gemini.md`](docs/providers/gemini.md:0) and [`docs/providers/vertex.md`](docs/providers/vertex.md:0) to include the `gemini-2.5-flash-preview-05-20` model. * **LM Studio:** Adds notes on token tracking and reasoning support to [`docs/providers/lmstudio.md`](docs/providers/lmstudio.md:0). * **Ollama:** Adds a note on token tracking to [`docs/providers/ollama.md`](docs/providers/ollama.md:0). * **Unbound:** Adds a note on the model list refresh button to [`docs/providers/unbound.md`](docs/providers/unbound.md:0). * docs: Clarify description of available tools in custom modes documentation * Update docs/features/custom-modes.mdx Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com> * Update docs/features/experimental/intelligent-context-condensing.mdx Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com> --------- Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
1 parent 613f568 commit f88bb75

23 files changed

+560
-152
lines changed

docs/advanced-usage/rate-limits-costs.md

Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,19 @@ Roo Code automatically calculates the estimated cost of each API request based o
2323
* Some providers may offer free tiers or credits. Check your provider's documentation for details.
2424
* Some providers offer prompt caching which greatly lowers cost.
2525

26+
### Limiting Auto-Approved Requests
27+
28+
To further help manage API costs and prevent unexpected expenses, Roo Code includes a "Max Requests" setting for auto-approved actions. This allows you to define a specific limit on how many consecutive API calls Roo Code can make without requiring your explicit re-approval during a task.
29+
30+
* **How it works:** If you set a limit (e.g., 5 requests), Roo Code will perform up to 5 auto-approved API calls. Before making the 6th call, it will pause and prompt you to "Reset and Continue," as shown below.
31+
<img src="/img/v3.18.0/v3.18.0-1.png" alt="Warning message indicating the auto-approved request limit has been reached." width="600" />
32+
*Notification when the auto-approved request limit is met.*
33+
* **Configuration:** This limit is configured within the "Auto-approve actions" settings. You can set a specific number or choose "Unlimited." For detailed steps on configuring this and other auto-approval settings, see the [Auto-Approving Actions documentation](/features/auto-approving-actions).
34+
<img src="/img/v3.18.0/v3.18.0.png" alt="Setting the Max Requests limit for auto-approved actions in Roo Code settings." width="600" />
35+
*Setting the "Max Requests" for auto-approved actions.*
36+
37+
This feature provides an additional safeguard, particularly for complex or long-running tasks where multiple API calls might be involved.
38+
2639
## Tips for Optimizing Token Usage
2740

2841
* **Be Concise:** Use clear and concise language in your prompts. Avoid unnecessary words or details.

docs/features/custom-modes.mdx

Lines changed: 250 additions & 102 deletions
Large diffs are not rendered by default.

docs/features/experimental/experimental-features.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ To enable or disable experimental features:
1616

1717
The following experimental features are currently available:
1818

19-
- [Intelligently Condense the Context Window](/features/experimental/intelligent-context-condensation)
19+
- [Intelligently Condense the Context Window](/features/experimental/intelligent-context-condensing)
2020
- [Power Steering](/features/experimental/power-steering)
2121

2222
## Providing Feedback

docs/features/experimental/intelligent-context-condensation.md

Lines changed: 0 additions & 45 deletions
This file was deleted.
Lines changed: 119 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,119 @@
1+
---
2+
sidebar_label: 'Intelligent Context Condensing'
3+
---
4+
import Codicon from '@site/src/components/Codicon';
5+
6+
# Intelligent Context Condensing (Experimental)
7+
8+
The Intelligent Context Condensing feature helps manage long conversations by summarizing earlier parts of the dialogue. This prevents important information from being lost when the context window nears its limit. This is an **experimental feature** and is **disabled by default**.
9+
10+
<div style={{width: '50%', margin: 'auto'}}>
11+
<div style={{position: 'relative', paddingBottom: '177.77%', height: 0, overflow: 'hidden'}}>
12+
<iframe
13+
src="https://www.youtube.com/embed/o5xgO9N8vVU"
14+
title="YouTube Short"
15+
frameBorder="0"
16+
allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share"
17+
allowFullScreen
18+
style={{position: 'absolute', top: 0, left: 0, width: '100%', height: '100%'}}
19+
></iframe>
20+
</div>
21+
</div>
22+
## How It Works
23+
24+
As your conversation with Roo Code grows, it might approach the context window limit of the underlying AI model. When this happens, older messages would typically be removed to make space. Intelligent Context Condensing aims to prevent this abrupt loss by:
25+
26+
1. **Summarizing:** Using an AI model, it condenses earlier parts of the conversation.
27+
2. **Retaining Essentials:** The goal is to reduce the overall token count while keeping the key information from the summarized messages.
28+
3. **Maintaining Flow:** This allows the AI to have a more coherent understanding of the entire conversation, even very long ones.
29+
30+
**Important Considerations:**
31+
* **Summarization Impact:** While original messages are preserved if you use [Checkpoints](/features/checkpoints) to rewind, the summarized version is what's used in ongoing LLM calls to keep the context manageable.
32+
* **Cost:** The AI call to perform the summarization incurs a cost. This cost is included in the context condensing metrics displayed in the UI.
33+
34+
## Enabling This Feature
35+
36+
As an experimental feature, Intelligent Context Condensing is **disabled by default**.
37+
38+
1. Open Roo Code settings (<Codicon name="gear" /> icon in the top right corner of the Roo Code panel).
39+
2. Navigate to the "Experimental" section.
40+
3. Toggle the "Automatically trigger intelligent context condensing" (`autoCondenseContext`) option to enable it.
41+
4. Optionally, adjust the "Threshold to trigger intelligent context condensing" (`autoCondenseContextPercent`) slider to control the trigger point for automatic context condensing.
42+
5. Save your changes.
43+
44+
<img src="/img/intelligent-context-condensation/intelligent-context-condensation-1.png" alt="Settings for Intelligent Context Condensing" width="600" />
45+
*The image above shows settings for Intelligent Context Condensing: the toggle to "Automatically trigger intelligent context condensing" and the "Threshold to trigger intelligent context condensing" slider.*
46+
47+
## Controlling and Understanding Context Condensing
48+
49+
Roo Code provides several ways to control and understand the Intelligent Context Condensing feature:
50+
51+
### Controlling Context Condensing
52+
* **Automatic Threshold:** In Roo Code Settings (<Codicon name="gear" />) > "Experimental," the `autoCondenseContextPercent` setting allows you to define a percentage (e.g., 80%). Roo Code will attempt to condense the context automatically when the conversation reaches this level of the context window's capacity.
53+
* **Manual Trigger:** A **Condense Context** button (<Codicon name="fold" /> icon) is available when a task is expanded, typically located at the bottom of the task view, next to other task action icons like the trash can. This allows you to initiate the context condensing process at any time.
54+
55+
<img src="/img/intelligent-context-condensation/intelligent-context-condensation-2.png" alt="Manual Condense Context button in expanded task view" width="600" />
56+
*The Manual Condense Context button (highlighted with a yellow arrow) appears in the expanded task view.*
57+
58+
### Understanding Context Condensing Activity
59+
* **Context Condensing Metrics:** When context condensing occurs, Roo Code displays:
60+
* The context token counts before and after context condensing.
61+
* The cost associated with the context condensing AI call.
62+
* An expandable summary detailing what was condensed (this information is part of the `ContextCondenseRow` component visible in the chat history).
63+
64+
<img src="/img/intelligent-context-condensation/intelligent-context-condensation-4.png" alt="Context condensed message in chat" width="600" />
65+
*After context condensing, a message indicates the context has been condensed, showing token changes and cost.*
66+
67+
* **Visual Indicators:**
68+
* A progress indicator ("Condensing context...") is shown in the chat interface while context condensing is active.
69+
70+
<img src="/img/intelligent-context-condensation/intelligent-context-condensation-3.png" alt="Condensing context progress indicator in chat" width="600" />
71+
*The "Condensing context..." indicator appears in the chat during the process.*
72+
73+
* The task header also displays the current context condensing status.
74+
* The `ContextWindowProgress` bar offers a visual representation of token distribution, including current usage, space reserved for the AI's output, available space, and raw token numbers.
75+
* **Interface Clarity:** The "Condense Context" button includes a tooltip explaining its function, available in all supported languages. The icon for context condensing-related actions is `codicon-compress`.
76+
77+
### Accurate Token Information
78+
* Roo Code employs accurate token counting methods, with some AI providers utilizing their native token counting endpoints. This ensures that context size and associated costs are calculated reliably.
79+
80+
### Internationalization
81+
* All user interface elements for this feature, such as button labels, tooltips, status messages, and settings descriptions, are available in multiple supported languages.
82+
83+
## Technical Implementation
84+
85+
### Token Counting
86+
Roo Code uses a sophisticated token counting system that:
87+
- Employs native token counting endpoints when available (e.g., Anthropic's API)
88+
- Falls back to tiktoken estimation if API calls fail
89+
- Provides accurate counting for different content types:
90+
- Text content: Uses word-based estimation with punctuation and newline overhead
91+
- Image content: Uses a conservative estimate of 300 tokens per image
92+
- System prompts: Includes additional overhead for structural elements
93+
94+
### Context Window Management
95+
- By default, 30% of the context window is reserved (20% for model output and 10% as a safety buffer), leaving 70% available for conversation history.
96+
- This reservation can be overridden by model-specific settings
97+
- The system automatically calculates available space while maintaining this reservation
98+
99+
## Performance Considerations
100+
101+
### Optimization
102+
- The system optimizes token counting to minimize performance impact
103+
- Token calculations are cached where possible
104+
- Background processing prevents UI blocking during context condensing
105+
106+
### Resource Usage
107+
- Context condensing operations are performed asynchronously
108+
- The UI remains responsive during the process
109+
- System resources are managed to prevent excessive memory usage
110+
111+
## Feedback
112+
113+
Your experience with experimental features is valuable. When reporting issues, please include:
114+
- The current threshold setting
115+
- The token counts before and after context condensing
116+
- Any error messages displayed
117+
- Steps to reproduce the issue
118+
119+
Please report any issues or suggestions regarding Intelligent Context Condensing on the [Roo Code GitHub Issues page](https://github.com/RooCodeInc/Roo-Code/issues).

docs/providers/gemini.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@ Roo Code supports Google's Gemini family of models through the Google AI Gemini
1919

2020
Roo Code supports the following Gemini models:
2121

22+
* `gemini-2.5-flash-preview-05-20`
2223
* `gemini-2.5-pro-exp-03-25`
2324
* `gemini-2.0-flash-001`
2425
* `gemini-2.0-flash-lite-preview-02-05`

docs/providers/lmstudio.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -38,3 +38,5 @@ Roo Code supports running models locally using LM Studio. LM Studio provides a
3838
* **Local Server:** The LM Studio local server must be running for Roo Code to connect to it.
3939
* **LM Studio Documentation:** Refer to the [LM Studio documentation](https://lmstudio.ai/docs) for more information.
4040
* **Troubleshooting:** If you see a "Please check the LM Studio developer logs to debug what went wrong" error, you may need to adjust the context length settings in LM Studio.
41+
* **Token Tracking:** Roo Code tracks token usage for models run via LM Studio, helping you monitor consumption.
42+
* **Reasoning Support:** For models that support it, Roo Code can parse "think" tags or similar reasoning indicators in LM Studio responses, offering more insight into the model's process.

docs/providers/ollama.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -72,4 +72,5 @@ Roo Code supports running models locally using Ollama. This provides privacy, of
7272
* **Resource Requirements:** Running large language models locally can be resource-intensive. Make sure your computer meets the minimum requirements for the model you choose.
7373
* **Model Selection:** Experiment with different models to find the one that best suits your needs.
7474
* **Offline Use:** Once you've downloaded a model, you can use Roo Code offline with that model.
75+
* **Token Tracking:** Roo Code tracks token usage for models run via Ollama, helping you monitor consumption.
7576
* **Ollama Documentation:** Refer to the [Ollama documentation](https://ollama.com/docs) for more information on installing, configuring, and using Ollama.

docs/providers/unbound.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -28,3 +28,4 @@ Unbound allows you configure a list of supported models in your application, and
2828
## Tips and Notes
2929

3030
* **Security Focus:** Unbound emphasizes security features for enterprise use. If your organization has strict security requirements for AI usage, Unbound might be a good option.
31+
* **Model List Refresh:** Roo Code includes a refresh button specifically for the Unbound provider in the settings. This allows you to easily update the list of available models from your Unbound application and get immediate feedback on your API key's validity.

docs/providers/vertex.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@ Roo Code supports accessing models through Google Cloud Platform's Vertex AI, a
2323
Roo Code supports the following models through Vertex AI (based on source code):
2424

2525
* **Google Gemini Models:**
26+
* `gemini-2.5-flash-preview-05-20`
2627
* `gemini-2.0-flash-001`
2728
* `gemini-2.5-pro-exp-03-25`
2829
* `gemini-2.0-pro-exp-02-05`

0 commit comments

Comments
 (0)