diff --git a/.changeset/v3.29.3.md b/.changeset/v3.29.3.md deleted file mode 100644 index 2a7cd9359d44..000000000000 --- a/.changeset/v3.29.3.md +++ /dev/null @@ -1,14 +0,0 @@ ---- -"roo-cline": patch ---- - -- Update Gemini models with latest 09-2025 versions including Gemini 2.5 Pro and Flash (#8485 by @cleacos, PR by @roomote) -- Add reasoning support for Z.ai GLM binary thinking mode (#8465 by @BeWater799, PR by @daniel-lxs) -- Enable reasoning in Roo provider (thanks @mrubens!) -- Add settings to configure time and cost display in system prompt (#8450 by @jaxnb, PR by @roomote) -- Fix: Use max_output_tokens when available in LiteLLM fetcher (#8454 by @fabb, PR by @roomote) -- Fix: Process queued messages after context condensing completes (#8477 by @JosXa, PR by @roomote) -- Fix: Use monotonic clock for rate limiting to prevent timing issues (#7770 by @intermarkec, PR by @chrarnoldus) -- Fix: Resolve checkpoint menu popover overflow (thanks @daniel-lxs!) -- Fix: LiteLLM test failures after merge (thanks @daniel-lxs!) -- Improve UX: Focus textbox and add newlines after adding to context (thanks @mrubens!) diff --git a/CHANGELOG.md b/CHANGELOG.md index e19703e5975a..eabb4622fc0c 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,18 @@ # Roo Code Changelog +## [3.29.3] - 2025-10-28 + +- Update Gemini models with latest 09-2025 versions including Gemini 2.5 Pro and Flash (#8485 by @cleacos, PR by @roomote) +- Add reasoning support for Z.ai GLM binary thinking mode (#8465 by @BeWater799, PR by @daniel-lxs) +- Enable reasoning in Roo provider (thanks @mrubens!) +- Add settings to configure time and cost display in system prompt (#8450 by @jaxnb, PR by @roomote) +- Fix: Use max_output_tokens when available in LiteLLM fetcher (#8454 by @fabb, PR by @roomote) +- Fix: Process queued messages after context condensing completes (#8477 by @JosXa, PR by @roomote) +- Fix: Use monotonic clock for rate limiting to prevent timing issues (#7770 by @intermarkec, PR by @chrarnoldus) +- Fix: Resolve checkpoint menu popover overflow (thanks @daniel-lxs!) +- Fix: LiteLLM test failures after merge (thanks @daniel-lxs!) +- Improve UX: Focus textbox and add newlines after adding to context (thanks @mrubens!) + ## [3.29.2] - 2025-10-27 - Add support for LongCat-Flash-Thinking-FP8 models in Chutes AI provider (#8425 by @leakless21, PR by @roomote) diff --git a/src/package.json b/src/package.json index 4217f1b8ef48..6f9d170c4d8c 100644 --- a/src/package.json +++ b/src/package.json @@ -3,7 +3,7 @@ "displayName": "%extension.displayName%", "description": "%extension.description%", "publisher": "RooVeterinaryInc", - "version": "3.29.2", + "version": "3.29.3", "icon": "assets/icons/icon.png", "galleryBanner": { "color": "#617A91",