Skip to content

Commit 685e027

Browse files
committed
docs Dynamic Rate Limiter v3
1 parent c947994 commit 685e027

File tree

1 file changed

+15
-8
lines changed
  • docs/my-website/release_notes/v1.77.7-stable

1 file changed

+15
-8
lines changed

docs/my-website/release_notes/v1.77.7-stable/index.md

Lines changed: 15 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -54,6 +54,7 @@ pip install litellm==1.77.7.rc.1
5454

5555
## Key Highlights
5656

57+
- **Dynamic Rate Limiter v3** - Automatically maximizes throughput when capacity is available (< 80% saturation) by allowing lower-priority requests to use unused capacity, then switches to fair priority-based allocation under high load (≥ 80%) to prevent blocking
5758
- **Major Performance Improvements** - Router optimization reducing P99 latency by 62.5%, cache improvements from O(n*log(n)) to O(log(n))
5859
- **Claude Sonnet 4.5** - Support for Anthropic's new Claude Sonnet 4.5 model family with 200K+ context and tiered pricing
5960
- **MCP Gateway Enhancements** - Fine-grained tool control, server permissions, and forwardable headers
@@ -85,14 +86,9 @@ pip install litellm==1.77.7.rc.1
8586
- Add cost tracking for /v1/messages in streaming response - [PR #15102](https://github.com/BerriAI/litellm/pull/15102)
8687
- Add /v1/messages/count_tokens to Anthropic routes for non-admin user access - [PR #15034](https://github.com/BerriAI/litellm/pull/15034)
8788
- **[Gemini](../../docs/providers/gemini)**
88-
- Add full support for native Gemini API translation - [PR #15029](https://github.com/BerriAI/litellm/pull/15029)
89-
- Add Gemini generateContent passthrough cost tracking - [PR #15014](https://github.com/BerriAI/litellm/pull/15014)
90-
- Add streamGenerateContent cost tracking in passthrough - [PR #15199](https://github.com/BerriAI/litellm/pull/15199)
9189
- Ignore type param for gemini tools - [PR #15022](https://github.com/BerriAI/litellm/pull/15022)
9290
- **[Vertex AI](../../docs/providers/vertex)**
9391
- Add LiteLLM Overhead metric for VertexAI - [PR #15040](https://github.com/BerriAI/litellm/pull/15040)
94-
- Add cost tracking for Vertex AI Passthrough `/predict` endpoint - [PR #15019](https://github.com/BerriAI/litellm/pull/15019)
95-
- Add cost tracking for Vertex AI Live API WebSocket Passthrough - [PR #14956](https://github.com/BerriAI/litellm/pull/14956)
9692
- Support googlemap grounding in vertex ai - [PR #15179](https://github.com/BerriAI/litellm/pull/15179)
9793
- **[Azure](../../docs/providers/azure)**
9894
- Add azure_ai grok-4 model family - [PR #15137](https://github.com/BerriAI/litellm/pull/15137)
@@ -140,9 +136,21 @@ pip install litellm==1.77.7.rc.1
140136
- **[Responses API](../../docs/response_api)**
141137
- Return Cost for Responses API Streaming requests - [PR #15053](https://github.com/BerriAI/litellm/pull/15053)
142138

139+
- **[/generateContent](../../docs/providers/gemini)**
140+
- Add full support for native Gemini API translation - [PR #15029](https://github.com/BerriAI/litellm/pull/15029)
141+
142+
- **Passthrough Gemini Routes**
143+
- Add Gemini generateContent passthrough cost tracking - [PR #15014](https://github.com/BerriAI/litellm/pull/15014)
144+
- Add streamGenerateContent cost tracking in passthrough - [PR #15199](https://github.com/BerriAI/litellm/pull/15199)
145+
146+
- **Passthrough Vertex AI Routes**
147+
- Add cost tracking for Vertex AI Passthrough `/predict` endpoint - [PR #15019](https://github.com/BerriAI/litellm/pull/15019)
148+
- Add cost tracking for Vertex AI Live API WebSocket Passthrough - [PR #14956](https://github.com/BerriAI/litellm/pull/14956)
149+
143150
- **General**
144151
- Preserve Whitespace Characters in Model Response Streams - [PR #15160](https://github.com/BerriAI/litellm/pull/15160)
145152
- Add provider name to payload specification - [PR #15130](https://github.com/BerriAI/litellm/pull/15130)
153+
- Ensure query params are forwarded from origin url to downstream request - [PR #15087](https://github.com/BerriAI/litellm/pull/15087)
146154

147155
---
148156

@@ -166,7 +174,6 @@ pip install litellm==1.77.7.rc.1
166174
#### Bugs
167175

168176
- **Dashboard** - Fix LiteLLM model name fallback in dashboard overview - [PR #14998](https://github.com/BerriAI/litellm/pull/14998)
169-
- **Passthrough API** - Ensure query params are forwarded from origin url to downstream request - [PR #15087](https://github.com/BerriAI/litellm/pull/15087)
170177

171178
---
172179

@@ -204,8 +211,6 @@ pip install litellm==1.77.7.rc.1
204211
- Dynamic Rate Limiter v3 - fixes for detecting saturation + fixes for post saturation behavior - [PR #15192](https://github.com/BerriAI/litellm/pull/15192)
205212
- **Teams**
206213
- Add model specific tpm/rpm limits to teams on LiteLLM - [PR #15044](https://github.com/BerriAI/litellm/pull/15044)
207-
- **Configuration**
208-
- Add max requests env var - [PR #15007](https://github.com/BerriAI/litellm/pull/15007)
209214

210215
---
211216

@@ -233,6 +238,8 @@ pip install litellm==1.77.7.rc.1
233238
- **Cache Optimizations**
234239
- Reduce complexity of InMemoryCache.evict_cache from O(n*log(n)) to O(log(n)) - [PR #15000](https://github.com/BerriAI/litellm/pull/15000)
235240
- Avoiding expensive operations when cache isn't available - [PR #15182](https://github.com/BerriAI/litellm/pull/15182)
241+
- **Worker Management**
242+
- Add proxy CLI option to recycle workers after N requests - [PR #15007](https://github.com/BerriAI/litellm/pull/15007)
236243
- **Metrics & Monitoring**
237244
- LiteLLM Overhead metric tracking - Add support for tracking litellm overhead on cache hits - [PR #15045](https://github.com/BerriAI/litellm/pull/15045)
238245

0 commit comments

Comments
 (0)