@@ -54,6 +54,7 @@ pip install litellm==1.77.7.rc.1
54
54
55
55
## Key Highlights
56
56
57
+ - ** Dynamic Rate Limiter v3** - Automatically maximizes throughput when capacity is available (< 80% saturation) by allowing lower-priority requests to use unused capacity, then switches to fair priority-based allocation under high load (≥ 80%) to prevent blocking
57
58
- ** Major Performance Improvements** - Router optimization reducing P99 latency by 62.5%, cache improvements from O(n* log(n)) to O(log(n))
58
59
- ** Claude Sonnet 4.5** - Support for Anthropic's new Claude Sonnet 4.5 model family with 200K+ context and tiered pricing
59
60
- ** MCP Gateway Enhancements** - Fine-grained tool control, server permissions, and forwardable headers
@@ -85,14 +86,9 @@ pip install litellm==1.77.7.rc.1
85
86
- Add cost tracking for /v1/messages in streaming response - [ PR #15102 ] ( https://github.com/BerriAI/litellm/pull/15102 )
86
87
- Add /v1/messages/count_tokens to Anthropic routes for non-admin user access - [ PR #15034 ] ( https://github.com/BerriAI/litellm/pull/15034 )
87
88
- ** [ Gemini] ( ../../docs/providers/gemini ) **
88
- - Add full support for native Gemini API translation - [ PR #15029 ] ( https://github.com/BerriAI/litellm/pull/15029 )
89
- - Add Gemini generateContent passthrough cost tracking - [ PR #15014 ] ( https://github.com/BerriAI/litellm/pull/15014 )
90
- - Add streamGenerateContent cost tracking in passthrough - [ PR #15199 ] ( https://github.com/BerriAI/litellm/pull/15199 )
91
89
- Ignore type param for gemini tools - [ PR #15022 ] ( https://github.com/BerriAI/litellm/pull/15022 )
92
90
- ** [ Vertex AI] ( ../../docs/providers/vertex ) **
93
91
- Add LiteLLM Overhead metric for VertexAI - [ PR #15040 ] ( https://github.com/BerriAI/litellm/pull/15040 )
94
- - Add cost tracking for Vertex AI Passthrough ` /predict ` endpoint - [ PR #15019 ] ( https://github.com/BerriAI/litellm/pull/15019 )
95
- - Add cost tracking for Vertex AI Live API WebSocket Passthrough - [ PR #14956 ] ( https://github.com/BerriAI/litellm/pull/14956 )
96
92
- Support googlemap grounding in vertex ai - [ PR #15179 ] ( https://github.com/BerriAI/litellm/pull/15179 )
97
93
- ** [ Azure] ( ../../docs/providers/azure ) **
98
94
- Add azure_ai grok-4 model family - [ PR #15137 ] ( https://github.com/BerriAI/litellm/pull/15137 )
@@ -140,9 +136,21 @@ pip install litellm==1.77.7.rc.1
140
136
- ** [ Responses API] ( ../../docs/response_api ) **
141
137
- Return Cost for Responses API Streaming requests - [ PR #15053 ] ( https://github.com/BerriAI/litellm/pull/15053 )
142
138
139
+ - ** [ /generateContent] ( ../../docs/providers/gemini ) **
140
+ - Add full support for native Gemini API translation - [ PR #15029 ] ( https://github.com/BerriAI/litellm/pull/15029 )
141
+
142
+ - ** Passthrough Gemini Routes**
143
+ - Add Gemini generateContent passthrough cost tracking - [ PR #15014 ] ( https://github.com/BerriAI/litellm/pull/15014 )
144
+ - Add streamGenerateContent cost tracking in passthrough - [ PR #15199 ] ( https://github.com/BerriAI/litellm/pull/15199 )
145
+
146
+ - ** Passthrough Vertex AI Routes**
147
+ - Add cost tracking for Vertex AI Passthrough ` /predict ` endpoint - [ PR #15019 ] ( https://github.com/BerriAI/litellm/pull/15019 )
148
+ - Add cost tracking for Vertex AI Live API WebSocket Passthrough - [ PR #14956 ] ( https://github.com/BerriAI/litellm/pull/14956 )
149
+
143
150
- ** General**
144
151
- Preserve Whitespace Characters in Model Response Streams - [ PR #15160 ] ( https://github.com/BerriAI/litellm/pull/15160 )
145
152
- Add provider name to payload specification - [ PR #15130 ] ( https://github.com/BerriAI/litellm/pull/15130 )
153
+ - Ensure query params are forwarded from origin url to downstream request - [ PR #15087 ] ( https://github.com/BerriAI/litellm/pull/15087 )
146
154
147
155
---
148
156
@@ -166,7 +174,6 @@ pip install litellm==1.77.7.rc.1
166
174
#### Bugs
167
175
168
176
- ** Dashboard** - Fix LiteLLM model name fallback in dashboard overview - [ PR #14998 ] ( https://github.com/BerriAI/litellm/pull/14998 )
169
- - ** Passthrough API** - Ensure query params are forwarded from origin url to downstream request - [ PR #15087 ] ( https://github.com/BerriAI/litellm/pull/15087 )
170
177
171
178
---
172
179
@@ -204,8 +211,6 @@ pip install litellm==1.77.7.rc.1
204
211
- Dynamic Rate Limiter v3 - fixes for detecting saturation + fixes for post saturation behavior - [ PR #15192 ] ( https://github.com/BerriAI/litellm/pull/15192 )
205
212
- ** Teams**
206
213
- Add model specific tpm/rpm limits to teams on LiteLLM - [ PR #15044 ] ( https://github.com/BerriAI/litellm/pull/15044 )
207
- - ** Configuration**
208
- - Add max requests env var - [ PR #15007 ] ( https://github.com/BerriAI/litellm/pull/15007 )
209
214
210
215
---
211
216
@@ -233,6 +238,8 @@ pip install litellm==1.77.7.rc.1
233
238
- ** Cache Optimizations**
234
239
- Reduce complexity of InMemoryCache.evict_cache from O(n* log(n)) to O(log(n)) - [ PR #15000 ] ( https://github.com/BerriAI/litellm/pull/15000 )
235
240
- Avoiding expensive operations when cache isn't available - [ PR #15182 ] ( https://github.com/BerriAI/litellm/pull/15182 )
241
+ - ** Worker Management**
242
+ - Add proxy CLI option to recycle workers after N requests - [ PR #15007 ] ( https://github.com/BerriAI/litellm/pull/15007 )
236
243
- ** Metrics & Monitoring**
237
244
- LiteLLM Overhead metric tracking - Add support for tracking litellm overhead on cache hits - [ PR #15045 ] ( https://github.com/BerriAI/litellm/pull/15045 )
238
245
0 commit comments