Skip to content

Commit ccb7ce0

Browse files
authored
[Docs] v1.76.1-stable (#14094)
* docs 1.76.1 * fix docs * docs 1.76.1 * New Model Support * docs fixes
1 parent 34275ab commit ccb7ce0

File tree

2 files changed

+274
-1
lines changed

2 files changed

+274
-1
lines changed

docs/my-website/docs/extras/gemini_img_migration.md

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44

55
Anyone using the following models with /chat/completions:
66
- `gemini/gemini-2.0-flash-exp-image-generation`
7-
- `vertex_ai/gemini-2.5-flash-image-preview`
7+
- `vertex_ai/gemini-2.0-flash-exp-image-generation`
88

99
## Key Change
1010

@@ -40,6 +40,10 @@ response = completion(
4040
image_url = response.choices[0].message.image["url"] # "data:image/png;base64,..."
4141
```
4242

43+
### Why the change?
44+
45+
Because the newer `gemini-2.5-flash-image-preview` model sends both text and image responses in the same response. This interface allows a developer to explicitly access the image or text components of the response. Before a developer would have needed to search through the message content to find the image generated by the model.
46+
4347
## Usage
4448

4549
### Using the Python SDK
Lines changed: 269 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,269 @@
1+
---
2+
title: "v1.76.1-stable - Gemini 2.5 Flash Image"
3+
slug: "v1-76-1"
4+
date: 2025-08-30T10:00:00
5+
authors:
6+
- name: Krrish Dholakia
7+
title: CEO, LiteLLM
8+
url: https://www.linkedin.com/in/krish-d/
9+
image_url: https://pbs.twimg.com/profile_images/1298587542745358340/DZv3Oj-h_400x400.jpg
10+
- name: Ishaan Jaffer
11+
title: CTO, LiteLLM
12+
url: https://www.linkedin.com/in/reffajnaahsi/
13+
image_url: https://pbs.twimg.com/profile_images/1613813310264340481/lz54oEiB_400x400.jpg
14+
15+
hide_table_of_contents: false
16+
---
17+
18+
import Image from '@theme/IdealImage';
19+
import Tabs from '@theme/Tabs';
20+
import TabItem from '@theme/TabItem';
21+
22+
## Deploy this version
23+
24+
<Tabs>
25+
<TabItem value="docker" label="Docker">
26+
27+
``` showLineNumbers title="docker run litellm"
28+
docker run \
29+
-e STORE_MODEL_IN_DB=True \
30+
-p 4000:4000 \
31+
ghcr.io/berriai/litellm:v1.76.1
32+
```
33+
</TabItem>
34+
35+
<TabItem value="pip" label="Pip">
36+
37+
``` showLineNumbers title="pip install litellm"
38+
pip install litellm==1.76.1
39+
```
40+
41+
</TabItem>
42+
</Tabs>
43+
44+
---
45+
46+
## Key Highlights
47+
48+
- **Major Performance Improvements** - 6.5x faster LiteLLM Python SDK completion with fastuuid integration.
49+
- **New Model Support** - Gemini 2.5 Flash Image Preview, Grok Code Fast, and GPT Realtime models
50+
- **Enhanced Provider Support** - DeepSeek-v3.1 pricing on Fireworks AI, Vercel AI Gateway, and improved Anthropic/GitHub Copilot integration
51+
- **MCP Improvements** - Better connection testing and SSE MCP tools bug fixes
52+
53+
## Major Changes
54+
- Added support for using Gemini 2.5 Flash Image Preview with /chat/completions. **🚨 Warning** If you were using `gemini-2.0-flash-exp-image-generation` please follow this migration guide.
55+
[Gemini Image Generation Migration Guide](../../docs/extras/gemini_img_migration)
56+
---
57+
58+
## Performance Improvements
59+
60+
This release includes significant performance optimizations:
61+
62+
- **6.5x faster LiteLLM Python SDK Completion** - Major performance boost for completion operations - [PR #13990](https://github.com/BerriAI/litellm/pull/13990)
63+
- **fastuuid Integration** - 2.1x faster UUID generation with +80 RPS improvement for /chat/completions and other LLM endpoints - [PR #13992](https://github.com/BerriAI/litellm/pull/13992), [PR #14016](https://github.com/BerriAI/litellm/pull/14016)
64+
- **Optimized Request Logging** - Don't print request params by default for +50 RPS improvement - [PR #14015](https://github.com/BerriAI/litellm/pull/14015)
65+
- **Cache Performance** - 21% speedup in InMemoryCache.evict_cache and 45% speedup in `_is_debugging_on` function - [PR #14012](https://github.com/BerriAI/litellm/pull/14012), [PR #13988](https://github.com/BerriAI/litellm/pull/13988)
66+
67+
---
68+
69+
## New Models / Updated Models
70+
71+
#### New Model Support
72+
73+
| Provider | Model | Context Window | Input ($/1M tokens) | Output ($/1M tokens) | Features |
74+
| ----------- | -------------------------------------- | -------------- | ------------------- | -------------------- | -------- |
75+
| Google | `gemini-2.5-flash-image-preview` | 1M | $0.30 | $2.50 | Chat completions + image generation ($0.039/image) |
76+
| X.AI | `xai/grok-code-fast` | 256K | $0.20 | $1.50 | Code generation |
77+
| OpenAI | `gpt-realtime` | 32K | $4.00 | $16.00 | Real-time conversation + audio |
78+
| Vercel AI Gateway | `vercel_ai_gateway/openai/o3` | 200K | $2.00 | $8.00 | Advanced reasoning |
79+
| Vercel AI Gateway | `vercel_ai_gateway/openai/o3-mini` | 200K | $1.10 | $4.40 | Efficient reasoning |
80+
| Vercel AI Gateway | `vercel_ai_gateway/openai/o4-mini` | 200K | $1.10 | $4.40 | Latest mini model |
81+
| DeepInfra | `deepinfra/zai-org/GLM-4.5` | 131K | $0.55 | $2.00 | Chat completions |
82+
| Perplexity | `perplexity/codellama-34b-instruct` | 16K | $0.35 | $1.40 | Code generation |
83+
| Fireworks AI | `fireworks_ai/accounts/fireworks/models/deepseek-v3p1` | 128K | $0.56 | $1.68 | Chat completions |
84+
85+
**Additional Models Added:** Various other Vercel AI Gateway models were added too. See [models.litellm.ai](https://models.litellm.ai) for the full list.
86+
87+
#### Features
88+
89+
- **[Google Gemini](../../docs/providers/gemini)**
90+
- Added support for `gemini-2.5-flash-image-preview` with image return capability - [PR #13979](https://github.com/BerriAI/litellm/pull/13979), [PR #13983](https://github.com/BerriAI/litellm/pull/13983)
91+
- Support for requests with only system prompt - [PR #14010](https://github.com/BerriAI/litellm/pull/14010)
92+
- Fixed invalid model name error for Gemini Imagen models - [PR #13991](https://github.com/BerriAI/litellm/pull/13991)
93+
- **[X.AI](../../docs/providers/xai)**
94+
- Added `xai/grok-code-fast` model family support - [PR #14054](https://github.com/BerriAI/litellm/pull/14054)
95+
- Fixed frequency_penalty parameter for grok-4 models - [PR #14078](https://github.com/BerriAI/litellm/pull/14078)
96+
- **[OpenAI](../../docs/providers/openai)**
97+
- Added support for gpt-realtime models - [PR #14082](https://github.com/BerriAI/litellm/pull/14082)
98+
- Support for reasoning and reasoning_effort parameters by default - [PR #12865](https://github.com/BerriAI/litellm/pull/12865)
99+
- **[Fireworks AI](../../docs/providers/fireworks_ai)**
100+
- Added DeepSeek-v3.1 pricing - [PR #13958](https://github.com/BerriAI/litellm/pull/13958)
101+
- **[DeepInfra](../../docs/providers/deepinfra)**
102+
- Fixed reasoning_effort setting for DeepSeek-V3.1 - [PR #14053](https://github.com/BerriAI/litellm/pull/14053)
103+
- **[GitHub Copilot](../../docs/providers/github_copilot)**
104+
- Added support for thinking and reasoning_effort parameters - [PR #13691](https://github.com/BerriAI/litellm/pull/13691)
105+
- Added image headers support - [PR #13955](https://github.com/BerriAI/litellm/pull/13955)
106+
- **[Anthropic](../../docs/providers/anthropic)**
107+
- Support for custom Anthropic-compatible API endpoints - [PR #13945](https://github.com/BerriAI/litellm/pull/13945)
108+
- Fixed /messages fallback from Anthropic API to Bedrock API - [PR #13946](https://github.com/BerriAI/litellm/pull/13946)
109+
- **[Nebius](../../docs/providers/nebius)**
110+
- Expanded provider models and normalized model IDs - [PR #13965](https://github.com/BerriAI/litellm/pull/13965)
111+
- **[Vertex AI](../../docs/providers/vertex)**
112+
- Fixed Vertex Mistral streaming issues - [PR #13952](https://github.com/BerriAI/litellm/pull/13952)
113+
- Fixed anyOf corner cases for Gemini tool calls - [PR #12797](https://github.com/BerriAI/litellm/pull/12797)
114+
- **[Bedrock](../../docs/providers/bedrock)**
115+
- Fixed structure output issues - [PR #14005](https://github.com/BerriAI/litellm/pull/14005)
116+
- **[OpenRouter](../../docs/providers/openrouter)**
117+
- Added GPT-5 family models pricing - [PR #13536](https://github.com/BerriAI/litellm/pull/13536)
118+
119+
#### New Provider Support
120+
121+
- **[Vercel AI Gateway](../../docs/providers/vercel_ai_gateway)**
122+
- New provider support added - [PR #13144](https://github.com/BerriAI/litellm/pull/13144)
123+
- **[DataRobot](../../docs/providers/datarobot)**
124+
- Added provider documentation - [PR #14038](https://github.com/BerriAI/litellm/pull/14038), [PR #14074](https://github.com/BerriAI/litellm/pull/14074)
125+
126+
---
127+
128+
## LLM API Endpoints
129+
130+
#### Features
131+
132+
- **[Images API](../../docs/image_generation)**
133+
- Support for multiple images in OpenAI images/edits endpoint - [PR #13916](https://github.com/BerriAI/litellm/pull/13916)
134+
- Allow using dynamic `api_key` for image generation requests - [PR #14007](https://github.com/BerriAI/litellm/pull/14007)
135+
- **[Responses API](../../docs/response_api)**
136+
- Fixed `/responses` endpoint ignoring extra_headers in GitHub Copilot - [PR #13775](https://github.com/BerriAI/litellm/pull/13775)
137+
- Added support for new web_search tool - [PR #14083](https://github.com/BerriAI/litellm/pull/14083)
138+
- **[Azure Passthrough](../../docs/providers/azure/azure)**
139+
- Fixed Azure Passthrough request with streaming - [PR #13831](https://github.com/BerriAI/litellm/pull/13831)
140+
141+
#### Bugs
142+
143+
- **General**
144+
- Fixed handling of None metadata in batch requests - [PR #13996](https://github.com/BerriAI/litellm/pull/13996)
145+
- Fixed token_counter with special token input - [PR #13374](https://github.com/BerriAI/litellm/pull/13374)
146+
- Removed incorrect web search support for azure/gpt-4.1 family - [PR #13566](https://github.com/BerriAI/litellm/pull/13566)
147+
148+
---
149+
150+
## [MCP Gateway](../../docs/mcp)
151+
152+
#### Features
153+
154+
- **SSE MCP Tools**
155+
- Bug fix for adding SSE MCP tools - improved connection testing when adding MCPs - [PR #14048](https://github.com/BerriAI/litellm/pull/14048)
156+
157+
[Read More](../../docs/mcp)
158+
159+
---
160+
161+
## Management Endpoints / UI
162+
163+
#### Features
164+
165+
- **Team Management**
166+
- Allow setting Team Member RPM/TPM limits when creating a team - [PR #13943](https://github.com/BerriAI/litellm/pull/13943)
167+
- **UI Improvements**
168+
- Fixed Next.js Security Vulnerabilities in UI Dashboard - [PR #14084](https://github.com/BerriAI/litellm/pull/14084)
169+
- Fixed collapsible navbar design - [PR #14075](https://github.com/BerriAI/litellm/pull/14075)
170+
171+
#### Bugs
172+
173+
- **Authentication**
174+
- Fixed Virtual keys with llm_api type causing Internal Server Error for /anthropic/* and other LLM passthrough routes - [PR #14046](https://github.com/BerriAI/litellm/pull/14046)
175+
176+
---
177+
178+
## Logging / Guardrail Integrations
179+
180+
#### Features
181+
182+
- **[Langfuse OTEL](../../docs/proxy/logging#langfuse)**
183+
- Allow using LANGFUSE_OTEL_HOST for configuring host - [PR #14013](https://github.com/BerriAI/litellm/pull/14013)
184+
- **[Braintrust](../../docs/proxy/logging#braintrust)**
185+
- Added span name metadata feature - [PR #13573](https://github.com/BerriAI/litellm/pull/13573)
186+
- Fixed tests to reference moved attributes in `braintrust_logging` module - [PR #13978](https://github.com/BerriAI/litellm/pull/13978)
187+
- **[OpenMeter](../../docs/proxy/logging#openmeter)**
188+
- Set user from token user_id for OpenMeter integration - [PR #13152](https://github.com/BerriAI/litellm/pull/13152)
189+
190+
#### New Guardrail Support
191+
192+
- **[Noma Security](../../docs/proxy/guardrails)**
193+
- Added Noma Security guardrail support - [PR #13572](https://github.com/BerriAI/litellm/pull/13572)
194+
- **[Pangea](../../docs/proxy/guardrails)**
195+
- Updated Pangea Guardrail to support new AIDR endpoint - [PR #13160](https://github.com/BerriAI/litellm/pull/13160)
196+
197+
---
198+
199+
## Performance / Loadbalancing / Reliability improvements
200+
201+
#### Features
202+
203+
- **Caching**
204+
- Verify if cache entry has expired prior to serving it to client - [PR #13933](https://github.com/BerriAI/litellm/pull/13933)
205+
- Fixed error saving latency as timedelta on Redis - [PR #14040](https://github.com/BerriAI/litellm/pull/14040)
206+
- **Router**
207+
- Refactored router to choose weights by 'weight', 'rpm', 'tpm' in one loop for simple_shuffle - [PR #13562](https://github.com/BerriAI/litellm/pull/13562)
208+
- **Logging**
209+
- Fixed LoggingWorker graceful shutdown to prevent CancelledError warnings - [PR #14050](https://github.com/BerriAI/litellm/pull/14050)
210+
- Enhanced logging for containers to log on files both with usual format and json format - [PR #13394](https://github.com/BerriAI/litellm/pull/13394)
211+
212+
#### Bugs
213+
214+
- **Dependencies**
215+
- Bumped `orjson` version to "3.11.2" - [PR #13969](https://github.com/BerriAI/litellm/pull/13969)
216+
217+
---
218+
219+
## General Proxy Improvements
220+
221+
#### Features
222+
223+
- **AWS**
224+
- Add support for AWS assume_role with a session token - [PR #13919](https://github.com/BerriAI/litellm/pull/13919)
225+
- **OCI Provider**
226+
- Added oci_key_file as an optional_parameter - [PR #14036](https://github.com/BerriAI/litellm/pull/14036)
227+
- **Configuration**
228+
- Allow configuration to set threshold before request entry in spend log gets truncated - [PR #14042](https://github.com/BerriAI/litellm/pull/14042)
229+
- Enhanced proxy_config configuration: add support for existing configmap in Helm charts - [PR #14041](https://github.com/BerriAI/litellm/pull/14041)
230+
- **Docker**
231+
- Added back supervisor to non-root image - [PR #13922](https://github.com/BerriAI/litellm/pull/13922)
232+
233+
234+
---
235+
236+
## New Contributors
237+
* @ArthurRenault made their first contribution in [PR #13922](https://github.com/BerriAI/litellm/pull/13922)
238+
* @stevenmanton made their first contribution in [PR #13919](https://github.com/BerriAI/litellm/pull/13919)
239+
* @uc4w6c made their first contribution in [PR #13914](https://github.com/BerriAI/litellm/pull/13914)
240+
* @nielsbosma made their first contribution in [PR #13573](https://github.com/BerriAI/litellm/pull/13573)
241+
* @Yuki-Imajuku made their first contribution in [PR #13567](https://github.com/BerriAI/litellm/pull/13567)
242+
* @codeflash-ai[bot] made their first contribution in [PR #13988](https://github.com/BerriAI/litellm/pull/13988)
243+
* @ColeFrench made their first contribution in [PR #13978](https://github.com/BerriAI/litellm/pull/13978)
244+
* @dttran-glo made their first contribution in [PR #13969](https://github.com/BerriAI/litellm/pull/13969)
245+
* @manascb1344 made their first contribution in [PR #13965](https://github.com/BerriAI/litellm/pull/13965)
246+
* @DorZion made their first contribution in [PR #13572](https://github.com/BerriAI/litellm/pull/13572)
247+
* @edwardsamuel made their first contribution in [PR #13536](https://github.com/BerriAI/litellm/pull/13536)
248+
* @blahgeek made their first contribution in [PR #13374](https://github.com/BerriAI/litellm/pull/13374)
249+
* @Deviad made their first contribution in [PR #13394](https://github.com/BerriAI/litellm/pull/13394)
250+
* @XSAM made their first contribution in [PR #13775](https://github.com/BerriAI/litellm/pull/13775)
251+
* @KRRT7 made their first contribution in [PR #14012](https://github.com/BerriAI/litellm/pull/14012)
252+
* @ikaadil made their first contribution in [PR #13991](https://github.com/BerriAI/litellm/pull/13991)
253+
* @timelfrink made their first contribution in [PR #13691](https://github.com/BerriAI/litellm/pull/13691)
254+
* @qidu made their first contribution in [PR #13562](https://github.com/BerriAI/litellm/pull/13562)
255+
* @nagyv made their first contribution in [PR #13243](https://github.com/BerriAI/litellm/pull/13243)
256+
* @xywei made their first contribution in [PR #12885](https://github.com/BerriAI/litellm/pull/12885)
257+
* @ericgtkb made their first contribution in [PR #12797](https://github.com/BerriAI/litellm/pull/12797)
258+
* @NoWall57 made their first contribution in [PR #13945](https://github.com/BerriAI/litellm/pull/13945)
259+
* @lmwang9527 made their first contribution in [PR #14050](https://github.com/BerriAI/litellm/pull/14050)
260+
* @WilsonSunBritten made their first contribution in [PR #14042](https://github.com/BerriAI/litellm/pull/14042)
261+
* @Const-antine made their first contribution in [PR #14041](https://github.com/BerriAI/litellm/pull/14041)
262+
* @dmvieira made their first contribution in [PR #14040](https://github.com/BerriAI/litellm/pull/14040)
263+
* @gotsysdba made their first contribution in [PR #14036](https://github.com/BerriAI/litellm/pull/14036)
264+
* @moshemorad made their first contribution in [PR #14005](https://github.com/BerriAI/litellm/pull/14005)
265+
* @joshualipman123 made their first contribution in [PR #13144](https://github.com/BerriAI/litellm/pull/13144)
266+
267+
---
268+
269+
## **[Full Changelog](https://github.com/BerriAI/litellm/compare/v1.76.0-nightly...v1.76.1)**

0 commit comments

Comments
 (0)