|
| 1 | +--- |
| 2 | +title: "v1.76.1-stable - Gemini 2.5 Flash Image" |
| 3 | +slug: "v1-76-1" |
| 4 | +date: 2025-08-30T10:00:00 |
| 5 | +authors: |
| 6 | + - name: Krrish Dholakia |
| 7 | + title: CEO, LiteLLM |
| 8 | + url: https://www.linkedin.com/in/krish-d/ |
| 9 | + image_url: https://pbs.twimg.com/profile_images/1298587542745358340/DZv3Oj-h_400x400.jpg |
| 10 | + - name: Ishaan Jaffer |
| 11 | + title: CTO, LiteLLM |
| 12 | + url: https://www.linkedin.com/in/reffajnaahsi/ |
| 13 | + image_url: https://pbs.twimg.com/profile_images/1613813310264340481/lz54oEiB_400x400.jpg |
| 14 | + |
| 15 | +hide_table_of_contents: false |
| 16 | +--- |
| 17 | + |
| 18 | +import Image from '@theme/IdealImage'; |
| 19 | +import Tabs from '@theme/Tabs'; |
| 20 | +import TabItem from '@theme/TabItem'; |
| 21 | + |
| 22 | +## Deploy this version |
| 23 | + |
| 24 | +<Tabs> |
| 25 | +<TabItem value="docker" label="Docker"> |
| 26 | + |
| 27 | +``` showLineNumbers title="docker run litellm" |
| 28 | +docker run \ |
| 29 | +-e STORE_MODEL_IN_DB=True \ |
| 30 | +-p 4000:4000 \ |
| 31 | +ghcr.io/berriai/litellm:v1.76.1 |
| 32 | +``` |
| 33 | +</TabItem> |
| 34 | + |
| 35 | +<TabItem value="pip" label="Pip"> |
| 36 | + |
| 37 | +``` showLineNumbers title="pip install litellm" |
| 38 | +pip install litellm==1.76.1 |
| 39 | +``` |
| 40 | + |
| 41 | +</TabItem> |
| 42 | +</Tabs> |
| 43 | + |
| 44 | +--- |
| 45 | + |
| 46 | +## Key Highlights |
| 47 | + |
| 48 | +- **Major Performance Improvements** - 6.5x faster LiteLLM Python SDK completion with fastuuid integration. |
| 49 | +- **New Model Support** - Gemini 2.5 Flash Image Preview, Grok Code Fast, and GPT Realtime models |
| 50 | +- **Enhanced Provider Support** - DeepSeek-v3.1 pricing on Fireworks AI, Vercel AI Gateway, and improved Anthropic/GitHub Copilot integration |
| 51 | +- **MCP Improvements** - Better connection testing and SSE MCP tools bug fixes |
| 52 | + |
| 53 | +## Major Changes |
| 54 | +- Added support for using Gemini 2.5 Flash Image Preview with /chat/completions. **🚨 Warning** If you were using `gemini-2.0-flash-exp-image-generation` please follow this migration guide. |
| 55 | + [Gemini Image Generation Migration Guide](../../docs/extras/gemini_img_migration) |
| 56 | +--- |
| 57 | + |
| 58 | +## Performance Improvements |
| 59 | + |
| 60 | +This release includes significant performance optimizations: |
| 61 | + |
| 62 | +- **6.5x faster LiteLLM Python SDK Completion** - Major performance boost for completion operations - [PR #13990](https://github.com/BerriAI/litellm/pull/13990) |
| 63 | +- **fastuuid Integration** - 2.1x faster UUID generation with +80 RPS improvement for /chat/completions and other LLM endpoints - [PR #13992](https://github.com/BerriAI/litellm/pull/13992), [PR #14016](https://github.com/BerriAI/litellm/pull/14016) |
| 64 | +- **Optimized Request Logging** - Don't print request params by default for +50 RPS improvement - [PR #14015](https://github.com/BerriAI/litellm/pull/14015) |
| 65 | +- **Cache Performance** - 21% speedup in InMemoryCache.evict_cache and 45% speedup in `_is_debugging_on` function - [PR #14012](https://github.com/BerriAI/litellm/pull/14012), [PR #13988](https://github.com/BerriAI/litellm/pull/13988) |
| 66 | + |
| 67 | +--- |
| 68 | + |
| 69 | +## New Models / Updated Models |
| 70 | + |
| 71 | +#### New Model Support |
| 72 | + |
| 73 | +| Provider | Model | Context Window | Input ($/1M tokens) | Output ($/1M tokens) | Features | |
| 74 | +| ----------- | -------------------------------------- | -------------- | ------------------- | -------------------- | -------- | |
| 75 | +| Google | `gemini-2.5-flash-image-preview` | 1M | $0.30 | $2.50 | Chat completions + image generation ($0.039/image) | |
| 76 | +| X.AI | `xai/grok-code-fast` | 256K | $0.20 | $1.50 | Code generation | |
| 77 | +| OpenAI | `gpt-realtime` | 32K | $4.00 | $16.00 | Real-time conversation + audio | |
| 78 | +| Vercel AI Gateway | `vercel_ai_gateway/openai/o3` | 200K | $2.00 | $8.00 | Advanced reasoning | |
| 79 | +| Vercel AI Gateway | `vercel_ai_gateway/openai/o3-mini` | 200K | $1.10 | $4.40 | Efficient reasoning | |
| 80 | +| Vercel AI Gateway | `vercel_ai_gateway/openai/o4-mini` | 200K | $1.10 | $4.40 | Latest mini model | |
| 81 | +| DeepInfra | `deepinfra/zai-org/GLM-4.5` | 131K | $0.55 | $2.00 | Chat completions | |
| 82 | +| Perplexity | `perplexity/codellama-34b-instruct` | 16K | $0.35 | $1.40 | Code generation | |
| 83 | +| Fireworks AI | `fireworks_ai/accounts/fireworks/models/deepseek-v3p1` | 128K | $0.56 | $1.68 | Chat completions | |
| 84 | + |
| 85 | +**Additional Models Added:** Various other Vercel AI Gateway models were added too. See [models.litellm.ai](https://models.litellm.ai) for the full list. |
| 86 | + |
| 87 | +#### Features |
| 88 | + |
| 89 | +- **[Google Gemini](../../docs/providers/gemini)** |
| 90 | + - Added support for `gemini-2.5-flash-image-preview` with image return capability - [PR #13979](https://github.com/BerriAI/litellm/pull/13979), [PR #13983](https://github.com/BerriAI/litellm/pull/13983) |
| 91 | + - Support for requests with only system prompt - [PR #14010](https://github.com/BerriAI/litellm/pull/14010) |
| 92 | + - Fixed invalid model name error for Gemini Imagen models - [PR #13991](https://github.com/BerriAI/litellm/pull/13991) |
| 93 | +- **[X.AI](../../docs/providers/xai)** |
| 94 | + - Added `xai/grok-code-fast` model family support - [PR #14054](https://github.com/BerriAI/litellm/pull/14054) |
| 95 | + - Fixed frequency_penalty parameter for grok-4 models - [PR #14078](https://github.com/BerriAI/litellm/pull/14078) |
| 96 | +- **[OpenAI](../../docs/providers/openai)** |
| 97 | + - Added support for gpt-realtime models - [PR #14082](https://github.com/BerriAI/litellm/pull/14082) |
| 98 | + - Support for reasoning and reasoning_effort parameters by default - [PR #12865](https://github.com/BerriAI/litellm/pull/12865) |
| 99 | +- **[Fireworks AI](../../docs/providers/fireworks_ai)** |
| 100 | + - Added DeepSeek-v3.1 pricing - [PR #13958](https://github.com/BerriAI/litellm/pull/13958) |
| 101 | +- **[DeepInfra](../../docs/providers/deepinfra)** |
| 102 | + - Fixed reasoning_effort setting for DeepSeek-V3.1 - [PR #14053](https://github.com/BerriAI/litellm/pull/14053) |
| 103 | +- **[GitHub Copilot](../../docs/providers/github_copilot)** |
| 104 | + - Added support for thinking and reasoning_effort parameters - [PR #13691](https://github.com/BerriAI/litellm/pull/13691) |
| 105 | + - Added image headers support - [PR #13955](https://github.com/BerriAI/litellm/pull/13955) |
| 106 | +- **[Anthropic](../../docs/providers/anthropic)** |
| 107 | + - Support for custom Anthropic-compatible API endpoints - [PR #13945](https://github.com/BerriAI/litellm/pull/13945) |
| 108 | + - Fixed /messages fallback from Anthropic API to Bedrock API - [PR #13946](https://github.com/BerriAI/litellm/pull/13946) |
| 109 | +- **[Nebius](../../docs/providers/nebius)** |
| 110 | + - Expanded provider models and normalized model IDs - [PR #13965](https://github.com/BerriAI/litellm/pull/13965) |
| 111 | +- **[Vertex AI](../../docs/providers/vertex)** |
| 112 | + - Fixed Vertex Mistral streaming issues - [PR #13952](https://github.com/BerriAI/litellm/pull/13952) |
| 113 | + - Fixed anyOf corner cases for Gemini tool calls - [PR #12797](https://github.com/BerriAI/litellm/pull/12797) |
| 114 | +- **[Bedrock](../../docs/providers/bedrock)** |
| 115 | + - Fixed structure output issues - [PR #14005](https://github.com/BerriAI/litellm/pull/14005) |
| 116 | +- **[OpenRouter](../../docs/providers/openrouter)** |
| 117 | + - Added GPT-5 family models pricing - [PR #13536](https://github.com/BerriAI/litellm/pull/13536) |
| 118 | + |
| 119 | +#### New Provider Support |
| 120 | + |
| 121 | +- **[Vercel AI Gateway](../../docs/providers/vercel_ai_gateway)** |
| 122 | + - New provider support added - [PR #13144](https://github.com/BerriAI/litellm/pull/13144) |
| 123 | +- **[DataRobot](../../docs/providers/datarobot)** |
| 124 | + - Added provider documentation - [PR #14038](https://github.com/BerriAI/litellm/pull/14038), [PR #14074](https://github.com/BerriAI/litellm/pull/14074) |
| 125 | + |
| 126 | +--- |
| 127 | + |
| 128 | +## LLM API Endpoints |
| 129 | + |
| 130 | +#### Features |
| 131 | + |
| 132 | +- **[Images API](../../docs/image_generation)** |
| 133 | + - Support for multiple images in OpenAI images/edits endpoint - [PR #13916](https://github.com/BerriAI/litellm/pull/13916) |
| 134 | + - Allow using dynamic `api_key` for image generation requests - [PR #14007](https://github.com/BerriAI/litellm/pull/14007) |
| 135 | +- **[Responses API](../../docs/response_api)** |
| 136 | + - Fixed `/responses` endpoint ignoring extra_headers in GitHub Copilot - [PR #13775](https://github.com/BerriAI/litellm/pull/13775) |
| 137 | + - Added support for new web_search tool - [PR #14083](https://github.com/BerriAI/litellm/pull/14083) |
| 138 | +- **[Azure Passthrough](../../docs/providers/azure/azure)** |
| 139 | + - Fixed Azure Passthrough request with streaming - [PR #13831](https://github.com/BerriAI/litellm/pull/13831) |
| 140 | + |
| 141 | +#### Bugs |
| 142 | + |
| 143 | +- **General** |
| 144 | + - Fixed handling of None metadata in batch requests - [PR #13996](https://github.com/BerriAI/litellm/pull/13996) |
| 145 | + - Fixed token_counter with special token input - [PR #13374](https://github.com/BerriAI/litellm/pull/13374) |
| 146 | + - Removed incorrect web search support for azure/gpt-4.1 family - [PR #13566](https://github.com/BerriAI/litellm/pull/13566) |
| 147 | + |
| 148 | +--- |
| 149 | + |
| 150 | +## [MCP Gateway](../../docs/mcp) |
| 151 | + |
| 152 | +#### Features |
| 153 | + |
| 154 | +- **SSE MCP Tools** |
| 155 | + - Bug fix for adding SSE MCP tools - improved connection testing when adding MCPs - [PR #14048](https://github.com/BerriAI/litellm/pull/14048) |
| 156 | + |
| 157 | +[Read More](../../docs/mcp) |
| 158 | + |
| 159 | +--- |
| 160 | + |
| 161 | +## Management Endpoints / UI |
| 162 | + |
| 163 | +#### Features |
| 164 | + |
| 165 | +- **Team Management** |
| 166 | + - Allow setting Team Member RPM/TPM limits when creating a team - [PR #13943](https://github.com/BerriAI/litellm/pull/13943) |
| 167 | +- **UI Improvements** |
| 168 | + - Fixed Next.js Security Vulnerabilities in UI Dashboard - [PR #14084](https://github.com/BerriAI/litellm/pull/14084) |
| 169 | + - Fixed collapsible navbar design - [PR #14075](https://github.com/BerriAI/litellm/pull/14075) |
| 170 | + |
| 171 | +#### Bugs |
| 172 | + |
| 173 | +- **Authentication** |
| 174 | + - Fixed Virtual keys with llm_api type causing Internal Server Error for /anthropic/* and other LLM passthrough routes - [PR #14046](https://github.com/BerriAI/litellm/pull/14046) |
| 175 | + |
| 176 | +--- |
| 177 | + |
| 178 | +## Logging / Guardrail Integrations |
| 179 | + |
| 180 | +#### Features |
| 181 | + |
| 182 | +- **[Langfuse OTEL](../../docs/proxy/logging#langfuse)** |
| 183 | + - Allow using LANGFUSE_OTEL_HOST for configuring host - [PR #14013](https://github.com/BerriAI/litellm/pull/14013) |
| 184 | +- **[Braintrust](../../docs/proxy/logging#braintrust)** |
| 185 | + - Added span name metadata feature - [PR #13573](https://github.com/BerriAI/litellm/pull/13573) |
| 186 | + - Fixed tests to reference moved attributes in `braintrust_logging` module - [PR #13978](https://github.com/BerriAI/litellm/pull/13978) |
| 187 | +- **[OpenMeter](../../docs/proxy/logging#openmeter)** |
| 188 | + - Set user from token user_id for OpenMeter integration - [PR #13152](https://github.com/BerriAI/litellm/pull/13152) |
| 189 | + |
| 190 | +#### New Guardrail Support |
| 191 | + |
| 192 | +- **[Noma Security](../../docs/proxy/guardrails)** |
| 193 | + - Added Noma Security guardrail support - [PR #13572](https://github.com/BerriAI/litellm/pull/13572) |
| 194 | +- **[Pangea](../../docs/proxy/guardrails)** |
| 195 | + - Updated Pangea Guardrail to support new AIDR endpoint - [PR #13160](https://github.com/BerriAI/litellm/pull/13160) |
| 196 | + |
| 197 | +--- |
| 198 | + |
| 199 | +## Performance / Loadbalancing / Reliability improvements |
| 200 | + |
| 201 | +#### Features |
| 202 | + |
| 203 | +- **Caching** |
| 204 | + - Verify if cache entry has expired prior to serving it to client - [PR #13933](https://github.com/BerriAI/litellm/pull/13933) |
| 205 | + - Fixed error saving latency as timedelta on Redis - [PR #14040](https://github.com/BerriAI/litellm/pull/14040) |
| 206 | +- **Router** |
| 207 | + - Refactored router to choose weights by 'weight', 'rpm', 'tpm' in one loop for simple_shuffle - [PR #13562](https://github.com/BerriAI/litellm/pull/13562) |
| 208 | +- **Logging** |
| 209 | + - Fixed LoggingWorker graceful shutdown to prevent CancelledError warnings - [PR #14050](https://github.com/BerriAI/litellm/pull/14050) |
| 210 | + - Enhanced logging for containers to log on files both with usual format and json format - [PR #13394](https://github.com/BerriAI/litellm/pull/13394) |
| 211 | + |
| 212 | +#### Bugs |
| 213 | + |
| 214 | +- **Dependencies** |
| 215 | + - Bumped `orjson` version to "3.11.2" - [PR #13969](https://github.com/BerriAI/litellm/pull/13969) |
| 216 | + |
| 217 | +--- |
| 218 | + |
| 219 | +## General Proxy Improvements |
| 220 | + |
| 221 | +#### Features |
| 222 | + |
| 223 | +- **AWS** |
| 224 | + - Add support for AWS assume_role with a session token - [PR #13919](https://github.com/BerriAI/litellm/pull/13919) |
| 225 | +- **OCI Provider** |
| 226 | + - Added oci_key_file as an optional_parameter - [PR #14036](https://github.com/BerriAI/litellm/pull/14036) |
| 227 | +- **Configuration** |
| 228 | + - Allow configuration to set threshold before request entry in spend log gets truncated - [PR #14042](https://github.com/BerriAI/litellm/pull/14042) |
| 229 | + - Enhanced proxy_config configuration: add support for existing configmap in Helm charts - [PR #14041](https://github.com/BerriAI/litellm/pull/14041) |
| 230 | +- **Docker** |
| 231 | + - Added back supervisor to non-root image - [PR #13922](https://github.com/BerriAI/litellm/pull/13922) |
| 232 | + |
| 233 | + |
| 234 | +--- |
| 235 | + |
| 236 | +## New Contributors |
| 237 | +* @ArthurRenault made their first contribution in [PR #13922](https://github.com/BerriAI/litellm/pull/13922) |
| 238 | +* @stevenmanton made their first contribution in [PR #13919](https://github.com/BerriAI/litellm/pull/13919) |
| 239 | +* @uc4w6c made their first contribution in [PR #13914](https://github.com/BerriAI/litellm/pull/13914) |
| 240 | +* @nielsbosma made their first contribution in [PR #13573](https://github.com/BerriAI/litellm/pull/13573) |
| 241 | +* @Yuki-Imajuku made their first contribution in [PR #13567](https://github.com/BerriAI/litellm/pull/13567) |
| 242 | +* @codeflash-ai[bot] made their first contribution in [PR #13988](https://github.com/BerriAI/litellm/pull/13988) |
| 243 | +* @ColeFrench made their first contribution in [PR #13978](https://github.com/BerriAI/litellm/pull/13978) |
| 244 | +* @dttran-glo made their first contribution in [PR #13969](https://github.com/BerriAI/litellm/pull/13969) |
| 245 | +* @manascb1344 made their first contribution in [PR #13965](https://github.com/BerriAI/litellm/pull/13965) |
| 246 | +* @DorZion made their first contribution in [PR #13572](https://github.com/BerriAI/litellm/pull/13572) |
| 247 | +* @edwardsamuel made their first contribution in [PR #13536](https://github.com/BerriAI/litellm/pull/13536) |
| 248 | +* @blahgeek made their first contribution in [PR #13374](https://github.com/BerriAI/litellm/pull/13374) |
| 249 | +* @Deviad made their first contribution in [PR #13394](https://github.com/BerriAI/litellm/pull/13394) |
| 250 | +* @XSAM made their first contribution in [PR #13775](https://github.com/BerriAI/litellm/pull/13775) |
| 251 | +* @KRRT7 made their first contribution in [PR #14012](https://github.com/BerriAI/litellm/pull/14012) |
| 252 | +* @ikaadil made their first contribution in [PR #13991](https://github.com/BerriAI/litellm/pull/13991) |
| 253 | +* @timelfrink made their first contribution in [PR #13691](https://github.com/BerriAI/litellm/pull/13691) |
| 254 | +* @qidu made their first contribution in [PR #13562](https://github.com/BerriAI/litellm/pull/13562) |
| 255 | +* @nagyv made their first contribution in [PR #13243](https://github.com/BerriAI/litellm/pull/13243) |
| 256 | +* @xywei made their first contribution in [PR #12885](https://github.com/BerriAI/litellm/pull/12885) |
| 257 | +* @ericgtkb made their first contribution in [PR #12797](https://github.com/BerriAI/litellm/pull/12797) |
| 258 | +* @NoWall57 made their first contribution in [PR #13945](https://github.com/BerriAI/litellm/pull/13945) |
| 259 | +* @lmwang9527 made their first contribution in [PR #14050](https://github.com/BerriAI/litellm/pull/14050) |
| 260 | +* @WilsonSunBritten made their first contribution in [PR #14042](https://github.com/BerriAI/litellm/pull/14042) |
| 261 | +* @Const-antine made their first contribution in [PR #14041](https://github.com/BerriAI/litellm/pull/14041) |
| 262 | +* @dmvieira made their first contribution in [PR #14040](https://github.com/BerriAI/litellm/pull/14040) |
| 263 | +* @gotsysdba made their first contribution in [PR #14036](https://github.com/BerriAI/litellm/pull/14036) |
| 264 | +* @moshemorad made their first contribution in [PR #14005](https://github.com/BerriAI/litellm/pull/14005) |
| 265 | +* @joshualipman123 made their first contribution in [PR #13144](https://github.com/BerriAI/litellm/pull/13144) |
| 266 | + |
| 267 | +--- |
| 268 | + |
| 269 | +## **[Full Changelog](https://github.com/BerriAI/litellm/compare/v1.76.0-nightly...v1.76.1)** |
0 commit comments