Skip to content

Commit 4bf4ff8

Browse files
Merge pull request #358 from Portkey-AI/may-changelog
2 parents ca87a95 + dbb577b commit 4bf4ff8

File tree

8 files changed

+228
-0
lines changed

8 files changed

+228
-0
lines changed

changelog/2025/may.mdx

Lines changed: 227 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,227 @@
1+
---
2+
title: "May"
3+
---
4+
5+
**May-king it production ready✨**
6+
7+
Infra shouldn’t slow you down. In May, we shipped the kind of upgrades that help you move fast into productiion and stay in control—whether you're scaling agents, securing AI behavior, or managing costs across teams.
8+
9+
From deeper integrations with agent frameworks to support for newer models Portkey keeps evolving as the AI infra layer teams can rely on. We also shipped observability upgrades, expanded our provider network, and added tighter controls for cost, access, and security.
10+
11+
Here’s everything new this month:
12+
13+
14+
## Summary
15+
16+
| Area | Key Updates |
17+
| :-- | :-- |
18+
| **AI agent infrastructure** | • PydanticAI integration for modular agent development<br/>• OpenAI Agents SDK support with monitoring, guardrails, and cost tracking<br/>• Strands Agents integration with observability, retries, and load balancing<br/>• Remote MCP server support via Responses API<br/>• Arize Phoenix tracing integration for unified agent observability |
19+
| **Platform** | • Deep integration into Azure AI ecosystem (OpenAI, Foundry, APIM, Marketplace)<br/>• Support for Claude Code with rate limits, observability, and access control<br/>• AI coding assistant integrations: Cline, Roo Code<br/>• Multimodal embedding support via Vertex AI (text, image, video)<br/>• Multi-label support for prompt versions<br/>• OpenAI Computer Use Tool routing and observability<br/>• Full support for `GET`, `PUT`, and `DELETE` HTTP methods<br/>• OTel analytics export to your existing observability stack |
20+
| **Improvements** | • Token cost tracking for gpt-image-1<br/>• Ping messages removed from streamed responses<br/>• Resizing metadata columns in logs |
21+
| **New models & providers** | • Claude 4 now live<br/>• PDF support for Claude via Anthropic and Bedrock<br/>• OpenAI’s Computer Use Tool supported via Responses API<br/>• Grok 3 and Grok 3 Mini on Azure<br/>• Gemini 2.5 Thinking Mode in Prompt Playground<br/>• Extended thinking for Claude 3.7 and Claude 4<br/>• Mistral supports function calling<br/>• WorkersAI supports image generation<br/>• Lepton AI, Nscale now integrated<br/>• Tool calling enabled for Mistral and OpenRouter<br/>• MIME type support for Vertex and Google<br/>• PDF support via Anthropic and Bedrock |
22+
| **Guardrails** | • Prompt Security guardrails for injection detection and sensitive data protection<br/>• JWT validator input guardrail<br/>• PANW Prisma AIRS plugin for real-time prompt/response risk blocking<br/>• Model whitelist guardrail for org/environment/request-level control |
23+
| **Documentation** | • Guardrail documentation moved under “Integrations”<br/>• New solution pages for AWS Bedrock and GovCloud<br/>• Cookbook: OpenAI Computer Use tool <br/>• Cookbook: Optimizing Prompts using LLama Prompt Ops |
24+
---
25+
26+
## AI agent infrastructure
27+
AI agent frameworks are helping teams prototype faster, but taking agents to production requires real infrastructure. Portkey integrates with leading frameworks to bring interoperability, observability, reliability, and cost management to your agent workflows.
28+
29+
**PydanticAI**
30+
31+
Portkey now integrates PydanticAI, a Python framework that brings FastAPI-like ergonomics to building AI agents. With Portkey, you can:
32+
33+
- Build modular, testable agents with a clean developer experience.
34+
- Route all agent calls through Portkey for observability and debugging.
35+
- Add retries, fallbacks, guardrails, and cost tracking without extra infra
36+
37+
See how it's done [here](https://portkey.ai/docs/integrations/agents/pydantic-ai#pydantic-ai)
38+
39+
**OpenAI Agents SDK**
40+
41+
Portkey integrates with the OpenAI Agents SDK to help teams ship production-grade agents with built-in planning, memory, and tool use. You can now:
42+
43+
- Monitor and debug each step of the agent’s reasoning and tool use.
44+
- Automatically track usage and cost for each agent call.
45+
- Apply guardrails to both agent input and output.
46+
- Scale agent-based workflows across environments with versioned control
47+
48+
Explore the integration [here](https://portkey.ai/docs/integrations/agents/openai-agents)
49+
50+
**Strands Agents**
51+
52+
Strands Agents is a lightweight agent framework built by AWS to simplify agent development.
53+
54+
Portkey now integrates seamlessly with Strands Agents to make them production-ready. With this integration, you get:
55+
56+
- Full observability into agent steps, tool calls, and interactions
57+
- Built-in reliability through fallbacks, retries, and load balancing
58+
- Cost tracking and spend optimization
59+
60+
See how it's done [here](https://portkey.ai/docs/integrations/agents/strands)
61+
62+
**Support for remote MCP servers!**
63+
64+
Portkey now supports calling a remote MCP server that is maintained by developers and organizations across the internet that expose these tools to MCP clients via the Responses API
65+
Read more about the integration[here](https://portkey.ai/docs/product/ai-gateway/remote-mcp)
66+
67+
**Tracing Integrations: Arize AI**
68+
69+
For teams consolidating observability into Arize, you can now view Portkey’s logs directly into Arize Phoenix to get unified trace views across your LLM workflows.
70+
71+
## Platform
72+
73+
**Azure AI ecosystem**
74+
<Frame>
75+
<img width="700" src="/images/changelog/azure-integration.jpeg" />
76+
</Frame>
77+
More than half of Fortune 500 companies use Azure OpenAI. But building GenAI apps in the enterprise is still messy, cost attribution, routing logic, usage tracking, model evaluation... all scattered.
78+
79+
With Portkey’s deep integration into the Azure AI ecosystem (OpenAI, Foundry, APIM, Marketplace), teams can now build, scale, and govern GenAI apps without leaving their existing cloud setup. Our customers are vouching for it!
80+
<Frame>
81+
<img width="700" src="/images/changelog/testimonial3.png" />
82+
</Frame>
83+
84+
85+
**Claude Code**
86+
87+
Bring visibility, governance, and control to Anthropic’s agentic coding assistant with Portkey.
88+
89+
With this integration, you can:
90+
91+
- Avoid system overload by enforcing rate limits
92+
- Monitor usage by tagging and filtering with metadata (e.g., user ID, workspace)
93+
- Debug and trace issues faster with detailed logs for every interaction
94+
- Share controlled, secure access by issuing virtual API keys per user
95+
- Use Claude code in your existing AWS Bedrock or Vertex AI setup, with granular governance and access control
96+
97+
**AI coding assistants**
98+
99+
Plug Portkey into [Cline](https://portkey.ai/docs/integrations/libraries/cline) or [Roo Code](https://portkey.ai/docs/integrations/libraries/roo-code) and enable:
100+
101+
- Access to the latest models from OpenAI, Anthropic, Mistral, and more
102+
- Full observability—log every prompt, tool use, and response with metadata
103+
- Access control with scoped API keys and JWT-based authentication
104+
- Built-in governance and cost tracking per user, project, or team
105+
106+
**Multilmodal embeddings**
107+
108+
Portkey now supports embedding APIs from Vertex AI for text, image, and video—across multiple languages.
109+
This unlocks the ability to:
110+
- Build cross-language search and retrieval
111+
- Power multimodal RAG pipelines
112+
- Track, route, and optimize embedding usage at scale
113+
114+
Read more about the implementation [here](https://portkey.ai/docs/integrations/llms/vertex-ai/embeddings)
115+
116+
**Multi-label support for prompts**
117+
118+
<Frame>
119+
<img width="700" src="/images/changelog/multi-label prompts.png" />
120+
</Frame>
121+
122+
You can now assign multiple labels to a single prompt version, making it easy to promote a version across environments like staging and production.
123+
124+
**OpenAI Computer Use Tool**
125+
126+
Build production-grade browser automation with enterprise-level controls using Portkey and:
127+
128+
- Route and monitor Computer Use API calls
129+
- Build a complete Playwright-based browser automation solution
130+
- Add observability, logging, and cost controls with Portkey
131+
132+
**Gateway to any API**
133+
134+
Portkey now supports `GET`, `PUT`, and `DELETE` HTTP methods in addition to `POST`, allowing you to route requests to any external or self-hosted provider endpoint. This means you can connect to custom APIs directly through Portkey with full observability for every call.
135+
136+
**OTel Integration (Analytics Data)**
137+
You can now export Portkey analytics to any OpenTelemetry (OTel)-compatible collector, integrating easily into your existing observability stack.
138+
139+
**Improvements**
140+
- Token cost tracking is now available for gpt-image-1.
141+
- Ping messages are removed from streamed responses.
142+
- Resizing metadata columns in logs
143+
144+
## New Models and Providers
145+
146+
- Claude 4 is now live on Portkey.
147+
- PDFs can be sent to Claude via Anthropic and Bedrock.
148+
- OpenAI’s Computer Use Tool works via the Responses API.
149+
- Grok 3 and Grok 3 Mini are available on Azure.
150+
- Gemini 2.5 supports Thinking Mode in Prompt Playground.
151+
- Extended thinking added for Claude 3.7 and Claude 4.
152+
- Mistral now supports function calling.
153+
- Image generation is now available on WorkersAI.
154+
- Lepton AI is now integrated with Portkey.
155+
- Nscale models can be accessed via Portkey.
156+
- Tool calling is live for Mistral and OpenRouter.
157+
- MIME types are now handled for Vertex and Google.
158+
- PDFs are supported via Anthropic and Bedrock routes.
159+
160+
161+
## Guardrails
162+
163+
- **Prompt Security guardrails**: Integrate with Prompt Security to detect prompt injection and prevent sensitive data exposure in both prompts and responses.
164+
165+
- **JWT validator guardrail**: Added as an input guardrail to validate incoming JWT tokens before requests are sent to the LLM.
166+
167+
- **PANW Prisma AIRS Plugin**:Portkey now integrates with Palo Alto Networks' AIRS (AI Runtime Security) to enforce guardrails that block risky prompts or model responses based on real-time security analysis.
168+
169+
- **Model whitelist guardrail**:Restrict or deny specific models at the org, environment, or request level using a flexible whitelist/blacklist guardrail.
170+
171+
## Documentation and Guides
172+
173+
**Optimizing Prompts using LLama Prompt Ops**
174+
175+
Need to try out or switch to the latest Llama models? There's an easier way to do it.
176+
Llama Prompt Ops transforms prompts that work well with other LLMs into ones that are optimized specifically for Llama models. This helps you get better performance and more reliable results without having to rewrite everything yourself.
177+
178+
If you work in customer support, we've put together a helpful guide that will show you how to build a system that analyzes support messages for urgency and sentiment, and helps categorize them properly.
179+
180+
Check it out [here](https://portkey.ai/docs/guides/prompts/llama-prompts)
181+
182+
**OpenAI’s Computer Use tool**
183+
184+
Build production-grade browser automation with enterprise-level controls. Our latest cookbook shows you how to:
185+
186+
- Route and monitor Computer Use API calls
187+
- Build a complete Playwright-based browser automation solution
188+
- Add observability, logging, and cost controls with Portkey
189+
190+
[Explore](https://portkey.ai/docs/guides/use-cases/openai-computer-use)
191+
192+
**Other updates**
193+
194+
- Guardrail documentation moved under “Integrations”.
195+
- Expanded guides for agent frameworks like CrewAI and LangGraph
196+
197+
198+
## Customer love!
199+
200+
<Frame>
201+
<img src="/images/changelog/testimonial1.png" width="100%"/>
202+
</Frame>
203+
<Frame>
204+
<img src="/images/changelog/lions-testimonial.png" width="100%"/>
205+
</Frame>
206+
207+
### Community Contributors
208+
209+
A special thanks to our community contributors this month:
210+
- [unsync](https://github.com/unsync)
211+
- [tomukmatthews](https://github.com/tomukmatthews)
212+
- [jroberts2600](https://github.com/jroberts2600)
213+
214+
## Coming this month!
215+
216+
Provision and manage LLM access across your entire org from a single admin panel. Centralized controls. Granular permissions. Enterprise-grade security. Stay tuned.
217+
218+
## Support
219+
220+
<CardGroup cols={2}>
221+
<Card title="Need Help?" icon="bug" href="https://github.com/Portkey-AI/gateway/issues">
222+
Open an issue on GitHub
223+
</Card>
224+
<Card title="Join Us" icon="discord" href="https://portkey.wiki/community">
225+
Get support in our Discord
226+
</Card>
227+
</CardGroup>

docs.json

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -866,6 +866,7 @@
866866
{
867867
"group": "2025",
868868
"pages": [
869+
"changelog/2025/may",
869870
"changelog/2025/apr",
870871
"changelog/2025/mar",
871872
"changelog/2025/feb",
110 KB
Loading
78.2 KB
Loading
335 KB
Loading

images/changelog/testimonial1.png

87.6 KB
Loading

images/changelog/testimonial3.png

49.5 KB
Loading

images/changelog/twstimonial3.png

137 KB
Loading

0 commit comments

Comments
 (0)