diff --git a/src/content/docs/agents/index.mdx b/src/content/docs/agents/index.mdx index c396697639ca9f..0744eac7e42616 100644 --- a/src/content/docs/agents/index.mdx +++ b/src/content/docs/agents/index.mdx @@ -24,6 +24,7 @@ import { Build AI-powered agents that can autonomously perform tasks, persist state, browse the web, and communicate back to users in real-time over any channel. +- ***Serverless inference that scales up _and_ down**: run AI directly on Cloudflre, without worrying about pre-provisioning VMs at peak, and worrying about utilization. Call the latest open-source models on [Workers AI](/workers-ai/), and pay just for what you use. - **Non I/O bound pricing:** don't pay for long-running processes when your code is not executing. Cloudflare Workers is designed to scale down and [only charge you for CPU time](https://blog.cloudflare.com/workers-pricing-scale-to-zero/), as opposed to wall-clock time. - **Designed for durable execution:** [Durable Objects](/durable-objects/) and [Workflows](/workflows) are built for a programming model that enables guaranteed execution for async tasks like long-running deep thinking LLM calls, human-in-the-loop, or unreliable API calls. - **Scalable, and reliable, without compromising on performance:** by running on Cloudflare's network, agents can execute tasks close to the user without introducing latency for real-time experiences. diff --git a/src/content/docs/workers-ai/index.mdx b/src/content/docs/workers-ai/index.mdx index fe961c81434031..982ec8fd9060f2 100644 --- a/src/content/docs/workers-ai/index.mdx +++ b/src/content/docs/workers-ai/index.mdx @@ -20,9 +20,12 @@ Run machine learning models, powered by serverless GPUs, on Cloudflare's global -Workers AI allows you to run machine learning models, on the Cloudflare network, from your own code -- whether that be from [Workers](/workers/), [Pages](/pages/), or anywhere via [the Cloudflare API](/api/resources/ai/methods/run/). +Workers AI allows you to run AI models in a serverless way, without having to worry about scaling, maintaining, or paying for unused infrastructure. You can invoke models running on GPUs on Cloudflare's network from your own code — from [Workers](/workers/), [Pages](/pages/), or anywhere via [the Cloudflare API](/api/resources/ai/methods/run/). -With the launch of Workers AI, Cloudflare is rolling out GPUs to its global network. This enables you to build and deploy ambitious AI applications that run near your users, wherever they are. +Workers AI gives you access to: +- **50+ [open-source models](/workers-ai/models/)**, available as a part of our model catalog +- Serverless, **pay-for-what-you-use** [pricing model](/workers-ai/platform/pricing/) +- All as part of a **fully-featured developer platform**, including [AI Gateway](/ai-gateway/), [Vectorize](/vectorize/), [Workers](/workers/) and more...
Get started