You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
If you're already using OpenAI's Python client, then you need a **drop-in OpenAI replacement**. Just swap-out the base URL to instantly access hundreds of additional open-weights models through our provider network.
130
128
131
-
Our system automatically routes your request to the optimal provider for the specified model:
129
+
Our system automatically routes your request to the most popular provider for the specified model. You can also order the available providers from your [user settings](https://huggingface.co/settings/inference-providers), in that case we will route your request to the first provider supporting the model.
For testing, debugging, or integrating with any HTTP client, here's the raw REST API format. Our intelligent routing automatically selects the optimal provider for your requested model:
283
+
For testing, debugging, or integrating with any HTTP client, here's the raw REST API format. Our intelligent routing automatically selects the most popular provider for your requested model,
284
+
or to your preferred provider if you've sorted the providers from your [user settings](https://huggingface.co/settings/inference-providers).
@@ -355,11 +352,10 @@ The Inference Providers API acts as a unified proxy layer that sits between your
355
352
When using Inference Providers, your requests go through Hugging Face's proxy infrastructure, which provides several key benefits:
356
353
357
354
-**Unified Authentication & Billing**: Use a single Hugging Face token for all providers
358
-
-**Automatic Failover**: If one provider is unavailable, requests can be routed to alternatives
359
-
-**Rate Limiting & Load Balancing**: Intelligent distribution of requests across providers
360
-
-**Consistent API Interface**: The same request format works across different providers
355
+
-**Automatic Failover**: When using automatic provider selection (`provider="auto"`), requests are automatically routed to alternative providers if the primary provider is flagged as unavailable by our validation system
356
+
-**Consistent Interface through client libraries**: When using our client libraries, the same request format works across different providers
361
357
362
-
Because the API acts as a proxy, the exact HTTP request may vary between providers as each provider has their own API requirements and response formats. The Hugging Face inference clients handle these provider-specific differences automatically when you use `provider="auto"` or specify a particular provider.
358
+
Because the API acts as a proxy, the exact HTTP request may vary between providers as each provider has their own API requirements and response formats. **When using our official client libraries** (JavaScript or Python), these provider-specific differences are handled automatically whether you use `provider="auto"` or specify a particular provider.
0 commit comments