Skip to content

Commit b18eddd

Browse files
committed
rename Inference Provider API to Inference Providers
1 parent 4e6ccd4 commit b18eddd

File tree

4 files changed

+10
-10
lines changed

4 files changed

+10
-10
lines changed

docs/api-inference/_toctree.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
- title: Get Started
22
sections:
33
- local: index
4-
title: Inference Providers API
4+
title: Inference Providers
55
- local: pricing
66
title: Pricing and Billing
77
- local: hub-integration

docs/api-inference/hub-integration.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# Hub Integration
22

3-
The Inference Providers API is tightly integrated with the Hugging Face Hub. No matter in which service you use it, the usage and billing will be centralized on your Hugging Face account.
3+
The Inference Providers is tightly integrated with the Hugging Face Hub. No matter in which service you use it, the usage and billing will be centralized on your Hugging Face account.
44

55
## Model search
66

@@ -20,7 +20,7 @@ It is also possible to select multiple providers or even all of them to filter a
2020

2121
## Features using Inference Providers
2222

23-
Several Hugging Face features utilize the Inference Providers API and count towards your monthly credits. The included monthly credits for PRO and Enterprise should cover moderate usage of these features for most users.
23+
Several Hugging Face features utilize the Inference Providers and count towards your monthly credits. The included monthly credits for PRO and Enterprise should cover moderate usage of these features for most users.
2424

2525
- [Inference Widgets](https://huggingface.co/deepseek-ai/DeepSeek-V3-0324): Interactive widgets available on model pages. This is the entry point to quickly test a model on the Hub.
2626

docs/api-inference/index.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
1-
# Inference Providers API
1+
# Inference Providers
22

3-
The Hugging Face Inference Providers API revolutionizes how developers access and run machine learning models by offering a unified, flexible interface to multiple serverless inference providers. This new approach extends our previous Serverless Inference API, providing more models, increased performances and better reliability thanks to our awesome partners.
3+
The Hugging Face Inference Providers revolutionizes how developers access and run machine learning models by offering a unified, flexible interface to multiple serverless inference providers. This new approach extends our previous Serverless Inference API, providing more models, increased performances and better reliability thanks to our awesome partners.
44

5-
To learn more about the launch of the Inference Providers API, check out our [announcement blog post](https://huggingface.co/blog/inference-providers).
5+
To learn more about the launch of the Inference Providers, check out our [announcement blog post](https://huggingface.co/blog/inference-providers).
66

77
## Why use the Inference Provider API?
88

@@ -33,13 +33,13 @@ To get started quickly with [Chat Completion models](http://huggingface.co/model
3333

3434
## Get Started
3535

36-
You can call the Inference Providers API with your preferred tools, such as Python, JavaScript, or cURL. To simplify integration, we offer both a Python SDK (`huggingface_hub`) and a JavaScript SDK (`huggingface.js`).
36+
You can call the Inference Providers with your preferred tools, such as Python, JavaScript, or cURL. To simplify integration, we offer both a Python SDK (`huggingface_hub`) and a JavaScript SDK (`huggingface.js`).
3737

3838
In this section, we will demonstrate a simple example using [deepseek-ai/DeepSeek-V3-0324](https://huggingface.co/deepseek-ai/DeepSeek-V3-0324), a conversational Large Language Model. For the example, we will use [Novita AI](https://novita.ai/) as Inference Provider with routed requests. You will learn what that means in the next chapters.
3939

4040
### Authentication
4141

42-
The Inference Providers API requires passing a user token in the request headers. You can generate a token by signing up on the Hugging Face website and going to the [settings page](https://huggingface.co/settings/tokens/new?ownUserPermissions=inference.serverless.write&tokenType=fineGrained). We recommend creating a `fine-grained` token with the scope to `Make calls to Inference Providers`.
42+
The Inference Providers requires passing a user token in the request headers. You can generate a token by signing up on the Hugging Face website and going to the [settings page](https://huggingface.co/settings/tokens/new?ownUserPermissions=inference.serverless.write&tokenType=fineGrained). We recommend creating a `fine-grained` token with the scope to `Make calls to Inference Providers`.
4343

4444
For more details about user tokens, check out [this guide](https://huggingface.co/docs/hub/en/security-tokens).
4545

docs/api-inference/pricing.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# Pricing and Billing
22

3-
The Inference Providers API is a production-ready service involving external partners and is therefore a paid-product. However, as an Hugging Face user you get monthly credits to run experiments. The amount of credits you get depends on your type of account:
3+
The Inference Providers is a production-ready service involving external partners and is therefore a paid-product. However, as an Hugging Face user you get monthly credits to run experiments. The amount of credits you get depends on your type of account:
44

55
| User Tier | Included monthly credits |
66
| ------------------------ | ---------------------------------- |
@@ -25,7 +25,7 @@ Hugging Face charges you the same rates as the provider, with no additional fees
2525

2626
The documentation above assumes you are making routed requests to external providers. In practice, there are 3 different ways to run inference, each with unique billing implications:
2727

28-
- **Routed Request**: This is the default method for using the Inference Providers API. Simply use the JavaScript or Python `InferenceClient`, or make raw HTTP requests with your Hugging Face User Access Token. Your request is automatically routed through Hugging Face to the provider's platform. No separate provider account is required, and billing is managed directly by Hugging Face. This approach lets you seamlessly switch between providers without additional setup.
28+
- **Routed Request**: This is the default method for using the Inference Providers. Simply use the JavaScript or Python `InferenceClient`, or make raw HTTP requests with your Hugging Face User Access Token. Your request is automatically routed through Hugging Face to the provider's platform. No separate provider account is required, and billing is managed directly by Hugging Face. This approach lets you seamlessly switch between providers without additional setup.
2929

3030
- **Routed Request with Custom Key**: In your [settings page](https://huggingface.co/settings/inference-providers) on the Hub, you can configure a custom key for each provider. To use this option, you'll need to create an account on the provider's platform, and billing will be handled directly by that provider. Hugging Face won't charge you for the call. This method gives you more control over billing when experimenting with models on the Hub. When making a routed request with a custom key, your code remains unchanged—you'll still pass your Hugging Face User Access Token. Hugging Face will automatically swap the authentication when routing the request.
3131

0 commit comments

Comments
 (0)