You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/api-inference/hub-integration.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
# Hub Integration
2
2
3
-
The Inference Providers is tightly integrated with the Hugging Face Hub. No matter which provider you use, the usage and billing will be centralized in your Hugging Face account.
3
+
Inference Providers is tightly integrated with the Hugging Face Hub. No matter which provider you use, the usage and billing will be centralized in your Hugging Face account.
4
4
5
5
## Model search
6
6
@@ -46,7 +46,7 @@ Several Hugging Face features utilize Inference Providers and count towards your
46
46
## User Settings
47
47
48
48
In your user account settings, you are able to:
49
-
- set your own API keys for the providers you’ve signed up with. Otherwise, you can still use them – your requests will be billed on your HF account. More details in the [billing section](./pricing#routed-requests-vs-direct-calls).
49
+
- set your own API keys for the providers you’ve signed up with. If you don't, your requests will be billed on your HF account. More details in the [billing section](./pricing#routed-requests-vs-direct-calls).
Copy file name to clipboardExpand all lines: docs/api-inference/index.md
+10-10Lines changed: 10 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,19 +1,19 @@
1
1
# Inference Providers
2
2
3
-
The Hugging Face Inference Providers revolutionizes how developers access and run machine learning models by offering a unified, flexible interface to multiple serverless inference providers. This new approach extends our previous Serverless Inference API, providing more models, increased performances and better reliability thanks to our awesome partners.
3
+
Hugging Face Inference Providers revolutionizes how developers access and run machine learning models by offering a unified, flexible interface to multiple serverless inference providers. This new approach extends our previous Serverless Inference API, providing more models, increased performances and better reliability thanks to our awesome partners.
4
4
5
-
To learn more about the launch of the Inference Providers, check out our [announcement blog post](https://huggingface.co/blog/inference-providers).
5
+
To learn more about the launch of Inference Providers, check out our [announcement blog post](https://huggingface.co/blog/inference-providers).
6
6
7
-
## Why use the Inference Providers?
7
+
## Why use Inference Providers?
8
8
9
-
The Inference Providers offers a fast and simple way to explore thousands of models for a variety of tasks. Whether you're experimenting with ML capabilities or building a new application, this API gives you instant access to high-performing models across multiple domains:
9
+
Inference Providers offers a fast and simple way to explore thousands of models for a variety of tasks. Whether you're experimenting with ML capabilities or building a new application, this API gives you instant access to high-performing models across multiple domains:
10
10
11
11
***Text Generation:** Including large language models and tool-calling prompts, generate and experiment with high-quality responses.
12
12
***Image and Video Generation:** Easily create customized images, including LoRAs for your own styles.
13
13
***Document Embeddings:** Build search and retrieval systems with SOTA embeddings.
14
14
***Classical AI Tasks:** Ready-to-use models for text classification, image classification, speech recognition, and more.
15
15
16
-
⚡ **Fast and Free to Get Started**: The Inference Providers comes with a free-tier and additional included credits for [PRO users](https://hf.co/subscribe/pro).
16
+
⚡ **Fast and Free to Get Started**: Inference Providers comes with a free-tier and additional included credits for [PRO users](https://hf.co/subscribe/pro), as well as [Enterprise Hub organizations](https://huggingface.co/enterprise).
17
17
18
18
## Key Features
19
19
@@ -33,13 +33,13 @@ To get started quickly with [Chat Completion models](http://huggingface.co/model
33
33
34
34
## Get Started
35
35
36
-
You can call the Inference Providers with your preferred tools, such as Python, JavaScript, or cURL. To simplify integration, we offer both a Python SDK (`huggingface_hub`) and a JavaScript SDK (`huggingface.js`).
36
+
You can use Inference Providers with your preferred tools, such as Python, JavaScript, or cURL. To simplify integration, we offer both a Python SDK (`huggingface_hub`) and a JavaScript SDK (`huggingface.js`).
37
37
38
38
In this section, we will demonstrate a simple example using [deepseek-ai/DeepSeek-V3-0324](https://huggingface.co/deepseek-ai/DeepSeek-V3-0324), a conversational Large Language Model. For the example, we will use [Novita AI](https://novita.ai/) as Inference Provider.
39
39
40
40
### Authentication
41
41
42
-
The Inference Providers requires passing a user token in the request headers. You can generate a token by signing up on the Hugging Face website and going to the [settings page](https://huggingface.co/settings/tokens/new?ownUserPermissions=inference.serverless.write&tokenType=fineGrained). We recommend creating a `fine-grained` token with the scope to `Make calls to Inference Providers`.
42
+
Inference Providers requires passing a user token in the request headers. You can generate a token by signing up on the Hugging Face website and going to the [settings page](https://huggingface.co/settings/tokens/new?ownUserPermissions=inference.serverless.write&tokenType=fineGrained). We recommend creating a `fine-grained` token with the scope to `Make calls to Inference Providers`.
43
43
44
44
For more details about user tokens, check out [this guide](https://huggingface.co/docs/hub/en/security-tokens).
45
45
@@ -140,7 +140,7 @@ const response = await fetch(
140
140
console.log(awaitresponse.json());
141
141
```
142
142
143
-
For convenience, the JS library `@huggingface/inference` provides an [`InferenceClient`](https://huggingface.co/docs/huggingface.js/inference/classes/InferenceClient) that handles inference for you. Make sure to install it with `npm install @huggingface/inference`.
143
+
For convenience, the JS library `@huggingface/inference` provides an [`InferenceClient`](https://huggingface.co/docs/huggingface.js/inference/classes/InferenceClient) that handles inference for you. You can install it with `npm install @huggingface/inference`.
Copy file name to clipboardExpand all lines: docs/api-inference/pricing.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
# Pricing and Billing
2
2
3
-
Inference Providers is a production-ready service involving external partners and is therefore a paid-product. However, as a Hugging Face user you get monthly credits to run experiments. The amount of credits you get depends on your type of account:
3
+
Inference Providers is a production-ready service involving external partners and is therefore a paidproduct. However, as a Hugging Face user, you get monthly credits to run experiments. The amount of credits you get depends on your type of account:
0 commit comments