Skip to content

Commit 0f7efe0

Browse files
Wauplinjulien-c
andauthored
Add 1 page per provider + list table (#1674)
* Draft to list providers * add logo * fix _toctree * logo * fix links * Semi-generated page for all providers * consistency * Add HF-Inference page * Update scripts/inference-providers/templates/providers/fal-ai.handlebars * Add badge to their Org page (will update the actual badge later) * Final badge + generate * make links clickable --------- Co-authored-by: Julien Chaumond <[email protected]>
1 parent b00d4bf commit 0f7efe0

25 files changed

+881
-10
lines changed

docs/inference-providers/_toctree.yml

Lines changed: 25 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,10 +6,33 @@
66
title: Pricing and Billing
77
- local: hub-integration
88
title: Hub integration
9-
- local: security
10-
title: Security
119
- local: register-as-a-provider
1210
title: Register as an Inference Provider
11+
- local: security
12+
title: Security
13+
14+
- title: Providers
15+
sections:
16+
- local: providers/cerebras
17+
title: Cerebras
18+
- local: providers/fal-ai
19+
title: Fal AI
20+
- local: providers/fireworks-ai
21+
title: Fireworks
22+
- local: providers/hyperbolic
23+
title: Hyperbolic
24+
- local: providers/hf-inference
25+
title: HF Inference
26+
- local: providers/nebius
27+
title: Nebius
28+
- local: providers/novita
29+
title: Novita
30+
- local: providers/replicate
31+
title: Replicate
32+
- local: providers/sambanova
33+
title: SambaNova
34+
- local: providers/together
35+
title: Together
1336
- title: API Reference
1437
sections:
1538
- local: tasks/index

docs/inference-providers/index.md

Lines changed: 17 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,23 @@ Hugging Face Inference Providers simplify and unify how developers access and ru
99

1010
To learn more about the launch of Inference Providers, check out our [announcement blog post](https://huggingface.co/blog/inference-providers).
1111

12+
## Partners
13+
14+
Here is the complete list of partners integrated with Inference Providers, and the supported tasks for each of them:
15+
16+
| Provider | Chat completion (LLM) | Chat completion (VLM) | Feature Extraction | Text to Image | Text to video |
17+
| ---------------------------------------- | :-------------------: | :-------------------: | :----------------: | :-----------: | :-----------: |
18+
| [Cerebras](./providers/cerebras) || | | | |
19+
| [Fal AI](./providers/fal-ai) | | | |||
20+
| [Fireworks](./providers/fireworks-ai) ||| | | |
21+
| [HF Inference](./providers/hf-inference) ||||| |
22+
| [Hyperbolic](./providers/hyperbolic) ||| | | |
23+
| [Nebius](./providers/nebius) ||| || |
24+
| [Novita](./providers/novita) ||| | ||
25+
| [Replicate](./providers/replicate) | | | |||
26+
| [SambaNova](./providers/sambanova) || || | |
27+
| [Together](./providers/together) ||| || |
28+
1229
## Why use Inference Providers?
1330

1431
Inference Providers offers a fast and simple way to explore thousands of models for a variety of tasks. Whether you're experimenting with ML capabilities or building a new application, this API gives you instant access to high-performing models across multiple domains:
@@ -28,7 +45,6 @@ Inference Providers offers a fast and simple way to explore thousands of models
2845
- **🔧 Developer-Friendly**: Simple requests, fast responses, and a consistent developer experience across Python and JavaScript clients.
2946
- **💰 Cost-Effective**: No extra markup on provider rates.
3047

31-
3248
## Inference Playground
3349

3450
To get started quickly with [Chat Completion models](http://huggingface.co/models?inference_provider=all&sort=trending&other=conversational), use the [Inference Playground](https://huggingface.co/playground) to easily test and compare models with your prompts.
Lines changed: 30 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,30 @@
1+
<!---
2+
WARNING
3+
4+
This markdown file has been generated from a script. Please do not edit it directly.
5+
6+
If you want to update the content related to cerebras's description, please edit the template file under `https://github.com/huggingface/hub-docs/tree/main/scripts/inference-providers/templates/providers/cerebras.handlebars`.
7+
8+
For more details, check out the `generate.ts` script: https://github.com/huggingface/hub-docs/blob/main/scripts/inference-providers/scripts/generate.ts.
9+
--->
10+
11+
# Cerebras
12+
13+
[![Cerebras Logo](https://upload.wikimedia.org/wikipedia/commons/thumb/1/15/Cerebras_logo.svg/512px-Cerebras_logo.svg.png)](https://www.cerebras.ai/)
14+
15+
[![Follow us on Hugging Face](https://huggingface.co/datasets/huggingface/badges/resolve/main/follow-us-on-hf-lg.svg)](https://huggingface.co/cerebras)
16+
17+
Cerebras stands alone as the world’s fastest AI inference and training platform. Organizations across fields like medical research, cryptography, energy, and agentic AI use our CS-2 and CS-3 systems to build on-premise supercomputers, while developers and enterprises everywhere can access the power of Cerebras through our pay-as-you-go cloud offerings.
18+
19+
## Supported tasks
20+
21+
22+
### Chat Completion (LLM)
23+
24+
Find out more about Chat Completion (LLM) [here](../tasks/chat-completion).
25+
26+
<InferenceSnippet
27+
pipeline=text-generation
28+
providersMapping={ {"cerebras":{"modelId":"meta-llama/Llama-3.3-70B-Instruct","providerModelId":"llama-3.3-70b"} } }
29+
conversational />
30+
Lines changed: 50 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,50 @@
1+
<!---
2+
WARNING
3+
4+
This markdown file has been generated from a script. Please do not edit it directly.
5+
6+
If you want to update the content related to fal-ai's description, please edit the template file under `https://github.com/huggingface/hub-docs/tree/main/scripts/inference-providers/templates/providers/fal-ai.handlebars`.
7+
8+
For more details, check out the `generate.ts` script: https://github.com/huggingface/hub-docs/blob/main/scripts/inference-providers/scripts/generate.ts.
9+
--->
10+
11+
# Fal
12+
13+
[![fal.ai logo](https://images.seeklogo.com/logo-png/61/1/fal-ai-logo-png_seeklogo-611592.png)](https://fal.ai/)
14+
15+
[![Follow us on Hugging Face](https://huggingface.co/datasets/huggingface/badges/resolve/main/follow-us-on-hf-lg.svg)](https://huggingface.co/fal)
16+
17+
Founded in 2021 by [Burkay Gur](https://huggingface.co/burkaygur) and [Gorkem Yurtseven](https://huggingface.co/gorkemyurt), fal.ai was born out of a shared passion for AI and a desire to address the challenges in AI infrastructure observed during their tenures at Coinbase and Amazon.
18+
19+
## Supported tasks
20+
21+
22+
### Automatic Speech Recognition
23+
24+
Find out more about Automatic Speech Recognition [here](../tasks/automatic_speech_recognition).
25+
26+
<InferenceSnippet
27+
pipeline=automatic-speech-recognition
28+
providersMapping={ {"fal-ai":{"modelId":"openai/whisper-large-v3","providerModelId":"fal-ai/whisper"} } }
29+
/>
30+
31+
32+
### Text To Image
33+
34+
Find out more about Text To Image [here](../tasks/text_to_image).
35+
36+
<InferenceSnippet
37+
pipeline=text-to-image
38+
providersMapping={ {"fal-ai":{"modelId":"black-forest-labs/FLUX.1-dev","providerModelId":"fal-ai/flux/dev"} } }
39+
/>
40+
41+
42+
### Text To Video
43+
44+
Find out more about Text To Video [here](../tasks/text_to_video).
45+
46+
<InferenceSnippet
47+
pipeline=text-to-video
48+
providersMapping={ {"fal-ai":{"modelId":"Wan-AI/Wan2.1-T2V-14B","providerModelId":"fal-ai/wan-t2v"} } }
49+
/>
50+
Lines changed: 39 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,39 @@
1+
<!---
2+
WARNING
3+
4+
This markdown file has been generated from a script. Please do not edit it directly.
5+
6+
If you want to update the content related to fireworks-ai's description, please edit the template file under `https://github.com/huggingface/hub-docs/tree/main/scripts/inference-providers/templates/providers/fireworks-ai.handlebars`.
7+
8+
For more details, check out the `generate.ts` script: https://github.com/huggingface/hub-docs/blob/main/scripts/inference-providers/scripts/generate.ts.
9+
--->
10+
11+
# Fireworks AI
12+
13+
[![fireworks.ai](https://d1.awsstatic.com/fireworks-ai-wordmark-color-dark.93b1f27fdf77899fa02afb949fb27317ee4081ad.png)](https://fireworks.ai/)
14+
15+
[![Follow us on Hugging Face](https://huggingface.co/datasets/huggingface/badges/resolve/main/follow-us-on-hf-lg.svg)](https://huggingface.co/fireworks-ai)
16+
17+
Fireworks AI is a developer-centric platform that delivers high-performance generative AI solutions, enabling efficient deployment and fine-tuning of large language models (LLMs) and image models.
18+
## Supported tasks
19+
20+
21+
### Chat Completion (LLM)
22+
23+
Find out more about Chat Completion (LLM) [here](../tasks/chat-completion).
24+
25+
<InferenceSnippet
26+
pipeline=text-generation
27+
providersMapping={ {"fireworks-ai":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"accounts/fireworks/models/deepseek-v3-0324"} } }
28+
conversational />
29+
30+
31+
### Chat Completion (VLM)
32+
33+
Find out more about Chat Completion (VLM) [here](../tasks/chat-completion).
34+
35+
<InferenceSnippet
36+
pipeline=image-text-to-text
37+
providersMapping={ {"fireworks-ai":{"modelId":"meta-llama/Llama-4-Scout-17B-16E-Instruct","providerModelId":"accounts/fireworks/models/llama4-scout-instruct-basic"} } }
38+
conversational />
39+
Lines changed: 201 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,201 @@
1+
<!---
2+
WARNING
3+
4+
This markdown file has been generated from a script. Please do not edit it directly.
5+
6+
If you want to update the content related to hf-inference's description, please edit the template file under `https://github.com/huggingface/hub-docs/tree/main/scripts/inference-providers/templates/providers/hf-inference.handlebars`.
7+
8+
For more details, check out the `generate.ts` script: https://github.com/huggingface/hub-docs/blob/main/scripts/inference-providers/scripts/generate.ts.
9+
--->
10+
11+
# HF Inference
12+
13+
[![Hugging Face](https://huggingface.co/datasets/huggingface/brand-assets/resolve/main/hf-logo-with-title.png)](https://huggingface.co/)
14+
15+
[![Follow us on Hugging Face](https://huggingface.co/datasets/huggingface/badges/resolve/main/follow-us-on-hf-lg.svg)](https://huggingface.co/hf-inference)
16+
17+
HF Inference is the serverless Inference API powered by Hugging Face. This service used to be called "Inference API (serverless)" prior to Inference Providers.
18+
If you are interested in deploying models to a dedicated and autoscaling infrastructure managed by Hugging Face, check out [Inference Endpoints](https://huggingface.co/docs/inference-endpoints/index) instead.
19+
20+
## Supported tasks
21+
22+
23+
### Audio Classification
24+
25+
Find out more about Audio Classification [here](../tasks/audio_classification).
26+
27+
<InferenceSnippet
28+
pipeline=audio-classification
29+
providersMapping={ {"hf-inference":{"modelId":"ehcalabres/wav2vec2-lg-xlsr-en-speech-emotion-recognition","providerModelId":"ehcalabres/wav2vec2-lg-xlsr-en-speech-emotion-recognition"} } }
30+
/>
31+
32+
33+
### Automatic Speech Recognition
34+
35+
Find out more about Automatic Speech Recognition [here](../tasks/automatic_speech_recognition).
36+
37+
<InferenceSnippet
38+
pipeline=automatic-speech-recognition
39+
providersMapping={ {"hf-inference":{"modelId":"openai/whisper-large-v3-turbo","providerModelId":"openai/whisper-large-v3-turbo"} } }
40+
/>
41+
42+
43+
### Chat Completion (LLM)
44+
45+
Find out more about Chat Completion (LLM) [here](../tasks/chat-completion).
46+
47+
<InferenceSnippet
48+
pipeline=text-generation
49+
providersMapping={ {"hf-inference":{"modelId":"Qwen/QwQ-32B","providerModelId":"Qwen/QwQ-32B"} } }
50+
conversational />
51+
52+
53+
### Chat Completion (VLM)
54+
55+
Find out more about Chat Completion (VLM) [here](../tasks/chat-completion).
56+
57+
<InferenceSnippet
58+
pipeline=image-text-to-text
59+
providersMapping={ {"hf-inference":{"modelId":"google/gemma-3-27b-it","providerModelId":"google/gemma-3-27b-it"} } }
60+
conversational />
61+
62+
63+
### Feature Extraction
64+
65+
Find out more about Feature Extraction [here](../tasks/feature_extraction).
66+
67+
<InferenceSnippet
68+
pipeline=feature-extraction
69+
providersMapping={ {"hf-inference":{"modelId":"intfloat/multilingual-e5-large-instruct","providerModelId":"intfloat/multilingual-e5-large-instruct"} } }
70+
/>
71+
72+
73+
### Fill Mask
74+
75+
Find out more about Fill Mask [here](../tasks/fill_mask).
76+
77+
<InferenceSnippet
78+
pipeline=fill-mask
79+
providersMapping={ {"hf-inference":{"modelId":"google-bert/bert-base-uncased","providerModelId":"google-bert/bert-base-uncased"} } }
80+
/>
81+
82+
83+
### Image Classification
84+
85+
Find out more about Image Classification [here](../tasks/image_classification).
86+
87+
<InferenceSnippet
88+
pipeline=image-classification
89+
providersMapping={ {"hf-inference":{"modelId":"Falconsai/nsfw_image_detection","providerModelId":"Falconsai/nsfw_image_detection"} } }
90+
/>
91+
92+
93+
### Image To Image
94+
95+
Find out more about Image To Image [here](../tasks/image_to_image).
96+
97+
<InferenceSnippet
98+
pipeline=image-to-image
99+
providersMapping={ {"hf-inference":{"modelId":"enhanceaiteam/Flux-Uncensored-V2","providerModelId":"black-forest-labs/FLUX.1-dev"} } }
100+
/>
101+
102+
103+
### Object Detection
104+
105+
Find out more about Object Detection [here](../tasks/object_detection).
106+
107+
<InferenceSnippet
108+
pipeline=object-detection
109+
providersMapping={ {"hf-inference":{"modelId":"facebook/detr-resnet-50","providerModelId":"facebook/detr-resnet-50"} } }
110+
/>
111+
112+
113+
### Question Answering
114+
115+
Find out more about Question Answering [here](../tasks/question_answering).
116+
117+
<InferenceSnippet
118+
pipeline=question-answering
119+
providersMapping={ {"hf-inference":{"modelId":"deepset/gelectra-large-germanquad","providerModelId":"deepset/gelectra-large-germanquad"} } }
120+
/>
121+
122+
123+
### Summarization
124+
125+
Find out more about Summarization [here](../tasks/summarization).
126+
127+
<InferenceSnippet
128+
pipeline=summarization
129+
providersMapping={ {"hf-inference":{"modelId":"facebook/bart-large-cnn","providerModelId":"facebook/bart-large-cnn"} } }
130+
/>
131+
132+
133+
### Text Classification
134+
135+
Find out more about Text Classification [here](../tasks/text_classification).
136+
137+
<InferenceSnippet
138+
pipeline=text-classification
139+
providersMapping={ {"hf-inference":{"modelId":"ProsusAI/finbert","providerModelId":"ProsusAI/finbert"} } }
140+
/>
141+
142+
143+
### Text Generation
144+
145+
Find out more about Text Generation [here](../tasks/text_generation).
146+
147+
<InferenceSnippet
148+
pipeline=text-generation
149+
providersMapping={ {"hf-inference":{"modelId":"Qwen/QwQ-32B","providerModelId":"Qwen/QwQ-32B"} } }
150+
/>
151+
152+
153+
### Text To Image
154+
155+
Find out more about Text To Image [here](../tasks/text_to_image).
156+
157+
<InferenceSnippet
158+
pipeline=text-to-image
159+
providersMapping={ {"hf-inference":{"modelId":"black-forest-labs/FLUX.1-dev","providerModelId":"black-forest-labs/FLUX.1-dev"} } }
160+
/>
161+
162+
163+
### Text To Video
164+
165+
Find out more about Text To Video [here](../tasks/text_to_video).
166+
167+
<InferenceSnippet
168+
pipeline=text-to-video
169+
providersMapping={ {"hf-inference":{"modelId":"AdamLucek/Wan2.1-T2V-14B-OldBookIllustrations","providerModelId":"black-forest-labs/FLUX.1-dev"} } }
170+
/>
171+
172+
173+
### Token Classification
174+
175+
Find out more about Token Classification [here](../tasks/token_classification).
176+
177+
<InferenceSnippet
178+
pipeline=token-classification
179+
providersMapping={ {"hf-inference":{"modelId":"dbmdz/bert-large-cased-finetuned-conll03-english","providerModelId":"dbmdz/bert-large-cased-finetuned-conll03-english"} } }
180+
/>
181+
182+
183+
### Translation
184+
185+
Find out more about Translation [here](../tasks/translation).
186+
187+
<InferenceSnippet
188+
pipeline=translation
189+
providersMapping={ {"hf-inference":{"modelId":"facebook/nllb-200-distilled-600M","providerModelId":"facebook/nllb-200-distilled-600M"} } }
190+
/>
191+
192+
193+
### Zero Shot Classification
194+
195+
Find out more about Zero Shot Classification [here](../tasks/zero_shot_classification).
196+
197+
<InferenceSnippet
198+
pipeline=zero-shot-classification
199+
providersMapping={ {"hf-inference":{"modelId":"facebook/bart-large-mnli","providerModelId":"facebook/bart-large-mnli"} } }
200+
/>
201+

0 commit comments

Comments
 (0)