You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: inference-providers-featherless-groq.md
+12-5Lines changed: 12 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -23,16 +23,21 @@ authors:
23
23
24
24
# Groq & Featherless AI on Hugging Face Inference Providers 🔥
25
25
26
-
We're thrilled to share that **Featherless AI**is now a supported Inference Provider on the Hugging Face Hub!
27
-
Featherless AI joins our growing ecosystem, enhancing the breadth and capabilities of serverless inference directly on the Hub’s model pages. Inference Providers are also seamlessly integrated into our client SDKs (for both JS and Python), making it super easy to use a wide variety of models with your preferred providers.
26
+
We're thrilled to share that **Featherless AI**and **Groq** are now supported Inference Providers on the Hugging Face Hub!
27
+
Featherless AI and Groq join our growing ecosystem, enhancing the breadth and capabilities of serverless inference directly on the Hub’s model pages. Inference Providers are also seamlessly integrated into our client SDKs (for both JS and Python), making it super easy to use a wide variety of models with your preferred providers.
28
28
29
29
[Featherless AI](https://featherless.ai) supports a wide variety of text and conversational models, including the latest open-source models from DeepSeek, Meta, Google, Qwen, and much more.
30
30
31
31
Featherless AI is a serverless AI inference provider with unique model loading and GPU orchestration abilities that makes an exceptionally large catalog of models available for users. Providers often offer either a low cost of access to a limited set of models, or an unlimited range of models with users managing servers and the associated costs of operation. Featherless provides the best of both worlds offering unmatched model range and variety but with serverless pricing. Find the full list of supported models on the [models page](https://huggingface.co/models?inference_provider=featherless-ai&sort=trending).
32
32
33
-
We're quite excited to see what you'll build with this new provider!
33
+
[Groq](https://groq.com) offers a fast Inference API, powered by the LPU (Language Processing Unit), their own AI hardware processor that meets the demand for instant speed, scalability, and low latency. By optimizing compute density, memory bandwidth, and scalability, LPUs overcome performance bottlenecks and deliver ultra-low latency inference, unlocking a new class of use cases.
34
+
35
+
Take advantage of Groq for fast AI inference performance for leading openly-available models from providers like Meta, DeepSeek, Qwen, Mistral, Google, OpenAI, and more.
36
+
37
+
We're quite excited to see what you'll build with those new providers!
34
38
35
39
Read more about how to use Featherless as Inference Provider in its dedicated [documentation page](https://huggingface.co/docs/inference-providers/providers/featherless-ai).
40
+
Read more about how to use Groq as Inference Provider in its dedicated [documentation page](https://huggingface.co/docs/inference-providers/providers/groq).
The following example shows how to use Qwen QWQ-32B using Groq as the inference provider. You can use a [Hugging Face token](https://huggingface.co/settings/tokens) for automatic routing through Hugging Face, or your own Groq cloud API key if you have one.
0 commit comments