You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/inference-providers/providers/groq.md
+11-1Lines changed: 11 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -33,7 +33,17 @@ For more details, check out the `generate.ts` script: https://github.com/hugging
33
33
</a>
34
34
</div>
35
35
36
-
Groq is fast AI inference. Their groundbreaking Language Processing Unit (LPU) technology delivers industry-leading speed and performance for large language models, achieving much faster performance than traditional GPU-based solutions. With custom chips specifically designed for AI inference workloads and a deterministic, software-first approach, Groq eliminates the bottlenecks of conventional hardware to enable real-time AI applications with predictable latency and exceptional throughput for developers building the next generation of AI-powered solutions.
36
+
Groq is fast AI inference. Their groundbreaking LPU technology delivers record-setting performance and efficiency for GenAI models. With custom chips specifically designed for AI inference workloads and a deterministic, software-first approach, Groq eliminates the bottlenecks of conventional hardware to enable real-time AI applications with predictable latency and exceptional throughput so developers can build fast.
37
+
38
+
For latest pricing, visit our [pricing page](https://groq.com/pricing/).
Copy file name to clipboardExpand all lines: docs/inference-providers/tasks/summarization.md
+2-1Lines changed: 2 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -24,6 +24,7 @@ For more details about the `summarization` task, check out its [dedicated page](
24
24
25
25
### Recommended models
26
26
27
+
-[facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn): A strong summarization model trained on English news articles. Excels at generating factual summaries.
27
28
-[Falconsai/medical_summarization](https://huggingface.co/Falconsai/medical_summarization): A summarization model trained on medical articles.
28
29
29
30
Explore all available models and find the one that suits you best [here](https://huggingface.co/models?inference=warm&pipeline_tag=summarization&sort=trending).
@@ -33,7 +34,7 @@ Explore all available models and find the one that suits you best [here](https:/
0 commit comments