From d1bc81238769c7bb8efc85e3508032b71d6e1858 Mon Sep 17 00:00:00 2001
From: Wauplin <11801849+Wauplin@users.noreply.github.com>
Date: Thu, 5 Jun 2025 03:28:31 +0000
Subject: [PATCH] Update Inference Providers documentation (automated)
---
docs/inference-providers/providers/hf-inference.md | 12 +-----------
docs/inference-providers/providers/sambanova.md | 2 +-
docs/inference-providers/providers/together.md | 10 ----------
docs/inference-providers/tasks/chat-completion.md | 2 +-
docs/inference-providers/tasks/image-segmentation.md | 3 +--
.../tasks/table-question-answering.md | 1 -
docs/inference-providers/tasks/text-generation.md | 2 +-
.../tasks/zero-shot-classification.md | 6 +-----
scripts/inference-providers/package.json | 2 +-
scripts/inference-providers/pnpm-lock.yaml | 10 +++++-----
10 files changed, 12 insertions(+), 38 deletions(-)
diff --git a/docs/inference-providers/providers/hf-inference.md b/docs/inference-providers/providers/hf-inference.md
index 4d253c83f..000e92667 100644
--- a/docs/inference-providers/providers/hf-inference.md
+++ b/docs/inference-providers/providers/hf-inference.md
@@ -105,7 +105,7 @@ Find out more about Image Segmentation [here](../tasks/image_segmentation).
@@ -198,13 +198,3 @@ Find out more about Translation [here](../tasks/translation).
providersMapping={ {"hf-inference":{"modelId":"google-t5/t5-base","providerModelId":"google-t5/t5-base"} } }
/>
-
-### Zero Shot Classification
-
-Find out more about Zero Shot Classification [here](../tasks/zero_shot_classification).
-
-
-
diff --git a/docs/inference-providers/providers/sambanova.md b/docs/inference-providers/providers/sambanova.md
index 0512d5d7a..03ba566e4 100644
--- a/docs/inference-providers/providers/sambanova.md
+++ b/docs/inference-providers/providers/sambanova.md
@@ -45,7 +45,7 @@ Find out more about Chat Completion (LLM) [here](../tasks/chat-completion).
diff --git a/docs/inference-providers/providers/together.md b/docs/inference-providers/providers/together.md
index 8ccce4cfa..677fdd9e8 100644
--- a/docs/inference-providers/providers/together.md
+++ b/docs/inference-providers/providers/together.md
@@ -58,16 +58,6 @@ Find out more about Chat Completion (VLM) [here](../tasks/chat-completion).
conversational />
-### Text Generation
-
-Find out more about Text Generation [here](../tasks/text_generation).
-
-
-
-
### Text To Image
Find out more about Text To Image [here](../tasks/text_to_image).
diff --git a/docs/inference-providers/tasks/chat-completion.md b/docs/inference-providers/tasks/chat-completion.md
index b4e695d6a..47c4f945b 100644
--- a/docs/inference-providers/tasks/chat-completion.md
+++ b/docs/inference-providers/tasks/chat-completion.md
@@ -61,7 +61,7 @@ The API supports:
diff --git a/docs/inference-providers/tasks/image-segmentation.md b/docs/inference-providers/tasks/image-segmentation.md
index 197800740..4467e4fd7 100644
--- a/docs/inference-providers/tasks/image-segmentation.md
+++ b/docs/inference-providers/tasks/image-segmentation.md
@@ -24,7 +24,6 @@ For more details about the `image-segmentation` task, check out its [dedicated p
### Recommended models
-- [openmmlab/upernet-convnext-small](https://huggingface.co/openmmlab/upernet-convnext-small): Solid semantic segmentation model trained on ADE20k.
- [facebook/mask2former-swin-large-coco-panoptic](https://huggingface.co/facebook/mask2former-swin-large-coco-panoptic): Panoptic segmentation model trained on the COCO (common objects) dataset.
Explore all available models and find the one that suits you best [here](https://huggingface.co/models?inference=warm&pipeline_tag=image-segmentation&sort=trending).
@@ -34,7 +33,7 @@ Explore all available models and find the one that suits you best [here](https:/
diff --git a/docs/inference-providers/tasks/table-question-answering.md b/docs/inference-providers/tasks/table-question-answering.md
index 397c69dd1..358287c69 100644
--- a/docs/inference-providers/tasks/table-question-answering.md
+++ b/docs/inference-providers/tasks/table-question-answering.md
@@ -24,7 +24,6 @@ For more details about the `table-question-answering` task, check out its [dedic
### Recommended models
-- [microsoft/tapex-base](https://huggingface.co/microsoft/tapex-base): A table question answering model that is capable of neural SQL execution, i.e., employ TAPEX to execute a SQL query on a given table.
- [google/tapas-base-finetuned-wtq](https://huggingface.co/google/tapas-base-finetuned-wtq): A robust table question answering model.
Explore all available models and find the one that suits you best [here](https://huggingface.co/models?inference=warm&pipeline_tag=table-question-answering&sort=trending).
diff --git a/docs/inference-providers/tasks/text-generation.md b/docs/inference-providers/tasks/text-generation.md
index 51558b221..4b94d45fd 100644
--- a/docs/inference-providers/tasks/text-generation.md
+++ b/docs/inference-providers/tasks/text-generation.md
@@ -41,7 +41,7 @@ Explore all available models and find the one that suits you best [here](https:/
diff --git a/docs/inference-providers/tasks/zero-shot-classification.md b/docs/inference-providers/tasks/zero-shot-classification.md
index 1c57edfb9..d046ba432 100644
--- a/docs/inference-providers/tasks/zero-shot-classification.md
+++ b/docs/inference-providers/tasks/zero-shot-classification.md
@@ -24,17 +24,13 @@ For more details about the `zero-shot-classification` task, check out its [dedic
### Recommended models
-- [facebook/bart-large-mnli](https://huggingface.co/facebook/bart-large-mnli): Powerful zero-shot text classification model.
Explore all available models and find the one that suits you best [here](https://huggingface.co/models?inference=warm&pipeline_tag=zero-shot-classification&sort=trending).
### Using the API
-
+There are currently no snippet examples for the **zero-shot-classification** task, as no providers support it yet.
diff --git a/scripts/inference-providers/package.json b/scripts/inference-providers/package.json
index 0803c680e..1da29b5e8 100644
--- a/scripts/inference-providers/package.json
+++ b/scripts/inference-providers/package.json
@@ -14,7 +14,7 @@
"author": "",
"license": "ISC",
"dependencies": {
- "@huggingface/inference": "^4.0.0",
+ "@huggingface/inference": "^4.0.2",
"@huggingface/tasks": "^0.19.11",
"@types/node": "^22.5.0",
"handlebars": "^4.7.8",
diff --git a/scripts/inference-providers/pnpm-lock.yaml b/scripts/inference-providers/pnpm-lock.yaml
index e1a7796f3..5957553a3 100644
--- a/scripts/inference-providers/pnpm-lock.yaml
+++ b/scripts/inference-providers/pnpm-lock.yaml
@@ -9,8 +9,8 @@ importers:
.:
dependencies:
'@huggingface/inference':
- specifier: ^4.0.0
- version: 4.0.0
+ specifier: ^4.0.2
+ version: 4.0.2
'@huggingface/tasks':
specifier: ^0.19.11
version: 0.19.11
@@ -189,8 +189,8 @@ packages:
cpu: [x64]
os: [win32]
- '@huggingface/inference@4.0.0':
- resolution: {integrity: sha512-NJD5qnYcOyG48qmqxIgmGZrgeR3+e/Sl+vGHTMvqgTK7vRXntLHzkgzSsHaP+qe7/etvVrr9yVy7vzSdA32dpg==}
+ '@huggingface/inference@4.0.2':
+ resolution: {integrity: sha512-XuWb8ocH7lA5kSdXrGnqshtRz3ocSBzEzxcp5xeAXLjgM1ocoIHq+RW8/Ti0xq3MeRGQWgUkYPCgDV/xgs8p4g==}
engines: {node: '>=18'}
'@huggingface/jinja@0.5.0':
@@ -415,7 +415,7 @@ snapshots:
'@esbuild/win32-x64@0.23.1':
optional: true
- '@huggingface/inference@4.0.0':
+ '@huggingface/inference@4.0.2':
dependencies:
'@huggingface/jinja': 0.5.0
'@huggingface/tasks': 0.19.11