From d1bc81238769c7bb8efc85e3508032b71d6e1858 Mon Sep 17 00:00:00 2001 From: Wauplin <11801849+Wauplin@users.noreply.github.com> Date: Thu, 5 Jun 2025 03:28:31 +0000 Subject: [PATCH] Update Inference Providers documentation (automated) --- docs/inference-providers/providers/hf-inference.md | 12 +----------- docs/inference-providers/providers/sambanova.md | 2 +- docs/inference-providers/providers/together.md | 10 ---------- docs/inference-providers/tasks/chat-completion.md | 2 +- docs/inference-providers/tasks/image-segmentation.md | 3 +-- .../tasks/table-question-answering.md | 1 - docs/inference-providers/tasks/text-generation.md | 2 +- .../tasks/zero-shot-classification.md | 6 +----- scripts/inference-providers/package.json | 2 +- scripts/inference-providers/pnpm-lock.yaml | 10 +++++----- 10 files changed, 12 insertions(+), 38 deletions(-) diff --git a/docs/inference-providers/providers/hf-inference.md b/docs/inference-providers/providers/hf-inference.md index 4d253c83f..000e92667 100644 --- a/docs/inference-providers/providers/hf-inference.md +++ b/docs/inference-providers/providers/hf-inference.md @@ -105,7 +105,7 @@ Find out more about Image Segmentation [here](../tasks/image_segmentation). @@ -198,13 +198,3 @@ Find out more about Translation [here](../tasks/translation). providersMapping={ {"hf-inference":{"modelId":"google-t5/t5-base","providerModelId":"google-t5/t5-base"} } } /> - -### Zero Shot Classification - -Find out more about Zero Shot Classification [here](../tasks/zero_shot_classification). - - - diff --git a/docs/inference-providers/providers/sambanova.md b/docs/inference-providers/providers/sambanova.md index 0512d5d7a..03ba566e4 100644 --- a/docs/inference-providers/providers/sambanova.md +++ b/docs/inference-providers/providers/sambanova.md @@ -45,7 +45,7 @@ Find out more about Chat Completion (LLM) [here](../tasks/chat-completion). diff --git a/docs/inference-providers/providers/together.md b/docs/inference-providers/providers/together.md index 8ccce4cfa..677fdd9e8 100644 --- a/docs/inference-providers/providers/together.md +++ b/docs/inference-providers/providers/together.md @@ -58,16 +58,6 @@ Find out more about Chat Completion (VLM) [here](../tasks/chat-completion). conversational /> -### Text Generation - -Find out more about Text Generation [here](../tasks/text_generation). - - - - ### Text To Image Find out more about Text To Image [here](../tasks/text_to_image). diff --git a/docs/inference-providers/tasks/chat-completion.md b/docs/inference-providers/tasks/chat-completion.md index b4e695d6a..47c4f945b 100644 --- a/docs/inference-providers/tasks/chat-completion.md +++ b/docs/inference-providers/tasks/chat-completion.md @@ -61,7 +61,7 @@ The API supports: diff --git a/docs/inference-providers/tasks/image-segmentation.md b/docs/inference-providers/tasks/image-segmentation.md index 197800740..4467e4fd7 100644 --- a/docs/inference-providers/tasks/image-segmentation.md +++ b/docs/inference-providers/tasks/image-segmentation.md @@ -24,7 +24,6 @@ For more details about the `image-segmentation` task, check out its [dedicated p ### Recommended models -- [openmmlab/upernet-convnext-small](https://huggingface.co/openmmlab/upernet-convnext-small): Solid semantic segmentation model trained on ADE20k. - [facebook/mask2former-swin-large-coco-panoptic](https://huggingface.co/facebook/mask2former-swin-large-coco-panoptic): Panoptic segmentation model trained on the COCO (common objects) dataset. Explore all available models and find the one that suits you best [here](https://huggingface.co/models?inference=warm&pipeline_tag=image-segmentation&sort=trending). @@ -34,7 +33,7 @@ Explore all available models and find the one that suits you best [here](https:/ diff --git a/docs/inference-providers/tasks/table-question-answering.md b/docs/inference-providers/tasks/table-question-answering.md index 397c69dd1..358287c69 100644 --- a/docs/inference-providers/tasks/table-question-answering.md +++ b/docs/inference-providers/tasks/table-question-answering.md @@ -24,7 +24,6 @@ For more details about the `table-question-answering` task, check out its [dedic ### Recommended models -- [microsoft/tapex-base](https://huggingface.co/microsoft/tapex-base): A table question answering model that is capable of neural SQL execution, i.e., employ TAPEX to execute a SQL query on a given table. - [google/tapas-base-finetuned-wtq](https://huggingface.co/google/tapas-base-finetuned-wtq): A robust table question answering model. Explore all available models and find the one that suits you best [here](https://huggingface.co/models?inference=warm&pipeline_tag=table-question-answering&sort=trending). diff --git a/docs/inference-providers/tasks/text-generation.md b/docs/inference-providers/tasks/text-generation.md index 51558b221..4b94d45fd 100644 --- a/docs/inference-providers/tasks/text-generation.md +++ b/docs/inference-providers/tasks/text-generation.md @@ -41,7 +41,7 @@ Explore all available models and find the one that suits you best [here](https:/ diff --git a/docs/inference-providers/tasks/zero-shot-classification.md b/docs/inference-providers/tasks/zero-shot-classification.md index 1c57edfb9..d046ba432 100644 --- a/docs/inference-providers/tasks/zero-shot-classification.md +++ b/docs/inference-providers/tasks/zero-shot-classification.md @@ -24,17 +24,13 @@ For more details about the `zero-shot-classification` task, check out its [dedic ### Recommended models -- [facebook/bart-large-mnli](https://huggingface.co/facebook/bart-large-mnli): Powerful zero-shot text classification model. Explore all available models and find the one that suits you best [here](https://huggingface.co/models?inference=warm&pipeline_tag=zero-shot-classification&sort=trending). ### Using the API - +There are currently no snippet examples for the **zero-shot-classification** task, as no providers support it yet. diff --git a/scripts/inference-providers/package.json b/scripts/inference-providers/package.json index 0803c680e..1da29b5e8 100644 --- a/scripts/inference-providers/package.json +++ b/scripts/inference-providers/package.json @@ -14,7 +14,7 @@ "author": "", "license": "ISC", "dependencies": { - "@huggingface/inference": "^4.0.0", + "@huggingface/inference": "^4.0.2", "@huggingface/tasks": "^0.19.11", "@types/node": "^22.5.0", "handlebars": "^4.7.8", diff --git a/scripts/inference-providers/pnpm-lock.yaml b/scripts/inference-providers/pnpm-lock.yaml index e1a7796f3..5957553a3 100644 --- a/scripts/inference-providers/pnpm-lock.yaml +++ b/scripts/inference-providers/pnpm-lock.yaml @@ -9,8 +9,8 @@ importers: .: dependencies: '@huggingface/inference': - specifier: ^4.0.0 - version: 4.0.0 + specifier: ^4.0.2 + version: 4.0.2 '@huggingface/tasks': specifier: ^0.19.11 version: 0.19.11 @@ -189,8 +189,8 @@ packages: cpu: [x64] os: [win32] - '@huggingface/inference@4.0.0': - resolution: {integrity: sha512-NJD5qnYcOyG48qmqxIgmGZrgeR3+e/Sl+vGHTMvqgTK7vRXntLHzkgzSsHaP+qe7/etvVrr9yVy7vzSdA32dpg==} + '@huggingface/inference@4.0.2': + resolution: {integrity: sha512-XuWb8ocH7lA5kSdXrGnqshtRz3ocSBzEzxcp5xeAXLjgM1ocoIHq+RW8/Ti0xq3MeRGQWgUkYPCgDV/xgs8p4g==} engines: {node: '>=18'} '@huggingface/jinja@0.5.0': @@ -415,7 +415,7 @@ snapshots: '@esbuild/win32-x64@0.23.1': optional: true - '@huggingface/inference@4.0.0': + '@huggingface/inference@4.0.2': dependencies: '@huggingface/jinja': 0.5.0 '@huggingface/tasks': 0.19.11