diff --git a/docs/inference-providers/register-as-a-provider.md b/docs/inference-providers/register-as-a-provider.md index 9952e161f..c4bffc4e7 100644 --- a/docs/inference-providers/register-as-a-provider.md +++ b/docs/inference-providers/register-as-a-provider.md @@ -124,6 +124,7 @@ Congratulations! You now have a JS implementation to successfully make inference First step is to use the Model Mapping API to register which HF models are supported. > [!TIP] +> The completion of step 1. and 2. are pre-requisites for this step. > To proceed with this step, we have to enable your account server-side. Make sure you have an organization on the Hub for your company, and upgrade it to a [Team or Enterprise plan](https://huggingface.co/pricing). ### Register a mapping item @@ -366,6 +367,8 @@ Content-Type: application/json } ``` +This API endpoint will be requested by our system every minute, in batches of up to 10,000 requests. + ### Price Unit We require the price to be a **non-negative integer** number of **nano-USDs** (10^-9 USD). @@ -393,6 +396,39 @@ Inference-Id: unique-id-00131 [response body] ``` +### Exposing pricing through OpenAI /models routes + +If your API is OpenAI-compatible, we expect that you expose LLM pricing information and context length through the [`/v1/models` endpoint](https://platform.openai.com/docs/api-reference/models/list). + +This powers our [provider comparison table](https://huggingface.co/inference/models) and other provider selection features like `:cheapest` (which selects the cheapest provider for a model). + + +
+
+