|
1 | | -# How to use OpenAI's GPT OSS |
| 1 | +# How to use OpenAI gpt-oss |
2 | 2 |
|
3 | 3 | <div class="flex justify-center"> |
4 | 4 | <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/inference-providers-guides/gpt-oss-thumbnail-light.png"/> |
5 | 5 | </div> |
6 | 6 |
|
7 | | -This guide walks you through using OpenAI's latest GPT OSS models with Hugging Face Inference Providers. GPT OSS is an open-weights family built for strong reasoning, agentic workflows and versatile developer use cases, and it comes in two sizes: a one with 120B parameters ([gpt-oss-120b](https://hf.co/openai/gpt-oss-120b)), and a smaller one with 20B parameters ([gpt-oss-20b](https://hf.co/openai/gpt-oss-120b)). |
| 7 | +This guide walks you through using OpenAI's latest gpt-oss models with Hugging Face Inference Providers which powers the official OpenAI playground ([gpt-oss.com](https://gpt-oss.com)). gpt-oss is an open-weights family built for strong reasoning, agentic workflows and versatile developer use cases, and it comes in two sizes: a one with 120B parameters [gpt-oss-120b](https://hf.co/openai/gpt-oss-120b), and a smaller one with 20B parameters ([gpt-oss-20b](https://hf.co/openai/gpt-oss-120b)). |
8 | 8 |
|
9 | 9 | Both models are supported on Inference Providers and can be accessed through either the OpenAI-compatible [Chat Completions API](https://platform.openai.com/docs/api-reference/chat/completions), or the more advanced [Responses API](https://platform.openai.com/docs/api-reference/responses). |
10 | 10 |
|
@@ -39,7 +39,7 @@ npm install openai |
39 | 39 | </hfoptions> |
40 | 40 |
|
41 | 41 | ## Chat Completion |
42 | | -Getting started with GPT OSS models on Inference Providers is simple and straightforward. The OpenAI-compatible Chat Completions API supports features like tool calling, structured outputs, streaming, and reasoning effort controls. |
| 42 | +Getting started with gpt-oss models on Inference Providers is simple and straightforward. The OpenAI-compatible Chat Completions API supports features like tool calling, structured outputs, streaming, and reasoning effort controls. |
43 | 43 |
|
44 | 44 | Here's a basic example using [gpt-oss-120b](https://hf.co/openai/gpt-oss-120b) through the fast Cerebras provider: |
45 | 45 |
|
@@ -282,7 +282,7 @@ console.log(parsedOutput); |
282 | 282 | </hfoption> |
283 | 283 | </hfoptions> |
284 | 284 |
|
285 | | -With just a few lines of code, you can start using GPT OSS models with Hugging Face Inference Providers, fully OpenAI API-compatible, easy to integrate, and ready out of the box! |
| 285 | +With just a few lines of code, you can start using gpt-oss models with Hugging Face Inference Providers, fully OpenAI API-compatible, easy to integrate, and ready out of the box! |
286 | 286 |
|
287 | 287 | ## Responses API |
288 | 288 |
|
@@ -566,5 +566,5 @@ response.output.forEach((item, index) => { |
566 | 566 | </hfoption> |
567 | 567 | </hfoptions> |
568 | 568 |
|
569 | | -That's it! With the Responses API on Inference Providers, you get fine-grained control over powerful open-weight models like GPT OSS, including streaming, tool calling, and remote MCP, making it ideal for building reliable, agent-driven applications. |
| 569 | +That's it! With the Responses API on Inference Providers, you get fine-grained control over powerful open-weight models like gpt-oss, including streaming, tool calling, and remote MCP, making it ideal for building reliable, agent-driven applications. |
570 | 570 |
|
0 commit comments