Skip to content

Commit b18426c

Browse files
authored
Update reference-model-inference-api.md
1 parent 7b85445 commit b18426c

File tree

1 file changed

+25
-1
lines changed

1 file changed

+25
-1
lines changed

articles/ai-studio/reference/reference-model-inference-api.md

Lines changed: 25 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -204,7 +204,7 @@ extra-parameters: pass-through
204204
---
205205

206206
> [!TIP]
207-
> Alternatively, you can set `extra-parameters: drop` to drop any unknown parameter in the request. Use this capability in case you happen to be sending requests with extra parameters that you know the model won't support but you want the request to completes anyway. A typical example of this is indicating `seed` parameter.
207+
> The default value for `extra-parameters` is `error` which returns an error if an extra parameter is indicated in the payload. Alternatively, you can set `extra-parameters: ignore` to drop any unknown parameter in the request. Use this capability in case you happen to be sending requests with extra parameters that you know the model won't support but you want the request to completes anyway. A typical example of this is indicating `seed` parameter.
208208
209209
### Models with disparate set of capabilities
210210

@@ -427,3 +427,27 @@ __Response__
427427
## Getting started
428428

429429
The Azure AI Model Inference API is currently supported in certain models deployed as [Serverless API endpoints](../how-to/deploy-models-serverless.md) and Managed Online Endpoints. Deploy any of the [supported models](#availability) and use the exact same code to consume their predictions.
430+
431+
# [Python](#tab/python)
432+
433+
The client library `azure-ai-inference` does inference, including chat completions, for AI models deployed by Azure AI Studio and Azure Machine Learning Studio. It supports Serverless API endpoints and Managed Compute endpoints (formerly known as Managed Online Endpoints).
434+
435+
Explore our [samples](https://github.com/Azure/azure-sdk-for-python/tree/main/sdk/ai/azure-ai-inference/samples) and read the [API reference documentation](https://aka.ms/azsdk/azure-ai-inference/python/reference) to get yourself started.
436+
437+
# [JavaScript](#tab/javascript)
438+
439+
The client library `@azure-rest/ai-inference` does inference, including chat completions, for AI models deployed by Azure AI Studio and Azure Machine Learning Studio. It supports Serverless API endpoints and Managed Compute endpoints (formerly known as Managed Online Endpoints).
440+
441+
Explore our [samples](https://github.com/Azure/azure-sdk-for-js/tree/main/sdk/ai/ai-inference-rest/samples) and read the [API reference documentation](https://aka.ms/AAp1kxa) to get yourself started.
442+
443+
# [REST](#tab/rest)
444+
445+
Explore the reference section of the Azure AI model inference API to see parameters and options to consume models, including chat completions models, deployed by Azure AI Studio and Azure Machine Learning Studio. It supports Serverless API endpoints and Managed Compute endpoints (formerly known as Managed Online Endpoints).
446+
447+
* [Get info](reference-model-inference-info.md): Returns the information about the model deployed under the endpoint.
448+
* [Text embeddings](reference-model-inference-embeddings.md): Creates an embedding vector representing the input text.
449+
* [Text completions](reference-model-inference-completions.md): Creates a completion for the provided prompt and parameters.
450+
* [Chat completions](reference-model-inference-chat-completions.md): Creates a model response for the given chat conversation.
451+
* [Image embeddings](reference-model-inference-images-embeddings.md): Creates an embedding vector representing the input text and image.
452+
453+
---

0 commit comments

Comments
 (0)