You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-studio/reference/reference-model-inference-chat-completions.md
-2Lines changed: 0 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -37,7 +37,6 @@ POST /chat/completions?api-version=2024-04-01-preview
37
37
| messages | True |[ChatCompletionRequestMessage](#chatcompletionrequestmessage)| A list of messages comprising the conversation so far. Returns a 422 error if at least some of the messages can't be understood by the model. |
38
38
| frequency\_penalty || number | Helps prevent word repetitions by reducing the chance of a word being selected if it has already been used. The higher the frequency penalty, the less likely the model is to repeat the same words in its output. Return a 422 error if value or parameter is not supported by model. |
39
39
| max\_tokens || integer | The maximum number of tokens that can be generated in the chat completion.<br><br>The total length of input tokens and generated tokens is limited by the model's context length. Passing null causes the model to use its max context length. |
40
-
| model || string | Kept for compatibility reasons. This parameter is ignored. |
41
40
| presence\_penalty || number | Helps prevent the same topics from being repeated by penalizing a word if it exists in the completion already, even just once. Return a 422 error if value or parameter is not supported by model. |
| seed || integer | If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result. Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend. |
@@ -288,7 +287,6 @@ The API call fails when the prompt triggers a content filter as configured. Modi
288
287
| frequency\_penalty | number | 0 | Helps prevent word repetitions by reducing the chance of a word being selected if it has already been used. The higher the frequency penalty, the less likely the model is to repeat the same words in its output. Return a 422 error if value or parameter is not supported by model. |
289
288
| max\_tokens | integer || The maximum number of tokens that can be generated in the chat completion.<br><br>The total length of input tokens and generated tokens is limited by the model's context length. Passing null causes the model to use its max context length. |
290
289
| messages | ChatCompletionRequestMessage\[\]|| A list of messages comprising the conversation so far. Returns a 422 error if at least some of the messages can't be understood by the model. |
291
-
| model | string || Kept for compatibility reasons. This parameter is ignored. |
292
290
| presence\_penalty | number | 0 | Helps prevent the same topics from being repeated by penalizing a word if it exists in the completion already, even just once. Return a 422 error if value or parameter is not supported by model. |
293
291
| response\_format |[ChatCompletionResponseFormat](#chatcompletionresponseformat)| text ||
294
292
| seed | integer || If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result. Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend. |
Copy file name to clipboardExpand all lines: articles/ai-studio/reference/reference-model-inference-completions.md
-2Lines changed: 0 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -37,7 +37,6 @@ POST /completions?api-version=2024-04-01-preview
37
37
| prompt | True || The prompts to generate completions for, encoded as a string, array of strings, array of tokens, or array of token arrays. Note that `<\|endoftext\|>` is the document separator that the model sees during training, so if a prompt is not specified the model generates as if from the beginning of a new document. |
38
38
| frequency\_penalty || number | Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim. |
39
39
| max\_tokens || integer | The maximum number of tokens that can be generated in the completion. The token count of your prompt plus `max_tokens` cannot exceed the model's context length. |
40
-
| model || string | Kept for compatibility reasons. This parameter is ignored. |
41
40
| presence\_penalty || number | Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics. |
42
41
| seed || integer | If specified, the model makes a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result.<br><br>Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend. |
43
42
| stop ||| Sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence. |
@@ -199,7 +198,6 @@ The API call fails when the prompt triggers a content filter as configured. Modi
199
198
| --- | --- | --- | --- |
200
199
| frequency\_penalty | number | 0 | Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim. |
201
200
| max\_tokens | integer | 256 | The maximum number of tokens that can be generated in the completion. The token count of your prompt plus `max_tokens` cannot exceed the model's context length. |
202
-
| model | string || Kept for compatibility reasons. This parameter is ignored. |
203
201
| presence\_penalty | number | 0 | Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics. |
204
202
| prompt ||`<\|endoftext\|>`| The prompts to generate completions for, encoded as a string, array of strings, array of tokens, or array of token arrays. Note that `<\|endoftext\|>` is the document separator that the model sees during training, so if a prompt is not specified the model generates as if from the beginning of a new document. |
205
203
| seed | integer || If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result.<br><br>Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend. |
Copy file name to clipboardExpand all lines: articles/ai-studio/reference/reference-model-inference-embeddings.md
+8-10Lines changed: 8 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,7 +38,6 @@ POST /embeddings?api-version=2024-04-01-preview
38
38
| dimensions || integer | The number of dimensions the resulting output embeddings should have. Returns a 422 error if the model doesn't support the value or parameter. |
39
39
| encoding_format ||[EmbeddingEncodingFormat](#embeddingencodingformat)| The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
40
40
| input_type ||[EmbeddingInputType](#embeddinginputtype)| The type of the input. Either `text`, `query`, or `document`. Returns a 422 error if the model doesn't support the value or parameter. |
41
-
| model || string | Kept for compatibility reasons. This parameter is ignored. |
| dimensions | integer || The number of dimensions the resulting output embeddings should have. Returns a 422 error if the model doesn't support the value or parameter. |
159
-
| encoding\_format |[EmbeddingEncodingFormat](#embeddingencodingformat)| float | The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
160
-
| input ||| Input text to embed, encoded as a string or array of tokens. To embed multiple inputs in a single request, pass an array of strings or array of token arrays. |
161
-
| input\_type |[EmbeddingInputType](#embeddinginputtype)| text | The type of the input. Either `text`, `query`, or `document`. Returns a 422 error if the model doesn't support the value or parameter. |
162
-
| model | string || Kept for compatibility reasons. This parameter is ignored. |
| input | True | string[]| Input text to embed, encoded as a string or array of tokens. To embed multiple inputs in a single request, pass an array of strings or array of token arrays. |
158
+
| dimensions || integer | The number of dimensions the resulting output embeddings should have. Returns a 422 error if the model doesn't support the value or parameter. |
159
+
| encoding_format ||[EmbeddingEncodingFormat](#embeddingencodingformat)| The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
160
+
| input_type ||[EmbeddingInputType](#embeddinginputtype)| The type of the input. Either `text`, `query`, or `document`. Returns a 422 error if the model doesn't support the value or parameter. |
Copy file name to clipboardExpand all lines: articles/ai-studio/reference/reference-model-inference-images-embeddings.md
-2Lines changed: 0 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,7 +38,6 @@ POST /images/embeddings?api-version=2024-04-01-preview
38
38
| input | True |[EmbeddingInput](#embeddinginput)\[\]| Input image to embed. To embed multiple inputs in a single request, pass an array. The input must not exceed the max input tokens for the model. |
39
39
| dimensions || integer | The number of dimensions the resulting output embeddings should have. Passing null causes the model to use its default value. Returns a 422 error if the model doesn't support the value or parameter. |
40
40
| encoding\_format ||[EmbeddingEncodingFormat](#embeddingencodingformat)| The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
41
-
| model || string | Kept for compatibility reasons. This parameter is ignored. |
42
41
43
42
44
43
## Responses
@@ -186,7 +185,6 @@ The API call fails when the prompt triggers a content filter as configured. Modi
186
185
| dimensions | integer || The number of dimensions the resulting output embeddings should have. Passing null causes the model to use its default value. Returns a 422 error if the model doesn't support the value or parameter. |
187
186
| encoding\_format |[EmbeddingEncodingFormat](#embeddingencodingformat)| float | The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
188
187
| input |[EmbeddingInput](#embeddinginput)\[\]|| Input image to embed. To embed multiple inputs in a single request, pass an array. The input must not exceed the max input tokens for the model. |
189
-
| model | string || Kept for compatibility reasons. This parameter is ignored. |
0 commit comments