Skip to content

Commit 3cc9f0c

Browse files
committed
revision
1 parent c09fe77 commit 3cc9f0c

File tree

4 files changed

+8
-16
lines changed

4 files changed

+8
-16
lines changed

articles/ai-studio/reference/reference-model-inference-chat-completions.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,6 @@ POST /chat/completions?api-version=2024-04-01-preview
3737
| messages | True | [ChatCompletionRequestMessage](#chatcompletionrequestmessage) | A list of messages comprising the conversation so far. Returns a 422 error if at least some of the messages can't be understood by the model. |
3838
| frequency\_penalty | | number | Helps prevent word repetitions by reducing the chance of a word being selected if it has already been used. The higher the frequency penalty, the less likely the model is to repeat the same words in its output. Return a 422 error if value or parameter is not supported by model. |
3939
| max\_tokens | | integer | The maximum number of tokens that can be generated in the chat completion.<br><br>The total length of input tokens and generated tokens is limited by the model's context length. Passing null causes the model to use its max context length. |
40-
| model | | string | Kept for compatibility reasons. This parameter is ignored. |
4140
| presence\_penalty | | number | Helps prevent the same topics from being repeated by penalizing a word if it exists in the completion already, even just once. Return a 422 error if value or parameter is not supported by model. |
4241
| response\_format | | [ChatCompletionResponseFormat](#chatcompletionresponseformat) | |
4342
| seed | | integer | If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result. Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend. |
@@ -288,7 +287,6 @@ The API call fails when the prompt triggers a content filter as configured. Modi
288287
| frequency\_penalty | number | 0 | Helps prevent word repetitions by reducing the chance of a word being selected if it has already been used. The higher the frequency penalty, the less likely the model is to repeat the same words in its output. Return a 422 error if value or parameter is not supported by model. |
289288
| max\_tokens | integer | | The maximum number of tokens that can be generated in the chat completion.<br><br>The total length of input tokens and generated tokens is limited by the model's context length. Passing null causes the model to use its max context length. |
290289
| messages | ChatCompletionRequestMessage\[\] | | A list of messages comprising the conversation so far. Returns a 422 error if at least some of the messages can't be understood by the model. |
291-
| model | string | | Kept for compatibility reasons. This parameter is ignored. |
292290
| presence\_penalty | number | 0 | Helps prevent the same topics from being repeated by penalizing a word if it exists in the completion already, even just once. Return a 422 error if value or parameter is not supported by model. |
293291
| response\_format | [ChatCompletionResponseFormat](#chatcompletionresponseformat) | text | |
294292
| seed | integer | | If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result. Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend. |

articles/ai-studio/reference/reference-model-inference-completions.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,6 @@ POST /completions?api-version=2024-04-01-preview
3737
| prompt | True | | The prompts to generate completions for, encoded as a string, array of strings, array of tokens, or array of token arrays. Note that `<\|endoftext\|>` is the document separator that the model sees during training, so if a prompt is not specified the model generates as if from the beginning of a new document. |
3838
| frequency\_penalty | | number | Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim. |
3939
| max\_tokens | | integer | The maximum number of tokens that can be generated in the completion. The token count of your prompt plus `max_tokens` cannot exceed the model's context length. |
40-
| model | | string | Kept for compatibility reasons. This parameter is ignored. |
4140
| presence\_penalty | | number | Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics. |
4241
| seed | | integer | If specified, the model makes a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result.<br><br>Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend. |
4342
| stop | | | Sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence. |
@@ -199,7 +198,6 @@ The API call fails when the prompt triggers a content filter as configured. Modi
199198
| --- | --- | --- | --- |
200199
| frequency\_penalty | number | 0 | Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim. |
201200
| max\_tokens | integer | 256 | The maximum number of tokens that can be generated in the completion. The token count of your prompt plus `max_tokens` cannot exceed the model's context length. |
202-
| model | string | | Kept for compatibility reasons. This parameter is ignored. |
203201
| presence\_penalty | number | 0 | Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics. |
204202
| prompt | | `<\|endoftext\|>` | The prompts to generate completions for, encoded as a string, array of strings, array of tokens, or array of token arrays. Note that `<\|endoftext\|>` is the document separator that the model sees during training, so if a prompt is not specified the model generates as if from the beginning of a new document. |
205203
| seed | integer | | If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result.<br><br>Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend. |

articles/ai-studio/reference/reference-model-inference-embeddings.md

Lines changed: 8 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,6 @@ POST /embeddings?api-version=2024-04-01-preview
3838
| dimensions | | integer | The number of dimensions the resulting output embeddings should have. Returns a 422 error if the model doesn't support the value or parameter. |
3939
| encoding_format | | [EmbeddingEncodingFormat](#embeddingencodingformat) | The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
4040
| input_type | | [EmbeddingInputType](#embeddinginputtype) | The type of the input. Either `text`, `query`, or `document`. Returns a 422 error if the model doesn't support the value or parameter. |
41-
| model | | string | Kept for compatibility reasons. This parameter is ignored. |
4241

4342
## Responses
4443

@@ -153,13 +152,12 @@ Status code: 200
153152

154153
### CreateEmbeddingRequest
155154

156-
| Name | Type | Default Value | Description |
157-
| ---------------- | --------------------------------------------------- | ------------- | ----------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
158-
| dimensions | integer | | The number of dimensions the resulting output embeddings should have. Returns a 422 error if the model doesn't support the value or parameter. |
159-
| encoding\_format | [EmbeddingEncodingFormat](#embeddingencodingformat) | float | The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
160-
| input | | | Input text to embed, encoded as a string or array of tokens. To embed multiple inputs in a single request, pass an array of strings or array of token arrays. |
161-
| input\_type | [EmbeddingInputType](#embeddinginputtype) | text | The type of the input. Either `text`, `query`, or `document`. Returns a 422 error if the model doesn't support the value or parameter. |
162-
| model | string | | Kept for compatibility reasons. This parameter is ignored. |
155+
| Name | Required | Type | Description |
156+
| --------------- | -------- | --------------------------------------------------- | ----------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
157+
| input | True | string[] | Input text to embed, encoded as a string or array of tokens. To embed multiple inputs in a single request, pass an array of strings or array of token arrays. |
158+
| dimensions | | integer | The number of dimensions the resulting output embeddings should have. Returns a 422 error if the model doesn't support the value or parameter. |
159+
| encoding_format | | [EmbeddingEncodingFormat](#embeddingencodingformat) | The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
160+
| input_type | | [EmbeddingInputType](#embeddinginputtype) | The type of the input. Either `text`, `query`, or `document`. Returns a 422 error if the model doesn't support the value or parameter. |
163161

164162
### CreateEmbeddingResponse
165163

@@ -180,11 +178,11 @@ Status code: 200
180178

181179
### Embedding
182180

183-
Represents an image generated.
181+
Represents the embedding generated.
184182

185183
| Name | Type | Description |
186184
| --------- | ----------------------------------- | ------------------------------------------------------------------------------------------------ |
187-
| embedding | number\[\] | The embedding vector. The length of vector depends on the model used and the type depends on the `encoding_format` used. |
185+
| embedding | \[\] | The embedding vector. The length of vector depends on the model used and the type depends on the `encoding_format` used. |
188186
| index | integer | The index of the embedding in the list of embeddings. |
189187
| object | [EmbeddingObject](#embeddingobject) | The object type, which is always "embedding". |
190188

articles/ai-studio/reference/reference-model-inference-images-embeddings.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,6 @@ POST /images/embeddings?api-version=2024-04-01-preview
3838
| input | True | [EmbeddingInput](#embeddinginput)\[\] | Input image to embed. To embed multiple inputs in a single request, pass an array. The input must not exceed the max input tokens for the model. |
3939
| dimensions | | integer | The number of dimensions the resulting output embeddings should have. Passing null causes the model to use its default value. Returns a 422 error if the model doesn't support the value or parameter. |
4040
| encoding\_format | | [EmbeddingEncodingFormat](#embeddingencodingformat) | The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
41-
| model | | string | Kept for compatibility reasons. This parameter is ignored. |
4241

4342

4443
## Responses
@@ -186,7 +185,6 @@ The API call fails when the prompt triggers a content filter as configured. Modi
186185
| dimensions | integer | | The number of dimensions the resulting output embeddings should have. Passing null causes the model to use its default value. Returns a 422 error if the model doesn't support the value or parameter. |
187186
| encoding\_format | [EmbeddingEncodingFormat](#embeddingencodingformat) | float | The format to return the embeddings in. Either base64, float, int8, uint8, binary, or ubinary. Returns a 422 error if the model doesn't support the value or parameter. |
188187
| input | [EmbeddingInput](#embeddinginput)\[\] | | Input image to embed. To embed multiple inputs in a single request, pass an array. The input must not exceed the max input tokens for the model. |
189-
| model | string | | Kept for compatibility reasons. This parameter is ignored. |
190188

191189

192190
### Detail

0 commit comments

Comments
 (0)