You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/openai/how-to/dall-e.md
+16-2Lines changed: 16 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -218,6 +218,10 @@ Use the *output_format* parameter to specify the format of the generated image.
218
218
219
219
Use the *output_compression* parameter to specify the compression level for the generated image. Input an integer between `0` and `100`, where `0` is no compression and `100` is maximum compression. The default is `100`.
220
220
221
+
#### Streaming
222
+
223
+
Use the *stream* parameter to enable streaming responses. When set to `true`, the API returns partial images as they are generated. This provides faster visual feedback for users and improves perceived latency. Set the *partial_images* parameter to control how many partial images are generated (1-3).
224
+
221
225
222
226
#### [DALL-E 3](#tab/dalle-3)
223
227
@@ -263,11 +267,12 @@ The format in which DALL-E 3 generated images are returned. Must be one of `url`
263
267
264
268
The Image Edit API allows you to modify existing images based on text prompts you provide. The API call is similar to the image generation API call, but you also need to provide an input image.
265
269
266
-
> [!IMPORTANT]
267
-
> The input image must be less than 20 MB in size and must be a PNG or JPG file.
268
270
269
271
#### [GPT-image-1](#tab/gpt-image-1)
270
272
273
+
> [!IMPORTANT]
274
+
> The input image must be less than 50 MB in size and must be a PNG or JPG file.
275
+
271
276
Send a POST request to:
272
277
273
278
```
@@ -326,11 +331,20 @@ The following API body parameters are available for image editing models, in add
326
331
327
332
The *image* value indicates the image file you want to edit.
328
333
334
+
#### Input fidelity
335
+
336
+
The *input_fidelity* parameter controls how much effort the model will exert to match the style and features, especially facial features, of input images
337
+
338
+
This allows you to make subtle edits to an image without altering unrelated areas. When you use high input fidelity, faces are preserved more accurately than in standard mode.
339
+
329
340
330
341
#### Mask
331
342
332
343
The *mask* parameter is the same type as the main *image* input parameter. It defines the area of the image that you want the model to edit, using fully transparent pixels (alpha of zero) in those areas. The mask must be a PNG file and have the same dimensions as the input image.
333
344
345
+
#### Streaming
346
+
347
+
Use the *stream* parameter to enable streaming responses. When set to `true`, the API returns partial images as they are generated. This provides faster visual feedback for users and improves perceived latency. Set the *partial_images* parameter to control how many partial images are generated (1-3).
| image | string or array | The image(s) to edit. Must be a supported image file or an array of images. Each image should be a png, or jpg file less than 25MB. | Yes ||
667
+
| image | string or array | The image(s) to edit. Must be a supported image file or an array of images. Each image should be a png, or jpg file less than 50MB. | Yes ||
668
+
| input_fidelity| string | Control how much effort the model will exert to match the style and features, especially facial features, of input images. This parameter is only supported for gpt-image-1. Supports `high` and `low`. | no |`low`. |
668
669
| mask | string | An additional image whose fully transparent areas (e.g., where alpha is zero) indicate where the image should be edited. If there are multiple images provided, the mask will be applied to the first image. Must be a valid PNG file, less than 4MB, and have the same dimensions as the image. | No ||
669
670
| n | integer | The number of images to generate. | No | 1 |
670
671
| prompt | string | A text description of the desired image(s). The maximum length is 32000 characters. | Yes ||
671
672
| quality | enum | The quality of the image that will be generated. `high`, `medium` and `low` are only supported for `gpt-image-1`. `dall-e-2` only supports `standard` quality. Defaults to `auto`.<br>Possible values: `standard`, `low`, `medium`, `high`, `auto`| No ||
673
+
|partial_images| integer | The number of partial images to generate. This parameter is used for streaming responses that return partial images. Value must be between 0 and 3. When set to 0, the response will be a single image sent in one streaming event. Note that the final image may be sent before the full number of partial images are generated if the full image is generated more quickly. | 0 |
674
+
| stream | boolean | Edit the image in streaming mode. | no |`false`|
672
675
| response_format | enum | The format in which the generated images are returned. Must be one of `url` or `b64_json`. URLs are only valid for 60 minutes after the image has been generated. This parameter is only supported for `dall-e-2`, as `gpt-image-1` will always return base64-encoded images.<br>Possible values: `url`, `b64_json`| No ||
673
676
| size | enum | The size of the generated images. Must be one of `1024x1024`, `1536x1024` (landscape), `1024x1536` (portrait), or `auto` (default value) for `gpt-image-1`, and one of `256x256`, `512x512`, or `1024x1024` for `dall-e-2`.<br>Possible values: `256x256`, `512x512`, `1024x1024`, `1536x1024`, `1024x1536`, `auto`| No ||
674
677
| user | string | A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. | No ||
@@ -693,7 +696,7 @@ Edits an image from a text caption on a given gpt-image-1 model deployment
POST {endpoint}/openai/v1/images/generations?api-version=preview
@@ -729,6 +732,8 @@ POST {endpoint}/openai/v1/images/generations?api-version=preview
729
732
| output_compression | integer | The compression level (0-100%) for the generated images. This parameter is only supported for `gpt-image-1` with the `webp` or `jpeg` output formats, and defaults to 100. | No | 100 |
730
733
| output_format | enum | The format in which the generated images are returned. This parameter is only supported for `gpt-image-1`. Must be one of `png`, `jpeg`, or `webp`.<br>Possible values: `png`, `jpeg`, `webp`| No ||
731
734
| prompt | string | A text description of the desired image(s). The maximum length is 32000 characters for `gpt-image-1`, 1000 characters for `dall-e-2` and 4000 characters for `dall-e-3`. | Yes ||
735
+
|partial_images| integer | The number of partial images to generate. This parameter is used for streaming responses that return partial images. Value must be between 0 and 3. When set to 0, the response will be a single image sent in one streaming event. Note that the final image may be sent before the full number of partial images are generated if the full image is generated more quickly. | 0 |
736
+
| stream | boolean | Edit the image in streaming mode. | no |`false`|
732
737
| quality | enum | The quality of the image that will be generated. <br><br>- `auto` (default value) will automatically select the best quality for the given model.<br>- `high`, `medium` and `low` are supported for `gpt-image-1`.<br>- `hd` and `standard` are supported for `dall-e-3`.<br>- `standard` is the only option for `dall-e-2`.<br>Possible values: `standard`, `hd`, `low`, `medium`, `high`, `auto`| No ||
733
738
| response_format | enum | The format in which generated images with `dall-e-2` and `dall-e-3` are returned. Must be one of `url` or `b64_json`. URLs are only valid for 60 minutes after the image has been generated. This parameter isn't supported for `gpt-image-1` which will always return base64-encoded images.<br>Possible values: `url`, `b64_json`| No ||
734
739
| size | enum | The size of the generated images. Must be one of `1024x1024`, `1536x1024` (landscape), `1024x1536` (portrait), or `auto` (default value) for `gpt-image-1`, one of `256x256`, `512x512`, or `1024x1024` for `dall-e-2`, and one of `1024x1024`, `1792x1024`, or `1024x1792` for `dall-e-3`.<br>Possible values: `auto`, `1024x1024`, `1536x1024`, `1024x1536`, `256x256`, `512x512`, `1792x1024`, `1024x1792`| No ||
@@ -2024,13 +2029,13 @@ Represents a completion response from the API. Note: both the streamed and non-s
2024
2029
| Name | Type | Description | Required | Default |
| background | enum | Allows to set transparency for the background of the generated image(s). <br>This parameter is only supported for `gpt-image-1`. Must be one of `transparent`, `opaque` or `auto` (default value). When `auto` is used, the model will automatically determine the best background for the image.<br><br>If `transparent`, the output format needs to support transparency, so it should be set to either `png` (default value) or `webp`.<br>Possible values: `transparent`, `opaque`, `auto`| No ||
2027
-
| image | string or array || Yes ||
2032
+
| image | string or array |The image(s) to edit. Must be a supported image file or an array of images. Each image should be a png, or jpg file less than 50MB.| Yes ||
2028
2033
| mask | string || No ||
2029
2034
| model | string | The model deployment to use for the image edit operation. | Yes ||
2030
2035
| n | integer | The number of images to generate. Must be between 1 and 10. | No | 1 |
2031
2036
| prompt | string | A text description of the desired image(s). The maximum length is 1000 characters for `dall-e-2`, and 32000 characters for `gpt-image-1`. | Yes ||
2032
2037
| quality | enum | The quality of the image that will be generated. `high`, `medium` and `low` are only supported for `gpt-image-1`. `dall-e-2` only supports `standard` quality. Defaults to `auto`.<br>Possible values: `standard`, `low`, `medium`, `high`, `auto`| No ||
2033
-
| response_format | enum | The format in which the generated images are returned. Must be one of `url` or `b64_json`. URLs are only valid for 60 minutes after the image has been generated. This parameter is only supported for `dall-e-2`, as `gpt-image-1` will always return base64-encoded images.<br>Possible values: `url`, `b64_json`| No ||
2038
+
| response_format | enum | The format in which the generated images are returned. Must be one of `url` or `b64_json`. <br>Possible values: `url`, `b64_json`| No ||
2034
2039
| size | enum | The size of the generated images. Must be one of `1024x1024`, `1536x1024` (landscape), `1024x1536` (portrait), or `auto` (default value) for `gpt-image-1`, and one of `256x256`, `512x512`, or `1024x1024` for `dall-e-2`.<br>Possible values: `256x256`, `512x512`, `1024x1024`, `1536x1024`, `1024x1536`, `auto`| No ||
2035
2040
| user | string | A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. | No ||
Copy file name to clipboardExpand all lines: articles/ai-foundry/openai/whats-new.md
+11Lines changed: 11 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,6 +18,17 @@ ms.custom:
18
18
19
19
This article provides a summary of the latest releases and major documentation updates for Azure OpenAI.
20
20
21
+
## July 2025
22
+
23
+
### GPT-image-1 update (preview)
24
+
25
+
- Input fidelity parameter: The `input_fidelity` parameter in the image edits API lets you control how closely the model conveys the style and features of the subjects in the original (input) image. This is useful for:
26
+
- Editing photos while preserving facial features; creating avatars that look like original person across different styles; combining faces from multiple people into one image.
27
+
- Maintaining brand identity in generated images for marketing assets, mockups, product photography.
28
+
- E-commerce and fashion, where you need to edit images of outfits or product details without compromising realism.
29
+
30
+
- Partial image streaming: The image generation and image edits APIs support partial image streaming, where they return images with partially rendered content throughout the image generation process. Display these images to the user to provide earlier visual feedback and show the progress of the image generation operation.
Copy file name to clipboardExpand all lines: articles/ai-foundry/responsible-ai/openai/data-privacy.md
+6-4Lines changed: 6 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -34,7 +34,7 @@ Azure OpenAI processes the following types of data:
34
34
35
35
-**Prompts and generated content**. Prompts are submitted by the user, and content is generated by the service, via the completions, chat completions, images, and embeddings operations.
36
36
-**Uploaded data**. You can provide your own data for use with certain service features (e.g., [fine-tuning](/azure/ai-services/openai/how-to/fine-tuning?pivots=programming-language-studio), [assistants API](/azure/ai-services/openai/how-to/batch?tabs=standard-input&pivots=programming-language-ai-studio), [batch processing](/azure/ai-services/openai/how-to/batch?tabs=standard-input&pivots=programming-language-ai-studio)) using the Files API or vector store.
37
-
-**Data for stateful entities**. When you use certain optional features of Azure OpenAI service, such as the Threads feature of the [Assistants API](/azure/ai-services/openai/how-to/assistant) and Stored completions, the service creates a data store to persist message history and other content, in accordance with how you configure the feature.
37
+
-**Data for stateful entities**. When you use certain optional features of Azure OpenAI, such as the [Responses API](/azure/ai-services/openai/how-to/responses), the Threads feature of the [Assistants API](/azure/ai-services/openai/how-to/assistant), and Stored completions, the service creates a data store to persist message history and other content, in accordance with how you configure the feature.
38
38
-**Augmented data included with or via prompts**. When you use data associated with stateful entities, the service retrieves relevant data from a configured data store and augments the prompt to produce generations that are grounded with your data. Prompts may also be augmented with data retrieved from a source included in the prompt itself, such as a URL.
39
39
-**Training & validation data**. You can provide your own training data consisting of prompt-completion pairs for the purposes of [fine-tuning an OpenAI model](/azure/ai-services/openai/how-to/fine-tuning?pivots=programming-language-studio).
40
40
@@ -44,11 +44,12 @@ The diagram below illustrates how your data is processed. This diagram covers se
44
44
45
45
1. How the Azure OpenAI Service processes your prompts via inferencing to generate content (including when additional data from a designated data source is added to a prompt using Azure OpenAI on your data, Assistants, or batch processing).
46
46
1. How the Assistants feature stores data in connection with Messages, Threads, and Runs.
47
+
1. How the Responses API feature stores data to persist message history.
47
48
1. How the Batch feature processes your uploaded data.
48
49
1. How the Azure OpenAI Service creates a fine-tuned (custom) model with your uploaded data.
49
50
1. How the Azure OpenAI Service and Microsoft personnel analyze prompts and completions (text and image) for harmful content and for patterns suggesting the use of the service in a manner that violates the Code of Conduct or other applicable product terms.
50
51
51
-
:::image type="content" source="media\flow.jpg" alt-text="Data flow diagram for the service." lightbox="media\flow.jpg":::
52
+
:::image type="content" source="media\flow-2.jpg" alt-text="Data flow diagram for the service." lightbox="media\flow-2.jpg":::
52
53
53
54
As depicted in the diagram above, managed customers may [apply to modify abuse monitoring](https://customervoice.microsoft.com/Pages/ResponsePage.aspx?id=v4j5cvGGr0GRqy180BHbR7en2Ais5pxKtso_Pz4b1_xUOE9MUTFMUlpBNk5IQlZWWkcyUEpWWEhGOCQlQCN0PWcu).
54
55
@@ -70,9 +71,9 @@ The Azure OpenAI "on your data" feature lets you connect data sources to ground
70
71
71
72
### Data storage for Azure OpenAI Service features
72
73
73
-
Some Azure OpenAI Service features store data in the service. This data is either uploaded by the customer, using the Files API or vector store, or is automatically stored in connection with certain stateful entities such as the Threads feature of the Assistants API and Stored completions. Data stored for Azure OpenAI Service features:
74
+
Some Azure OpenAI Service features store data in the service. This data is either uploaded by the customer, using the Files API or vector store, or is automatically stored in connection with certain stateful entities such as the Responses API, the Threads feature of the Assistants API, and Stored completions. Data stored for Azure OpenAI Service features:
74
75
- Is stored at rest in the Azure OpenAI resource in the customer's Azure tenant, within the same [geography](https://azure.microsoft.com/explore/global-infrastructure/geographies/) as the Azure OpenAI resource;
75
-
- Is always encrypted at rest with Microsoft’s AES-256-encryption by default, with the option of using a customer managed key (certain preview features may not support customermanaged keys).
76
+
- Is always encrypted at rest with Microsoft’s AES-256-encryption by default, with the option of using a customer managed key (certain preview features may not support customer-managed keys). Microsoft-managed keys are always used to ensure baseline encryption for all stored data.
76
77
- Can be deleted by the customer at any time.
77
78
78
79
> [!NOTE]
@@ -82,6 +83,7 @@ Stored data may be used with the following service features/capabilities:
82
83
-**Creating a customized (fine-tuned) model**. Learn more about [how fine-tuning works](/azure/ai-services/openai/how-to/fine-tuning?tabs=turbo%2Cpython-new&pivots=programming-language-studio). Fine-tuned models are exclusively available to the customer whose data was used to create the fine-tuned model, are encrypted at rest (when not deployed for inferencing), and can be deleted by the customer at any time. Training data uploaded for fine-tuning is not used to train, retrain, or improve any Microsoft or third party base models.
83
84
-**Batch processing**. Learn more about [how batch processing works](https://aka.ms/aoai-batch-how-to). Batch processing is a Global deployment type; data stored at rest remains in the designated Azure geography until processing capacity becomes available; processing may occur in any geography where the relevant Azure OpenAI model is deployed (learn more about [region availability of models](/azure/ai-services/openai/concepts/models#model-summary-table-and-region-availability)).
84
85
-**Assistants API (preview)**. Learn more about [how the Assistants API works](/azure/ai-services/openai/concepts/assistants). Some features of Assistants, such as Threads, store message history and other content.
86
+
-**Responses API (preview)**. Learn more about [how the Responses API works](/azure/ai-services/openai/how-to/responses?tabs=python-secure). This API stores message history and other content related to message history. This is required for multi-turn conversations and workflows.
85
87
-**Stored completions (preview)**. <!--Learn more about Stored completions. (link)--> Stored completions stores input-output pairs from the customer’s deployed Azure OpenAI models such as GPT-4o through the chat completions API and displays the pairs in the [Azure AI Foundry portal](https://ai.azure.com/). This allows customers to build datasets with their production data, which can then be used for evaluating or fine-tuning models (as permitted in applicable Product Terms).
0 commit comments