Skip to content

Commit 912d9c7

Browse files
committed
Learn Editor: Update prompt-caching.md
1 parent 80b31e2 commit 912d9c7

File tree

1 file changed

+5
-2
lines changed

1 file changed

+5
-2
lines changed

articles/ai-services/openai/how-to/prompt-caching.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,10 +22,13 @@ Currently only the following models support prompt caching with Azure OpenAI:
2222

2323
- `o1-preview-2024-09-12`
2424
- `o1-mini-2024-09-12`
25+
- `gpt-4o-2024-05-13`
26+
- `gpt-4o-2024-08-06`
27+
- `gpt-4o-mini-2024-07-18`
2528

2629
## API support
2730

28-
Official support for prompt caching was first added in API version `2024-10-01-preview`.
31+
Official support for prompt caching was first added in API version `2024-10-01-preview`. At this time, only `o1-preview-2024-09-12` and `o1-mini-2024-09-12` models support the `cached_tokens` API response parameter.
2932

3033
## Getting started
3134

@@ -67,7 +70,7 @@ A single character difference in the first 1,024 tokens will result in a cache m
6770

6871
The o1-series models are text only and don't support system messages, images, tool use/function calling, or structured outputs. This limits the efficacy of prompt caching for these models to the user/assistant portions of the messages array which are less likely to have an identical 1024 token prefix.
6972

70-
Once prompt caching is enabled for other supported models prompt caching will expand to support:
73+
For `gpt-4o` and `gpt-4o-mini` models, prompt caching is supported for:
7174

7275
| **Caching Supported** | **Description** |
7376
|--------|--------|

0 commit comments

Comments
 (0)