You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/openai/how-to/prompt-caching.md
+1-2Lines changed: 1 addition & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -36,7 +36,6 @@ For a request to take advantage of prompt caching the request must be:
36
36
37
37
When a match is found between a prompt and the current content of the prompt cache it is referred to a cache hit. Cache hits will show up as [`cached_tokens`](/azure/ai-services/openai/reference-preview#cached_tokens) under [`prompt_token_details`](/azure/ai-services/openai/reference-preview#properties-for-prompt_tokens_details) in the chat completions response.
38
38
39
-
40
39
```json
41
40
{
42
41
"created": 1729227448,
@@ -57,7 +56,7 @@ When a match is found between a prompt and the current content of the prompt cac
57
56
"cached_tokens": 1408
58
57
}
59
58
}
60
-
59
+
}
61
60
```
62
61
63
62
After the first 1024 tokens cache hits will occur for every 128 additional identical tokens.
0 commit comments