Skip to content

Commit 0a8022e

Browse files
committed
Temporarily adding back to all documentation.
1 parent bce810b commit 0a8022e

File tree

3 files changed

+5
-0
lines changed

3 files changed

+5
-0
lines changed

articles/api-management/azure-openai-enable-semantic-caching.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -127,6 +127,7 @@ Configure the following policies to enable semantic caching for Azure OpenAI API
127127
<azure-openai-semantic-cache-lookup
128128
score-threshold="0.8"
129129
embeddings-backend-id="embeddings-deployment"
130+
embeddings-backend-auth="system-assigned"
130131
ignore-system-messages="true"
131132
max-message-count="10">
132133
<vary-by>@(context.Subscription.Id)</vary-by>

articles/api-management/azure-openai-semantic-cache-lookup-policy.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,7 @@ Use the `azure-openai-semantic-cache-lookup` policy to perform cache lookup of r
3434
<azure-openai-semantic-cache-lookup
3535
score-threshold="similarity score threshold"
3636
embeddings-backend-id ="backend entity ID for embeddings API"
37+
embeddings-backend-auth ="system-assigned"
3738
ignore-system-messages="true | false"
3839
max-message-count="count" >
3940
<vary-by>"expression to partition caching"</vary-by>
@@ -46,6 +47,7 @@ Use the `azure-openai-semantic-cache-lookup` policy to perform cache lookup of r
4647
| ----------------- | ------------------------------------------------------ | -------- | ------- |
4748
| score-threshold | Similarity score threshold used to determine whether to return a cached response to a prompt. Value is a decimal between 0.0 and 1.0. [Learn more](../azure-cache-for-redis/cache-tutorial-semantic-cache.md#change-the-similarity-threshold). | Yes | N/A |
4849
| embeddings-backend-id | [Backend](backends.md) ID for OpenAI embeddings API call. | Yes | N/A |
50+
| embeddings-backend-auth | Authentication used for Azure OpenAI embeddings API backend. | Yes. Must be set to `system-assigned`. | N/A |
4951
| ignore-system-messages | Boolean. If set to `true`, removes system messages from a GPT chat completion prompt before assessing cache similarity. | No | false |
5052
| max-message-count | If specified, number of remaining dialog messages after which caching is skipped. | No | N/A |
5153

articles/api-management/llm-semantic-cache-lookup-policy.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,7 @@ Use the `llm-semantic-cache-lookup` policy to perform cache lookup of responses
3434
<llm-semantic-cache-lookup
3535
score-threshold="similarity score threshold"
3636
embeddings-backend-id ="backend entity ID for embeddings API"
37+
embeddings-backend-auth ="system-assigned"
3738
ignore-system-messages="true | false"
3839
max-message-count="count" >
3940
<vary-by>"expression to partition caching"</vary-by>
@@ -46,6 +47,7 @@ Use the `llm-semantic-cache-lookup` policy to perform cache lookup of responses
4647
| ----------------- | ------------------------------------------------------ | -------- | ------- |
4748
| score-threshold | Similarity score threshold used to determine whether to return a cached response to a prompt. Value is a decimal between 0.0 and 1.0. [Learn more](../azure-cache-for-redis/cache-tutorial-semantic-cache.md#change-the-similarity-threshold). | Yes | N/A |
4849
| embeddings-backend-id | [Backend](backends.md) ID for OpenAI embeddings API call. | Yes | N/A |
50+
| embeddings-backend-auth | Authentication used for Azure OpenAI embeddings API backend. | Yes. Must be set to `system-assigned`. | N/A |
4951
| ignore-system-messages | Boolean. If set to `true`, removes system messages from a GPT chat completion prompt before assessing cache similarity. | No | false |
5052
| max-message-count | If specified, number of remaining dialog messages after which caching is skipped. | No | N/A |
5153

0 commit comments

Comments
 (0)