Skip to content

Commit 747ff1e

Browse files
committed
review
1 parent ddf8494 commit 747ff1e

File tree

2 files changed

+6
-0
lines changed

2 files changed

+6
-0
lines changed

articles/api-management/llm-emit-token-metric-policy.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -18,6 +18,9 @@ ms.custom:
1818

1919
The `llm-emit-token-metric` policy sends metrics to Application Insights about consumption of large language model (LLM) tokens through LLM APIs. Token count metrics include: Total Tokens, Prompt Tokens, and Completion Tokens.
2020

21+
> [!NOTE]
22+
> Currently, this policy is in preview.
23+
2124
[!INCLUDE [api-management-policy-generic-alert](../../includes/api-management-policy-generic-alert.md)]
2225

2326
[!INCLUDE [api-management-llm-models](../../includes/api-management-llm-models.md)]

articles/api-management/llm-token-limit-policy.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,9 @@ The `llm-token-limit` policy prevents large language model (LLM) API usage spike
2020

2121
By relying on token usage metrics returned from the LLM endpoint, the policy can accurately monitor and enforce limits in real time. The policy also enables precalculation of prompt tokens by API Management, minimizing unnecessary requests to the LLM backend if the limit is already exceeded.
2222

23+
> [!NOTE]
24+
> Currently, this policy is in preview.
25+
2326
[!INCLUDE [api-management-policy-generic-alert](../../includes/api-management-policy-generic-alert.md)]
2427

2528
[!INCLUDE [api-management-llm-models](../../includes/api-management-llm-models.md)]

0 commit comments

Comments
 (0)