Skip to content

Commit 77cfaad

Browse files
committed
edit pass: deploy-and-monitor-flows
1 parent 86c47a9 commit 77cfaad

File tree

3 files changed

+11
-13
lines changed

3 files changed

+11
-13
lines changed

articles/ai-foundry/how-to/develop/trace-production-sdk.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -75,14 +75,14 @@ The **Dependency** type event records calls from your deployments. The name of t
7575

7676
| Metrics name | Type | Dimensions | Description |
7777
|--------------------------------------|-----------|-------------------------------------------|---------------------------------------------------------------------------------|
78-
| token_consumption | counter | - flow <br> - node<br> - llm_engine<br> - token_type: `prompt_tokens`: LLM API input tokens; `completion_tokens`: LLM API response tokens; `total_tokens` = `prompt_tokens + completion tokens` | OpenAI token consumption metrics |
79-
| flow_latency | histogram | flow, response_code, streaming, response_type | request execution cost, response_type means whether it's full/firstbyte/lastbyte|
80-
| flow_request | counter | flow, response_code, exception, streaming | flow request count |
81-
| node_latency | histogram | flow, node, run_status | node execution cost |
82-
| node_request | counter | flow, node, exception, run_status | node execution count |
83-
| rpc_latency | histogram | flow, node, api_call | rpc cost |
84-
| rpc_request | counter | flow, node, api_call, exception | rpc count |
85-
| flow_streaming_response_duration | histogram | flow | streaming response sending cost, from sending first byte to sending last byte |
78+
| `token_consumption` | counter | - flow <br> - node<br> - `llm_engine`<br> - `token_type`: `prompt_tokens`: LLM API input tokens; `completion_tokens`: LLM API response tokens; `total_tokens` = `prompt_tokens + completion tokens` | OpenAI token consumption metrics |
79+
| `flow_latency` | histogram | flow, `response_code`, streaming, `response_type` | request execution cost, `response_type` means whether it's full/firstbyte/lastbyte|
80+
| `flow_request` | counter | flow, `response_code`, exception, streaming | flow request count |
81+
| `node_latency` | histogram | flow, node, `run_status` | node execution cost |
82+
| `node_request` | counter | flow, node, exception, `run_status` | node execution count |
83+
| `rpc_latency` | histogram | flow, node, `api_call` | rpc cost |
84+
| `rpc_request` | counter | flow, node, `api_call`, exception | rpc count |
85+
| `flow_streaming_response_duration` | histogram | flow | streaming response sending cost, from sending first byte to sending last byte |
8686

8787
You can find the workspace default Application Insights metrics on your workspace overview page in the Azure portal.
8888

articles/ai-foundry/how-to/flow-deploy.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -237,8 +237,6 @@ If you aren't going to use the endpoint after you finish this tutorial, delete t
237237

238238
## Related content
239239

240-
## Next steps
241-
242240
- Learn more about what you can do in [Azure AI Foundry](../what-is-ai-foundry.md).
243241
- Get answers to frequently asked questions in the [Azure AI Foundry FAQ](../faq.yml).
244242
- [Enable trace and collect feedback for your deployment](./develop/trace-production-sdk.md).

articles/ai-foundry/how-to/prompt-flow-troubleshoot.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ Errors related to compute session failures that use a custom base image are disc
3737

3838
Flow run-related issues are discussed.
3939

40-
### How do I find the raw inputs and outputs of in the LLM tool for further investigation?
40+
### How do I find the raw inputs and outputs of the LLM tool for further investigation?
4141

4242
In a prompt flow, on a **Flow** page with a successful run and run detail page, you can find the raw inputs and outputs of the LLM tool in the output section. Select **View full output** to view the full output.
4343

@@ -87,7 +87,7 @@ You might encounter a 409 error from Azure OpenAI. This error means that you rea
8787

8888
Flow deployment-related issues are discussed.
8989

90-
### Upstream request time-out issue when consuming the endpoint
90+
### How do I resolve an upstream request time-out issue?
9191

9292
If you use the Azure CLI or SDK to deploy the flow, you might encounter a time-out error. By default, `request_timeout_ms` is 5000. You can specify a maximum of five minutes, which is 300,000 ms. The following example shows how to specify a `request timeout` in the deployment yaml file. To learn more, see [deployment schema](/azure/machine-learning/reference-yaml-deployment-managed-online).
9393

@@ -96,7 +96,7 @@ request_settings:
9696
request_timeout_ms: 300000
9797
```
9898
99-
### What do I do when OpenAI API hits an authentication error?
99+
### What do I do when OpenAI API generates an authentication error?
100100
101101
If you regenerate your Azure OpenAI key and manually update the connection used in a prompt flow, you might encounter errors like "Unauthorized. Access token is missing, invalid, audience is incorrect or have expired." You might see these messages when you invoke an existing endpoint that was created before the key was regenerated.
102102

0 commit comments

Comments
 (0)