You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-studio/how-to/prompt-flow-troubleshoot.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -40,17 +40,17 @@ Compute session support to use `requirements.txt` or custom base image in `flow.
40
40
41
41
In prompt flow, on flow page with successful run and run detail page, you can find the raw inputs and outputs of LLM tool in the output section. Select the `view full output` button to view full output.
42
42
43
-
:::image type="content" source="./media/prompt-flow/view-full-output.png" alt-text="Screenshot that shows view full output on LLM node." lightbox = "./media/prompt-flow/view-full-output.png":::
43
+
:::image type="content" source="../media/prompt-flow/view-full-output.png" alt-text="Screenshot that shows view full output on LLM node." lightbox = "../media/prompt-flow/view-full-output.png":::
44
44
45
45
`Trace` section includes each request and response to the LLM tool. You can check the raw message sent to the LLM model and the raw response from the LLM model.
46
46
47
-
:::image type="content" source="./media/prompt-flow/trace-large-language-model-tool.png" alt-text="Screenshot that shows raw request send to LLM model and response from LLM model." lightbox = "./media/prompt-flow/trace-large-language-model-tool.png":::
47
+
:::image type="content" source="../media/prompt-flow/trace-large-language-model-tool.png" alt-text="Screenshot that shows raw request send to LLM model and response from LLM model." lightbox = "../media/prompt-flow/trace-large-language-model-tool.png":::
48
48
49
49
### How to fix 409 error from Azure OpenAI?
50
50
51
51
You may encounter 409 error from Azure OpenAI, it means you have reached the rate limit of Azure OpenAI. You can check the error message in the output section of LLM node. Learn more about [Azure OpenAI rate limit](../../ai-services/openai/quotas-limits.md).
52
52
53
-
:::image type="content" source="./media/prompt-flow/429-rate-limit.png" alt-text="Screenshot that shows 429 rate limit error from Azure OpenAI." lightbox = "./media/prompt-flow/429-rate-limit.png":::
53
+
:::image type="content" source="../media/prompt-flow/429-rate-limit.png" alt-text="Screenshot that shows 429 rate limit error from Azure OpenAI." lightbox = "../media/prompt-flow/429-rate-limit.png":::
54
54
55
55
### Identify which node consumes the most time
56
56
@@ -64,13 +64,13 @@ You may encounter 409 error from Azure OpenAI, it means you have reached the rat
64
64
65
65
-**Case 1:** Python script node runs for a long time.
66
66
67
-
:::image type="content" source="./media/prompt-flow/runtime-timeout-running-for-long-time.png" alt-text="Screenshot that shows a timeout run sign." lightbox = "./media/prompt-flow/runtime-timeout-running-for-long-time.png":::
67
+
:::image type="content" source="../media/prompt-flow/runtime-timeout-running-for-long-time.png" alt-text="Screenshot that shows a timeout run sign." lightbox = "../media/prompt-flow/runtime-timeout-running-for-long-time.png":::
68
68
69
69
In this case, you can find that `PythonScriptNode` was running for a long time (almost 300 seconds). Then you can check the node details to see what's the problem.
70
70
71
71
-**Case 2:** LLM node runs for a long time.
72
72
73
-
:::image type="content" source="./media/prompt-flow/rruntime-timeout-by-language-model-timeout.png" alt-text="Screenshot that shows timeout logs caused by an LLM timeout." lightbox = "./media/prompt-flow/how-to-create-manage-runtime/runtime-timeout-by-language-model-timeout.png":::
73
+
:::image type="content" source="../media/prompt-flow/rruntime-timeout-by-language-model-timeout.png" alt-text="Screenshot that shows timeout logs caused by an LLM timeout." lightbox = "../media/prompt-flow/how-to-create-manage-runtime/runtime-timeout-by-language-model-timeout.png":::
74
74
75
75
In this case, if you find the message `request canceled` in the logs, it might be because the OpenAI API call is taking too long and exceeding the timeout limit.
76
76
@@ -176,7 +176,7 @@ Consider optimizing the endpoint with above considerations to improve the perfor
176
176
177
177
After you deploy the endpoint and want to test it in the **Test** tab in the deployment detail page, if the **Test** tab shows **Unable to fetch deployment schema**, you can try the following two methods to mitigate this issue:
178
178
179
-
:::image type="content" source="./media/prompt-flow//unable-to-fetch-deployment-schema.png" alt-text="Screenshot of the error unable to fetch deployment schema in Test tab in deployment detail page. " lightbox = "./media/prompt-flow/unable-to-fetch-deployment-schema.png":::
179
+
:::image type="content" source="../media/prompt-flow//unable-to-fetch-deployment-schema.png" alt-text="Screenshot of the error unable to fetch deployment schema in Test tab in deployment detail page. " lightbox = "../media/prompt-flow/unable-to-fetch-deployment-schema.png":::
180
180
181
181
- Make sure you have granted the correct permission to the endpoint identity. Learn more about [how to grant permission to the endpoint identity](./flow-deploy.md#grant-permissions-to-the-endpoint).
182
182
- It might be because you ran your flow in an old version runtime and then deployed the flow, the deployment used the environment of the runtime that was in old version as well. To update the runtime, follow [Update a runtime on the UI](./create-manage-compute-session.md#upgrade-compute-instance-runtime) and rerun the flow in the latest runtime and then deploy the flow again.
0 commit comments