You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- `DocumentKnowledgeBase`: Enable document knowledge base functionality
160
159
- `KnowledgeBaseModelId`: Bedrock model for knowledge base queries
161
160
- `PostProcessingLambdaHookFunctionArn`: Optional Lambda ARN for custom post-processing (see [post-processing-lambda-hook.md](post-processing-lambda-hook.md) for detailed implementation guidance)
Description: Model to use for evaluation reports (e.g., "us.anthropic.claude-3-7-sonnet-20250219-v1:0")
95
91
```
96
92
93
+
### Runtime Configuration
94
+
95
+
Control evaluation behavior through the configuration file (no stack redeployment needed):
96
+
97
+
```yaml
98
+
evaluation:
99
+
enabled: true # Set to false to disable evaluation processing
100
+
llm_method:
101
+
model: "us.anthropic.claude-3-haiku-20240307-v1:0"# Model for evaluation reports
102
+
temperature: "0.0"
103
+
top_p: "0.1"
104
+
max_tokens: "4096"
105
+
# Additional model parameters...
106
+
```
107
+
108
+
**Benefits of Configuration-Based Control:**
109
+
- Enable/disable evaluation without stack redeployment
110
+
- Runtime control similar to summarization and assessment features
111
+
- Zero LLM costs when disabled (step executes but skips processing)
112
+
- Consistent feature control pattern across the solution
113
+
114
+
### Attribute-Specific Evaluation Methods
115
+
97
116
You can also configure evaluation methods for specific document classes and attributes through the solution's configuration. The framework supports three types of attributes with different evaluation approaches:
Copy file name to clipboardExpand all lines: docs/idp-cli.md
+2-12Lines changed: 2 additions & 12 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -709,20 +709,10 @@ idp-cli run-inference \
709
709
710
710
Download the evaluation results to analyze accuracy:
711
711
712
-
**⏱️ Important Timing Note:** Evaluation processing runs as a separate step after the main document processing completes. This takes an additional 2-3 minutes per document. If you download results immediately after the batch shows "Complete", the evaluation data may not be ready yet.
713
-
714
-
**Best practice:**
715
-
1. Wait 5-10 minutes after batch completion before downloading evaluation results
716
-
2. Check that the downloaded files include the `evaluation/` directory
717
-
3. If evaluation data is missing, wait a few more minutes and download again
712
+
**✓ Synchronous Evaluation:** Evaluation runs as the final step in the workflow before completion. When a document shows status "COMPLETE", all processing including evaluation is finished - results are immediately available for download.
**Note:** Evaluation is now controlled via configuration file (`evaluation.enabled: true/false`) rather than stack parameters. See the [evaluation.md](./evaluation.md) documentation for details.
1368
+
1375
1369
### Evaluation Methods Configuration
1376
1370
1377
1371
Configure evaluation methods for specific document classes and attributes:
0 commit comments