You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+17-6Lines changed: 17 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -185,7 +185,7 @@ See the [Configuration Guide](configs/default_config.yaml) for a full list of op
185
185
186
186
## Artifacts Channel
187
187
188
-
OpenEvolve includes a **artifacts side-channel** that allows evaluators to capture build errors, profiling results, etc. to provide better feedback to the LLM in subsequent generations. This feature enhances the evolution process by giving the LLM context about what went wrong and how to fix it.
188
+
OpenEvolve includes an **artifacts side-channel** that allows evaluators to capture build errors, profiling results, etc. to provide better feedback to the LLM in subsequent generations. This feature enhances the evolution process by giving the LLM context about what went wrong and how to fix it.
189
189
190
190
The artifacts channel operates alongside the traditional fitness metrics.
191
191
@@ -205,17 +205,28 @@ return EvaluationResult(
205
205
```
206
206
207
207
The next generation prompt will include:
208
-
```
208
+
```markdown
209
209
## Last Execution Output
210
210
### Stderr
211
-
```
212
211
SyntaxError: invalid syntax (line 15)
213
-
```
212
+
214
213
### Traceback
215
-
```
216
214
...
217
215
```
216
+
217
+
## Example: LLM Feedback
218
+
219
+
An example for an LLM artifact side channel is part of the default evaluation prompt template, which ends with
220
+
```markdown
221
+
Return your evaluation as a JSON object with the following format:
222
+
{{
223
+
"readability": [score],
224
+
"maintainability": [score],
225
+
"efficiency": [score],
226
+
"reasoning": "[brief explanation of scores]"
227
+
}}
218
228
```
229
+
The non-float values, in this case the "reasoning" key of the json response that the evaluator LLM generates, will be available within the next generation prompt.
219
230
220
231
### Configuration
221
232
@@ -240,7 +251,7 @@ export ENABLE_ARTIFACTS=false
240
251
### Benefits
241
252
242
253
- **Faster convergence** - LLMs can see what went wrong and fix it directly
243
-
-**Better error handling** - Compilation and runtime failures become learning opportunities
254
+
- **Better error handling** - Compilation and runtime failures become learning opportunities
244
255
- **Rich debugging context** - Full stack traces and error messages guide improvements
245
256
- **Zero overhead** - When disabled, no performance impact on evaluation
0 commit comments