You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
optillm will now be running on `http://localhost:8000` with its built-in local inference server.
66
66
67
-
**Note for Non-Mac Users**: This example uses `Qwen/Qwen3-0.6B-MLX-bf16` which is optimized for Apple Silicon (M1/M2/M3 chips). If you're not using a Mac, you should:
67
+
**Note for Non-Mac Users**: This example uses `Qwen/Qwen3-1.7B-MLX-bf16` which is optimized for Apple Silicon (M1/M2/M3 chips). If you're not using a Mac, you should:
68
68
69
69
1.**For NVIDIA GPUs**: Use a CUDA-compatible model like:
70
70
-`Qwen/Qwen2.5-32B-Instruct` (best quality, high VRAM)
@@ -81,9 +81,9 @@ optillm will now be running on `http://localhost:8000` with its built-in local i
1.**optillm Enhanced Early Discovery**: The best solution was found by iteration 3, suggesting optillm's test-time compute (MoA) and documentation access (readurls) helped quickly identify effective parsing strategies.
241
+
242
+
2.**Smaller Models Can Excel**: The 1.7B Qwen model with optillm achieved significant improvements (+8.6%), proving that test-time compute can make smaller models highly effective.
212
243
213
-
1.**Documentation Access Matters**: The readurls plugin significantly improves the LLM's ability to generate correct parsing code by providing actual HTML structure
244
+
3.**Local Optimization Works**: Fast inference times (<100ms after initial) show that local models with optillm provide both efficiency and quality.
214
245
215
-
2.**Test-Time Compute Works**: MoA's multiple generation and critique approach produces more robust solutions than single-shot generation
246
+
4.**Pattern: Quick Discovery, Then Plateau**: Evolution found a strong local optimum quickly. This suggests the current test cases were well-solved by the table parsing innovation.
216
247
217
-
3.**Powerful Local Models**: Large models like Qwen-32B with 4-bit quantization provide excellent results while being memory efficient when enhanced with optillm techniques
248
+
5.**optillm Plugin Value**: The evolved program's sophisticated multi-strategy approach (especially table parsing) likely benefited from optillm's enhanced reasoning capabilities.
218
249
219
-
## Customization
250
+
## Available optillm Plugins and Techniques
220
251
221
-
You can experiment with different optillm features by modifying `config.yaml`:
252
+
optillm offers many plugins and optimization techniques. Here are the most useful for code evolution:
222
253
223
-
1.**Different Plugins**: Try the `executecode` plugin for runtime validation
224
-
2.**Other Techniques**: Experiment with `cot_reflection`, `rstar`, or `bon`
225
-
3.**Model Combinations**: Adjust weights or try different technique combinations
254
+
### Core Plugins
255
+
-**`readurls`**: Automatically fetches web content when URLs are detected in prompts
256
+
-**`executecode`**: Runs code and includes output in the response (great for validation)
257
+
258
+
### Optimization Techniques
259
+
-**`moa`** (Mixture of Agents): Generates multiple responses, critiques them, and synthesizes the best
260
+
-**`cot_reflection`**: Uses chain-of-thought reasoning with self-reflection
261
+
-**`rstar`**: Advanced reasoning technique for complex problems
262
+
-**`bon`** (Best of N): Generates N responses and selects the best one
263
+
-**`z3_solver`**: Uses Z3 theorem prover for logical reasoning
264
+
-**`rto`** (Round Trip Optimization): Optimizes responses through iterative refinement
0 commit comments