⭐ If you like this, please give the repo a star – it helps!
The fully-open successor to Google DeepMind’s AlphaEvolve for automated algorithm discovery. First released 📅 Nov 2024 • MIT License • 100 % reproducible. Also read the Documentation. 🥈 Winner of the Silver Humies 2025 at GECCO!
LLaMEA couples large-language-model reasoning with an evolutionary loop to invent, mutate and benchmark algorithms fully autonomously.
LLaMEA (Large Language Model Evolutionary Algorithm) is an innovative framework that leverages the power of large language models (LLMs) such as GPT-4 for the automated generation and refinement of metaheuristic optimization algorithms. The framework utilizes a novel approach to evolve and optimize algorithms iteratively based on performance metrics and runtime evaluations without requiring extensive prior algorithmic knowledge. This makes LLaMEA an ideal tool for both research and practical applications in fields where optimization is crucial.
Key Features:
- Automated Algorithm Generation: Automatically generates and refines algorithms using GPT-based or similar LLM models.
- Performance Evaluation: Integrates seamlessly with the IOHexperimenter for real-time performance feedback, guiding the evolutionary process.
- LLaMEA-HPO: Provides an in-the-loop hyper-parameter optimization mechanism (via SMAC) to offload numerical tuning, so that LLM queries focus on novel structural improvements.
- Extensible & Modular: You can easily integrate additional models and evaluation tools.
- Niching for Diversity: Fitness sharing and clearing strategies maintain a diverse set of solutions.
- Unified Diff Mode: Evolve code through patch-based edits for efficient token usage.
- Population Evaluation Mode: Optionally evaluate whole populations in a single call to speed up expensive fitness functions.
Example use-cases:
- Problem specific optimization algorithms: Easily generate and fine-tune optimization algorithms to work on your specific problem. By leveraging problem knowledge in the prompt the generated optimized algorithms can perform even better.
- Efficient new Bayesian Optimization algorithms: Generate optimized and novel Bayesian optimization algorithms, specifically for optimizing very expensive problems such as auto-motive crash worthiness or car shape design optimization tasks.
- Machine Learning Pipelines: Without any ML knowledge, you can use LLaMEA to generate optimized machine learning pipelines for any task. Just insert the task description and provide the dataset and evaluation metric and start LLaMEA.
- 2025.08 🥉 A LLaMEA generated agent won the 3rd price at the IEEE COG Planet Wars RTS competition
- 2025.07 🥇 "LLaMEA" won the Silver award at the Humies @GECCO2025
- 2025.07 🥈 "LLaMEA" won the Any-time Performancy on Many-Affine BBOB 2025 competition
- 2025.06 🎉 "LLaMEA-BO: A Large Language Model Evolutionary Algorithm for Automatically Generating Bayesian Optimization Algorithms" published on Arxiv
- 2025.05 🎉 "Optimizing Photonic Structures with Large Language Model Driven Algorithm Discovery" and "BLADE: Benchmark Suite for LLM-Driven Automated Design and Evolution of iterative optimisation heuristics" accepted as workshop papers at GECCO 2025
- 2025.04 🎉 LLaMEA-HPO paper accepted in ACM TELO “In-the-loop Hyper-Parameter Optimization for LLM-Based Automated Design of Heuristics"!
- 2025.04 🎉 "Code Evolution Graphs" accepted as full paper at GECCO 2025!
- 2025.03 🎉 LLaMEA v1.0.0 released!
- 2025.01 🎉 LLaMEA paper accepted in IEEE TEVC “Llamea: A large language model evolutionary algorithm for automatically generating metaheuristics"!
Important
Ensure that SWIG is installed on your system by running swig --version
If it is not installed, use the appropriate method for your platform:
- macOS:
brew install swig - Ubuntu/Linux:
sudo apt-get install swig(or your distribution’s package manager) - Windows: Download and install from the official SWIG website and make sure to add it to your system PATH.
It is the easiest to use LLaMEA from the pypi package.
pip install llameaImportant
The Python version must be higher or equal to 3.11. You need an OpenAI/Gemini/Ollama API key for using LLM models.
For a slimmed-down installation without bundled LLM or HPO dependencies, install the lite package:
pip install llamea-liteHPO features in the lite package require installing ConfigSpace separately.
You can also install the package from source using uv (0.7.19).
make sure you have uv installed.
- Clone the repository:
git clone https://github.com/XAI-liacs/LLaMEA.git cd LLaMEA - Install the required dependencies via uv:
uv sync
- Optional install dev or/and example dependencies:
uv sync --dev --group examples
-
Set up an OpenAI API key:
- Obtain an API key from OpenAI.
- Set the API key in your environment variables:
export OPENAI_API_KEY='your_api_key_here'
-
Running an Experiment
To run an optimization experiment using LLaMEA:
from llamea import LLaMEA # Define your evaluation function def your_evaluation_function(solution): # Implementation of your function # return feedback, quality score, error information return "feedback for LLM", 0.1, "" # Initialize LLaMEA with your API key and other parameters optimizer = LLaMEA(f=your_evaluation_function, api_key="your_api_key_here") # Run the optimizer best_solution, best_fitness = optimizer.run() print(f"Best Solution: {best_solution}, Fitness: {best_fitness}")
Key hyper-parameters of LLaMEA:
| Parameter | Description |
|---|---|
n_parents, n_offspring |
Population sizes controlling selection and mutation |
budget |
Number of generations to run |
niching |
Diversity management strategy (None, "sharing", "clearing") with niche_radius and related options |
evaluate_population |
If True, the evaluation function f operates on a list of solutions |
diff_mode |
Requests unified diff patches instead of full code for mutations |
HPO |
Enable in-the-loop hyper-parameter optimization |
eval_timeout, max_workers, parallel_backend |
Control evaluation time and parallelism |
adaptive_mutation, adaptive_prompt |
Adaptive control of mutation strength and task prompt |
See the Documentation for a complete description of all parameters.
Below are two example scripts from the examples directory demonstrating LLaMEA in action for black-box optimization with a BBOB (24 noiseless) function suite. One script (examples/black-box-optimization.py) runs basic LLaMEA, while the other (examples/black-box-opt-with-HPO.py) incorporates a hyper-parameter optimization pipeline—known as LLaMEA-HPO—that employs SMAC to tune the algorithm’s parameters in the loop.
black-box-optimization.py showcases a straightforward use-case of LLaMEA. It:
- Defines an evaluation function
evaluateBBOBthat runs generated algorithms on a standard set of BBOB problems (24 functions). - Initializes LLaMEA with a specific model (e.g., GPT-4, GPT-3.5) and prompts the LLM to generate metaheuristic code.
- Iterates over a
(1+1)-style evolutionary loop, refining the code until a certain budget is reached.
How to run:
uv run python examples/black-box-optimization.pyThe script will:
- Query the specified LLM with a prompt describing the black-box optimization task.
- Dynamically execute each generated algorithm on BBOB problems.
- Log performance data such as AOCC (Area Over the Convergence Curve).
- Iteratively refine the best-so-far algorithms.
black-box-opt-with-HPO.py extends LLaMEA with in-the-loop hyper-parameter optimization—termed LLaMEA-HPO. Instead of having the LLM guess or refine hyper-parameters directly, the code:
- Allows the LLM to generate a Python class representing the metaheuristic plus a ConfigSpace dictionary describing hyper-parameters.
- Passes these hyper-parameters to SMAC, which then searches for good parameter settings on a BBOB training set.
- Evaluates the best hyper-parameters found by SMAC on the full BBOB suite.
- Feeds back the final performance (and errors) to the LLM, prompting it to mutate the algorithm’s structure (rather than simply numeric settings).
Why LLaMEA-HPO? Offloading hyper-parameter search to SMAC significantly reduces LLM query overhead and encourages the LLM to focus on novel structural improvements.
How to run:
uv run python examples/black-box-opt-with-HPO.pyScript outline:
- Prompt & Generation: Script sets up a role/task prompt, along with hyper-parameter config space templates.
- HPO Step: For each newly generated algorithm, SMAC tries different parameter values within a budget.
- Evaluation: The final best configuration from SMAC is tested across BBOB instances.
- Refinement: The script returns the performance to LLaMEA, prompting the LLM to mutate the algorithm design.
Note
Adjust the model name (ai_model) or API key as needed in the script.
Changing budget or the HPO budget can drastically affect runtime and cost.
Additional arguments (e.g., logging directories) can be set if desired.
automl_example.py uses LLaMEA to showcase that it can not only evolve and generate metaheuristics but also all kind of other algorithms, such as Machine Learning pipelines.
In this example, a basic classification task on the breast-cancer dataset from sklearn is solved by generating and evolving open-ended ML pipelines.
- We define the evaluate function (accuracy score on a hold-out test set)
- We provide a very basic example code to get the algorithm started.
- We run a few iterations and observe the excellent performance of our completely automatic ML pipeline.
How to run:
uv run python examples/automl_example.pyNote
Adjust the model name (ai_model) or API key as needed in the script.
The repository provides a minimal Flask app in logreader/app.py to explore
conversation logs stored as JSON Lines files. Start the server with a log file
path:
uv run python logreader/app.py --logfile path/to/conversationlog.jsonlYou can also set the environment variable CONVERSATION_LOG instead of passing
--logfile. If neither is given, the app defaults to conversationlog.jsonl in
the current working directory. Navigate to http://localhost:5001 to browse the
messages.
Contributions to LLaMEA are welcome! Here are a few ways you can help:
- Report Bugs: Use GitHub Issues to report bugs.
- Feature Requests: Suggest new features or improvements.
- Pull Requests: Submit PRs for bug fixes or feature additions.
Please refer to CONTRIBUTING.md for more details on contributing guidelines.
Distributed under the MIT License. See LICENSE for more information.
Each paper we published also has an accompanying Zenodo repository for full reproducability of all our results.
- van Stein, N. (2025). BLADE - Code and Results for the paper [Data set]. Zenodo. https://doi.org/10.5281/zenodo.15119985
- van Stein, N., Vermetten, D., & Bäck, T. (2025). LLaMEA-HPO: code, generated algorithms and IOH logging data. https://doi.org/10.5281/zenodo.14917719
- van Stein, N., Kononova, A. V., Kotthoff, L., & Bäck, T. (2025). Figures and code for Code Evolution Graphs. Zenodo. https://doi.org/10.5281/zenodo.14770672
- van Stein, N., & Bäck, T. (2024). LLaMEA. Zenodo. https://doi.org/10.5281/zenodo.13842144
If you use LLaMEA in your research, please consider citing the associated paper:
@ARTICLE{van2025llamea,
author={Stein, Niki van and Bäck, Thomas},
journal={IEEE Transactions on Evolutionary Computation},
title={LLaMEA: A Large Language Model Evolutionary Algorithm for Automatically Generating Metaheuristics},
year={2025},
volume={29},
number={2},
pages={331-345},
keywords={Benchmark testing;Evolutionary computation;Metaheuristics;Codes;Large language models;Closed box;Heuristic algorithms;Mathematical models;Vectors;Systematics;Automated code generation;evolutionary computation (EC);large language models (LLMs);metaheuristics;optimization},
doi={10.1109/TEVC.2024.3497793}
}If you only want to cite the LLaMEA-HPO variant use the folllowing:
@article{van2024intheloop,
author = {van Stein, Niki and Vermetten, Diederick and B\"{a}ck, Thomas},
title = {In-the-loop Hyper-Parameter Optimization for LLM-Based Automated Design of Heuristics},
year = {2025},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3731567},
doi = {10.1145/3731567},
note = {Just Accepted},
journal = {ACM Trans. Evol. Learn. Optim.},
month = apr,
keywords = {Code Generation, Heuristic Optimization, Large Language Models, Evolutionary Computation, Black-Box Optimization, Traveling Salesperson Problems}
}Other works about extensions or integrations of LLaMEA:
@InProceedings{yin2024controlling,
author="Yin, Haoran and Kononova, Anna V. and B{\"a}ck, Thomas and van Stein, Niki",
editor="Garc{\'i}a-S{\'a}nchez, Pablo and Hart, Emma and Thomson, Sarah L.",
title="Controlling the Mutation in Large Language Models for the Efficient Evolution of Algorithms",
booktitle="Applications of Evolutionary Computation",
year="2025",
publisher="Springer Nature Switzerland",
address="Cham",
pages="403--417",
isbn="978-3-031-90065-5"
}For more details, please refer to the documentation and tutorials available in the repository.
flowchart LR
A[Initialization] -->|Starting prompt| B{Stop? fa:fa-hand}
B -->|No| C(Generate Algorithm - LLM )
B --> |Yes| G{{Return best so far fa:fa-code}}
C --> |fa:fa-code|D(Evaluate)
D -->|errors, scores| E[Store session history fa:fa-database]
E --> F(Construct Refinement Prompt)
F --> B
CodeCov test coverage
