LLaMEA (Large Language Model Evolutionary Algorithm) is an innovative framework that leverages the power of large language models (LLMs) such as GPT-4 for the automated generation and refinement of metaheuristic optimization algorithms. The framework utilizes a novel approach to evolve and optimize algorithms iteratively based on performance metrics and runtime evaluations without requiring extensive prior algorithmic knowledge. This makes LLaMEA an ideal tool for both research and practical applications in fields where optimization is crucial.
- Automated Algorithm Generation: Automatically generates and refines algorithms using GPT models.
- Performance Evaluation: Integrates with the IOHexperimenter for real-time performance feedback, guiding the evolutionary process to generate metaheuristic optimization algorithms.
- Customizable Evolution Strategies: Supports configuration of evolutionary strategies to explore algorithmic design spaces effectively.
- Extensible and Modular: Designed to be flexible, allowing users to integrate other models and evaluation tools.
- Python 3.8 or later
- OpenAI API key for accessing GPT models
It is the easiest to use LLaMEA from the pypi package.
pip install llamea
You can also install the package from source using Poetry.
- Clone the repository:
git clone https://github.com/nikivanstein/LLaMEA.git cd LLaMEA
- Install the required dependencies via Poetry:
poetry install
-
Set up an OpenAI API key:
- Obtain an API key from OpenAI.
- Set the API key in your environment variables:
export OPENAI_API_KEY='your_api_key_here'
-
Running an Experiment
To run an optimization experiment using LLaMEA:
from llamea import LLaMEA # Define your evaluation function def your_evaluation_function(solution): # Implementation of your function # return feedback, quality score, error information return "feedback for LLM", 0.1, "" # Initialize LLaMEA with your API key and other parameters optimizer = LLaMEA(f=your_evaluation_function, api_key="your_api_key_here") # Run the optimizer best_solution, best_fitness = optimizer.run() print(f"Best Solution: {best_solution}, Fitness: {best_fitness}")
Contributions to LLaMEA are welcome! Here are a few ways you can help:
- Report Bugs: Use GitHub Issues to report bugs.
- Feature Requests: Suggest new features or improvements.
- Pull Requests: Submit PRs for bug fixes or feature additions.
Please refer to CONTRIBUTING.md for more details on contributing guidelines.
Distributed under the MIT License. See LICENSE
for more information.
If you use LLaMEA in your research, please consider citing the associated paper:
@misc{vanstein2024llamea,
title={LLaMEA: A Large Language Model Evolutionary Algorithm for Automatically Generating Metaheuristics},
author={Niki van Stein and Thomas Bäck},
year={2024},
eprint={2405.20132},
archivePrefix={arXiv},
primaryClass={cs.NE}
}
For more details, please refer to the documentation and tutorials available in the repository.
flowchart LR
A[Initialization] -->|Starting prompt| B{Stop? fa:fa-hand}
B -->|No| C(Generate Algorithm - LLM )
B --> |Yes| G{{Return best so far fa:fa-code}}
C --> |fa:fa-code|D(Evaluate)
D -->|errors, scores| E[Store session history fa:fa-database]
E --> F(Construct Refinement Prompt)
F --> B
CodeCov test coverage