A curated list of frameworks that combine Large Language Models (LLMs) with evolutionary algorithms or search-based optimization techniques (e.g. genetic algorithms, reinforcement learning, MCTS). Categorized by application domain.
-
FunSearch (DeepMind, 2024) โ Pairs an LLM with an evaluator to evolve code for open math and algorithm problems. Achieved new records in combinatorics and discovered high-performance heuristics.
-
AlphaEvolve (DeepMind, 2025) โ Gemini-based LLM agent that evolves entire programs, outperforming previous algorithm benchmarks, including a breakthrough in matrix multiplication.
-
OpenEvolve โ Open-source version of AlphaEvolve. Modular LLM+EA framework with support for multi-language generation, ensemble agents, and custom evaluation loops.
-
Evolution of Heuristics (EoH) โ Evolves natural-language 'thoughts' and corresponding heuristics with an LLM loop. Beats previous methods on combinatorial optimization problems.
-
ReEvo โ Reflective evolution loop where the LLM both mutates and critiques heuristic strategies, leading to general-purpose solvers.
-
LLaMEA โ Uses GPT-4 to iteratively evolve optimization algorithm code; beats classic optimizers on benchmark suites.
-
MPaGE โ Evolves diverse multi-objective heuristics guided by a Pareto grid. Achieves strong performance across several domains.
-
SOAR โ Self-improving LLM that fine-tunes itself from successful generations. Dominates ARC symbolic reasoning benchmark.
-
LAEA โ Zero-shot LLMs act as surrogate models to rank solutions, replacing learned predictors in classic EA.
- LLM4EC โ Community-sourced list of all papers at the intersection of LLMs and Evolutionary Computation.
-
LLM-MCTS โ GPT-4 guides Monte Carlo Tree Search for high-level planning in robotics and games.
-
MC-DML โ Dynamic memory + GPT integrated into MCTS for interactive fiction. Significantly boosts single-shot win rates.
-
LERO โ Evolves reward shaping functions and partial observation strategies in multi-agent RL with GPT assistance.
-
Voyager โ GPT-4 powered autonomous Minecraft agent that iteratively improves its own skills and codebase.
-
Tree-of-Thoughts โ Search-based reasoning framework that treats LLM generations as a search tree of "thoughts".
-
ASI-Arch โ Gemini-based autonomous AI researcher that discovered 100+ transformer architectures outperforming human designs.
-
LLMatic โ GPT-4 powered neural architecture mutation guided by quality-diversity search. Efficient CIFAR and NAS-Bench exploration.
-
DesignGPT โ Framework where GPT-4 recommends network architecture improvements iteratively. Early-stage AutoML system.
-
MOLLEO โ GPT-4 proposes chemical modifications as evolutionary mutations. Strong results on drug design and property optimization.
-
MOLLM โ LLM framework for multi-objective molecule generation using in-context experience replay.
-
MultiMol โ Two-agent LLM architecture: one learns from data, one reads literature to guide mutation of drug candidates.
-
LLM-Evolver for Polymers โ Claude-3.5 driven optimizer outperforms traditional Bayesian methods for designing polymer sequences.
-
ChemLatica โ Family of small chemistry LLMs trained on 100M property-labeled molecules. Combines with evolutionary prompts.
-
VALID-Mol โ Incorporates chemical validation filtering into LLM-driven evolutionary molecular design.
-
EvoPrompt โ Uses evolutionary search with LLM-informed mutation to discover superior prompts for >30 NLP tasks.
-
Promptbreeder โ Self-mutating prompt generation. LLM breeds new prompts using self-descriptive meta prompts.
-
Vision-Language EvoPrompt โ GA optimized prompts induced emergent tool-use behavior in multimodal LLMs.
-
Bayesian Prompt Optimization โ BO in prompt space, using embeddings + surrogate models for search.
-
LLM Policy Evolvers (various) โ Early-stage research into using LLMs to generate, mutate, or hybridize RL policies.
Have a framework to add? Submit a PR with the canonical link, a 1โ3 sentence summary, and a domain tag.
MIT.