Autonomous Evolution
Autonomous evolution is a paradigm for LLM-based search where agents control the evolutionary process — what to explore, when to test, and what knowledge to preserve — rather than following externally fixed heuristics.
Contrast with fixed evolutionary search
In fixed evolutionary search (FunSearch, AlphaEvolve, EvoX), an external algorithm governs all key decisions: parent selection uses predetermined rules (MAP-Elites, island models), prompt construction follows a fixed template, and population updates apply predefined criteria. The LLM acts primarily as a mutation operator — it proposes candidates from a constructed context in a single forward pass but has no agency over the search trajectory.
Autonomous evolution delegates these decisions to the agent. The same four stages apply — retrieve, propose, evaluate, update — but their timing and realization are agent-controlled. The agent decides what evidence to inspect, whether to run local tests before external evaluation, how to react to failure, and what knowledge to externalize for future reuse.
Evidence for the paradigm
CORAL provides the first empirical validation (Qu et al., 2026). Single-agent autonomous evolution achieves 3-10x higher improvement rates than fixed search baselines while converging in 5-20 evaluations vs. 60-100. The performance gap is attributed directly to agent autonomy: agents analyze prior attempts and evaluation feedback to choose modifications, rather than following predetermined heuristics (source).
Multi-agent autonomous evolution extends the gains further. Four co-evolving agents discover solutions no single agent finds, even with 4x compute. This is enabled by cross-agent knowledge transfer — agents inspire each other through shared artifacts without explicit communication protocols.
Key enablers
- Shared persistent memory — agents need a durable knowledge store to accumulate insights across attempts
- Heartbeat mechanism — periodic interventions prevent agents from drifting into local minima
- Local verification — agents test candidates locally before consuming expensive evaluations
- Knowledge accumulation — capturing reusable insights (not just progress logs) is causally linked to performance; disabling it degrades results by up to 18.6%
Relationship to agent autonomy
Autonomous evolution applies a specific claim from the broader AI agents discourse: that agents perform better when given more control over execution. The controller pattern in infrastructure separates desired state from execution — similarly, autonomous evolution separates the objective (evaluator) from the search strategy (agent-controlled). The evaluator defines what to optimize; the agent decides how.