The PyTorch for Agentic AI. A-Evolve is an open-source infrastructure that evolves any agent, across any domain, using any evolution algorithm β with zero human intervention.
Quick Start | News | Benchmark Highlights | Architecture & Design | Contribution
You provide a Base Agent. A-Evolve returns a SOTA Agent. 3 lines of code. 0 hours of manual harness engineering. One infra, any domain, any evolution algorithm.
import agent_evolve as ae
evolver = ae.Evolver(agent="./my_agent", benchmark="swe-verified")
results = evolver.run(cycles=10)By applying our open-source reference evolution algorithms to a base Claude Opus-4.6 model with zero manual harness engineering, A-Evolve pushed agents into top-tier performance across four diverse benchmarks:
We are evolving fast! Support our research by leaving a β. A-Evolve mutates real files in the workspace. Here's a before/after from our MCP-Atlas evolution:
5 targeted skills outperformed 10 generic ones. Every mutation is git-tagged ( # PyPI (recommended)
pip install a-evolve # core
pip install a-evolve[anthropic] # Claude support
pip install a-evolve[mcp] # MCP-Atlas benchmark
pip install a-evolve[swe] # SWE-bench benchmark
pip install a-evolve[all] # everything
# From source (for development)
git clone https://github.com/A-EVO-Lab/a-evolve.git && cd a-evolve
pip install -e ".[all,dev]"import agent_evolve as ae
evolver = ae.Evolver(
agent="swe-verified", # built-in seed workspace (or path to yours)
benchmark="swe-verified", # built-in benchmark adapter
)
results = evolver.run(cycles=10)
print(f"Final score: {results.final_score:.3f}")
print(f"Converged: {results.converged}")A-Evolve ships with built-in seed workspaces ( To make any agent evolvable, implement one method β from agent_evolve.protocol.base_agent import BaseAgent
from agent_evolve.types import Task, Trajectory
class MyAgent(BaseAgent):
def solve(self, task: Task) -> Trajectory:
return Trajectory(task_id=task.id, output="result")Then evolve it: evolver = ae.Evolver(agent=MyAgent("./my_workspace"), benchmark="mcp-atlas")
results = evolver.run(cycles=10)Your agent's evolvable state (prompts, skills, memory) lives as a standard directory β the Agent Workspace. A-Evolve mutates these files; your agent reloads. See Architecture & Design for the full picture. For benchmark-specific walkthroughs, see SWE-bench Demo Guide, MCP-Atlas Demo Guide, and SkillBench Setup Guide. A-Evolve's core insight: all evolvable agent state lives on the file system as a standard directory structure. This lets the evolution engine mutate any agent via LLM-driven file operations β without knowing the agent's internals. The evolution engine reads these files, analyzes performance logs, and writes mutations back. The agent reloads. That's the entire contract. Every cycle follows five phases:
The loop converges when EGL (Evolutionary Generality Loss) stabilizes or A-Evolve ships with ready-to-use benchmark adapters and seed workspaces:
A-Evolve is a framework, not a standalone agent. Every axis is pluggable:
A-Evolve ships with 4 reference evolution algorithms, each targeting different domains and strategies:
Each algorithm lives in its own directory under from agent_evolve.engine.base import EvolutionEngine
from agent_evolve.types import StepResult
class MyEvolutionEngine(EvolutionEngine):
def step(self, workspace, observations, history, trial) -> StepResult:
# Analyze observations, mutate workspace files, optionally run trial tasks
...
return StepResult(accepted=True, score=new_score)Then pass it to the Evolver: evolver = ae.Evolver(
agent="swe-verified",
benchmark="swe-verified",
engine=MyEvolutionEngine(config),
)The engine has full access to shared primitives β A-Evolve is built for the research community. We welcome contributions across every axis of the framework. If you work in LLM self-optimization, reinforcement learning, or agent architectures β implement the
Drop your algorithm into Implement
If you use A-Evolve in your research, please cite our position paper: @article{lin2026position,
title={Position: Agentic Evolution is the Path to Evolving LLMs},
author={Lin, Minhua and Lu, Hanqing and Shi, Zhan and He, Bing and Mao, Rui and Zhang, Zhiwei and Wu, Zongyu and Tang, Xianfeng and Liu, Hui and Dai, Zhenwei and others},
journal={arXiv preprint arXiv:2602.00359},
year={2026}
}Release History
Dependencies & License AuditLoading dependencies... Similar PackagesAutoGPTAutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.autogpt-platform-beta-v0.6.56 agenticSeekFully Local Manus AI. No APIs, No $200 monthly bills. Enjoy an autonomous agent that thinks, browses the web, and code for the sole cost of electricity. π Official updates only via twitter @Martin993main@2026-04-11 cyber-pilotCyber Pilot is a traceable delivery system for requirements, design, plans, and code.v3.7.0-beta ToolAgentsToolAgents is a lightweight and flexible framework for creating function-calling agents with various language models and APIs.0.3.0 |



