i built
for those of you who are autoresearch pilled , or have been meaning to get into autoresearch but dont know how. Its an opensource Claude Code & Codex plugin that optimizes code through experiments
you hand it a codebase. it finds a benchmark, runs the baseline, then fires off parallel agents to try to beat it. kept if better, discarded if worse.
inspired by karpathy's autoresearch, but with structure on top:
- tree search over greedy hill-climb — multiple forks from any committed node
- N parallel agents in git worktrees
- shared failure traces so agents don't repeat each other's mistakes
- regression gates
under the hood: each experiment is a git worktree branching from its parent. commits on score improvement + gate pass. discards + worktree cleanup on regression. everything observable in a local dashboard
Apache 2.0, no signup, no API keys beyond what Claude Code already has:
/plugin marketplace add evo-hq/evo
/plugin install evo@evo-hq-evo
United States
NORTH AMERICA
Related News
How Braze’s CTO is rethinking engineering for the agentic area
10h ago
Amazon Employees Are 'Tokenmaxxing' Due To Pressure To Use AI Tools
21h ago

Implementing Multicloud Data Sharding with Hexagonal Storage Adapters
15h ago

DeepMind’s CEO Says AGI May Be ~4 Years Away. The Last Three Missing Pieces Are Not What Most People Think.
15h ago

CCSnapshot - A Claude Code Configs Transfer Tool
21h ago