Thanks for your interest in contributing! This project provides companion tools for Karpathy's autoresearch.
git clone https://github.com/YOUR_USERNAME/autolab.git
cd autolab
# Install all tools in development mode
pip install -e ./auto-judge -e ./auto-steer -e ./auto-evolve
# Install dev dependencies
pip install pytest ruff mypy
# Run tests
pytest
# Run linting
ruff check .autolab/
├── auto-judge/ # autojudge CLI — experiment evaluation
├── auto-steer/ # autosteer CLI — research direction suggestions
├── auto-evolve/ # autoevolve CLI — multi-agent competitions
├── skills/ # Claude Code skill definitions
├── templates/ # User-facing integration templates
└── test-data/ # Sample data for testing
Each tool is an independent Python package with its own pyproject.toml.
- Fork the repository
- Create a feature branch:
git checkout -b my-feature - Make your changes
- Run tests:
pytest - Run linting:
ruff check . - Commit with a clear message
- Open a pull request
- No fake solutions — all code must work against real autoresearch data
- Result types — use
Ok/Errfor error handling, never bare exceptions in business logic - Test behaviors — tests should validate what the tool does, not how it does it
- Keep tools independent — each tool should work standalone with only
clickas a dependency - Backward compatible — don't break existing CLI flags or output formats
- Create a directory:
auto-<name>/ - Add
pyproject.tomlwithhatchlingbuild backend - Add the tool's README
- Add a Claude Code skill in
skills/autoresearch-<name>/ - Update the root README
Please include:
- Which tool (
autojudge,autosteer, orautoevolve) - Python version (
python --version) - A sample
results.tsvif relevant (anonymize if needed) - The full command you ran
- The full error output