Skip to content

wan-huiyan/agent-review-panel

Repository files navigation

GitHub release Claude Code license last commit Tests Research Papers

Agent Review Panel

Multiple AI reviewers independently evaluate your code, plans, or docs — then debate each other's findings. A judge renders the final verdict. You get a structured report with consensus, disagreements, and prioritized action items.

A Claude Code plugin that orchestrates multi-agent adversarial review panels backed by 9 research papers on multi-agent debate.

Packaged as a Claude Code plugin (containing the agent-review-panel skill). Install once via marketplace; it activates automatically on slash command or natural-language request. Requires Claude Code — the CLI, an IDE extension, or the Code tab inside the Claude Desktop app. Does not work with the regular claude.ai web chat or Claude API direct (details below).

Agent Review Panel — pipeline architecture

Agent Review Panel — terminal demo

Agent Review Panel — interactive HTML dashboard with expandable issue cards

The v2.15 HTML report: expandable 10-section issue cards with narrative, code evidence (Prism.js highlighted), debate transcripts, judge rulings, fix recommendations, and cross-references — all deep-linkable, keyboard-navigable, and print-friendly.

Quick Start

Install (recommended — Claude Code marketplace). Run these in your terminal (bash/zsh):

claude plugin marketplace add wan-huiyan/agent-review-panel
claude plugin install roundtable@agent-review-panel
claude plugin install plan-review-integrator@agent-review-panel    # optional companion: integrate review findings into plans
Already inside a Claude Code session? Use the slash-command form instead

Type these at the REPL prompt (note the leading / and no claude prefix):

/plugin marketplace add wan-huiyan/agent-review-panel
/plugin install roundtable@agent-review-panel
/plugin install plan-review-integrator@agent-review-panel

Both forms do the same thing. Pick whichever matches where you are: shell-form claude plugin … for terminal, REPL-form /plugin … for inside Claude Code.

The marketplace bundles two plugins: roundtable (this project) and plan-review-integrator (the review→integrate companion). See Bundled plugins below.

Use:

> Review this implementation plan from multiple perspectives: docs/my_plan.md

> /roundtable:agent-review-panel

⚠️ Plugin slash commands are namespaced as /<plugin>:<skill>. The plugin is named roundtable (the install name); the skill inside it is agent-review-panel. So the slash form is /roundtable:agent-review-panel. Natural-language invocation (the first example above — "Review this implementation plan from multiple perspectives: ...") works without the slash command because it triggers by the skill's description, not by exact slash match. Use whichever feels more natural.

What you get: Three output files:

  • review_panel_report.md — executive summary, consensus, disagreements (with judge rulings), prioritized action items tagged with epistemic labels
  • review_panel_process.md — full "director's cut" log of every agent's verbatim output with persona profiles
  • review_panel_report.html — interactive dashboard with expandable 10-section issue cards (Narrative, Code Evidence, Debate, Judge Ruling, Fix Recommendation, and more — new in v2.15), filterable issue cards, charts, and a Panel Gallery
Example report output (truncated)
# Review Panel Report
**Work reviewed:** src/auth/middleware.ts  |  **Date:** 2026-03-28
**Panel:** 4 reviewers + Auditor + Judge
**Verdict:** Approve with Revisions  |  **Confidence:** High
**Review mode:** Precise (auto-detected from content type: code)

## Executive Summary
The authentication middleware is well-structured with proper token validation
and rate limiting. Two substantive issues emerged: the session store lacks
TTL enforcement (P1, [VERIFIED]) and the CORS configuration is overly
permissive for production (P1, [CONSENSUS]). Score: 7/10.

## Consensus Points
- Token rotation logic is correct and handles edge cases well
- Error responses follow RFC 7807 format consistently

## Disagreement Points
**Session store TTL:** Security Auditor (P0) vs Architecture Critic (P2)
Judge ruling: P1 — the risk is real but mitigated by the upstream API gateway
timeout. [VERIFIED] against actual code.

## Action Items
- [P1] [VERIFIED] Add TTL to session store entries (src/auth/session.ts:47)
- [P1] [CONSENSUS] Restrict CORS origins in production config
- [P2] [SINGLE-SOURCE] Consider adding request signing for internal APIs

Installation

Requires Claude Code

This plugin only works on Claude Code surfaces — or on the Claude Agent SDK — because it needs the Agent tool for subagent spawning, local-filesystem access for output files, and a plugin-loader. Supported surfaces:

Works ✅

  • CLIclaude command in your terminal
  • VS Code extension — Claude Code extension from the VS Code marketplace
  • JetBrains IDE extension — IntelliJ, PyCharm, WebStorm, GoLand, Rider, etc.
  • Claude Desktop app → Code tab — the Desktop app bundles a Claude Code surface in its dedicated "Code" tab; the plugin installs and runs there the same way it does in the CLI. (official docs)
  • Claude Agent SDK — programmatic agent-building library on top of the Anthropic API. Load this plugin with options.plugins: [{ type: "local", path: "./agent-review-panel" }] in TypeScript or Python; subagents, skills, slash commands, and filesystem tools all work. See Plugins in the SDK.

Does not work ❌

  • Claude Desktop app → regular chat tabs — the chat surface has no /plugin marketplace; and although Agent Skills can run there, this plugin's 4–6 parallel reviewers plus three local-file outputs need Claude Code's subagent + filesystem infrastructure. Use the Code tab instead.
  • claude.ai web chat — same reason: no /plugin marketplace, and Skills on the web surface can't spawn the parallel subagents or write the review_panel_*.md/.html files the plugin produces.
  • Anthropic Messages API called directly (without the Agent SDK) — the raw API is a prompt→response interface; it has no plugin loader, no subagent orchestration, and no filesystem. If you want to run the plugin against the API, use the Claude Agent SDK entry above.

Why: the panel spawns 4–6 reviewer subagents in parallel via Claude Code's Agent tool, reads/writes files on your local filesystem to generate the three output reports, and responds to the /agent-review-panel:agent-review-panel slash command (or any natural-language "review panel" request). Only Claude Code surfaces expose those capabilities.

Don't have Claude Code yet? Install it from claude.ai/code, then come back and run the Quick Start commands above.

Claude Code marketplace (recommended)

The two install commands are shown in Quick Start above. This section explains what they do, covers the shell/CLI equivalent, and links to Updating + Manual clone alternatives.

Shell / CLI equivalent (instead of typing slash commands in the Claude Code REPL):

claude plugin marketplace add wan-huiyan/agent-review-panel
claude plugin install roundtable@agent-review-panel

Claude Code downloads the plugin to its cache, loads the agent-review-panel skill inside it, and activates the trigger phrases automatically. The plugin then activates when you ask for multi-perspective reviews, panel reviews, adversarial reviews, or invoke /agent-review-panel:agent-review-panel.

Command format: @<marketplace-name>, not @<repo-name>. The marketplace name is plugin (defined in .claude-plugin/marketplace.json), which is distinct from the plugin name agent-review-panel. Pre-v2.16.1 releases used @wan-huiyan-agent-review-panel; pre-v2.16 used @agent-review-panel. If you installed under an older marketplace name, see the Migration section to switch.

Why the marketplace path? The repo ships with .claude-plugin/marketplace.json + plugins/agent-review-panel/.claude-plugin/plugin.json manifests (v2.16+ canonical layout) that Claude Code reads to register the plugin. The marketplace install handles caching, version tracking, and automatic activation in one step. The manual clone path below still works but doesn't use the manifests — the marketplace flow is the canonical path for v2.14+.

Updating to the latest version

New releases land on main; Claude Code does not auto-pull. Run the update flow after each release (or any time you want the newest features) in your terminal:

claude plugin marketplace update agent-review-panel
claude plugin update roundtable@agent-review-panel
Or, if you're already in a Claude Code session, use the slash-command form
/plugin marketplace update agent-review-panel
/plugin update roundtable@agent-review-panel

Verify the update worked:

cat ~/.claude/plugins/cache/agent-review-panel/roundtable/*/.claude-plugin/plugin.json | grep version

The version should match the latest entry in the Version History table below. (The cache layout is cache/<marketplace-name>/<plugin-name>/<version>/ — note that the plugins/ intermediate directory from the repo is flattened out during install, and a version segment is added. The * glob above matches whatever version is installed so you don't have to look it up first.)

If the update appears to work but you're still getting old behavior (e.g. missing the v2.12 HTML report, missing the v2.15 expandable cards, or missing the v2.14 data-flow trace phase), check for a stale local clone that shadows the marketplace install:

ls ~/.claude/skills/agent-review-panel 2>/dev/null

If that directory exists, it's loaded before the marketplace cache and will pin you to whatever version was cloned. Remove it:

rm -rf ~/.claude/skills/agent-review-panel

Then restart Claude Code. The marketplace install in ~/.claude/plugins/cache/agent-review-panel/ will take over.

Fallback — clean reinstall: If the update commands misbehave, uninstall and reinstall from scratch. From your terminal:

claude plugin uninstall roundtable@agent-review-panel
claude plugin marketplace remove agent-review-panel
claude plugin marketplace add wan-huiyan/agent-review-panel
claude plugin install roundtable@agent-review-panel
REPL-form equivalent (inside a Claude Code session)
/plugin uninstall roundtable@agent-review-panel
/plugin marketplace remove agent-review-panel
/plugin marketplace add wan-huiyan/agent-review-panel
/plugin install roundtable@agent-review-panel

Manual clone (development / custom setup)

For local development, forking, or air-gapped environments:

git clone https://github.com/wan-huiyan/agent-review-panel.git ~/.claude/skills/agent-review-panel

Or load a cloned repo as a local plugin for testing without committing to marketplace install:

claude --plugin-dir ./agent-review-panel

Claude Code version requirement

Claude Code v1.0+ (the skill uses the Agent tool for parallel subagent spawning and model: "opus" overrides — v2.14+ forces opus on all launches including VoltAgent specialist agents).

Cursor (experimental)

Cursor installation options

This skill was built for Claude Code's Agent tool (parallel subagent spawning, model selection). Cursor has its own mechanisms that may require adaptation.

Per-project rule (most reliable):

mkdir -p .cursor/rules
# Create .cursor/rules/agent-review-panel.mdc with the content of SKILL.md
# Add frontmatter: alwaysApply: true

Manual global install:

git clone https://github.com/wan-huiyan/agent-review-panel.git ~/.cursor/skills/agent-review-panel

The core pattern is straightforward — one subagent/task per reviewer in Phase 3, collect results, then one per reviewer in Phase 5 (debate), then single agents for verification and judge. If you adapt it, PRs are welcome.

Why Use a Panel Instead of a Single Reviewer?

When you ask Claude to "review this code," you get one perspective. It won't argue with itself, catch its own blind spots, or tell you "I'm not sure about this."

The panel spawns independent reviewers that genuinely engage:

Feasibility Analyst: "The data_available_through hardcoding is minor — it's documented."

Risk Assessor: "Disagree. If stale, the lookforward extends past actual data — model trains on incomplete outcomes — silent false-negative bias."

Feasibility Analyst (Round 2): "Valid point. I upgrade this to IMPORTANT."

A single reviewer gives you a list. The panel gives you a deliberation — with structured disagreements, judge rulings, and confidence levels.

How It Works

16 phases + optional multi-run merge. Phase numbers are sequential integers (v2.14 cleanup — old decimal numbering like Phase 4.55 retired).

Stage Phase Action
Gather 1. Setup — scan sibling dirs, trace references, discover safeguards, detect signals, select personas
2. Data Flow Trace (v2.14, code only) — trace critical path(s), document schemas at each function boundary, flag composition/seam bugs
Review 3. Independent Review — 4-6 reviewers evaluate in parallel (no cross-talk)
4. Private Reflection — each reviewer re-reads and rates own confidence
Debate 5. Adversarial Debate (1-3 rounds) — reviewers engage + find new issues
6. Round Summarization — distill resolved/unresolved points between rounds
7. Blind Final — each reviewer gives final score independently
Verify 8. Completeness Audit — dedicated agent scans for what the panel missed
9. Verify Commands — run reviewer grep/read commands for P0/P1 findings (advisory)
10. Claim Verification — verify all line-number citations against source
11. Severity Verification — read actual code for every P0/P1; downgrade if overstated; web-verify external domain claims (v2.16.3)
12. Tier Assignment — confidence-based draft → judge-advised refinement per dispute
13. Targeted Verification — persona-matched agents investigate each dispute point
Adjudicate 14. Supreme Judge — Opus arbitrates everything including verification round evidence
Output 15. Triple output: Primary Report (.md) + Process History (.md) + Expandable-card Dashboard (.html) (v2.15)
Merge 16. Multi-Run Merge (v2.14, optional) — deduplicate findings across runs, score stability, resolve judge divergence

Features

Review process:

  • 4-6 reviewers with distinct personas evaluate in parallel, then debate across 1-3 rounds
  • Auto-selects personas based on content type (code, plan, docs, mixed) and technology signals across 10 signal groups (SQL, ML, Terraform, Auth, API, Frontend, Cost, Pipeline, Portability, Repo Hygiene)
  • Each reviewer uses a different reasoning strategy (systematic enumeration, adversarial simulation, backward reasoning, etc.)
  • Auto Precise/Exhaustive mode: code requires line citations; plans allow broader risk identification

Verification layer:

  • Claim verification checks all reviewer citations against actual source code
  • Severity verification reads the codebase to confirm P0/P1 findings before the judge sees them (v2.6 benchmark: 2/3 P0 findings were overstated). External domain claims (product limits, regulatory jurisdiction, API behavior) are automatically web-searched and tagged [WEB-VERIFIED], [WEB-CONTRADICTED], or [WEB-INCONCLUSIVE] (v2.16.3)
  • Verification commands: runs read-only grep/cat commands from reviewers to confirm or contradict claims
  • Defect classification: findings labeled [EXISTING_DEFECT] or [PLAN_RISK] — P0 requires existing defect evidence
  • Completeness audit: post-debate agent re-reads source line-by-line for what everyone missed
  • Targeted verification round (v2.11): each unresolved dispute gets a tiered (Light ~2k / Standard ~8k / Deep ~32k tokens) verification agent matched to the claim type (statistician for stats claims, security auditor for security claims, etc.) — verdicts feed directly into the judge's rulings
  • Data Flow Trace (v2.14): a dedicated agent traces data through critical paths before reviewers begin, flagging composition/seam bugs (two individually-correct functions producing incorrect results together). Three tiers: Standard (default, single path), Thorough (top 3 paths + transform-completeness checks), Exhaustive (all paths, no token limit — aims to catch all bugs). Uses Meta's semi-formal certificate prompting (2026, 78%→93% accuracy). Skipped for pure docs/plans or code with no data transforms.
  • Force opus on all launches (v2.14): every subagent_type launch — including VoltAgent specialist agents — must pass model: "opus". Fixes an invisible source of cross-run variance where VoltAgent agents silently fell through to their frontmatter-declared default model (potentially sonnet or haiku), producing different reasoning depths across otherwise identical runs.

Anti-groupthink safeguards:

  • Blind final scoring, private reflection, calibrated skepticism levels (20-60%)
  • Sycophancy detection intervenes when >50% of position changes lack new evidence
  • Anti-rhetoric assessment flags position changes driven by eloquence rather than evidence
  • Judge confidence gating: low-confidence verdicts flag "HUMAN REVIEW RECOMMENDED"
  • Correlated-bias warning when all reviewers converge (unanimous agreement is the most dangerous failure mode)

Output (three files per review):

  • Primary report (review_panel_report.md): executive summary, consensus, disagreements (with judge rulings), prioritized action items with epistemic labels ([VERIFIED], [CONSENSUS], [SINGLE-SOURCE], [UNVERIFIED], [DISPUTED], [WEB-VERIFIED], [WEB-CONTRADICTED], [WEB-INCONCLUSIVE])
  • Process history (review_panel_process.md): verbatim "director's cut" of every agent's output with persona profiles at each entry point — full transparency into the panel's reasoning
  • Interactive HTML dashboard (review_panel_report.html) with expandable 10-section issue cards (v2.15): each card expands to reveal a nested accordion with 📖 Narrative (full reviewer reasoning), 📄 Code Evidence (Prism.js-highlighted snippets with file:line headers), 👥 Raised by (per-reviewer rating + reasoning), 🔍 Verification Trail (full VR agent output), 💬 Debate (round-by-round transcript), ⚖️ Judge Ruling, 🛠️ Fix Recommendation (proposed change + before/after code + regression test + blast radius + effort), 🔗 Cross-references, 🏷️ Epistemic Tags (with hover tooltips), and 📊 Prior Runs. Plus: deep-link support (report.html#issue-A1), keyboard navigation, expand all/collapse all controls, print-friendly @media print CSS. Dashboard also includes a filterable/sortable issue list, Panel Gallery with avatar cards for every agent, and confidence/tier/verdict charts (Tailwind CSS + Chart.js + Prism.js via CDN).
  • Scope & limitations disclosure — every report states what the panel cannot evaluate

Advanced:

  • VoltAgent integration — maps personas to 127+ specialist agents for deeper domain-specific reviews when installed (all launches forced to opus in v2.14)
  • Multi-Run Union Protocol (v2.14) — invoke --runs N or "run 3 times and merge" to execute the panel N times with rotated persona compositions (Run 1: standard base; Run 2: complementary set; Run 3: adversarial-heavy 3 DAs + Correctness Hawk). Phase 16 merges findings by location + bug class, scores stability as [K/N RUNS], uses highest severity when runs disagree, resolves judge divergence. Eliminates the ~30% single-run blind spot documented in the v2.10→v2.14 consistency analysis.
  • Codebase state check — detects worktree/branch divergence to prevent false "missing code" findings
  • Tiered knowledge mining (L0/L1/L2) — scans index lines first, then summaries, then full content only for relevant items
  • Deep research mode — opt-in web research for domain best practices

Usage Examples

> Review this implementation plan from multiple perspectives: docs/my_plan.md

> /agent-review-panel:agent-review-panel

> Get a panel review of the authentication module — I want to stress-test the design

> Red team this deployment strategy

> Have agents debate whether this refactor is worth the complexity

> /agent-review-panel:agent-review-panel deep   # adds web research for domain best practices

> Do a deep review of this ML pipeline          # also triggers deep research mode

The skill auto-detects content type and selects appropriate personas and review mode. You can also specify custom reviewers.

Cost & Performance

Metric Value
Duration ~6-10 minutes (4-reviewer panel); ~10-12 minutes (6-reviewer with auto-persona)
Token usage Varies by content length and panel size. Typical range: 150k-350k total tokens across all subagent calls (input + output). Higher than single-agent review due to parallel reviewers + debate rounds.
Best for High-stakes reviews where you need structured disagreement tracking
Not for Quick code reviews, style checks, or single-opinion feedback

Known Limitations

  • Same base model: All reviewers are Claude instances. Unanimous agreement may reflect shared model biases rather than genuine quality. The correlated-bias warning flags this, but cannot eliminate it.
  • No runtime analysis: The panel reviews static code and documents. It cannot evaluate runtime behavior, production data patterns, or performance under load.
  • Token cost: Multi-agent review costs more than single-agent. Use for high-stakes reviews, not routine checks.
  • Temporal reasoning: Despite explicit checks, temporal scope verification (e.g., "excludes Christmas" with multi-year data) remains the hardest class of bug for panels to catch reliably.

Research Foundations

Agent Review Panel is grounded in 9 peer-reviewed papers on multi-agent debate and evaluation quality (ChatEval/ICLR 2024, Du et al./ICML 2024, MachineSoM/ACL 2024, and more). Additionally inspired by MiroFish (multi-agent prediction engine with heterogeneous agent personalities) — MiroFish's research patterns influenced the v2.1 auto-persona detection and the v2.11 persona-matched verification agent design. See ROADMAP.md for the full research roadmap.

Prerequisites

  • Claude Code v1.0+ (the skill uses the Agent tool for parallel subagent spawning)
  • Works with Claude Pro, Max, or API access
  • Optional: VoltAgent specialist agents for stronger domain-specific reviews

Tests

The project includes a comprehensive test suite (393 tests) using Node.js built-in test runner (zero dependencies):

npm test                    # run all 393 tests
npm run test:triggers       # trigger classification (55+ prompts)
npm run test:manifest       # manifest consistency + phase/opus enforcement
npm run test:eval-suite     # eval suite integrity + v2.14/v2.15 coverage
npm run test:report         # report structure validation
npm run test:behavioral     # behavioral assertion framework
npm run test:golden         # golden-file structural snapshots

Manifest tests enforce key invariants introduced in v2.14/v2.15:

  • All 16 phases present in SKILL.md (Phase 1 through Phase 16, no decimal numbering)
  • Every subagent_type: launch co-occurs with model: "opus" (force-opus enforcement)
  • Phase 15.3 spec documents all 10 expandable-card accordion sections
  • The canonical SKILL.md lives at plugins/agent-review-panel/SKILL.md (v2.16+ plugin layout, see PR #18)

Bundled plugins

This marketplace ships two plugins in one repository. They are independently installable; install only what you need.

Plugin Source What It Does When to Use
agent-review-panel plugins/agent-review-panel/ Multi-agent adversarial review panel — 4–6 reviewers debate, judge renders final verdict (this project, v2.16.1) When you need a structured review of code, plans, docs, or configs
plan-review-integrator plugins/plan-review-integrator/ Takes review panel output and integrates findings into an implementation plan — classifies each finding, applies concrete edits, produces a traceability summary (v2.0.0) After a panel review of a plan document, when you need findings reflected in the plan with traceability

Install both for the full review→integrate pipeline. From your terminal:

claude plugin install roundtable@agent-review-panel
claude plugin install plan-review-integrator@agent-review-panel

(Or, inside a Claude Code session, use the REPL form: /plugin install roundtable@agent-review-panel and /plugin install plan-review-integrator@agent-review-panel.)

plan-review-integrator was previously published as a standalone repo at wan-huiyan/plan-review-integrator. That repo is now archived in favor of the bundled distribution here. See Migration for upgrade instructions.

Migration from previous marketplaces

If you installed before v2.16.1, you used one of the old marketplace names. Migrate from your terminal:

# Old agent-review-panel install
claude plugin uninstall agent-review-panel@wan-huiyan-agent-review-panel
claude plugin marketplace remove wan-huiyan-agent-review-panel

# Old plan-review-integrator standalone install (if applicable)
claude plugin uninstall plan-review-integrator@wan-huiyan-plan-review-integrator
claude plugin marketplace remove wan-huiyan-plan-review-integrator

# New bundled install
claude plugin marketplace add wan-huiyan/agent-review-panel
claude plugin install roundtable@agent-review-panel
claude plugin install plan-review-integrator@agent-review-panel
REPL-form equivalent (inside a Claude Code session)
/plugin uninstall agent-review-panel@wan-huiyan-agent-review-panel
/plugin marketplace remove wan-huiyan-agent-review-panel
/plugin uninstall plan-review-integrator@wan-huiyan-plan-review-integrator
/plugin marketplace remove wan-huiyan-plan-review-integrator
/plugin marketplace add wan-huiyan/agent-review-panel
/plugin install roundtable@agent-review-panel
/plugin install plan-review-integrator@agent-review-panel

Verify both are loaded under the new marketplace:

ls ~/.claude/plugins/cache/agent-review-panel/
# expected: roundtable  plan-review-integrator

Contributing

Contributions welcome! Areas where help is especially useful:

  • Cursor adaptation — adapting the Agent tool calls to Cursor's subagent mechanism
  • New domain checklists — adding signal groups beyond the current 10
  • Benchmark cases — real-world review scenarios for the eval suite

Please open an issue to discuss before submitting large PRs.

Uninstalling

If installed via marketplace, from your terminal:

claude plugin uninstall roundtable@agent-review-panel
claude plugin marketplace remove agent-review-panel

(REPL-form equivalent: /plugin uninstall roundtable@agent-review-panel and /plugin marketplace remove agent-review-panel.)

If installed via manual clone:

rm -rf ~/.claude/skills/agent-review-panel

Version History

See CHANGELOG.md for detailed version history. See ROADMAP.md for research sources and deferred items.

Version Highlights
v2.15 Expandable 10-section issue cards in HTML dashboard (narrative, code evidence, debate, judge ruling, fix, cross-refs, prior runs); Prism.js syntax highlighting; deep-linking; keyboard nav; print-friendly
v2.14 Phase 2 Data Flow Trace (composition bug detector, 3 tiers); Multi-Run Union Protocol + Phase 16 Merge; force model: "opus" on all launches; integer phase renumbering (1–16)
v2.13 Persona profiles in process history + Panel Gallery in HTML dashboard
v2.12 Triple output: primary report + process history + interactive HTML dashboard
v2.11 Verification round: tiered (Light/Standard/Deep) persona-matched agents per dispute
v2.10 Codebase state check — prevents false "missing code" findings in worktrees
v2.9 VoltAgent specialist agent integration (127+ agents, 10 families)
v2.8 Auto Precise/Exhaustive mode, verification commands, tiered knowledge mining
v2.7 Severity verification, defect classification, temporal scope checks
v2.6 Schliff optimization (75 → 86), reference extraction, A/B validated
v2.5 Trust layer: claim verification, epistemic labels, scope disclosure
v2.4 Portability signal group
v2.3 Knowledge mining, domain checklists, deep research mode
v2.2 DMAD reasoning strategies, context gathering, anti-rhetoric guard
v2.1 Auto-persona from content signals, source-grounded debate
v2.0 Completeness auditor, new discovery requirement
v1.0 Initial release: multi-agent review with debate and judge

License

MIT — Huiyan Wan

Acknowledgements

  • Inspired by MiroFish — multi-agent prediction engine with heterogeneous agent personalities and memory; influenced auto-persona detection and persona-matched verification agents
  • Eval suite improved using schliff
  • See HOW_WE_BUILT_THIS.md for the design journey

Contributors

External contributions — thank you!

  • @okuuva#30 restructured the plugin to the canonical nested skills layout for Claude Code ≥2.1.112 manifest validation (resolves #28)

About

Claude Code skill: Multi-agent adversarial review panel — 4-6 AI reviewers debate your code/plans, then a supreme judge delivers the verdict. 9 auto-detected signal groups, built-in domain checklists, anti-groupthink mechanisms.

Topics

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Contributors