Zum Inhalt springen
← Back to English edition

Anthropic Code Review: How AI Agent Teams Analyze Your Pull Requests

deep-dives

Anthropic has launched a multi‑agent system for Claude Code that automatically checks AI‑generated code, detects logic errors and relieves developer teams. We explain how it works, what it costs and for whom it is worthwhile.

Anthropic Claude Code Review Multi-Agent KI‑Development GitHub Automation

🔍 Deep Dive – In this article we dive into Anthropic’s new multi‑agent system for code reviews. We look at how it works, what problems it is supposed to solve and whether it is worth the hype.

Since AI coding tools like Claude Code, GitHub Copilot or Cursor have multiplied the productivity of developer teams, there is a new problem: the flood of AI‑generated code. Every pull request suddenly contains hundreds of lines that a human can hardly review thoroughly in a reasonable time. At the same time subtle logic errors creep in that traditional linters and static analyzers don’t detect.

Anthropic now has an answer: Code Review in Claude Code – a multi‑agent system that automatically analyzes AI‑generated pull requests, prioritizes errors and provides concrete improvement suggestions. According to Anthropic, review quality thereby increases from an average of 16 % to over 54 % substantive comments.

In this Deep Dive we look at how the system works under the hood, what architecture is behind it and for whom the use is worthwhile.

The problem: AI‑generated code overwhelms human reviewers

“We have seen teams whose pull request volume has tripled due to AI tools – with constant reviewer capacity,” says Anthropic product lead Maya Chen in the TechCrunch interview. The consequence: superficial reviews, overlooked bugs, technical debt that will pay back later.

Traditional automated code reviews (e.g., through linters, SonarQube or self‑written scripts) mainly detect syntax problems and simple patterns. Logic errors, race conditions or semantic inconsistencies often remain undiscovered – precisely those error classes that occur especially frequently with AI‑generated code because the model doesn’t fully capture the context of the entire codebase.

The solution: A team of specialized AI agents

Anthropic’s Code Review does not rely on a single omniscient AI reviewer, but on multiple parallel working agents each with different task focus:

  1. Syntax agent – checks for obvious syntax errors, formatting and naming conventions.
  2. Logic agent – analyzes control flow, conditions and possible race conditions.
  3. Security agent – searches for known security vulnerabilities (e.g., injection, unsafe deserialization).
  4. Architecture agent – keeps an overview of the entire codebase and recognizes deviations from established patterns.
  5. Verification agent – filters false positives by checking potential problems against actual runtime behavior (simulated test run on Anthropic infrastructure).

Each agent works in parallel on the same code diff, exchanges information via a central orchestration layer and passes its findings to the verification agent. This sorts the results by severity and finally creates a consolidated review list that can be posted directly in GitHub (or other platforms).

Scaling by PR size

Interesting is the dynamic scaling of the system: For small pull requests (under 1,000 changed lines) only a reduced agent set runs. From 1,000 lines the system adds additional agents and analyzes also the context of the entire codebase, not just the diff. For very large PRs (10,000+ lines) up to twelve specialized agents can work simultaneously – a human team would need days for that.

What does a review look like in practice?

A typical workflow:

  1. Developer opens a pull request in GitHub (e.g., with changes generated by Claude Code).
  2. The CI pipeline triggers code review via GitHub App or CLI tool.
  3. Within 2–5 minutes (depending on PR size) the first comments appear directly in the PR – prioritized by severity (critical, high, medium, low).
  4. Each comment contains:
    • A short description of the problem
    • The affected code snippet
    • A suggestion for correction (often directly as a patch snippet)
    • References to similar problems in other parts of the codebase
  5. The developer can accept, discuss or ignore the suggestions – just like with human reviews.

According to heise.de, a beta team at a fintech company achieved 54 % substantive comments (before 16 %). At the same time, the time senior developers spend on routine reviews decreased by about 70 %.

Costs: When is the use worthwhile?

Code review is a paid add‑on to Claude Code Enterprise. Billing is per “review unit”, where a unit roughly corresponds to 1,000 characters diff size. Anthropic gives the following list prices (as of March 2026):

  • Small PRs (≤ 500 lines): ca. $0.50–$2 per review
  • Medium PRs (500–2,000 lines): $2–$10
  • Large PRs (2,000–10,000 lines): $10–$50

For companies that already use Claude Code Enterprise, the reviews come as a natural extension. For small teams or open‑source projects the price is likely a hurdle – here one remains temporarily with human reviewers or simpler automation tools.

💡 Tip: Anyone who thinks about it should first plan a pilot month with limited budget and measure how many human reviewer hours are actually saved. Often the tool pays off already with 10–15 PRs per week.

Limits and pitfalls

No system is perfect – not even code review. Important to know:

  • No substitute for human expertise: Architectural decisions, team conventions or domain‑specific knowledge cannot be evaluated by the system.
  • False positives: Despite the verification agent, about 5–10 % of comments remain false alarms (but can be reduced with feedback loops).
  • Slow introduction in legacy codebases: If the codebase has hardly any tests, the verification agent can work less reliably.
  • Cost spike in large refactorings: A PR with 20,000 lines can quickly cost over $100 – here one must weigh whether the automated review is worth the effort.

Conclusion: A step towards autonomous software development

Anthropic’s Code Review is more than just another tool – it is a concrete example of how multi‑agent systems work in practice. Instead of having a single AI assistant that should be able to do everything, Anthropic relies on division of labor, specialization and verification – just like a human team.

For companies that already produce a lot of AI‑generated code, the tool is likely a genuine relief. For small teams or hobby projects it is (still) too expensive. Yet the trend is clear: The future of code review will be hybrid – human reviewers concentrate on the big picture, while AI agent teams take over the routine work.

If you already use Claude Code, a look at the official documentation and the 30‑day trial period is worthwhile. And if you already have experience: feel free to write us on Mastodon or by email – we are curious about your feedback.


This article is part of our Deep‑Dive series on current AI topics. Next week we will look at how agent frameworks like LangChain, CrewAI and AutoGen differ in practice. Stay tuned!

Translated to English with AI assistance.