Skip to content
Zheng Nian's Blog
Go back

The Evolution of AI Debate: From Rigid Workflows to Autonomous Deep Agents

In the quest for more reliable AI reasoning, “Single Shot” prompting often falls short. Hallucinations creep in, and biases go unchecked. To solve this, we turned to Multi-Agent Systems (MAS).

Our goal was simple: reduce hallucinations and enhance realism by making agents check each other’s work. Our journey evolved through three distinct architectural phases over the last year.

Phase 1: The “Scripted Play” (May 2025)

Our first attempt was a Workflow-based system. We read numerous papers and implemented a standard structure: a Moderator, a Reviewer, and Red/Blue teams.

Phase 2: The “Relay Station” (August 2025)

In August, we rebuilt the system in our experimental labs project, basing it on a flow-based architecture (referencing ByteDance’s Deer Flow).

Phase 3: The “Autonomous Council” (Dec 2025 - Jan 2026)

With the release of our DeepAgents V3 engine, we reached our current architecture. The system now supports three distinct reasoning frameworks, all powered by the same underlying A2A (Agent-to-Agent) engine:

1. Red/Blue Team (Adversarial Analysis)

Used for critical decision making.

2. The Delphi Method (Expert Consensus)

Used for forecasting and complex estimation.

Delphi Method

3. SCAMPER (Creative Ideation)

Used for brainstorming and product innovation.

Future Roadmap: Six Thinking Hats

We are also exploring Edward de Bono’s “Six Thinking Hats” to further structure the emotional and logical dimensions of agent reasoning.

The Core Insight: The “A2A” Philosophy

The biggest technical takeaway from this evolution is what we call the A2A (Agent-to-Agent) Philosophy, even without a formal protocol.

In V3, we achieved a delicate balance:

  1. Context Isolation: Each agent runs in its own memory space.
  2. Collaboration via Tools: The “Debate” is essentially one Master Agent calling other Agents as Tools.

Shared Context

Call Tool

Spawn

Spawn

Spawn

Tool Output

Tool Output

Master Agent

DebateMiddleware

Agent A (e.g., Red Team)

Agent B (e.g., Blue Team)

Moderator Agent

Files & Search Results

By wrapping Agents as Tools, we achieve Context Isolation (clean reasoning) with Collaborative Output (shared goals). This “Implicit A2A” pattern has become the backbone of our system stability.


Share this post on:

Next Post
Solving the Context Window Problem: Resilient Context Management in Long-Running Agents