arXiv 2603.28488v1Mar 30, 2026

Courtroom-Style Multi-Agent Debate with Progressive RAG and Role-Switching for Controversial Claim Verification

Masnun Nuha Chowdhury et al.

Brief context

Publication timing, weekly edition context, and source links for this brief.

Published

Mar 30, 2026, 2:23 PM

Current score

85

Original paper

The executive brief below is grounded in the source paper and linked back to the arXiv abstract.

Large language models (LLMs) remain unreliable for high-stakes claim verification due to hallucinations and shallow reasoning. While retrieval-augmented generation (RAG) and multi-agent debate (MAD) address this, they are limited by one-pass retrieval and unstructured debate dynamics. We propose a courtroom-style multi-agent framework, PROClaim, that reformulates verification as a structured, adversarial deliberation. Our approach integrates specialized roles (e.g., Plaintiff, Defense, Judge) with Progressive RAG (P-RAG) to dynamically expand and refine the evidence pool during the debate. Furthermore, we employ evidence negotiation, self-reflection, and heterogeneous multi-judge aggregation to enforce calibration, robustness, and diversity. In zero-shot evaluations on the Check-COVID benchmark, PROClaim achieves 81.7% accuracy, outperforming standard multi-agent debate by 10.0 percentage points, with P-RAG driving the primary performance gains (+7.5 pp). We ultimately demonstrate that structural deliberation and model heterogeneity effectively mitigate systematic biases, providing a robust foundation for reliable claim verification. Our code and data are publicly available at https://github.com/mnc13/PROClaim.

Score 85Full-paper briefagentsinferencemodelsdata

Executive brief

A short business-reader brief that explains why the paper matters now and what to watch or do next.

Why this is worth your attention

This paper’s real claim is not that “more agents” magically fix fact-checking, but that structured process matters: dynamic retrieval during the argument, forced role reversal, and mixed-model judging can make verification systems meaningfully more reliable than a standard debate setup. If that holds outside this benchmark, trust-sensitive workflows in compliance, policy, medical, legal, and enterprise search could shift from single-answer chatbots toward auditable deliberation systems that actively look for missing evidence before deciding. The catch is readiness: the gains are credible on this COVID claim benchmark, but they come with very high inference cost and only light proof that the same design generalizes cleanly to broader domains.

  • The most important result here is that dynamic evidence expansion during the debate appears to matter more than the debate theater itself: removing P-RAG costs 7.5 points of accuracy, while role-switching and multi-judge voting add smaller gains. For teams building review, compliance, or due-diligence workflows, that shifts the design question from “which model answered best?” to “can the system notice evidence gaps and go fetch more before committing?”
  • If a vendor claims ‘reliable verification,’ ask whether confidence comes from genuine evidence refresh or just model consensus. This paper shows agreement can rise when retrieval is removed even as accuracy falls, meaning a system can become more confidently wrong if it keeps reusing the same evidence pool.
  • The architecture buys auditability and some robustness, but it is expensive: about 210.9K tokens per claim versus 18.9K for standard multi-agent debate. That makes it plausible first for low-volume, high-consequence review queues—not frontline customer support, high-throughput moderation, or other latency-sensitive uses.
  • The heterogeneous judge panel added 3.3 points of accuracy over a single judge, which is a useful signal for buyers: mixing models may be more valuable than squeezing one more upgrade from a single flagship model. Watch for verification and agent platforms to expose model diversity, adjudication logic, and evidence provenance as configurable controls rather than hidden implementation details.
  • The current evidence is strongest for a narrow setup: COVID claims, a custom PubMed retrieval corpus, and mostly proof-of-concept external checks on small samples. A meaningful next signal would be replication in regulated enterprise domains with fresher corpora, stronger throughput reporting, and clear semantics for ‘inconclusive’ decisions, because this framework explicitly maps some inconclusive outcomes toward support under its burden-of-refutation logic.

Evidence ledger

The strongest claims in the brief, along with the confidence and citation depth behind them.

capabilityhighp.1

PROClaim reaches 81.7% zero-shot accuracy on Check-COVID, beating standard multi-agent debate by 10.0 points.

stackhighp.8p.8

Progressive retrieval is the main source of improvement; removing P-RAG reduces accuracy by 7.5 points.

caveathighp.2

Higher judge agreement can coincide with lower correctness when retrieval is weaker, so consensus is not a safe proxy for truth.

strategichighp.8

Heterogeneous judging improves outcomes over a single judge by 3.3 points.

inferencehighp.8

Self-reflection mainly improves efficiency, cutting rounds by 29% and token usage by 17% with little accuracy loss.

inferencehighp.32

The full system is expensive at roughly 210.9K tokens per claim, around 11× standard MAD.

Related briefs

More plain-English summaries from the archive with nearby topics or operator relevance.

cs.CL

From Anchors to Supervision: Memory-Graph Guided Corpus-Free Unlearning for Large Language Models

Wenxuan Li et al.

cs.SE

WebTestBench: Evaluating Computer-Use Agents towards End-to-End Automated Web Testing

Fanheng Kong et al.

cs.CL

Marco DeepResearch: Unlocking Efficient Deep Research Agents via Verification-Centric Design

Bin Zhu et al.

cs.AI

SkillClaw: Let Skills Evolve Collectively with Agentic Evolver

Ziyu Ma et al.

Thank you to arXiv for use of its open access interoperability. This product was not reviewed or approved by, nor does it necessarily express or reflect the policies or opinions of, arXiv.
LightDark