arXiv 2603.25226v1Mar 26, 2026

WebTestBench: Evaluating Computer-Use Agents towards End-to-End Automated Web Testing

Fanheng Kong et al.

Brief context

Publication timing, weekly edition context, and source links for this brief.

Published

Mar 26, 2026, 9:27 AM

Current score

87

Original paper

The executive brief below is grounded in the source paper and linked back to the arXiv abstract.

The emergence of Large Language Models (LLMs) has catalyzed a paradigm shift in programming, giving rise to "vibe coding", where users can build complete projects and even control computers using natural language instructions. This paradigm has driven automated webpage development, but it introduces a new requirement about how to automatically verify whether the web functionalities are reliably implemented. Existing works struggle to adapt, relying on static visual similarity or predefined checklists that constrain their utility in open-ended environments. Furthermore, they overlook a vital aspect of software quality, namely latent logical constraints. To address these gaps, we introduce WebTestBench, a benchmark for evaluating end-to-end automated web testing. WebTestBench encompasses comprehensive dimensions across diverse web application categories. We decompose the testing process into two cascaded sub-tasks, checklist generation and defect detection, and propose WebTester, a baseline framework for this task. Evaluating popular LLMs with WebTester reveals severe challenges, including insufficient test completeness, detection bottlenecks, and long-horizon interaction unreliability. These findings expose a substantial gap between current computer-use agent capabilities and industrial-grade deployment demands. We hope that WebTestBench provides valuable insights and guidance for advancing end-to-end automated web testing. Our dataset and code are available at https://github.com/friedrichor/WebTestBench.

Score 87Full-paper briefagentsmodelsinferencedata

Executive brief

A short business-reader brief that explains why the paper matters now and what to watch or do next.

Why this is worth your attention

If AI-generated web apps keep getting easier to produce, QA becomes the gating function—and this paper says current computer-use agents are nowhere near ready to take that job over end to end. On this benchmark, every tested model stayed below 30% F1, with the best at 26.4%, and the main failure is not just missing bugs but failing to generate complete test plans in the first place. For engineering leaders, product teams, and anyone buying “AI software testing” tools, the practical takeaway is that autonomous web testing still looks like a supervised co-pilot workflow, not a lights-out replacement for QA.

  • If your working assumption is that browser-using agents can now replace manual or scripted web QA, this paper argues that is premature. The biggest bottleneck is upstream test completeness—models usually failed to cover even 70% of the gold checklist—so many defects are missed before detection even starts.
  • A credible vendor should be able to tell you whether their gains come from better test-case generation, better browser execution, or both. This paper shows those are different problems: when models get the human-written checklist in the oracle setting, detection improves sharply, which means “agent can browse a site” is not the same as “agent can design a reliable test plan.”
  • Even weak results can be expensive: some runs require dozens of turns and millions of tokens per sample, and the paper ties longer interaction histories to state-tracking failures and redundant actions. That makes autonomous testing a cost-control and reliability issue for engineering operations, not just a model-quality issue.
  • The practical path from here is likely human- or rule-guided testing, not full autonomy. If products start combining generated checklists, browser automation, and human review in a tightly scoped workflow, that would fit this paper’s evidence better than claims of fully automated production-grade web testing today.
  • The warning sign is real, but the setup is still synthetic: the apps were generated with Lovable.dev and annotators sometimes modified instructions or apps to ensure enough defects. That makes this a strong stress test for autonomous QA agents, not a clean readout of how they will perform on every real production web stack.

Evidence ledger

The strongest claims in the brief, along with the confidence and citation depth behind them.

capabilityhighp.2p.6

Current computer-use agents are far from industrial-grade autonomous web testing on this benchmark.

stackhighp.6p.8

Checklist generation is a major bottleneck, not just browser interaction or defect classification.

inferencehighp.7p.7

Long-horizon browser testing can be operationally expensive and unstable.

caveatmediump.12p.12

Results may not transfer cleanly to production websites because the benchmark is synthesized and defect density is curated.

Related briefs

More plain-English summaries from the archive with nearby topics or operator relevance.

cs.LG

AgentOpt v0.1 Technical Report: Client-Side Optimization for LLM-Based Agent

Wenyue Hua et al.

cs.AI

Nurture-First Agent Development: Building Domain-Expert AI Agents Through Conversational Knowledge Crystallization

Linghao Zhang

cs.SE

PostTrainBench: Can LLM Agents Automate LLM Post-Training?

Ben Rank et al.

cs.SE

SWE-CI: Evaluating Agent Capabilities in Maintaining Codebases via Continuous Integration

Jialong Chen et al.

Thank you to arXiv for use of its open access interoperability. This product was not reviewed or approved by, nor does it necessarily express or reflect the policies or opinions of, arXiv.
LightDark