arXiv 2604.07429v1Apr 8, 2026

GameWorld: Towards Standardized and Verifiable Evaluation of Multimodal Game Agents

Mingyu Ouyang et al.

Brief context

Publication timing, weekly edition context, and source links for this brief.

Published

Apr 8, 2026, 5:49 PM

Current score

82

Original paper

The executive brief below is grounded in the source paper and linked back to the arXiv abstract.

Towards an embodied generalist for real-world interaction, Multimodal Large Language Model (MLLM) agents still suffer from challenging latency, sparse feedback, and irreversible mistakes. Video games offer an ideal testbed with rich visual observations and closed-loop interaction, demanding fine-grained perception, long-horizon planning, and precise control. However, systematically evaluating these capabilities is currently hindered by heterogeneous action interfaces and heuristic verification. To this end, we introduce GameWorld, a benchmark designed for standardized and verifiable evaluation of MLLMs as generalist game agents in browser environments. Two game agent interfaces are studied: (i) computer-use agents that directly emit keyboard and mouse controls, and (ii) generalist multimodal agents that act in a semantic action space via deterministic Semantic Action Parsing. GameWorld contains 34 diverse games and 170 tasks, each paired with state-verifiable metrics for outcome-based evaluation. The results across 18 model-interface pairs suggest that even the best performing agent is far from achieving human capabilities on video games. Extensive experiments of repeated full-benchmark reruns demonstrate the robustness of the benchmark, while further studies on real-time interaction, context-memory sensitivity, and action validity expose more challenges ahead for game agents. Together, by offering a standardized, verifiable, and reproducible evaluation framework, GameWorld lays a robust foundation for advancing research on multimodal game agents and beyond. The project page is at https://gameworld-bench.github.io.

Score 82Full-paper briefagentsinferenceinframodels

Executive brief

A short business-reader brief that explains why the paper matters now and what to watch or do next.

Why this is worth your attention

The useful shift here is not that game-playing AI suddenly works; it is that the field now has a more credible way to compare multimodal agents on closed-loop, visual, action-taking tasks without leaning on fuzzy “VLM-as-judge” scoring. That matters for anyone betting on computer-use agents, UI automation, or embodied AI, because it makes vendor claims easier to audit and exposes where current systems actually break: timing, navigation, memory, and converting partial progress into reliable completion. The paper’s own results are sobering — best agents are still well below a novice human — but that is precisely why this benchmark matters now: it pressures the market to compete on grounded execution and reproducible evaluation, not just polished demos.

  • The paper makes clear that today’s multimodal agents can often make partial progress but rarely finish reliably: overall success rates sit around 12.4%–21.2%, and the best agent still trails a novice human by a wide margin. For product, ops, and automation teams, that means flashy browser or game demos should not yet be read as proof of dependable end-to-end task execution.
  • A practical contribution here is state-verifiable scoring from instrumented game state rather than OCR or model-judged outputs. If you are evaluating computer-use or UI agents, a good vendor question is whether they can show deterministic task completion checks and reproducible reruns, because that is much harder to game than anecdotal success videos.
  • The paused benchmark isolates decision quality, but the real-time variant shows that once the environment keeps moving, slow reasoning becomes part of the failure mode. In other words, bigger or more deliberative models may look better in controlled scoring yet still miss real-world interactive workloads where 2.4–6.4 seconds per step is unusable.
  • One of the more important implications is that model performance is being shaped by the control layer around the model: memory rounds, tool-calling, action parsing, and low-level execution validity all materially affect results. That creates room for platforms to differentiate on orchestration and control reliability, not just on the underlying model, especially since memory can improve one interface while hurting another and evaluation cost already varies sharply by model/interface choice.
  • The benchmark is reproducible enough to be useful, but it is not free: the authors estimate $815.19 to run all 170 tasks across listed closed models, with large per-model variation and open-weight costs excluded. For teams building internal eval loops, that means benchmark design, caching, and model selection will directly affect how often you can afford to test and how quickly you can iterate.

Evidence ledger

The strongest claims in the brief, along with the confidence and citation depth behind them.

stackhighp.3p.25

GameWorld standardizes and verifies evaluation across browser games using instrumented state rather than judge models or visual heuristics.

capabilityhighp.14p.14

Current top agents are still materially below a novice human baseline on the benchmark.

inferencehighp.17p.17

Realtime interaction remains a distinct problem because inference latency and action timing are tightly coupled when the environment does not pause.

strategicmediump.6p.18

Model interface and harness design materially influence outcomes, creating strategic importance for orchestration layers beyond the base model.

caveathighp.48p.47

Benchmark use has a real operating cost, and pricing varies widely by model/interface choice.

Related briefs

More plain-English summaries from the archive with nearby topics or operator relevance.

cs.LG

ClawGUI: A Unified Framework for Training, Evaluating, and Deploying GUI Agents

Fei Tang et al.

cs.CV

SkinGPT-X: A Self-Evolving Collaborative Multi-Agent System for Transparent and Trustworthy Dermatological Diagnosis

Zhangtianyi Chen et al.

cs.LG

MOON3.0: Reasoning-aware Multimodal Representation Learning for E-commerce Product Understanding

Junxian Wu et al.

cs.CR

The Salami Slicing Threat: Exploiting Cumulative Risks in LLM Systems

Yihao Zhang et al.

Thank you to arXiv for use of its open access interoperability. This product was not reviewed or approved by, nor does it necessarily express or reflect the policies or opinions of, arXiv.
LightDark