arXiv 2603.28986v1Mar 30, 2026

Mimosa Framework: Toward Evolving Multi-Agent Systems for Scientific Research

Martin Legrand et al.

Brief context

Publication timing, weekly edition context, and source links for this brief.

Published

Mar 30, 2026, 8:35 PM

Current score

87

Original paper

The executive brief below is grounded in the source paper and linked back to the arXiv abstract.

Current Autonomous Scientific Research (ASR) systems, despite leveraging large language models (LLMs) and agentic architectures, remain constrained by fixed workflows and toolsets that prevent adaptation to evolving tasks and environments. We introduce Mimosa, an evolving multi-agent framework that automatically synthesizes task-specific multi-agent workflows and iteratively refines them through experimental feedback. Mimosa leverages the Model Context Protocol (MCP) for dynamic tool discovery, generates workflow topologies via a meta-orchestrator, executes subtasks through code-generating agents that invoke available tools and scientific software libraries, and scores executions with an LLM-based judge whose feedback drives workflow refinement. On ScienceAgentBench, Mimosa achieves a success rate of 43.1% with DeepSeek-V3.2, surpassing both single-agent baselines and static multi-agent configurations. Our results further reveal that models respond heterogeneously to multi-agent decomposition and iterative learning, indicating that the benefits of workflow evolution depend on the capabilities of the underlying execution model. Beyond these benchmarks, Mimosa modular architecture and tool-agnostic design make it readily extensible, and its fully logged execution traces and archived workflows support auditability by preserving every analytical step for inspection and potential replication. Combined with domain-expert guidance, the framework has the potential to automate a broad range of computationally accessible scientific tasks across disciplines. Released as a fully open-source platform, Mimosa aims to provide an open foundation for community-driven ASR.

Score 87Full-paper briefagentsinframodelsinference

Executive brief

A short business-reader brief that explains why the paper matters now and what to watch or do next.

Why this is worth your attention

This paper pushes multi-agent AI a step closer from demoware to a usable automation pattern for scientific and other tool-heavy knowledge work: instead of hard-coding one workflow, the system builds and revises its own workflow as tasks change. The practical shift is not just better benchmark performance, but a more credible path to automating messy, multi-step analysis with audit trails, dynamic tool access, and model choice at each stage—features ops, R&D, platform, and compliance teams will all care about. The evidence is promising rather than decisive: the best result reaches 43.1% success on ScienceAgentBench, but gains are highly model-dependent, the judge that steers improvement is only loosely validated, and the current search loop gets expensive fast.

  • If this result holds up, the important product shift is that agent systems may no longer need one fixed prompt chain per use case; they can search for a better division of labor on the fly. That would matter most in research, analytics, and operations workflows where tasks vary enough that brittle automations keep breaking.
  • This paper implies some performance gains come from bounded context, better task decomposition, and tool routing—not only from model quality. A useful vendor question is whether their agent product can explain when orchestration improves outcomes, for which models, and at what token and runtime cost.
  • One of the more production-relevant ideas is full workflow logging and archived execution traces. If agents are going to touch regulated or high-stakes analysis, the ability to inspect every step may become as important as raw task success.
  • The long-term business case gets stronger if prior successful workflows can be reused across similar tasks, cutting setup time and cost. But that is not demonstrated here: in the reported benchmark, no task reused an archived workflow, so the headline results come from fresh workflow synthesis every time.
  • The system improves itself using an LLM judge, but the paper admits that judge is only a directional signal and may reward plausible but wrong answers. Until judge scores are shown to track real-world correctness more tightly, this is better viewed as an experimental orchestration layer than a drop-in autonomous worker.

Evidence ledger

The strongest claims in the brief, along with the confidence and citation depth behind them.

capabilityhighp.1

Mimosa reaches 43.1% success on ScienceAgentBench with DeepSeek-V3.2, beating single-agent and static multi-agent baselines in that setup.

stackhighp.1p.16

The framework dynamically discovers tools, synthesizes workflows, executes code-generating agents, and uses an LLM judge to refine workflows.

strategichighp.1p.21

Model choice materially affects whether multi-agent evolution helps or hurts.

inferencehighp.24p.19

The current iterative loop raises cost substantially for some models and tasks.

caveathighp.9p.14

The evaluation signal used to drive refinement is only partially validated and may reward plausible but incorrect outputs.

Related briefs

More plain-English summaries from the archive with nearby topics or operator relevance.

cs.LG

AutoSurrogate: An LLM-Driven Multi-Agent Framework for Autonomous Construction of Deep Learning Surrogate Models in Subsurface Flow

Jiale Liu, Nanzhe Wang

cs.LG

Bridging MARL to SARL: An Order-Independent Multi-Agent Transformer via Latent Consensus

Zijian Zhao, Jing Gao, Sen Li

cs.CL

From Anchors to Supervision: Memory-Graph Guided Corpus-Free Unlearning for Large Language Models

Wenxuan Li et al.

cs.CR

The Salami Slicing Threat: Exploiting Cumulative Risks in LLM Systems

Yihao Zhang et al.

Thank you to arXiv for use of its open access interoperability. This product was not reviewed or approved by, nor does it necessarily express or reflect the policies or opinions of, arXiv.
LightDark