arXiv 2603.15341v1Mar 16, 2026

Intelligent Co-Design: An Interactive LLM Framework for Interior Spatial Design via Multi-Modal Agents

Ren Jian Lim, Rushi Dai

Brief context

Publication timing, weekly edition context, and source links for this brief.

Published

Mar 16, 2026, 2:28 PM

Current score

73

Original paper

The executive brief below is grounded in the source paper and linked back to the arXiv abstract.

In architectural interior design, miscommunication frequently arises as clients lack design knowledge, while designers struggle to explain complex spatial relationships, leading to delayed timelines and financial losses. Recent advancements in generative layout tools narrow the gap by automating 3D visualizations. However, prevailing methodologies exhibit limitations: rule-based systems implement hard-coded spatial constraints that restrict participatory engagement, while data-driven models rely on extensive training datasets. Recent large language models (LLMs) bridge this gap by enabling intuitive reasoning about spatial relationships through natural language. This research presents an LLM-based, multimodal, multi-agent framework that dynamically converts natural language descriptions and imagery into 3D designs. Specialized agents (Reference, Spatial, Interactive, Grader), operating via prompt guidelines, collaboratively address core challenges: the agent system enables real-time user interaction for iterative spatial refinement, while Retrieval-Augmented Generation (RAG) reduces data dependency without requiring task-specific model training. This framework accurately interprets spatial intent and generates optimized 3D indoor design, improving productivity, and encouraging nondesigner participation. Evaluations across diverse floor plans and user questionnaires demonstrate effectiveness. An independent LLM evaluator consistently rated participatory layouts higher in user intent alignment, aesthetic coherence, functionality, and circulation. Questionnaire results indicated 77% satisfaction and a clear preference over traditional design software. These findings suggest the framework enhances user-centric communication and fosters more inclusive, effective, and resilient design processes. Project page: https://rsigktyper.github.io/AICodesign/

Score 73PDF-backedagentstraininginferencemodels

Executive brief

A short business-reader brief that explains why the paper matters now and what to watch or do next.

Why this is worth your attention

This paper matters because it pushes generative design from a one-shot image or layout trick toward a usable co-design workflow: non-designers can steer a room layout in plain English, and the system translates that into constraints, optimization, and 3D output without task-specific model training. If that holds up in production, it could lower the labor needed for early-stage space planning, client alignment, and design iteration for real estate, interiors, hospitality, workplace, and renovation teams. The interesting shift is not just better layouts, but cheaper communication between experts and non-experts; the caution is that the evidence is still modest, with a small user study and heavy reliance on LLM-based grading rather than hard operational metrics.

  • The real value is upstream of final rendering: briefs, client comments, and reference images can become machine-readable layout rules that a solver can optimize, which could compress the slowest part of design work—back-and-forth clarification and revision. That matters for teams selling design services, fit-outs, furnishings, or space planning where delays and rework are expensive.
  • You may not need a heavily trained, domain-specific generative model to get useful interior layouts. This paper argues a lighter stack—general LLMs, retrieval of design rules, and classical optimization—can handle diverse room types without additional training, which is strategically important if you want to support new catalogs, brands, or geographies without rebuilding models each time.
  • Ask whether their product can show its work: what constraints it generated, what the user approved, what optimization criteria it used, and why one layout beat another. This paper’s stronger idea is not just generation, but an auditable loop with plain-language translations, scoring logic, and logs—features that matter if these tools are going into client-facing or regulated workflows.
  • A meaningful next step would be evidence on cycle time, revision count, cost per design, and conversion or approval rates in live projects. The current results are directionally encouraging—77% satisfaction and better LLM-judged layouts in test rooms—but they do not yet prove the economics that operations, procurement, or finance teams would need.
  • The evaluation is not yet strong enough to settle the case: only 53 users were studied, 42% had no strong tool preference, and the independent LLM grader needed a hand-built rubric after a simpler prompt produced nonsensical outputs. That makes this more credible as a product direction than as proof that AI co-design is already ready to replace established design workflows.

Evidence ledger

stackhighp.4p.8

The system uses four specialized agents to turn images and natural-language input into spatial rules, user-facing explanations, and final layout evaluation.

strategichighp.5p.8

Participatory interaction is central: users can review and modify object selection, constraints, and score terms during generation.

trainingmediump.7p.9

The approach reduces dependence on task-specific training by using RAG plus optimization instead of a trained end-to-end layout model.

capabilitymediump.11p.12

Interactive layouts scored better than non-interactive layouts in example room evaluations.

caveathighp.16p.11

The evidence base remains limited by small sample size and fragile LLM-based evaluation.

Related briefs

More plain-English summaries from the archive with nearby topics or operator relevance.

cs.AI

XSkill: Continual Learning from Experience and Skills in Multimodal Agents

Guanyu Jiang et al.

cs.LG

The PokeAgent Challenge: Competitive and Long-Context Learning at Scale

Seth Karten et al.

cs.AI

Memento-Skills: Let Agents Design Agents

Huichi Zhou et al.

cs.AI

When OpenClaw Meets Hospital: Toward an Agentic Operating System for Dynamic Clinical Workflows

Wenxian Yang et al.

Thank you to arXiv for use of its open access interoperability. This product was not reviewed or approved by, nor does it necessarily express or reflect the policies or opinions of, arXiv.
LightDark