arXiv 2604.08123v1Apr 9, 2026

LegoDiffusion: Micro-Serving Text-to-Image Diffusion Workflows

Lingyun Yang et al.

Brief context

Publication timing, weekly edition context, and source links for this brief.

Published

Apr 9, 2026, 11:44 AM

Current score

87

Original paper

The executive brief below is grounded in the source paper and linked back to the arXiv abstract.

Text-to-image generation executes a diffusion workflow comprising multiple models centered on a base diffusion model. Existing serving systems treat each workflow as an opaque monolith, provisioning, placing, and scaling all constituent models together, which obscures internal dataflow, prevents model sharing, and enforces coarse-grained resource management. In this paper, we make a case for micro-serving diffusion workflows with LegoDiffusion, a system that decomposes a workflow into loosely coupled model-execution nodes that can be independently managed and scheduled. By explicitly managing individual model inference, LegoDiffusion unlocks cluster-scale optimizations, including per-model scaling, model sharing, and adaptive model parallelism. Collectively, LegoDiffusion outperforms existing diffusion workflow serving systems, sustaining up to 3x higher request rates and tolerating up to 8x higher burst traffic.

Score 87Full-paper briefinferenceinframodels

Executive brief

A short business-reader brief that explains why the paper matters now and what to watch or do next.

Why this is worth your attention

This paper argues that text-to-image serving is hitting an infrastructure bottleneck, not just a model bottleneck: today’s systems often scale whole image-generation pipelines as one unit, even when only one model inside the workflow is overloaded. If LegoDiffusion’s results hold up, image platforms could handle meaningfully more traffic with fewer GPUs by treating diffusion workflows more like composable services than sealed apps, which would pressure vendors on scheduler quality, model-sharing, and GPU data movement rather than just raw model support. The evidence is stronger on systems efficiency than market readiness: the gains are substantial in the authors’ H800-based setup, but they depend on specialized interconnect-aware engineering and haven’t yet shown broad, real-world deployment economics.

  • If you run or buy text-to-image infrastructure, the default assumption that each workflow needs its own full replica looks increasingly wasteful. The paper’s core claim is that scaling only the bottleneck model can avoid loading non-bottleneck components and materially improve utilization, which matters for capacity planning and margin as workflows pile on ControlNets, LoRAs, and other adapters.
  • A practical buying question is whether a serving stack can reuse loaded backbones and adapters across tenants and workflows, or whether it silently keeps separate copies. That distinction can become a direct cost and latency lever when a few popular models dominate demand; the paper cites production traces where the top 5 ControlNets serve 95% of requests and reports up to 60% lower GPU memory footprint from model sharing.
  • The architecture is persuasive, but the implementation leans hard on GPU-native communication: the paper says over 99% of transferred data is CUDA tensors, with one SDXL-plus-ControlNet workflow moving 5.3 GiB between nodes. That means the business impact is biggest for operators with NVLink/RDMA-class clusters; on weaker interconnects, some of the advantage could shrink fast.
  • If this direction is right, image-serving platforms will compete less on simply exposing more models and more on whether they can schedule model-level work intelligently under bursty demand. The paper’s strongest end-to-end claim is not just higher throughput but better SLO survival under pressure—up to 3× higher request rates and 8× more burstiness—which is exactly where user-facing services tend to fail in practice.
  • The evidence here is meaningful but still bounded: the evaluation centers on 12 workflows built from SD3, SD3.5-Large, and Flux variants on H800-based setups, and some operational pieces rely on offline latency profiles. The next adoption signal is replication on more heterogeneous clusters, broader model families, and real production traffic where scheduling estimates drift and interconnect quality varies.

Affiliations

Institution names extracted from the brief's PDF summary call.

Hong Kong University of Science and Technology

Author marker †

From PDF summary

Alibaba Group

No author markers parsed

From PDF summary

Evidence ledger

The strongest claims in the brief, along with the confidence and citation depth behind them.

capabilityhighp.10p.10

LegoDiffusion sustains materially higher text-to-image serving throughput under latency targets than monolithic baselines.

stackhighp.2p.4

Per-model scaling and model sharing are the main architectural levers, reducing redundant replication and GPU waste.

inferencehighp.7p.7

The design relies on a specialized GPU-native data plane to move large tensors fast enough for micro-serving to work.

strategicmediump.11p.11

If replicated broadly, the approach could reduce GPU capacity needs for commercial image services and make bursty demand easier to absorb.

caveatmediump.12p.8

Results may not transfer cleanly to other clusters or model mixes because the evaluation is concentrated on specific workflows and H800-class infrastructure.

Related briefs

More plain-English summaries from the archive with nearby topics or operator relevance.

No related briefs found yet.

Thank you to arXiv for use of its open access interoperability. This product was not reviewed or approved by, nor does it necessarily express or reflect the policies or opinions of, arXiv.
LightDark