arXiv 2604.26951v1Apr 29, 2026

Turning the TIDE: Cross-Architecture Distillation for Diffusion Large Language Models

Gongbo Zhang et al.

Brief context

Publication timing, weekly edition context, and source links for this brief.

Published

Apr 29, 2026, 5:59 PM

Current score

91

Original paper

The executive brief below is grounded in the source paper and linked back to the arXiv abstract.

Diffusion large language models (dLLMs) offer parallel decoding and bidirectional context, but state-of-the-art dLLMs require billions of parameters for competitive performance. While existing distillation methods for dLLMs reduce inference steps within a single architecture, none address cross-architecture knowledge transfer, in which the teacher and student differ in architecture, attention mechanism, and tokenizer. We present TIDE, the first framework for cross-architecture dLLM distillation, comprising three modular components: (1) TIDAL, which jointly modulates distillation strength across training progress and diffusion timestep to account for the teacher's noise-dependent reliability; (2) CompDemo, which enriches the teacher's context via complementary mask splitting to improve predictions under heavy masking; and (3) Reverse CALM, a cross-tokenizer objective that inverts chunk-level likelihood matching, yielding bounded gradients and dual-end noise filtering. Distilling 8B dense and 16B MoE teachers into a 0.6B student via two heterogeneous pipelines outperforms the baseline by an average of 1.53 points across eight benchmarks, yielding notable gains in code generation, where HumanEval scores reach 48.78 compared to 32.3 for the AR baseline.

Score 91Full-paper briefmodelstraininginferenceinfra

Executive brief

A short business-reader brief that explains why the paper matters now and what to watch or do next.

Why this is worth your attention

If this paper is right, diffusion LLMs become more plausible as small, fast deployment models rather than just an interesting alternative decoding scheme. The authors show a way to transfer capability from much larger, even incompatible, teachers into a 0.6B diffusion student, with reported gains in benchmark average, code generation, memory, and throughput. The business implication is cheaper inference and less vendor-stack lock-in; the caveat is that the evidence is still narrow, with one small student, short training context, and controlled hardware measurements.

  • The practical assumption this challenges is that small diffusion LLMs must be trained or distilled inside a tightly matched model family. If cross-architecture and cross-tokenizer transfer keeps working, companies with access to strong teacher models may have more freedom to build smaller deployment models without copying the teacher stack.
  • The paper reports 22× lower memory and 5× faster throughput for the small distilled model, but those figures come from a controlled H100/bfloat16 setup and best-of-five runs. Ask vendors whether gains survive your sequence lengths, hardware, batching, and quality thresholds—not just a benchmark demo.
  • The most commercially interesting result is not the modest eight-benchmark average gain; it is the HumanEval jump from 32.30 for a same-size autoregressive baseline to 48.78 for the distilled diffusion model. If similar gains appear in real code-assist workflows, small diffusion models become more credible for latency-sensitive developer tools.
  • This is not a free compression trick. One component doubles teacher forward passes and adds roughly 50% to training duration, so the business case depends on whether repeated cheaper inference offsets a heavier one-time distillation job.
  • The main unresolved question is whether this still works when the student is larger, the context is longer, and the tasks look more like production workloads. Evidence at 0.6B parameters and a 512-token training window is promising, but not enough to underwrite platform-level bets.

Evidence ledger

The strongest claims in the brief, along with the confidence and citation depth behind them.

capabilityhighp.2p.7

TIDE improves a 0.6B diffusion LLM over a non-distilled baseline across eight benchmarks, with the best reported average rising from 32.67 to 34.20.

capabilityhighp.1p.7

The largest visible capability gain is in code generation, where the distilled diffusion model substantially outperforms a same-size autoregressive baseline on HumanEval.

inferencemediump.2p.14

The paper reports large inference-efficiency gains for the small distilled model, though measured in a controlled hardware and precision setting.

caveathighp.16p.16

The evidence is limited to specific teacher-student pipelines, one 0.6B student architecture, and a short training context window.

Related briefs

More plain-English summaries from the archive with nearby topics or operator relevance.

cs.LG

KV Cache Offloading for Context-Intensive Tasks

Andrey Bocharnikov et al.

cs.AI

Don't Overthink It: Inter-Rollout Action Agreement as a Free Adaptive-Compute Signal for LLM Agents

Khushal Sethi

cs.CL

DIVERSED: Relaxed Speculative Decoding via Dynamic Ensemble Verification

Ziyi Wang et al.

cs.CR

The Salami Slicing Threat: Exploiting Cumulative Risks in LLM Systems

Yihao Zhang et al.

Thank you to arXiv for use of its open access interoperability. This product was not reviewed or approved by, nor does it necessarily express or reflect the policies or opinions of, arXiv.
LightDark