arXiv 2604.13777v1Apr 15, 2026

From Anchors to Supervision: Memory-Graph Guided Corpus-Free Unlearning for Large Language Models

Wenxuan Li et al.

Brief context

Publication timing, weekly edition context, and source links for this brief.

Published

Apr 15, 2026, 12:07 PM

Current score

84

Original paper

The executive brief below is grounded in the source paper and linked back to the arXiv abstract.

Large language models (LLMs) may memorize sensitive or copyrighted content, raising significant privacy and legal concerns. While machine unlearning has emerged as a potential remedy, prevailing paradigms rely on user-provided forget sets, making unlearning requests difficult to audit and exposing systems to secondary leakage and malicious abuse. We propose MAGE, a Memory-grAph Guided Erasure framework for user-minimized, corpus-free unlearning. Given only a lightweight user anchor that identifies a target entity, MAGE probes the target LLM to recover target-related memorization, organizes it into a weighted local memory graph, and synthesizes scoped supervision for unlearning. MAGE is model-agnostic, can be plugged into standard unlearning methods, and requires no access to the original training corpus. Experiments on two benchmarks, TOFU and RWKU, demonstrate that MAGE's self-generated supervision achieves effective unlearning performance comparable to supervision generated with external reference, while preserving overall utility. These results support a practical and auditable unlearning workflow driven by minimal anchors rather than user-supplied forget corpora.

Score 84Full-paper briefmodelstrainingdatainfra

Executive brief

A short business-reader brief that explains why the paper matters now and what to watch or do next.

Why this is worth your attention

This paper pushes unlearning a step closer to something enterprises could actually operationalize: instead of asking a user or rights holder to hand over a full “forget corpus,” it claims you can start with just a name or short description and have the model help surface what needs to be removed. If that holds up, compliance, legal, and model-ops teams get a cheaper and more auditable path for handling privacy or copyright takedown requests without retaining more sensitive data just to delete it later. The evidence is stronger on benchmarked feasibility than on real-world deployment, but the practical signal is important: unlearning may become a workflow and tooling problem, not just a data-access problem.

  • The core shift here is operational: the paper claims you can generate usable unlearning supervision from a lightweight anchor plus model probing, not from a user-supplied forget set or the original training corpus. If true, that lowers friction for copyright, privacy, and contract-driven removal requests and reduces the awkward practice of collecting more sensitive data just to process deletion.
  • What is interesting is not just deletion, but boundary control. MAGE explicitly builds a neighbor set of related but non-target knowledge to preserve, which is a practical answer to the business problem of removing risky content without degrading adjacent capabilities; any vendor claiming unlearning should be able to explain an equivalent mechanism and show locality metrics.
  • Because MAGE outputs standard fine-tuning data and plugs into existing unlearning pipelines, the commercial opening is likely workflow software around request intake, auditability, supervision generation, and retraining orchestration. That matters for platform teams and model providers because differentiation may shift from raw model training access to who can run defensible, low-friction remediation pipelines.
  • The paper shows meaningful benchmark performance, including much stronger forgetting than a simple DirectQA baseline on Llama-2-7b-chat under GA+GD, but the evaluations are still concentrated on TOFU and RWKU and performance varies by objective and model. The main failure mode is straightforward: if the probing stage misses memorized facts or generates noisy supervision, forgetting will be incomplete or inefficient.
  • The paper’s implementation detail matters: the authors report LoRA-based runs on a single RTX 4090, with bounded probing budgets and limited graph expansion. If vendors can show similar low-overhead remediation on customer-specific models, this stops looking like a research-only idea and starts looking like a viable operational control.

Evidence ledger

The strongest claims in the brief, along with the confidence and citation depth behind them.

strategichighp.1p.3

A minimal user anchor can be used to drive corpus-free unlearning supervision generation instead of requiring a forget corpus.

stackhighp.5

MAGE is model-agnostic and compatible with existing unlearning pipelines because it outputs standard fine-tuning data.

capabilityhighp.5

The method explicitly tries to limit collateral forgetting by constructing a neighbor set of related non-target knowledge.

inferencemediump.11

The implementation appears lightweight enough to run on modest hardware using LoRA.

caveathighp.4p.5

Results are promising but benchmark-limited, and effectiveness depends on recovering the right memorized facts during probing.

Related briefs

More plain-English summaries from the archive with nearby topics or operator relevance.

cs.LG

KV Cache Offloading for Context-Intensive Tasks

Andrey Bocharnikov et al.

cs.AI

SkillClaw: Let Skills Evolve Collectively with Agentic Evolver

Ziyu Ma et al.

cs.AI

Don't Overthink It: Inter-Rollout Action Agreement as a Free Adaptive-Compute Signal for LLM Agents

Khushal Sethi

cs.CR

The Salami Slicing Threat: Exploiting Cumulative Risks in LLM Systems

Yihao Zhang et al.

Thank you to arXiv for use of its open access interoperability. This product was not reviewed or approved by, nor does it necessarily express or reflect the policies or opinions of, arXiv.
LightDark