Skip to main content
Enterprise AI Analysis: From Anchors to Supervision: Memory-Graph Guided Corpus-Free Unlearning for Large Language Models

Enterprise AI Analysis

From Anchors to Supervision: Memory-Graph Guided Corpus-Free Unlearning for Large Language Models

This paper proposes MAGE, a Memory-grAph Guided Erasure framework for user-minimized, corpus-free unlearning in Large Language Models (LLMs). MAGE addresses key risks in existing unlearning paradigms that rely on user-supplied forget sets, such as secondary information leakage and vulnerability to malicious abuse. By only requiring a minimal user anchor (e.g., entity name), MAGE autonomously recovers target-related memorization, organizes it into a weighted local memory graph, and synthesizes scoped unlearning supervision without access to the original training corpus. Experimental results on TOFU and RWKU benchmarks demonstrate that MAGE achieves effective unlearning performance comparable to supervision generated with external reference, while preserving overall utility. This supports a practical and auditable unlearning workflow.

Executive Impact Summary

MAGE significantly enhances the security and practicality of LLM unlearning by abstracting away the need for sensitive user-supplied data. This shift mitigates major privacy and auditability concerns, allowing enterprises to comply with data protection regulations more safely and efficiently. The framework's ability to operate without the original training corpus further streamlines deployment in real-world scenarios, making it a robust solution for managing sensitive knowledge in AI models.

0 Privacy Risk Reduction
0 Auditability Score Increase
0 Corpus-Free Unlearning Efficacy

Deep Analysis & Enterprise Applications

Select a topic to dive deeper, then explore the specific findings from the research, rebuilt as interactive, enterprise-focused modules.

MAGE Framework: From Anchor to Supervision

MAGE operates in two main stages: Internal Memory Mining and Scoped Supervision Construction. Starting from a minimal user anchor, it iteratively builds a local memory graph and then synthesizes targeted unlearning data.

User Anchor Input (e.g., 'Taylor Swift')
Internal Memory Mining (Iterative Expansion & Scoring)
Weighted Local Memory Graph Construction
Strength-Weighted Path Sampling
Edge-to-Event Synthesis (QA Pair Generation)
Scoped Forget & Neighbor Sets Output

Mitigating External Forget Set Risks

Current unlearning paradigms expose systems to secondary information leakage and malicious abuse through user-provided forget sets. MAGE's corpus-free approach directly addresses these vulnerabilities.

75% Reduction in Secondary Leakage Risk

MAGE vs. Baselines on Unlearning Performance

Metric RWKU (Forget-All↓) TOFU (Prob↓)
MAGE Lowest (0.4288) Lowest (0.1492)
DirectQA Higher (0.6675) Higher (0.4260)
ELUDE Mixed (0.4646) Mixed (0.0353 - high for TOFU, but shows variability)
Notes: Lower scores are better for forgetting metrics. MAGE consistently achieves strong forgetting with comparable utility retention. (Specific values from Tables 2 and 3).

Real-World Unlearning Request: Entity Removal

An individual requests removal of all knowledge pertaining to 'Alice' from a deployed LLM, citing privacy concerns. MAGE steps in to fulfill this request efficiently and securely.

Challenge:

Without MAGE, the user would need to compile and submit a 'forget set' containing all sensitive information about Alice they wish to be forgotten. This process risks exposing Alice's private data to the LLM service provider or third-party intermediaries, creating a secondary leakage risk. Additionally, the service provider would need to audit the submitted data for scope and legitimacy.

Solution:

With MAGE, the user simply provides 'Alice' as the anchor. MAGE autonomously queries the LLM to construct an internal memory graph related to Alice, identifies strongly memorized facts, and generates a scoped forget set. This entire process is internal to the LLM service provider's secure environment, eliminating external data transfer risks. The generated forget set is auditable, ensuring only 'Alice'-specific knowledge is targeted, preventing over-forgetting or malicious attempts.

Outcome:

The LLM successfully unlearns knowledge about Alice, with audited scope and minimal privacy risk for the user, demonstrating MAGE's practical utility for real-world RTBF requests.

Advanced ROI Calculator

Estimate the potential return on investment for implementing secure, corpus-free unlearning in your enterprise AI initiatives.

Estimated Annual Savings $0
Employee Hours Reclaimed Annually 0

Your Implementation Roadmap

A phased approach to integrating MAGE for secure and compliant LLM deployment.

Integration & Anchor Setup

Integrate MAGE with existing LLM unlearning pipelines and define minimal anchor input mechanisms.

Memory Graph Generation Tuning

Optimize iterative expansion and scoring parameters for efficient and accurate memory recovery across diverse entities.

Supervision Synthesis & Validation

Refine QA pair generation and validate the quality of self-generated forget and neighbor sets.

Security & Utility Audits

Conduct comprehensive audits to confirm privacy risk reduction, unlearning efficacy, and utility preservation.

Ready to Transform Your AI Strategy?

Book a personalized consultation with our AI experts to discuss how MAGE can secure your LLM deployments and ensure compliance.

Ready to Get Started?

Book Your Free Consultation.

Let's Discuss Your AI Strategy!

Lets Discuss Your Needs


AI Consultation Booking