A Cryptographic Framework for Evaluating Machine Unlearning
Machine unlearning addresses data protection regulations like GDPR’s “Right to be Forgotten” by updating machine learning models to remove specific training data influence. While retraining from scratch is ideal, it’s computationally prohibitive, leading to numerous approximate unlearning algorithms. However, a fundamental challenge threatens the entire field: there is no reliable way to evaluate whether these algorithms actually work. The central problem is measuring data removal efficacy—determining whether information from deleted data has truly been removed from the model. Without trustworthy evaluation metrics, researchers cannot identify effective algorithms, practitioners cannot confidently deploy unlearning systems, and users’ privacy may remain inadequately protected.
The Cause: Fundamental Flaws in MIA-Based Metrics
The most common evaluation approach uses Membership Inference Attacks (MIAs), which attempt to determine whether specific data points were in the training dataset. While seemingly natural for measuring privacy leakage, MIA-based metrics suffer critical flaws. MIA performance is poorly calibrated for unlearning evaluation—even theoretically optimal retraining doesn’t guarantee the lowest MIA scores because the attacks themselves are imperfect. Results are highly sensitive to data composition and the specific MIA algorithm chosen, varying significantly between different attacks. This makes results incomparable and prevents drawing definitive conclusions about which unlearning algorithms truly protect privacy.
Context: A Fragmented Evaluation Landscape
The unlearning literature offers three metric categories, each with significant limitations. Attack-based methods directly measure privacy risks but lack theoretical grounding and reliability. Theory-based approaches like certified removal provide rigorous guarantees but require strong model assumptions (convexity/linearity) and inefficient white-box access, limiting practical applicability. Retraining-based metrics compare unlearned models to retrained ones through parameter differences, but these measurements are sensitive to random training factors like batch ordering and initialization. This fragmentation creates a crisis of confidence—no consensus exists on how to reliably evaluate unlearning algorithms.
The Solution: A Game-Theoretic Framework
The authors introduce the unlearning sample inference game, inspired by cryptographic security games. The evaluation is modeled as a game between a challenger (the unlearning algorithm) and an adversary (MIA attacker). The dataset is randomly split into retain (for training), forget (to be unlearned), and test (never seen) sets. The challenger produces an unlearned model by removing the forget set from a model trained on retain plus forget data. The adversary attempts to determine whether given data points come from the forget or test set by analyzing the unlearned model. The key innovation is the advantage metric—the difference in adversary success rates between these scenarios, averaged symmetrically across all possible dataset splits. This symmetry ensures biases toward specific data points cancel out when those points appear in both forget and test sets across different splits.
Theoretical Guarantees: Zero Grounding and Certified Removal
The advantage metric possesses desirable theoretical properties that existing metrics lack. Most importantly, it achieves zero grounding—any adversary has exactly zero advantage against retraining, formally certifying it as theoretically optimal. This provides a well-calibrated baseline impossible with standard MIA metrics. The framework also connects to certified removal from differential privacy. For algorithms with (ϵ,δ)-certified removal guarantees, the authors prove an upper bound on adversary advantage that scales with privacy budget parameters. As privacy guarantees strengthen (smaller ϵ), Unlearning Quality provably increases, aligning empirical measurement with theoretical guarantees. The framework naturally accommodates multiple MIA adversaries, resolving conflicts from different attack methods.
The SWAP Test: Practical Implementation
While the theoretical framework requires enumerating all dataset splits—computationally infeasible—the SWAP test provides an elegant approximation. It uses only two splits: an original division into retain, forget, and test sets, and a swapped version exchanging forget and test roles. Averaging advantages across these symmetric splits preserves zero grounding while requiring minimal computation. The authors prove the SWAP test maintains theoretical guarantees and demonstrate that naive random splitting leads to pathological cases where even perfect retraining shows high advantage.
Experimental Validation: Robustness Across Dimensions
Extensive experiments on CIFAR10 with ResNet architectures evaluated multiple unlearning algorithms including RETRAIN, FISHER, SALUN, and state-of-the-art SSD. Results demonstrate remarkable scalability—consistent algorithm rankings across dataset sizes from 10% to 100%, enabling reliable small-scale evaluation. The metric remains robust as the unlearning portion varies from 0.1 to 0.67. Critically, experiments validate theory: RETRAIN achieves Q ≈ 0.993-0.999 (near-perfect grounding), while approximate methods show appropriately lower scores. Differential privacy experiments confirmed theoretical predictions—Unlearning Quality decreases monotonically as privacy budget ϵ increases, with consistent rankings across all budgets. SSD achieved Q ≈ 0.928-0.996, validating its superiority.
The Best Solution: A New Standard for the Field
The SWAP test establishes the optimal practical solution, uniquely bridging theoretical rigor and practical usability. It provides provable guarantees—perfect zero grounding and alignment with certified removal—while requiring only two dataset splits. The metric demonstrates robustness across random seeds, dataset sizes, model architectures (ResNet variants), and domains (vision and language tasks). Most importantly, it correctly identifies real algorithmic differences, recognizing SSD as superior to other methods and providing actionable insights for practitioners. Comparison with existing metrics revealed their failures: both IC test and MIA AUC showed inconsistent correlations with privacy budgets and unstable rankings, while Unlearning Quality maintained theoretically correct patterns. By providing the first evaluation metric that is simultaneously theoretically sound, computationally practical, and empirically reliable, the SWAP test framework enables meaningful progress in developing effective machine unlearning techniques that truly protect user privacy.
Reference
Tu, Yiwen, Pingbang Hu, and Jiaqi W. Ma. “A Reliable Cryptographic Framework for Empirical Machine Unlearning Evaluation.” In 39th Conference on Neural Information Processing Systems (NeurIPS 2025). arXiv preprint arXiv:2404.11577v4 (2025).
All Founding Subscribers receive a full Enterprise License to Risk Anchorincluded with their subscription.
That includes:
The full local package (HTML/JS/CSS) to run on your own infrastructure.
Unlimited assessments: Use it across as many models, business units, or portfolio companies as you need.
Ongoing upgrades: New modules (for example, Business Continuity, Drift Monitoring, or sector-specific controls) are included as they are released.
For more ideas consider purchasing “Shaping the Decade: Governance, Sustainability, and AI 2026–2036,” a guide for boards at the crossroads of governance, technology, and stakeholder capitalism. Available Here.
Tanya Matanda is a governance strategist bridging institutional oversight, AI governance, and fiduciary resilience. Her work supports boards, LPs, and regulators in designing governance systems fit for the AI era.
Copyright © 2025 Matanda Advisory Services
Research and Audio Supported by AI Systems





Totally agree — you nailed the production reality.
If “retraining” can’t reliably anchor the baseline, then we’re not measuring unlearning so much as measuring the quirks of whichever attack we picked that day. That calibration drift is exactly what makes these evaluations hard to defend in compliance reviews and audit trails.
That’s why I like the “zero-grounding” property here: it gives you a stable reference point, so improvements (or regressions) are actually interpretable — not just attack-dependent noise.
Appreciate you calling out the theory → privacy engineering bridge.