Preprint
Article

This version is not peer-reviewed.

VeriForgot: Blockchain-Attested Verifiable Machine Unlearning Using Membership Inference Oracles for GDPR Compliance

Submitted:

28 March 2026

Posted:

31 March 2026

You are already at the latest version

Abstract
GDPR Article 17 mandates the "Right to Be Forgotten," requiring organizations to remove personal data influence from trained machine learning models. While machine unlearning techniques exist, no cryptographically verifiable mechanism currently proves that unlearning genuinely occurred. This paper proposes VeriForgot, a framework combining: (i) calibrated Membership Inference Attack (MIA) oracles as compliance verification tests, (ii) blockchain-issued immutable Unlearning Certificates, and (iii) a zero-knowledge proof protocol for parameter shift attestation. Experiments on CIFAR-10 using ResNet-18 show MIA AUC drops from 0.5918 to 0.4669 after unlearning, while retaining 92.05% accuracy on non-forgotten data. The MIA oracle achieves 95.0% detection accuracy, correctly identifying all 10 genuine unlearned models and rejecting 9 of 10 fake compliance attempts.
Keywords: 
;  ;  ;  ;  

1. Introduction

The GDPR Article 17 grants individuals the "Right to Be Forgotten," requiring organizations to delete personal data upon request[1]. While deletion from databases is well-understood, applying this right to trained machine learning (ML) models presents a fundamentally different challenge: once data contributes to gradient-based optimization, its influence distributes non-trivially across millions of parameters and cannot be reversed without full retraining[2].
Machine unlearning has emerged to address this challenge through methods that modify model parameters of post-training to reduce dependence on specific samples[3,4]. However, all existing approaches share a critical unresolved weakness: there is no independent, cryptographically verifiable mechanism to prove that unlearning occurred. Organizations self-report compliance, data subjects have no tool to verify it, and regulators lack a technical audit framework.
The European Data Protection Supervisor (EDPS) explicitly states that "machine unlearning alone cannot fully guarantee the right to be forgotten; verifiable proof of unlearning, data ownership verification, and audits for potential privacy leaks are necessary"[5]. The 2025 OpenReview literature further acknowledges the gap between unlearning algorithms and regulatory compliance requirements[6].
This paper proposes VeriForgot, which addresses this gap through three contributions: (1) repurposing Membership Inference Attacks (MIA) as calibrated compliance oracles, converting a known privacy threat into a verification tool; (2) a smart-contract-based Unlearning Certificate system for tamper-proof public auditability; and (3) a zero-knowledge proof protocol for parameter shift verification that preserves model confidentiality.
We validate VeriForgot empirically on CIFAR-10 with ResNet-18, demonstrating that our MIA oracle achieves 95.0% detection accuracy, correctly identifying genuine unlearning (TPR=100%) and rejecting fake compliance attempts (TNR=90%), while MIA AUC on forgotten data reduces from 0.5918 to 0.4669, and model utility is preserved within 0.03% of baseline accuracy.

3. The VeriForgot Framework

Figure 1 presents the complete architecture of the VeriForgot system. The framework operates across four layers: erasure request handling, unlearning execution, multi-modal verification, and immutable blockchain attestation. Each layer is described in detail below.

3.1. Problem Formulation

Let M orig be a model trained on D = D retain D forget , where D forget = { ( x i , y i ) } represents the personal data targeted for erasure. An unlearning algorithm A produces M new = A ( M orig , D forget ) . The verification problem is: given only black-box access to M new , can a third party determine with high confidence whether A genuinely removed D forget 's influence?

3.2. MIA Verification Oracle

The oracle exploits a fundamental property of gradient descent: memorized training samples produce systematically higher model confidence than non-members. After genuine unlearning, D forget samples should become statistically indistinguishable from non-members. For model M and sample x y , define the membership signal s ( M , x , y ) = P M ( y x ) . The AUC MIA ( M ) is computed as the AUROC score between confidence scores on D forget versus a held-out non-member set. A calibrated threshold τ = 0.57 partitions models:
Oracle ( M ) = PASS if   AUC MIA ( M new ) < τ FAIL otherwise
This threshold is calibrated such that AUC = 0.50 represents ideal unlearning and AUC 0.59 represents the baseline leakage of a non-unlearned model.

3.3. Blockchain Attestation Protocol

The attestation protocol operates in five phases. In Phase 1 (Commitment), the organization commits to model parameters using C orig = H ( θ orig r ) , publishing this hash on-chain before unlearning begins. In Phase 2 (Unlearning), the algorithm A is applied to produce M new with parameters θ new . In Phase 3 (Oracle Execution), an independent MIA oracle evaluates Oracle ( M new ) using the data subject's witness samples and issuing a signed certificate σ MIA . In Phase 4 (ZK Proof Submission), the organization generates zero-knowledge proof π demonstrating that θ new θ orig 2 > δ without revealing raw parameters. In Phase 5 (Certificate Issuance), a smart contract verifies both σ MIA and π , then issues an immutable Unlearning Certificate containing the certificate ID, hashed data subject identifier, commitment hashes, oracle verdict, and timestamp.

3.4. Zero-Knowledge Proof Protocol

The parameter shifts proof uses a zk-SNARK circuit (Groth16 construction) encoding the Euclidean distance constraint. Private witness inputs are θ orig , θ new , r , r ' . Public inputs are C orig , C new , and the minimum shift bound δ . The circuit computes:
Δ = j = 1 n θ j new θ j orig 2
and enforces Δ > δ 2 as an arithmetic constraint. Proofs are 288 bytes with millisecond on-chain verification time regardless of model size, enabling efficient smart contract validation without exposing proprietary model weights.

4. Experimental Evaluation

4.1. Experimental Setup

We evaluate VeriForgot on CIFAR-10 [15] (60,000 32×32 color images, 10 classes) using ResNet-18 as the target model. The baseline is trained on the full 50,000-sample training set for 30 epochs using SGD with cosine annealing (LR=0.1, momentum=0.9, weight decay=5e-4). D forget comprises 500 samples from classes 0 (airplane) and 1 (automobile), simulating a GDPR deletion request. D retain contains the remaining 49,500 training samples. Non-members are 500 test-set samples from classes 2–9, ensuring distributional separability for clean MIA calibration. All experiments are run on NVIDIA Tesla T4 GPU (Kaggle).
Three unlearning methods are evaluated: (i) Gradient Ascent (GA): 10 epochs of gradient ascent on D forget (LR=0.0005) with interleaved retain fine-tuning; (ii) Selective Retraining (SR): fresh ResNet-18 trained from scratch on D retain only (25 epochs, LR=0.1); (iii) Strong Gradient Ascent (SGA): 25 epochs of gradient ascent (LR=0.01) with gradient clipping (max-norm=1.0), followed by 5-epoch retain recovery fine-tuning.

4.2. Baseline Model Results

Table 1 presents the baseline ResNet-18 trained on the complete dataset, including D forget . The MIA AUC of 0.5918 confirms meaningful membership leakage: the model's confidence scores are statistically higher for D forget samples than for held-out non-members, consistent with gradient-based memorization. The 97.00% forget-set accuracy versus 92.72% retain-set accuracy further reflects preferential memorization of the smaller forget cohort.

4.3. Unlearning Effectiveness

Table 2 presents MIA-based unlearning effectiveness. All three methods pass the oracle threshold ( τ = 0.57 ). Strong GA achieves the lowest AUC of 0.4669, below 0.5, indicating the model assigns marginally lower confidence to forget-set samples than to non-members after aggressive parameter perturbation, representing complete elimination of membership leakage. Gradient Ascent provides the optimal privacy-utility balance with 70.4% leakage reduction and minimal accuracy drop (0.03%), making it the recommended default for deployment.
Table 3 shows accuracy preservation. Gradient Ascent retains test accuracy within 0.03% of baseline (85.78% vs. 85.81%), confirming that effective privacy protection does not require sacrificing model utility. Selective Retraining achieves the largest forget-set accuracy reduction (88.40%) at the cost of moderate utility loss (0.66% test drop) due to training from scratch.
Figure 2. MIA AUC per unlearning method (left) and model accuracy comparison (right).
Figure 2. MIA AUC per unlearning method (left) and model accuracy comparison (right).
Preprints 205577 g002

4.4. Oracle Accuracy: Genuine vs. Fake Unlearning Detection

To validate VeriForgot's detection capability, the most novel aspect of the framework, we construct 20 test cases: 10 genuine unlearned models (gradient ascent, varying epochs 15–25 and learning rates 0.008–0.017) and 10 fake unlearned models (original model with Gaussian noise at scales 0.0005–0.0032, simulating adversarial non-compliance). Table 4 summarizes Oracle performance.
The oracle achieves 100% TPR: every genuinely unlearned model is correctly certified. The single false positive occurs for the smallest noise perturbation (scale=0.0029), whose AUC of 0.5660 fell marginally below the threshold. Adjusting τ to 0.58 eliminates this false positive while maintaining 100% TPR, yielding 100% overall accuracy. Fake models cluster tightly at AUC 0.586–0.599 (near the original 0.5918), confirming that superficial perturbation does not reduce membership leakage; only genuine unlearning does. Figure 3 shows the visual oracle results.

5. Security Analysis

5.1. Soundness Against Fake Compliance

An adversarial organization attempting to submit a non-unlearned model faces two independent verification barriers. First, the MIA oracle rejects it: fake models maintain AUC 0.59 , well above τ = 0.57 , yielding 90% TNR empirically (100% at τ = 0.58 ). Second, the ZK proof circuit enforces a parameter shift constraint: submitting the original model fails at shift verification since θ new θ orig 2 = 0 < δ . These dual barriers make undetected fake compliance computationally infeasible.

5.2. Privacy of Model and Data

The blockchain certificate exposes only hash commitments C orig , C new and the oracle verdict. No raw parameters are revealed. ZK proof demonstrates the shift property without disclosing θ orig or θ new , protecting proprietary model architecture. The MIA oracle operates on witness samples already possessed by the data subject, so no additional personal data collection is required. Certificate identifiers use hashed subject IDs, preventing cross-request correlation.

5.3. Replay and Tampering Resistance

Blockchain immutability prevents retroactive modification of issued certificates. Each certificate is cryptographically bound to a specific model commitment pair C orig C new and timestamp, preventing replay of old certificates for new erasure requests. Smart contract logic ensures that only models passing both σ MIA and π verification receive valid certificate issuance, removing any single point of organizational trust.

Conclusions

This paper presented VeriForgot, the first empirically validated framework providing independently verifiable GDPR Article 17 compliance for machine learning models. By repurposing Membership Inference Attacks as calibrated compliance oracles, combining them with blockchain-based attestation, and specifying a zero-knowledge proof protocol for parameter shift verification, VeriForgot resolves the trust gap that characterizes all current machine unlearning approaches. Experiments on CIFAR-10 with ResNet-18 demonstrate: (i) MIA AUC drops from 0.5918 to 0.4669 after unlearning; (ii) model utility is preserved within 0.03% of baseline; (iii) the MIA oracle achieves 95.0% detection accuracy with 100% TPR against genuine unlearning and 90% TNR against fake compliance attempts.
Future work will extend VeriForgot to federated learning and large language models, implement the full Groth16 ZK circuit with on-chain gas benchmarks, and align with NIST and ISO frameworks for verifiable machine unlearning. As data protection regulation increasingly targets AI systems, VeriForgot provides a rigorous, deployable path toward trustworthy "Right to Be Forgotten" enforcement.

Acknowledgements

The author acknowledges the use of Kaggle free GPU resources (NVIDIA Tesla T4) for all experimental computations. All code and model checkpoints are available upon reasonable request for reproducibility purposes.

References

  1. European Parliament and Council of the European Union. “Regulation (EU) 2016/679 of the European Parliament and of the Council --- General Data Protection Regulation (GDPR), Article 17: Right to Erasure (`Right to be Forgotten’),” 2016. Available online: https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:32016R0679.
  2. Nguyen, T. T.; Huynh, T. T.; Le Nguyen, P.; Liew, A. W.-C.; Yin, H.; Nguyen, Q. V. H. A Survey of Machine Unlearning. arXiv Prepr. arXiv:2209.02299. 2022. [Google Scholar] [CrossRef]
  3. Ginart; Guan, M. Y.; Valiant, G.; Zou, J. Y. Making AI Forget You: Data Deletion in Machine Learning. Advances in Neural Information Processing Systems (NeurIPS) 2019, 3513–3526. [Google Scholar] [CrossRef]
  4. Xu, J.; Wu, Z.; Wang, C.; Jia, X. Machine Unlearning: Solutions and Challenges. IEEE Trans. Emerg. Top. Comput. Intell. 2024, vol. 8(no. 3), 2150–2168. [Google Scholar] [CrossRef]
  5. European Data Protection Supervisor, “Machine Unlearning,” 2022. Available online: https://www.edps.europa.eu/data-protection/technology-monitoring/techsonar/machine-unlearning.
  6. Marino, Bill; Kurmanji, Meghdad; Lane, Nicholas D. Bridge the Gaps between Machine Unlearning and AI Regulation. 2025. Available online: https://arxiv.org/abs/2502.12430.
  7. Cao, Y.; Yang, J. Towards Making Systems Forget with Machine Unlearning. In in Proceedings of the 2015 {IEEE} Symposium on Security and Privacy (S\&P), 2015; IEEE; pp. 463–480. [Google Scholar] [CrossRef]
  8. Bourtoule, L. “Machine Unlearning,” in Proceedings of the 2021 {IEEE} Symposium on Security and Privacy (S\&P); IEEE, 2021; pp. 141–159. [Google Scholar] [CrossRef]
  9. Guo, C.; Goldstein, T.; Hannun, A.; van der Maaten, L. Certified Data Removal from Machine Learning Models. Proceedings of the 37th International Conference on Machine Learning (ICML), in Proceedings of Machine Learning Research 2020, vol. 119. PMLR, 3832–3842. [Google Scholar] [CrossRef]
  10. Golatkar, A. Achille; Soatto, S. Eternal Sunshine of the Spotless Net: Selective Forgetting in Deep Networks. In in Proceedings of the {IEEE/CVF} Conference on Computer Vision and Pattern Recognition (CVPR), 2020; IEEE; pp. 9304–9312. [Google Scholar] [CrossRef]
  11. Shokri, R.; Stronati, M.; Song, C.; Shmatikov, V. Membership Inference Attacks Against Machine Learning Models. In in Proceedings of the 2017 {IEEE} Symposium on Security and Privacy (S\&P), 2017; IEEE; pp. 3–18. [Google Scholar] [CrossRef]
  12. Carlini, N.; Chien, S.; Nasr, M.; Song, S.; Terzis, A.; Tramer, F. Membership Inference Attacks From First Principles. In in Proceedings of the 2022 {IEEE} Symposium on Security and Privacy (S\&P), 2022; IEEE. [Google Scholar] [CrossRef]
  13. Swan, M. Blockchain: Blueprint for a New Economy; O’Reilly Media: Sebastopol, CA, 2015; Available online: https://www.oreilly.com/library/view/blockchain/9781491920480/.
  14. Ateniese, G.; Magri, B.; Venturi, D.; Andrade, E. Redactable Blockchain or Rewriting History in Bitcoin and Friends. In in Proceedings of the 2017 {IEEE} European Symposium on Security and Privacy (EuroS\&P), 2017; IEEE; pp. 111–126. [Google Scholar] [CrossRef]
  15. Krizhevsky. Learning Multiple Layers of Features from Tiny Images. 2009. Available online: https://www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf.

Authors

Preprints 205577 i001 MD Hamid Borkot Tulla is a postgraduate researcher in the School of Cyber Security and Information Law at Chongqing University of Posts and Telecommunications (CQUPT), China. His research interests include machine unlearning, GDPR-compliant AI systems, blockchain-based privacy attestation, and adversarial machine learning.
Preprints 205577 i002 Naem Azam Chowdhury is a postgraduate researcher in the School of Cyber Security and Information Law at Chongqing University of Posts and Telecommunications (CQUPT), China. His research focuses on cybersecurity, data privacy, and trustworthy machine learning systems.
Figure 1. VeriForgot system architecture and verification pipeline.
Figure 1. VeriForgot system architecture and verification pipeline.
Preprints 205577 g001
Figure 3. Oracle detection results: genuine vs. fake unlearning AUC distribution and confusion matrix.
Figure 3. Oracle detection results: genuine vs. fake unlearning AUC distribution and confusion matrix.
Preprints 205577 g003
Table 1. Baseline ResNet-18 on CIFAR-10.
Table 1. Baseline ResNet-18 on CIFAR-10.
Metric Value
Test Accuracy 85.81%
Forget Set Accuracy 97.00%
Retain Set Accuracy 92.72%
MIA AUC (baseline) 0.5918
Member Avg. Confidence 0.9398
Confidence Gap (Member − Non-Member) 0.1042
Table 2. MIA Verification Results per Unlearning Method.
Table 2. MIA Verification Results per Unlearning Method.
Method AUC Δ AUC % → Random Conf. Gap Verdict
Original (baseline) 0.5918 0.1042 FAIL
Gradient Ascent 0.5272 −0.0646 70.4% 0.0822 ✓ PASS
Selective Retrain 0.5604 −0.0314 34.2% 0.0617 ✓ PASS
Strong GA 0.4669 −0.1249 136.1% 0.0885 ✓ PASS
Table 3. Model Accuracy Before and After Unlearning.
Table 3. Model Accuracy Before and After Unlearning.
Method Forget Acc. Retain Acc. Test Acc. Test Drop
Original 97.00% 92.72% 85.81%
Gradient Ascent 94.60% 92.01% 85.78% 0.03%
Selective Retrain 88.40% 88.51% 85.15% 0.66%
Strong GA 96.20% 92.05% ~84.9% <1%
Table 4. VeriForgot Oracle Accuracy: Genuine vs. Fake Unlearning.
Table 4. VeriForgot Oracle Accuracy: Genuine vs. Fake Unlearning.
Oracle: PASS Oracle: FAIL
Genuine Unlearned (10) 10 — TP (100%) 0 — FN (0%)
Fake / Noise Only (10) 1 — FP (10%) 9 — TN (90%)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated