Articles producció científicaEnginyeria Informàtica i Matemàtiques

DP2Unlearning: An efficient and guaranteed unlearning framework for LLMs

  • Dades identificatives

    Identificador:  imarina:9463649
    Autors:  Al Mahmud, T; Jebreel, N; Domingo-Ferrer, J; Sánchez, D
    Resum:
    Large language models (LLMs) have recently revolutionized language processing tasks but have also brought ethical and legal issues. LLMs have a tendency to memorize potentially private or copyrighted information present in the training data, which might then be delivered to end users at inference time. When this happens, a naive solution is to retrain the model from scratch after excluding the undesired data. Although this guarantees that the target data have been forgotten, it is also prohibitively expensive for LLMs. Approximate unlearning offers a more efficient alternative, as it consists of expost modifications of the trained model itself to prevent undesirable results, but it lacks forgetting guarantees because it relies solely on empirical evidence. In this work, we present DP2Unlearning, a novel LLM unlearning framework that offers formal forgetting guarantees at a significantly lower cost than retraining from scratch on the data to be retained. DP2Unlearning involves training LLMs on textual data protected using epsilon-differential privacy (DP), which later enables efficient unlearning with the guarantees against disclosure associated with the chosen epsilon. Our experiments demonstrate that DP2Unlearning achieves similar model performance post-unlearning, compared to an LLM retraining from scratch on retained data-the gold standard exact unlearning-but at approximately half the unlearning cost. In addition, with a reasonable computational cost, it outperforms approximate unlearning methods at both preserving the utility of the model post-unlearning and effectively forgetting the targeted information. The code of our experiments is available at https://github.com/tamimalmahmud/LLM-Unlearning/tree/main/ DP2Unlearning.
  • Altres:

    Autor segons l'article: Al Mahmud, T; Jebreel, N; Domingo-Ferrer, J; Sánchez, D
    Departament: Enginyeria Informàtica i Matemàtiques
    Autor/s de la URV: Domingo Ferrer, Josep / Sánchez Ruenes, David
    Paraules clau: Approximate unlearning; Differential privacy; Divergence; Exact unlearning; Llm unlearning; Mode; Privacy-preserving ll; Privacy-preserving llm
    Resum: Large language models (LLMs) have recently revolutionized language processing tasks but have also brought ethical and legal issues. LLMs have a tendency to memorize potentially private or copyrighted information present in the training data, which might then be delivered to end users at inference time. When this happens, a naive solution is to retrain the model from scratch after excluding the undesired data. Although this guarantees that the target data have been forgotten, it is also prohibitively expensive for LLMs. Approximate unlearning offers a more efficient alternative, as it consists of expost modifications of the trained model itself to prevent undesirable results, but it lacks forgetting guarantees because it relies solely on empirical evidence. In this work, we present DP2Unlearning, a novel LLM unlearning framework that offers formal forgetting guarantees at a significantly lower cost than retraining from scratch on the data to be retained. DP2Unlearning involves training LLMs on textual data protected using epsilon-differential privacy (DP), which later enables efficient unlearning with the guarantees against disclosure associated with the chosen epsilon. Our experiments demonstrate that DP2Unlearning achieves similar model performance post-unlearning, compared to an LLM retraining from scratch on retained data-the gold standard exact unlearning-but at approximately half the unlearning cost. In addition, with a reasonable computational cost, it outperforms approximate unlearning methods at both preserving the utility of the model post-unlearning and effectively forgetting the targeted information. The code of our experiments is available at https://github.com/tamimalmahmud/LLM-Unlearning/tree/main/ DP2Unlearning.
    Àrees temàtiques: Artificial intelligence; Astronomia / física; Biotecnología; Ciência da computação; Ciências agrárias i; Ciencias sociales; Cognitive neuroscience; Computer science, artificial intelligence; Economia; Engenharias iii; Engenharias iv; Filosofia/teologia:subcomissão filosofia; General medicine; Interdisciplinar; Matemática / probabilidade e estatística; Neurosciences; Psicología; Psychology; Química
    Accès a la llicència d'ús: https://creativecommons.org/licenses/by/3.0/es/
    Adreça de correu electrònic de l'autor: josep.domingo@urv.cat; josep.domingo@urv.cat; josep.domingo@urv.cat; david.sanchez@urv.cat
    Data d'alta del registre: 2026-02-09
    Versió de l'article dipositat: info:eu-repo/semantics/publishedVersion
    Enllaç font original: https://www.sciencedirect.com/science/article/pii/S0893608025007592?via%3Dihub
    Referència a l'article segons font original: Neural Networks. 192 107879-
    Referència de l'ítem segons les normes APA: Al Mahmud, T; Jebreel, N; Domingo-Ferrer, J; Sánchez, D (2025). DP2Unlearning: An efficient and guaranteed unlearning framework for LLMs. Neural Networks, 192(), 107879-. DOI: 10.1016/j.neunet.2025.107879
    URL Document de llicència: https://repositori.urv.cat/ca/proteccio-de-dades/
    DOI de l'article: 10.1016/j.neunet.2025.107879
    Entitat: Universitat Rovira i Virgili
    Any de publicació de la revista: 2025-12-01
    Tipus de publicació: Journal Publications
  • Paraules clau:

    Artificial Intelligence,Cognitive Neuroscience,Computer Science, Artificial Intelligence,Neurosciences
    Approximate unlearning
    Differential privacy
    Divergence
    Exact unlearning
    Llm unlearning
    Mode
    Privacy-preserving ll
    Privacy-preserving llm
    Artificial intelligence
    Astronomia / física
    Biotecnología
    Ciência da computação
    Ciências agrárias i
    Ciencias sociales
    Cognitive neuroscience
    Computer science, artificial intelligence
    Economia
    Engenharias iii
    Engenharias iv
    Filosofia/teologia:subcomissão filosofia
    General medicine
    Interdisciplinar
    Matemática / probabilidade e estatística
    Neurosciences
    Psicología
    Psychology
    Química
  • Documents:

  • Cerca a google

    Search to google scholar