| Tian et al. |
DeRDaVa: Deletion-Robust Data Valuation for Machine Learning |
AAAI |
| Ni et al. |
ORES: open-vocabulary responsible visual synthesis |
AAAI |
| Moon et al. |
Feature Unlearning for Pre-trained GANs and VAEs |
AAAI |
| Rashid et al. |
Forget to Flourish: Leveraging Machine-Unlearning on Pretrained Language Models for Privacy Leakage |
AAAI |
| Cha et al. |
Learning to Unlearn: Instance-wise Unlearning for Pre-trained Classifiers |
AAAI |
| Hong et al. |
All but One: Surgical Concept Erasing with Model Preservation in Text-to-Image Diffusion Models |
AAAI |
| Kim et al. |
Layer Attack Unlearning: Fast and Accurate Machine Unlearning via Layer Level Attack and Knowledge Distillation |
AAAI |
| Foster et al. |
Fast Machine Unlearning Without Retraining Through Selective Synaptic Dampening |
AAAI |
| Hu et al. |
Separate the Wheat from the Chaff: Model Deficiency Unlearning via Parameter-Efficient Module Operation |
AAAI |
| Li et al. |
Towards Effective and General Graph Unlearning via Mutual Evolution |
AAAI |
| Liu et al. |
Backdoor Attacks via Machine Unlearning |
AAAI |
| You et al. |
RRL: Recommendation Reverse Learning |
AAAI |
| Moon et al. |
Feature Unlearning for Generative Models via Implicit Feedback |
AAAI |
| Li et al. |
SafeGen: Mitigating Sexually Explicit Content Generation in Text-to-Image Models |
ACM CCS |
| Lin et al. |
GDR-GMA: Machine Unlearning via Direction-Rectified and Magnitude-Adjusted Gradients |
ACM MM |
| Huang et al. |
Your Code Secret Belongs to Me: Neural Code Completion Tools Can Memorize Hard-Coded Credentials |
ACM SE |
| Feng et al. |
Fine-grained Pluggable Gradient Ascent for Knowledge Unlearning in Language Models |
ACL |
| Arad et al. |
ReFACT: Updating Text-to-Image Models by Editing the Text Encoder |
ACL |
| Wu et al. |
Universal Prompt Optimizer for Safe Text-to-Image Generation |
ACL |
| Liu et al. |
Towards Safer Large Language Models through Machine Unlearning |
ACL |
| Kim et al. |
Towards Robust and Generalized Parameter-Efficient Fine-Tuning for Noisy Label Learning |
ACL |
| Lee et al. |
Protecting Privacy Through Approximating Optimal Parameters for Sequence Unlearning in Language Models |
ACL |
| Choi et al. |
Cross-Lingual Unlearning of Selective Knowledge in Multilingual Language Models |
ACL |
| Isonuma et al. |
Unlearning Traces the Influential Training Data of Language Models |
ACL |
| Zhou et al. |
Visual In-Context Learning for Large Vision-Language Models |
ACL |
| Xing et al. |
EFUF: Efficient Fine-Grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language Models |
ACL |
| Yao et al. |
Machine Unlearning of Pre-trained Large Language Models |
ACL |
| Zhao et al. |
Deciphering the Impact of Pretraining Data on Large Language Models through Machine Unlearning |
ACL |
| Ni et al. |
Forgetting before Learning: Utilizing Parametric Arithmetic for Knowledge Updating in Large Language Models |
ACL |
| Zhou et al. |
Making Harmful Behaviors Unlearnable for Large Language Models |
ACL |
| Yamashita et al. |
One-Shot Machine Unlearning with Mnemonic Code |
ACML |
| Fraboni et al. |
SIFU: Sequential Informed Federated Unlearning for Efficient and Provable Client Unlearning in Federated Optimization |
AISTATS |
| Alshehri and Zhang |
Forgetting User Preference in Recommendation Systems with Label-Flipping |
BigData |
| Qiu et al. |
FedCIO: Efficient Exact Federated Unlearning with Clustering, Isolation, and One-shot Aggregation |
BigData |
| Yang and Li |
When Contrastive Learning Meets Graph Unlearning: Graph Contrastive Unlearning for Link Prediction |
BigData |
| Hu et al. |
ERASER: Machine Unlearning in MLaaS via an Inference Serving-Aware Approach |
CCS |
| Zhang et al. |
Negative Preference Optimization: From Catastrophic Collapse to Effective Unlearning |
COLM |
| Maini et al. |
TOFU: A Task of Fictitious Unlearning for LLMs |
COLM |
| Abbasi et al. |
Brainwash: A Poisoning Attack to Forget in Continual Learning |
CVPR |
| Chen et al. |
Towards Memorization-Free Diffusion Models |
CVPR |
| Lyu et al. |
One-Dimensional Adapter to Rule Them All: Concepts, Diffusion Models and Erasing Applications |
CVPR |
| Wallace et al. |
Diffusion Model Alignment Using Direct Preference Optimization |
CVPR |
| Lu et al. |
MACE: Mass Concept Erasure in Diffusion Models |
CVPR |
| Chen et al. |
WPN: An Unlearning Method Based on N-pair Contrastive Learning in Language Models |
ECAI |
| Fan et al. |
Challenging Forgets: Unveiling the Worst-Case Forget Sets in Machine Unlearning |
ECCV |
| Gong et al. |
Reliable and Efficient Concept Erasure of Text-to-Image Diffusion Models |
ECCV |
| Kim et al. |
R.A.C.E. : Robust Adversarial Concept Erasure for Secure Text-to-Image Diffusion Model |
ECCV |
| Kim et al. |
Safeguard Text-to-Image Diffusion Models with Human Feedback Inversion |
ECCV |
| Wu et al. |
Scissorhands: Scrub Data Influence via Connection Sensitivity in Networks |
ECCV |
| Zhang et al. |
To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Unsafe Images ... For Now |
ECCV |
| Liu et al. |
Implicit Concept Removal of Diffusion Models |
ECCV |
| Ban et al. |
Understanding the Impact of Negative Prompts: When and How Do They Take Effect? |
ECCV |
| Zhang et al. |
IMMA: Immunizing Text-to-Image Models Against Malicious Adaptation |
ECCV |
| Poppi et al. |
Removing NSFW Concepts from Vision-and-Language Models for Text-to-Image Retrieval and Generation |
ECCV |
| Liu et al. |
Latent Guard: A Safety Framework for Text-to-Image Generation |
ECCV |
| Huang et al. |
Receler: Reliable Concept Erasing of Text-to-Image Diffusion Models via Lightweight Erasers |
ECCV |
| Cheng et al. |
MultiDelete for Multimodal Machine Unlearning |
ECCV |
| Wang et al. |
How to Forget Clients in Federated Online Learning to Rank? |
ECIR |
| Jia et al. |
SOUL: Unlocking the Power of Second-Order Optimization for LLM Unlearning |
EMNLP |
| Joshi et al. |
Towards Robust Evaluation of Unlearning in LLMs via Data Transformations |
EMNLP |
| Tian et al. |
To Forget or Not? Towards Practical Knowledge Unlearning for Large Language Models |
arxiv |
| Chakraborty et al. |
Can Textual Unlearning Solve Cross-Modality Safety Alignment? |
EMNLP |
| Huang et al. |
Demystifying Verbatim Memorization in Large Language Models |
EMNLP |
| Liu et al. |
Revisiting Who's Harry Potter: Towards Targeted Unlearning from a Causal Intervention Perspective |
EMNLP |
| Chen et al. |
Unlearn What You Want to Forget: Efficient Unlearning for LLMs |
EMNLP |
| Liu et al. |
Forgetting Private Textual Sequences in Language Models Via Leave-One-Out Ensemble |
ICASSP |
| Liu et al. |
Learning to Refuse: Towards Mitigating Privacy Risks in LLMs |
ICCL |
| Fan et al. |
SalUn: Empowering Machine Unlearning via Gradient-based Weight Saliency in Both Image Classification and Generation |
ICLR |
| Liu et al. |
Tangent Transformers for Composition, Privacy and Removal |
ICLR |
| Li et al. |
Machine Unlearning for Image-to-Image Generative Models |
ICLR |
| Shen et al. |
Label-Agnostic Forgetting: A Supervision-Free Unlearning in Deep Models |
ICLR |
| Li et al. |
Get What You Want, Not What You Don't: Image Content Suppression for Text-to-Image Diffusion Models |
ICLR |
| Tsai et al. |
Ring-A-Bell! How Reliable are Concept Removal Methods for Diffusion Models? |
ICLR |
| Wang et al. |
A Unified and General Framework for Continual Learning |
ICLR |
| Shi et al. |
Detecting Pretraining Data from Large Language Models |
ICLR |
| Eldan et al. |
Who’s Harry Potter? Approximate Unlearning in LLMs |
ICLR |
| Wang et al. |
LLM Unlearning via Loss Adjustment with Only Forget Data |
ICLR |
| Chavhan et al. |
ConceptPrune: Concept Editing in Diffusion Models via Skilled Neuron Pruning |
ICLR |
| Zhao et al. |
Rethinking Adversarial Robustness in the Context of the Right to be Forgotten |
ICML |
| Pawelczyk et al. |
In-Context Unlearning: Language Models As Few Shot Unlearners |
ICML |
| Barbulescu et al. |
To each (textual sequence) its own: improving memorized-data unlearning in large language models |
ICML |
| Li et al. |
The WMDP benchmark: measuring and reducing malicious use with unlearning |
ICML |
| Das et al. |
Larimar: large language models with episodic memory control |
ICML |
| Barbulescu et al. |
To each (textual sequence) its own: improving memorized-data unlearning in large language models |
ICML |
| Zhao et al. |
Learning and forgetting unsafe examples in large language models |
ICML |
| Basu et al. |
On mechanistic knowledge localization in text-to-image generative models |
ICML |
| Zhang et al. |
SecureCut: Federated Gradient Boosting Decision Trees with Efficient Machine Unlearning |
ICPR |
| Cai et al. |
Where have you been? A Study of Privacy Risk for Point-of-Interest Recommendation |
KDD |
| Gong et al. |
A Population-to-individual Tuning Framework for Adapting Pretrained LM to On-device User Intent Prediction |
KDD |
| Xue et al. |
Erase to Enhance: Data-Efficient Machine Unlearning in MRI Reconstruction |
MIDL |
| Gao et al. |
Ethos: Rectifying Language Models in Orthogonal Parameter Space |
NAACL |
| Park et al. |
Direct Unlearning Optimization for Robust and Safe Text-to-Image Models |
NeurIPS |
| Ko et al. |
Boosting Alignment for Post-Unlearning Text-to-Image Generative Models |
NeurIPS |
| Yang et al. |
GuardT2I: Defending Text-to-Image Models from Adversarial Prompts |
NeurIPS |
| Li et al. |
Single Image Unlearning: Efficient Machine Unlearning in Multimodal Large Language Models |
NeurIPS |
| Jain et al. |
What Makes and Breaks Safety Fine-tuning? A Mechanistic Study |
NeurIPS |
| Wu et al. |
Cross-model Control: Improving Multiple Large Language Models in One-time Training |
NeurIPS |
| Bui et al. |
Erasing Undesirable Concepts in Diffusion Models with Adversarial Preservation |
NeurIPS |
| Zhao et al. |
What makes unlearning hard and what to do about it |
NeurIPS |
| Zhang et al. |
Defensive Unlearning with Adversarial Training for Robust Concept Erasure in Diffusion Models |
NeurIPS |
| Yao et al. |
Large Language Model Unlearning |
NeurIPS |
| Ji et al. |
Reversing the Forget-Retain Objectives: An Efficient LLM Unlearning Framework from Logit Difference |
NeurIPS |
| Liu et al. |
Large Language Model Unlearning via Embedding-Corrupted Prompts |
NeurIPS |
| Jia et al. |
WAGLE: Strategic Weight Attribution for Effective and Modular Unlearning in Large Language Models |
NeurIPS |
| Zhang et al. |
UnlearnCanvas: A Stylized Image Dataset to Benchmark Machine Unlearning for Diffusion Models |
NeurIPS D&B |
| Jin et al. |
RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models |
NeurIPS D&B |
| Kurmanji et al. |
Machine Unlearning in Learned Databases: An Experimental Analysis |
SIGMOD |
| Shen et al. |
CaMU: Disentangling Causal Effects in Deep Model Unlearning |
SDM |
| Yoon et al. |
Few-Shot Unlearning |
SP |
| Hu et al. |
Learn What You Want to Unlearn: Unlearning Inversion Attacks against Machine Unlearning |
SP |
| Hoang et al. |
Learn To Unlearn for Deep Neural Networks: Minimizing Unlearning Interference With Gradient Projection |
WACV |
| Gandikota et al. |
Unified Concept Editing in Diffusion Models |
WACV |
| Malnick et al. |
Taming Normalizing Flows |
WACV |
| Xin et al. |
On the Effectiveness of Unlearning in Session-Based Recommendation |
WSDM |
| Zhang |
Graph Unlearning with Efficient Partial Retraining |
WWW |
| Liu et al. |
Breaking the Trilemma of Privacy, Utility, Efficiency via Controllable Machine Unlearning |
WWW |
|
|
|
| Liu et al. |
A Survey on Federated Unlearning: Challenges, Methods, and Future Directions |
ACM Computing Surveys |
| Zhang et al. |
Right to be Forgotten in the Era of Large Language Models: Implications, Challenges, and Solutions |
AI and Ethics |
| Zha et al. |
To Be Forgotten or To Be Fair: Unveiling Fairness Implications of Machine Unlearning Methods |
AI and Ethics |
| Zhang et al. |
Recommendation Unlearning via Influence Function |
ACM Transactions on Recommender Systems |
| Schoepf et al. |
Potion: Towards Poison Unlearning |
DMLR |
| Wang et al. |
Towards efficient and effective unlearning of large language models for recommendation |
Frontiers of Computer Science |
| Poppi et al. |
Multi-Class Explainable Unlearning for Image Classification via Weight Filtering |
IEEE Intelligent Systems |
| Panda and AP |
FAST: Feature Aware Similarity Thresholding for Weak Unlearning in Black-Box Generative Models |
IEEE Transactions on Artificial Intelligence |
| Alam et al. |
Get Rid Of Your Trail: Remotely Erasing Backdoors in Federated Learning |
IEEE Transactions on Artificial Intelligence |
| Shaik et al. |
FRAMU: Attention-based Machine Unlearning using Federated Reinforcement Learning |
IEEE Transactions on Knowledge and Data Engineering |
| Shaik et al. |
Exploring the Landscape of Machine Unlearning: A Comprehensive Survey and Taxonomy |
IEEE Transactions on Neural Networks and Learning Systems |
| Romandini et al. |
Federated Unlearning: A Survey on Methods, Design Guidelines, and Evaluation Metrics |
IEEE Transactions on Neural Networks and Learning Systems |
| Xu and Teng |
Task-Aware Machine Unlearning and Its Application in Load Forecasting |
IEEE Transactions on Power Systems |
| Li et al. |
Pseudo Unlearning via Sample Swapping with Hash |
Information Science |
|
|
|
| Fore et al. |
Unlearning Climate Misinformation in Large Language Models |
ClimateNLP |
| Zhang et al. |
Forget-Me-Not: Learning to Forget in Text-to-Image Diffusion Models |
CVPR Workshop |
| Shi et al. |
DeepClean: Machine Unlearning on the Cheap by Resetting Privacy Sensitive Weights using the Fisher Diagonal |
ECCV Workshop |
| Sridhar et al. |
Prompt Sliders for Fine-Grained Control, Editing and Erasing of Concepts in Diffusion Models |
ECCV Workshop |
| Schoepf et al. |
Loss-Free Machine Unlearning |
ICLR Tiny Paper |
| Tamirisa et al. |
Toward Robust Unlearning for LLMs |
ICLR Workshop |
| Sun et al. |
Learning and Unlearning of Fabricated Knowledge in Language Models |
ICML Workshop |
| Wang et al. |
Alignment Calibration: Machine Unlearning for Contrastive Learning under Auditing |
ICML Workshop |
| Kadhe et al. |
Split, Unlearn, Merge: Leveraging Data Attributes for More Effective Unlearning in LLMs |
ICML Workshop |
| Zhao et al. |
Scalability of memorization-based machine unlearning |
NeurIPS Workshop |
| Wu et al. |
CodeUnlearn: Amortized Zero-Shot Machine Unlearning in Language Models Using Discrete Concept |
NeurIPS Workshop |
| Cheng et al. |
MU-Bench: A Multitask Multimodal Benchmark for Machine Unlearning |
NeurIPS Workshop |
| Seyitoğlu et al. |
Extracting Unlearned Information from LLMs with Activation Steering |
NeurIPS Workshop |
| Wei et al. |
Provable unlearning in topic modeling and downstream tasks |
NeurIPS Workshop |
| Lucki et al. |
An Adversarial Perspective on Machine Unlearning for AI Safety |
NeurIPS Workshop |
| Li et al. |
LLM Defenses Are Not Robust to Multi-Turn Human Jailbreaks Yet |
NeurIPS Workshop |
| Smirnov et al. |
Classifier-free guidance in LLMs Safety |
NeurIPS Workshop |
|
|
|
| Liu et al. |
Machine Unlearning in Generative AI: A Survey |
arxiv |
| Xu |
Machine Unlearning for Traditional Models and Large Language Models: A Short Survey |
arxiv |
| Lynch et al. |
Eight Methods to Evaluate Robust Unlearning in LLMs |
arxiv |
| Dontsov et al. |
CLEAR: Character Unlearning in Textual and Visual Modalities |
arXiv |
| Hong et al. |
Intrinsic Evaluation of Unlearning Using Parametric Knowledge Traces |
arXiv |
| Jung et al. |
Attack and Reset for Unlearning: Exploiting Adversarial Noise toward Machine Unlearning through Parameter Re-initialization |
arXiv |
| Pham et al. |
Robust Concept Erasure Using Task Vectors |
arXiv |
| Qian et al. |
Exploring Fairness in Educational Data Mining in the Context of the Right to be Forgotten |
arXiv |
| Schoepf et al. |
An Information Theoretic Approach to Machine Unlearning |
arxiv |
| Schoepf et al. |
Parameter-tuning-free data entry error unlearning with adaptive selective synaptic dampening |
arXiv |
| Zhao et al. |
Separable Multi-Concept Erasure from Diffusion Models |
arXiv |
| Dige et al. |
Mitigating Social Biases in Language Models through Unlearning |
arxiv |
| Hong et al. |
Intrinsic Evaluation of Unlearning Using Parametric Knowledge Traces |
arxiv |
| Wang et al. |
Towards Effective Evaluations and Comparisons for LLM Unlearning Methods |
arxiv |
| Ashuach et al. |
REVS: Unlearning Sensitive Information in Language Models via Rank Editing in the Vocabulary Space |
arxiv |
| Zuo et al. |
Federated TrustChain: Blockchain-Enhanced LLM Training and Unlearning |
arxiv |
| Wang et al. |
RKLD: Reverse KL-Divergence-based Knowledge Distillation for Unlearning Personal Information in Large Language Models |
arxiv |
| Chen e tal. |
Machine Unlearning in Large Language Models |
arxiv |
| Lu et al. |
Eraser: Jailbreaking Defense in Large Language Models via Unlearning Harmful Knowledge |
arxiv |
| Stoehr et al. |
Localizing Paragraph Memorization in Language Models |
arxiv |
| Pochinkov et al. |
Dissecting Language Models: Machine Unlearning via Selective Pruning |
arxiv |
| Gu et al. |
Second-Order Information Matters: Revisiting Machine Unlearning for Large Language Models |
arxiv |
| Thaker et al. |
Guardrail Baselines for Unlearning in LLMs |
arxiv |
| Wang et al. |
When Machine Unlearning Meets Retrieval-Augmented Generation (RAG): Keep Secret or Forget Knowledge? |
arxiv |
| Muresanu et al. |
Unlearnable Algorithms for In-context Learning |
arxiv |
| Zhao et al. |
Unlearning Backdoor Attacks for LLMs with Weak-to-Strong Knowledge Distillation |
arxiv |
| Choi et al. |
Breaking Chains: Unraveling the Links in Multi-Hop Knowledge Unlearning |
arxiv |
| Guo et al. |
Mechanistic Unlearning: Robust Knowledge Unlearning and Editing via Mechanistic Localization |
arxiv |
| Deeb et al. |
Do Unlearning Methods Remove Information from Language Model Weights? |
arxiv |
| Takashiro et al. |
Answer When Needed, Forget When Not: Language Models Pretend to Forget via In-Context Knowledge Unlearning |
arxiv |
| Veldanda et al. |
LLM Surgery: Efficient Knowledge Unlearning and Editing in Large Language Models |
arxiv |
| Gu et al. |
MEOW: MEMOry Supervised LLM Unlearning Via Inverted Facts |
arxiv |
| Zhang et al. |
Unforgettable Generalization in Language Models |
arxiv |
| Kazemi et al. |
Unlearning Trojans in Large Language Models: A Comparison Between Natural Language and Source Code |
arxiv |
| Huu-Tien et al. |
On Effects of Steering Latent Representation for Large Language Model Unlearning |
arxiv |
| Yang et al. |
Hotfixing Large Language Models for Code |
arxiv |
| Lizzo et al. |
UNLEARN Efficient Removal of Knowledge in Large Language Models |
arxiv |
| Tamirisa et al. |
Tamper-Resistant Safeguards for Open-Weight LLMs |
arxiv |
| Zhou et al. |
On the Limitations and Prospects of Machine Unlearning for Generative AI |
arxiv |
| Tang et al. |
Learn while Unlearn: An Iterative Unlearning Framework for Generative Language Models |
arxiv |
| Lu et al. |
Towards Transfer Unlearning: Empirical Evidence of Cross-Domain Bias Mitigation |
arxiv |
| Gao et al. |
On Large Language Model Continual Unlearning |
arxiv |
| Kolbeinsson et al. |
Composable Interventions for Language Models |
arxiv |
| Hernandez et al. |
If You Don't Understand It, Don't Use It: Eliminating Trojans with Filters Between Layers |
arxiv |
| Zhang et al. |
From Theft to Bomb-Making: The Ripple Effect of Unlearning in Defending Against Jailbreak Attacks |
arxiv |
| Scaria et al. |
Can Small Language Models Learn, Unlearn, and Retain Noise Patterns? |
arxiv |
| Shumailov et al. |
UnUnlearning: Unlearning is not sufficient for content regulation in advanced generative AI |
arxiv |
| Qiu et al. |
How Data Inter-connectivity Shapes LLMs Unlearning: A Structural Unlearning Perspective |
arxiv |
| Lu et al. |
Learn and Unlearn in Multilingual LLMs |
arxiv |
| Ma et al. |
Benchmarking Vision Language Model Unlearning via Fictitious Facial Identity Dataset |
arxiv |
| Rezaei et al. |
RESTOR: Knowledge Recovery in Machine Unlearning |
arxiv |
| Baluta et al. |
Unlearning in- vs. out-of-distribution data in LLMs under gradient-based method |
arxiv |
| Doshi et al. |
Does Unlearning Truly Unlearn? A Black Box Evaluation of LLM Unlearning Methods |
arxiv |
| Wei et al. |
Underestimated Privacy Risks for Minority Populations in Large Language Model Unlearning |
arxiv |
| Zuo et al. |
Large Language Model Federated Learning with Blockchain and Unlearning for Cross-Organizational Collaboration |
arxiv |
| Dou et al. |
Investigating the Feasibility of Mitigating Potential Copyright Infringement via Large Language Model Unlearning |
arxiv |
| Ren et al. |
Copyright Protection in Generative AI: A Technical Perspective, 2024 |
arxiv |
| Gu et al. |
Second-Order Information Matters: Revisiting Machine Unlearning for Large Language Models |
arxiv |
| Chakraborty et al. |
Cross-Modal Safety Alignment: Is textual unlearning all you need? |
arxiv |
| Liang et al. |
Unlearning Backdoor Threats: Enhancing Backdoor Defense in Multimodal Contrastive Learning via Local Token Unlearning |
arxiv |
| Wu et al. |
Erasing Undesirable Influence in Diffusion Models |
arxiv |
| Gao et al. |
Meta-Unlearning on Diffusion Models: Preventing Relearning Unlearned Concepts |
arxiv |
| Huang et al. |
Enhancing User-Centric Privacy Protection: An Interactive Framework through Diffusion Models and Machine Unlearning |
arxiv |
| Liu et al. |
Unlearning Concepts from Text-to-Video Diffusion Models |
arxiv |
| Gandikota et al. |
Erasing Conceptual Knowledge from Language Models |
arxiv |
| Tu et al. |
Towards Reliable Empirical Machine Unlearning Evaluation: A Cryptographic Game Perspective |
arxiv |
| Zhuang et al. |
UOE: Unlearning One Expert is Enough for Mixture-of-Experts LLMs |
arxiv |
|
|
|
| Liu |
Machine Unlearning in 2024 |
Blog Post |