a-F1 commited on
Commit
9d1c75d
·
verified ·
1 Parent(s): 715f5bd

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -0
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - cais/wmdp
5
+ language:
6
+ - en
7
+ base_model:
8
+ - HuggingFaceH4/zephyr-7b-beta
9
+ pipeline_tag: text-generation
10
+ library_name: transformers
11
+ tags:
12
+ - unlearn
13
+ - machine-unlearning
14
+ - llm-unlearning
15
+ - data-privacy
16
+ - large-language-models
17
+ - trustworthy-ai
18
+ - trustworthy-machine-learning
19
+ - language-model
20
+ ---
21
+
22
+ # NPO-Unlearned w/ GP Model on Task "WMDP"
23
+
24
+ ## Model Details
25
+
26
+ - **Unlearning**:
27
+ - **Task**: [🤗datasets/cais/wmdp wmdp-bio](https://huggingface.co/datasets/cais/wmdp)
28
+ - **Method**: NPO
29
+ - **Smoothness Optimization**: Gradient Penalty (GP)
30
+ - **Origin Model**: [🤗HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta)
31
+ - **Code Base**: [github.com/OPTML-Group/Unlearn-Smooth](https://github.com/OPTML-Group/Unlearn-Smooth)
32
+ - **Research Paper**: ["Towards LLM Unlearning Resilient to Relearning Attacks: A Sharpness-Aware Minimization Perspective and Beyond"](https://arxiv.org/abs/2502.05374)
33
+
34
+ ## Loading the Model
35
+
36
+ ```python
37
+ import torch
38
+ from transformers import AutoModelForCausalLM
39
+
40
+ model = AutoModelForCausalLM.from_pretrained("OPTML-Group/NPO-GP-WMDP", torch_dtype=torch.bfloat16, trust_remote_code=True)
41
+ ```
42
+
43
+ ## Citation
44
+
45
+ If you use this model in your research, please cite:
46
+ ```
47
+ @article{fan2025towards,
48
+ title={Towards LLM Unlearning Resilient to Relearning Attacks: A Sharpness-Aware Minimization Perspective and Beyond},
49
+ author={Fan, Chongyu and Jia, Jinghan and Zhang, Yihua and Ramakrishna, Anil and Hong, Mingyi and Liu, Sijia},
50
+ journal={arXiv preprint arXiv:2502.05374},
51
+ year={2025}
52
+ }
53
+ ```
54
+
55
+ ## Reporting Issues
56
+
57
+ Reporting issues with the model: [github.com/OPTML-Group/Unlearn-Smooth](https://github.com/OPTML-Group/Unlearn-Smooth)