a-F1 commited on
Commit
c222dcb
·
verified ·
1 Parent(s): b29fb81

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -15
README.md CHANGED
@@ -1,18 +1,40 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
4
 
5
- # LLaMA-2-chat 7B unlearned using SimNPO on TOFU Forget10
6
 
7
  ## Model Details
8
 
9
- - **Base Model**: LLaMA-2-chat 7B
10
- - **Training**: Fine-tuned on TOFU dataset
11
- - **Unlearning**: SimNPO on TOFU Forget10
 
 
 
12
 
13
  ## Unlearning Algorithm
14
 
15
- This model uses the `SimNPO` unlearning algorithm with the following parameters:
 
 
16
  - Learning Rate: `1e-5`
17
  - beta: `4.5`
18
  - lambda: `0.125`
@@ -27,21 +49,26 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
27
  model = AutoModelForCausalLM.from_pretrained("OPTML-Group/SimNPO-TOFU-forget10-Llama-2-7b-chat", use_flash_attention_2=True, torch_dtype=torch.bfloat16, trust_remote_code=True)
28
  ```
29
 
 
 
 
 
 
 
 
 
30
  ## Citation
31
 
32
  If you use this model in your research, please cite:
33
  ```
34
- @misc{fan2024simplicityprevailsrethinkingnegative,
35
- title={Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning},
36
- author={Chongyu Fan and Jiancheng Liu and Licong Lin and Jinghan Jia and Ruiqi Zhang and Song Mei and Sijia Liu},
37
- year={2024},
38
- eprint={2410.07163},
39
- archivePrefix={arXiv},
40
- primaryClass={cs.CL},
41
- url={https://arxiv.org/abs/2410.07163},
42
  }
43
  ```
44
 
45
- ## Contact
46
 
47
- For questions or issues regarding this model, please contact chongyu.fan93@gmail.com.
 
1
  ---
2
  license: mit
3
+ datasets:
4
+ - locuslab/TOFU
5
+ language:
6
+ - en
7
+ base_model:
8
+ - OPTML-Group/TOFU-origin-Llama-2-7b-chat
9
+ pipeline_tag: text-generation
10
+ library_name: transformers
11
+ tags:
12
+ - unlearn
13
+ - machine-unlearning
14
+ - llm-unlearning
15
+ - data-privacy
16
+ - large-language-models
17
+ - trustworthy-ai
18
+ - trustworthy-machine-learning
19
+ - language-model
20
  ---
21
 
22
+ # SimNPO-Unlearned Model on Task "TOFU - Forget05"
23
 
24
  ## Model Details
25
 
26
+ - **Unlearning**:
27
+ - **Task**: [🤗datasets/locuslab/TOFU - Forget10](https://huggingface.co/datasets/locuslab/TOFU)
28
+ - **Method**: [SimNPO](https://arxiv.org/abs/2410.07163)
29
+ - **Origin Model**: [🤗OPTML-Group/TOFU-origin-Llama-2-7b-chat](https://huggingface.co/OPTML-Group/TOFU-origin-Llama-2-7b-chat)
30
+ - **Code Base**: [github.com/OPTML-Group/Unlearn-Simple](https://github.com/OPTML-Group/Unlearn-Simple)
31
+ - **Research Paper**: ["Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning"](https://arxiv.org/abs/2410.07163)
32
 
33
  ## Unlearning Algorithm
34
 
35
+ This model uses the `SimNPO` unlearning algorithm with the following optimization objective:
36
+ $$\ell_{SimNPO}(\mathbf{\theta}) = \mathbb{E}_{(x, y) \in \mathcal{D}_f}\left[-\frac{2}{\beta}\log\sigma\left(-\frac{\beta}{|y|}\log\pi_{\mathbf{\theta}}(y|x) - \gamma\right)\right] + \lambda \mathbb{E}_{(x, y) \in \mathcal{D}_r}[-\log\pi_{\mathbf{\theta}} (y|x)]$$
37
+ Unlearning hyper-parameters:
38
  - Learning Rate: `1e-5`
39
  - beta: `4.5`
40
  - lambda: `0.125`
 
49
  model = AutoModelForCausalLM.from_pretrained("OPTML-Group/SimNPO-TOFU-forget10-Llama-2-7b-chat", use_flash_attention_2=True, torch_dtype=torch.bfloat16, trust_remote_code=True)
50
  ```
51
 
52
+ ## Evaluation Results
53
+ ||Forgeting Quality (FQ)|Model Utility (MU)|
54
+ |---|---|---|
55
+ |Origin|0.00|0.62|
56
+ |Retrain|1.00|0.62|
57
+ |NPO|0.29|0.55|
58
+ |**SimNPO**|0.45|0.62|
59
+
60
  ## Citation
61
 
62
  If you use this model in your research, please cite:
63
  ```
64
+ @article{fan2024simplicity,
65
+ title={Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning},
66
+ author={Fan, Chongyu and Liu, Jiancheng and Lin, Licong and Jia, Jinghan and Zhang, Ruiqi and Mei, Song and Liu, Sijia},
67
+ journal={arXiv preprint arXiv:2410.07163},
68
+ year={2024}
 
 
 
69
  }
70
  ```
71
 
72
+ ## Reporting Issues
73
 
74
+ Reporting issues with the model: [github.com/OPTML-Group/Unlearn-Simple](https://github.com/OPTML-Group/Unlearn-Simple)