a-F1 commited on
Commit
b29fb81
·
verified ·
1 Parent(s): 2516497

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -3
README.md CHANGED
@@ -1,3 +1,47 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ ---
4
+
5
+ # LLaMA-2-chat 7B unlearned using SimNPO on TOFU Forget10
6
+
7
+ ## Model Details
8
+
9
+ - **Base Model**: LLaMA-2-chat 7B
10
+ - **Training**: Fine-tuned on TOFU dataset
11
+ - **Unlearning**: SimNPO on TOFU Forget10
12
+
13
+ ## Unlearning Algorithm
14
+
15
+ This model uses the `SimNPO` unlearning algorithm with the following parameters:
16
+ - Learning Rate: `1e-5`
17
+ - beta: `4.5`
18
+ - lambda: `0.125`
19
+ - gamma: `0.0`
20
+
21
+ ## Loading the Model
22
+
23
+ ```python
24
+ import torch
25
+ from transformers import AutoModelForCausalLM, AutoTokenizer
26
+
27
+ model = AutoModelForCausalLM.from_pretrained("OPTML-Group/SimNPO-TOFU-forget10-Llama-2-7b-chat", use_flash_attention_2=True, torch_dtype=torch.bfloat16, trust_remote_code=True)
28
+ ```
29
+
30
+ ## Citation
31
+
32
+ If you use this model in your research, please cite:
33
+ ```
34
+ @misc{fan2024simplicityprevailsrethinkingnegative,
35
+ title={Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning},
36
+ author={Chongyu Fan and Jiancheng Liu and Licong Lin and Jinghan Jia and Ruiqi Zhang and Song Mei and Sijia Liu},
37
+ year={2024},
38
+ eprint={2410.07163},
39
+ archivePrefix={arXiv},
40
+ primaryClass={cs.CL},
41
+ url={https://arxiv.org/abs/2410.07163},
42
+ }
43
+ ```
44
+
45
+ ## Contact
46
+
47
+ For questions or issues regarding this model, please contact [email protected].