rebeccaqian commited on
Commit
c2f9b04
1 Parent(s): a17dee7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +79 -1
README.md CHANGED
@@ -1,3 +1,81 @@
1
  ---
2
- license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ license:
5
+ - mit
6
+ library_name: "pytorch"
7
+ multilinguality:
8
+ - monolingual
9
+ pretty_name: FairBERTa
10
+ tags:
11
+ - counterfactual
12
+ - perturb
13
+ - fairness
14
+ - nlp
15
+ - demographic
16
+ - diverse
17
+ - gender
18
+ - non-binary
19
+ - race
20
+ - age
21
+ metrics:
22
+ - bleu
23
  ---
24
+
25
+ # FairBERTa
26
+ FairBERTa is the first large language model trained on demographically perturbed corpora. Compared to RoBERTa, FairBERTa's fairness is improved, while its performance does not degrade on downstream tasks.
27
+
28
+ - **Repository:** https://github.com/facebookresearch/ResponsibleNLP/
29
+ - **Paper:** https://aclanthology.org/2022.emnlp-main.646/
30
31
+ - **License:** MIT
32
+
33
+ ## Model Description
34
+ FairBERTa is a transformers model pretrained on a large corpus of English data with the Masked language modeling (MLM) objective. The model randomly masks 15% of words in the input sequence then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence.
35
+
36
+ The model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks.
37
+
38
+ ### Model Summary
39
+
40
+ FairBERTa can be finetuned on a variety of downstream tasks.
41
+
42
+ FairBERTa is trained using the FairSeq library, with the same parameters as the RoBERTa-base model.
43
+
44
+ ### Uses
45
+
46
+ The perturber is intended for use by fairness researchers and engineers working on demographic debiasing applications. The perturber is a controllable generation model that given a word, target demographic attribute and input text, outputs text where the selected word and associated references are rewritten to the target demographic attribute. Control variables and the input text are separated by a <PERT_SEP> token.
47
+
48
+ ## Training
49
+ The FairBERTa model was pretrained in a similar manner to [RoBERTa](https://huggingface.co/roberta-base). It was pretrained on 160GB of perturbed datasets. The training data consists of five sources: BookCorpus, English Wikipedia, CC-News, OpenWebText, Stories.
50
+
51
+ We sample chunks of 512 token sequences (the perturber's max context window), select a
52
+
53
+ ## Bias, Risks & Limitations
54
+ FairBERTa shows improved performance compared to RoBERTa on a variety of fairness metrics.
55
+
56
+ For an in-depth discussion of bias, risks and limitations, see the Results and Limitations sections of [our paper](https://aclanthology.org/2022.emnlp-main.646/).
57
+
58
+ ## Citation
59
+ ```
60
+ @inproceedings{qian-etal-2022-perturbation,
61
+ title = "Perturbation Augmentation for Fairer {NLP}",
62
+ author = "Qian, Rebecca and
63
+ Ross, Candace and
64
+ Fernandes, Jude and
65
+ Smith, Eric Michael and
66
+ Kiela, Douwe and
67
+ Williams, Adina",
68
+ booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",
69
+ month = dec,
70
+ year = "2022",
71
+ address = "Abu Dhabi, United Arab Emirates",
72
+ publisher = "Association for Computational Linguistics",
73
+ url = "https://aclanthology.org/2022.emnlp-main.646",
74
+ pages = "9496--9521",
75
+ abstract = "Unwanted and often harmful social biases are becoming ever more salient in NLP research, affecting both models and datasets. In this work, we ask whether training on demographically perturbed data leads to fairer language models. We collect a large dataset of human annotated text perturbations and train a neural perturbation model, which we show outperforms heuristic alternatives. We find that (i) language models (LMs) pre-trained on demographically perturbed corpora are typically more fair, and (ii) LMs finetuned on perturbed GLUE datasets exhibit less demographic bias on downstream tasks, and (iii) fairness improvements do not come at the expense of performance on downstream tasks. Lastly, we discuss outstanding questions about how best to evaluate the (un)fairness of large language models. We hope that this exploration of neural demographic perturbation will help drive more improvement towards fairer NLP.",
76
+ }
77
+ ```
78
+
79
+ ### Model Card Contact
80
+
81
+ Thanks to [@Rebecca-Qian](https://github.com/Rebecca-Qian) for adding this model.