system HF staff commited on
Commit
4d63426
·
1 Parent(s): c0c0f3f

Update files from the datasets library (from 1.4.0)

Browse files

Release notes: https://github.com/huggingface/datasets/releases/tag/1.4.0

Files changed (1) hide show
  1. README.md +19 -19
README.md CHANGED
@@ -48,7 +48,7 @@ task_ids:
48
  - [Citation Information](#citation-information)
49
  - [Contributions](#contributions)
50
 
51
- ## [Dataset Description](#dataset-description)
52
 
53
  - **Homepage:** [https://homes.cs.washington.edu/~msap/social-bias-frames/](https://homes.cs.washington.edu/~msap/social-bias-frames/)
54
  - **Repository:** [https://homes.cs.washington.edu/~msap/social-bias-frames/](https://homes.cs.washington.edu/~msap/social-bias-frames/)
@@ -59,27 +59,27 @@ task_ids:
59
  - **Size of the generated dataset:** 42.41 MB
60
  - **Total amount of disk used:** 48.45 MB
61
 
62
- ### [Dataset Summary](#dataset-summary)
63
 
64
  Warning: this document and dataset contain content that may be offensive or upsetting.
65
 
66
  Social Bias Frames is a new way of representing the biases and offensiveness that are implied in language. For example, these frames are meant to distill the implication that "women (candidates) are less qualified" behind the statement "we shouldn’t lower our standards to hire more women." The Social Bias Inference Corpus (SBIC) supports large-scale learning and evaluation of social implications with over 150k structured annotations of social media posts, spanning over 34k implications about a thousand demographic groups.
67
 
68
- ### [Supported Tasks](#supported-tasks)
69
 
70
  This dataset supports both classification and generation. Sap et al. developed several models using the SBIC. They report an F1 score of 78.8 in predicting whether the posts in the test set were offensive, an F1 score of 78.6 in predicting whether the posts were intending to be offensive, an F1 score of 80.7 in predicting whether the posts were lewd, and an F1 score of 69.9 in predicting whether the posts were targeting a specific group.
71
 
72
  Another of Sap et al.’s models performed better in the generation task. They report a BLUE score of 77.9, a Rouge-L score of 68.7, and a WMD score of 0.74 in generating a description of the targeted group given a post as well as a BLUE score of 52.6, a Rouge-L score of 44.9, and a WMD score of 2.79 in generating a description of the implied offensive statement given a post. See the paper for further details.
73
 
74
- ### [Languages](#languages)
75
 
76
  The language in SBIC is predominantly white-aligned English (78%, using a lexical dialect detector, [Blodgett et al., 2016](https://www.aclweb.org/anthology/D16-1120)). The curators find less than 10% of posts in SBIC are detected to have the AAE dialect category. The BCP-47 language tag is, presumably, en-US.
77
 
78
- ## [Dataset Structure](#dataset-structure)
79
 
80
  We show detailed information for up to 5 configurations of the dataset.
81
 
82
- ### [Data Instances](#data-instances)
83
 
84
  Each instance contains a post that may contain an offensive statement and annotated information concerning the nature of the offensive implication as well as the demographics of the annotator and origin of the post. See the [Social Bias Frames dataset viewer](https://huggingface.co/datasets/viewer/?dataset=social_bias_frames) to explore more examples.
85
 
@@ -114,7 +114,7 @@ An example of 'validation' looks as follows.
114
  }
115
  ```
116
 
117
- ### [Data Fields](#data-fields)
118
 
119
  The data fields are the same among all splits.
120
 
@@ -140,7 +140,7 @@ The data fields are the same among all splits.
140
  - _dataSource_: a string indicating the source of the post (`t/...`: means Twitter, `r/...`: means a subreddit)
141
 
142
 
143
- ### [Data Splits Sample Size](#data-splits-sample-size)
144
 
145
  To ensure that no post appeared in multiple splits, the curators defined a training instance as the post and its three sets of annotations. They then split the dataset into train, validation, and test sets (75%/12.5%/12.5%).
146
 
@@ -148,13 +148,13 @@ To ensure that no post appeared in multiple splits, the curators defined a train
148
  |-------|-----:|---------:|----:|
149
  |default|112900| 16738|17501|
150
 
151
- ## [Dataset Creation](#dataset-creation)
152
 
153
- ### [Curation Rationale](#curation-rationale)
154
 
155
  The main aim for this dataset is to cover a wide variety of social biases that are implied in text, both subtle and overt, and make the biases representative of real world discrimination that people experience [RWJF 2017](https://web.archive.org/web/20200620105955/https://www.rwjf.org/en/library/research/2017/10/discrimination-in-america--experiences-and-views.html). The curators also included some innocuous statements, to balance out biases, offensive, or harmful content.
156
 
157
- ### [Source Data](#source-data)
158
 
159
  The curators included online posts from the following sources sometime between 2014-2019:
160
  - r/darkJokes, r/meanJokes, r/offensiveJokes
@@ -170,7 +170,7 @@ The curators wanted posts to be as self-contained as possible, therefore, they a
170
 
171
  Due to the nature of this corpus, there is no way to know who the speakers are. But, the speakers of the Reddit, Gab, and Stormfront posts are likely white men (see [Gender by subreddit](http://bburky.com/subredditgenderratios/), [Gab users](https://en.wikipedia.org/wiki/Gab_(social_network)#cite_note-insidetheright-22), [Stormfront description](https://en.wikipedia.org/wiki/Stormfront_(website))).
172
 
173
- ### [Annotations](#annotations)
174
 
175
  #### Annotation process
176
 
@@ -186,13 +186,13 @@ The annotators are Amazon Mechanical Turk workers aged 36±10 years old. The ann
186
 
187
  Usernames are not included with the data, but the site where the post was collected is, so the user could potentially be recovered.
188
 
189
- ## [Considerations for Using the Data](#considerations-for-using-the-data)
190
 
191
- ### [Social Impact of Dataset](#social-impact-of-dataset)
192
 
193
  The curators recognize that studying Social Bias Frames necessarily requires confronting online content that may be offensive or disturbing but argue that deliberate avoidance does not eliminate such problems. By assessing social media content through the lens of Social Bias Frames, automatic flagging or AI-augmented writing interfaces may be analyzed for potentially harmful online content with detailed explanations for users or moderators to consider and verify. In addition, the collective analysis over large corpora can also be insightful for educating people on reducing unconscious biases in their language by encouraging empathy towards a targeted group.
194
 
195
- ### [Discussion of Biases](#discussion-of-biases)
196
 
197
  Because this is a corpus of social biases, a lot of posts contain implied or overt biases against the following groups (in decreasing order of prevalence):
198
  - gender/sexuality
@@ -204,13 +204,13 @@ Because this is a corpus of social biases, a lot of posts contain implied or ove
204
 
205
  The curators warn that technology trained on this dataset could have side effects such as censorship and dialect-based racial bias.
206
 
207
- ### [Other Known Limitations](#other-known-limitations)
208
 
209
  Because the curators found that the dataset is predominantly written in White-aligned English, they caution researchers to consider the potential for dialect or identity-based biases in labelling ([Davidson et al.,2019](https://www.aclweb.org/anthology/W19-3504.pdf); [Sap et al., 2019a](https://www.aclweb.org/anthology/P19-1163.pdf)) before deploying technology based on SBIC.
210
 
211
- ## [Additional Information](#additional-information)
212
 
213
- ### [Dataset Curators](#dataset-curators)
214
 
215
  This dataset was developed by Maarten Sap of the Paul G. Allen School of Computer Science & Engineering at the University of Washington, Saadia Gabriel, Lianhui Qin, Noah A Smith, and Yejin Choi of the Paul G. Allen School of Computer Science & Engineering and the Allen Institute for Artificial Intelligence, and Dan Jurafsky of the Linguistics & Computer Science Departments of Stanford University.
216
 
@@ -218,7 +218,7 @@ This dataset was developed by Maarten Sap of the Paul G. Allen School of Compute
218
 
219
  The SBIC is licensed under the [Creative Commons 4.0 License](https://creativecommons.org/licenses/by/4.0/)
220
 
221
- ### [Citation Information](#citation-information)
222
 
223
  ```
224
  @inproceedings{sap-etal-2020-social,
 
48
  - [Citation Information](#citation-information)
49
  - [Contributions](#contributions)
50
 
51
+ ## Dataset Description
52
 
53
  - **Homepage:** [https://homes.cs.washington.edu/~msap/social-bias-frames/](https://homes.cs.washington.edu/~msap/social-bias-frames/)
54
  - **Repository:** [https://homes.cs.washington.edu/~msap/social-bias-frames/](https://homes.cs.washington.edu/~msap/social-bias-frames/)
 
59
  - **Size of the generated dataset:** 42.41 MB
60
  - **Total amount of disk used:** 48.45 MB
61
 
62
+ ### Dataset Summary
63
 
64
  Warning: this document and dataset contain content that may be offensive or upsetting.
65
 
66
  Social Bias Frames is a new way of representing the biases and offensiveness that are implied in language. For example, these frames are meant to distill the implication that "women (candidates) are less qualified" behind the statement "we shouldn’t lower our standards to hire more women." The Social Bias Inference Corpus (SBIC) supports large-scale learning and evaluation of social implications with over 150k structured annotations of social media posts, spanning over 34k implications about a thousand demographic groups.
67
 
68
+ ### Supported Tasks
69
 
70
  This dataset supports both classification and generation. Sap et al. developed several models using the SBIC. They report an F1 score of 78.8 in predicting whether the posts in the test set were offensive, an F1 score of 78.6 in predicting whether the posts were intending to be offensive, an F1 score of 80.7 in predicting whether the posts were lewd, and an F1 score of 69.9 in predicting whether the posts were targeting a specific group.
71
 
72
  Another of Sap et al.’s models performed better in the generation task. They report a BLUE score of 77.9, a Rouge-L score of 68.7, and a WMD score of 0.74 in generating a description of the targeted group given a post as well as a BLUE score of 52.6, a Rouge-L score of 44.9, and a WMD score of 2.79 in generating a description of the implied offensive statement given a post. See the paper for further details.
73
 
74
+ ### Languages
75
 
76
  The language in SBIC is predominantly white-aligned English (78%, using a lexical dialect detector, [Blodgett et al., 2016](https://www.aclweb.org/anthology/D16-1120)). The curators find less than 10% of posts in SBIC are detected to have the AAE dialect category. The BCP-47 language tag is, presumably, en-US.
77
 
78
+ ## Dataset Structure
79
 
80
  We show detailed information for up to 5 configurations of the dataset.
81
 
82
+ ### Data Instances
83
 
84
  Each instance contains a post that may contain an offensive statement and annotated information concerning the nature of the offensive implication as well as the demographics of the annotator and origin of the post. See the [Social Bias Frames dataset viewer](https://huggingface.co/datasets/viewer/?dataset=social_bias_frames) to explore more examples.
85
 
 
114
  }
115
  ```
116
 
117
+ ### Data Fields
118
 
119
  The data fields are the same among all splits.
120
 
 
140
  - _dataSource_: a string indicating the source of the post (`t/...`: means Twitter, `r/...`: means a subreddit)
141
 
142
 
143
+ ### Data Splits Sample Size
144
 
145
  To ensure that no post appeared in multiple splits, the curators defined a training instance as the post and its three sets of annotations. They then split the dataset into train, validation, and test sets (75%/12.5%/12.5%).
146
 
 
148
  |-------|-----:|---------:|----:|
149
  |default|112900| 16738|17501|
150
 
151
+ ## Dataset Creation
152
 
153
+ ### Curation Rationale
154
 
155
  The main aim for this dataset is to cover a wide variety of social biases that are implied in text, both subtle and overt, and make the biases representative of real world discrimination that people experience [RWJF 2017](https://web.archive.org/web/20200620105955/https://www.rwjf.org/en/library/research/2017/10/discrimination-in-america--experiences-and-views.html). The curators also included some innocuous statements, to balance out biases, offensive, or harmful content.
156
 
157
+ ### Source Data
158
 
159
  The curators included online posts from the following sources sometime between 2014-2019:
160
  - r/darkJokes, r/meanJokes, r/offensiveJokes
 
170
 
171
  Due to the nature of this corpus, there is no way to know who the speakers are. But, the speakers of the Reddit, Gab, and Stormfront posts are likely white men (see [Gender by subreddit](http://bburky.com/subredditgenderratios/), [Gab users](https://en.wikipedia.org/wiki/Gab_(social_network)#cite_note-insidetheright-22), [Stormfront description](https://en.wikipedia.org/wiki/Stormfront_(website))).
172
 
173
+ ### Annotations
174
 
175
  #### Annotation process
176
 
 
186
 
187
  Usernames are not included with the data, but the site where the post was collected is, so the user could potentially be recovered.
188
 
189
+ ## Considerations for Using the Data
190
 
191
+ ### Social Impact of Dataset
192
 
193
  The curators recognize that studying Social Bias Frames necessarily requires confronting online content that may be offensive or disturbing but argue that deliberate avoidance does not eliminate such problems. By assessing social media content through the lens of Social Bias Frames, automatic flagging or AI-augmented writing interfaces may be analyzed for potentially harmful online content with detailed explanations for users or moderators to consider and verify. In addition, the collective analysis over large corpora can also be insightful for educating people on reducing unconscious biases in their language by encouraging empathy towards a targeted group.
194
 
195
+ ### Discussion of Biases
196
 
197
  Because this is a corpus of social biases, a lot of posts contain implied or overt biases against the following groups (in decreasing order of prevalence):
198
  - gender/sexuality
 
204
 
205
  The curators warn that technology trained on this dataset could have side effects such as censorship and dialect-based racial bias.
206
 
207
+ ### Other Known Limitations
208
 
209
  Because the curators found that the dataset is predominantly written in White-aligned English, they caution researchers to consider the potential for dialect or identity-based biases in labelling ([Davidson et al.,2019](https://www.aclweb.org/anthology/W19-3504.pdf); [Sap et al., 2019a](https://www.aclweb.org/anthology/P19-1163.pdf)) before deploying technology based on SBIC.
210
 
211
+ ## Additional Information
212
 
213
+ ### Dataset Curators
214
 
215
  This dataset was developed by Maarten Sap of the Paul G. Allen School of Computer Science & Engineering at the University of Washington, Saadia Gabriel, Lianhui Qin, Noah A Smith, and Yejin Choi of the Paul G. Allen School of Computer Science & Engineering and the Allen Institute for Artificial Intelligence, and Dan Jurafsky of the Linguistics & Computer Science Departments of Stanford University.
216
 
 
218
 
219
  The SBIC is licensed under the [Creative Commons 4.0 License](https://creativecommons.org/licenses/by/4.0/)
220
 
221
+ ### Citation Information
222
 
223
  ```
224
  @inproceedings{sap-etal-2020-social,