Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
parquet-converter commited on
Commit
c345e53
·
1 Parent(s): 33b7078

Update parquet files

Browse files
.gitattributes DELETED
@@ -1,27 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bin.* filter=lfs diff=lfs merge=lfs -text
5
- *.bz2 filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.model filter=lfs diff=lfs merge=lfs -text
12
- *.msgpack filter=lfs diff=lfs merge=lfs -text
13
- *.onnx filter=lfs diff=lfs merge=lfs -text
14
- *.ot filter=lfs diff=lfs merge=lfs -text
15
- *.parquet filter=lfs diff=lfs merge=lfs -text
16
- *.pb filter=lfs diff=lfs merge=lfs -text
17
- *.pt filter=lfs diff=lfs merge=lfs -text
18
- *.pth filter=lfs diff=lfs merge=lfs -text
19
- *.rar filter=lfs diff=lfs merge=lfs -text
20
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
21
- *.tar.* filter=lfs diff=lfs merge=lfs -text
22
- *.tflite filter=lfs diff=lfs merge=lfs -text
23
- *.tgz filter=lfs diff=lfs merge=lfs -text
24
- *.xz filter=lfs diff=lfs merge=lfs -text
25
- *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md DELETED
@@ -1,217 +0,0 @@
1
- ---
2
- annotations_creators:
3
- - crowdsourced
4
- language:
5
- - en
6
- language_creators:
7
- - found
8
- - crowdsourced
9
- license:
10
- - mit
11
- multilinguality:
12
- - monolingual
13
- pretty_name: CommonGen
14
- size_categories:
15
- - 10K<n<100K
16
- source_datasets:
17
- - original
18
- task_categories:
19
- - text2text-generation
20
- task_ids: []
21
- paperswithcode_id: commongen
22
- tags:
23
- - concepts-to-text
24
- dataset_info:
25
- features:
26
- - name: concept_set_idx
27
- dtype: int32
28
- - name: concepts
29
- sequence: string
30
- - name: target
31
- dtype: string
32
- splits:
33
- - name: train
34
- num_bytes: 6724250
35
- num_examples: 67389
36
- - name: validation
37
- num_bytes: 408752
38
- num_examples: 4018
39
- - name: test
40
- num_bytes: 77530
41
- num_examples: 1497
42
- download_size: 1845699
43
- dataset_size: 7210532
44
- ---
45
-
46
- # Dataset Card for "common_gen"
47
-
48
- ## Table of Contents
49
- - [Dataset Description](#dataset-description)
50
- - [Dataset Summary](#dataset-summary)
51
- - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
52
- - [Languages](#languages)
53
- - [Dataset Structure](#dataset-structure)
54
- - [Data Instances](#data-instances)
55
- - [Data Fields](#data-fields)
56
- - [Data Splits](#data-splits)
57
- - [Dataset Creation](#dataset-creation)
58
- - [Curation Rationale](#curation-rationale)
59
- - [Source Data](#source-data)
60
- - [Annotations](#annotations)
61
- - [Personal and Sensitive Information](#personal-and-sensitive-information)
62
- - [Considerations for Using the Data](#considerations-for-using-the-data)
63
- - [Social Impact of Dataset](#social-impact-of-dataset)
64
- - [Discussion of Biases](#discussion-of-biases)
65
- - [Other Known Limitations](#other-known-limitations)
66
- - [Additional Information](#additional-information)
67
- - [Dataset Curators](#dataset-curators)
68
- - [Licensing Information](#licensing-information)
69
- - [Citation Information](#citation-information)
70
- - [Contributions](#contributions)
71
-
72
- ## Dataset Description
73
-
74
- - **Homepage:** [https://inklab.usc.edu/CommonGen/index.html](https://inklab.usc.edu/CommonGen/index.html)
75
- - **Repository:** https://github.com/INK-USC/CommonGen
76
- - **Paper:** [CommonGen: A Constrained Text Generation Challenge for Generative Commonsense Reasoning](https://arxiv.org/abs/1911.03705)
77
- - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
78
- - **Size of downloaded dataset files:** 1.76 MB
79
- - **Size of the generated dataset:** 6.88 MB
80
- - **Total amount of disk used:** 8.64 MB
81
-
82
- ### Dataset Summary
83
-
84
- CommonGen is a constrained text generation task, associated with a benchmark dataset,
85
- to explicitly test machines for the ability of generative commonsense reasoning. Given
86
- a set of common concepts; the task is to generate a coherent sentence describing an
87
- everyday scenario using these concepts.
88
-
89
- CommonGen is challenging because it inherently requires 1) relational reasoning using
90
- background commonsense knowledge, and 2) compositional generalization ability to work
91
- on unseen concept combinations. Our dataset, constructed through a combination of
92
- crowd-sourcing from AMT and existing caption corpora, consists of 30k concept-sets and
93
- 50k sentences in total.
94
-
95
- ### Supported Tasks and Leaderboards
96
-
97
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
98
-
99
- ### Languages
100
-
101
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
102
-
103
- ## Dataset Structure
104
-
105
- ### Data Instances
106
-
107
- #### default
108
-
109
- - **Size of downloaded dataset files:** 1.76 MB
110
- - **Size of the generated dataset:** 6.88 MB
111
- - **Total amount of disk used:** 8.64 MB
112
-
113
- An example of 'train' looks as follows.
114
- ```
115
- {
116
- "concept_set_idx": 0,
117
- "concepts": ["ski", "mountain", "skier"],
118
- "target": "Three skiers are skiing on a snowy mountain."
119
- }
120
- ```
121
-
122
- ### Data Fields
123
-
124
- The data fields are the same among all splits.
125
-
126
- #### default
127
- - `concept_set_idx`: a `int32` feature.
128
- - `concepts`: a `list` of `string` features.
129
- - `target`: a `string` feature.
130
-
131
- ### Data Splits
132
-
133
- | name |train|validation|test|
134
- |-------|----:|---------:|---:|
135
- |default|67389| 4018|1497|
136
-
137
- ## Dataset Creation
138
-
139
- ### Curation Rationale
140
-
141
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
142
-
143
- ### Source Data
144
-
145
- #### Initial Data Collection and Normalization
146
-
147
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
148
-
149
- #### Who are the source language producers?
150
-
151
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
152
-
153
- ### Annotations
154
-
155
- #### Annotation process
156
-
157
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
158
-
159
- #### Who are the annotators?
160
-
161
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
162
-
163
- ### Personal and Sensitive Information
164
-
165
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
166
-
167
- ## Considerations for Using the Data
168
-
169
- ### Social Impact of Dataset
170
-
171
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
172
-
173
- ### Discussion of Biases
174
-
175
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
176
-
177
- ### Other Known Limitations
178
-
179
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
180
-
181
- ## Additional Information
182
-
183
- ### Dataset Curators
184
-
185
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
186
-
187
- ### Licensing Information
188
-
189
- The dataset is licensed under [MIT License](https://github.com/INK-USC/CommonGen/blob/master/LICENSE).
190
-
191
- ### Citation Information
192
-
193
- ```bib
194
- @inproceedings{lin-etal-2020-commongen,
195
- title = "{C}ommon{G}en: A Constrained Text Generation Challenge for Generative Commonsense Reasoning",
196
- author = "Lin, Bill Yuchen and
197
- Zhou, Wangchunshu and
198
- Shen, Ming and
199
- Zhou, Pei and
200
- Bhagavatula, Chandra and
201
- Choi, Yejin and
202
- Ren, Xiang",
203
- booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020",
204
- month = nov,
205
- year = "2020",
206
- address = "Online",
207
- publisher = "Association for Computational Linguistics",
208
- url = "https://www.aclweb.org/anthology/2020.findings-emnlp.165",
209
- doi = "10.18653/v1/2020.findings-emnlp.165",
210
- pages = "1823--1840"
211
- }
212
- ```
213
-
214
-
215
- ### Contributions
216
-
217
- Thanks to [@JetRunner](https://github.com/JetRunner), [@yuchenlin](https://github.com/yuchenlin), [@thomwolf](https://github.com/thomwolf), [@lhoestq](https://github.com/lhoestq) for adding this dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
common_gen.py DELETED
@@ -1,109 +0,0 @@
1
- import json
2
- import os
3
- import random
4
-
5
- import datasets
6
-
7
-
8
- random.seed(42) # This is important, to ensure the same order for concept sets as the official script.
9
-
10
- _CITATION = """\
11
- @inproceedings{lin-etal-2020-commongen,
12
- title = "{C}ommon{G}en: A Constrained Text Generation Challenge for Generative Commonsense Reasoning",
13
- author = "Lin, Bill Yuchen and
14
- Zhou, Wangchunshu and
15
- Shen, Ming and
16
- Zhou, Pei and
17
- Bhagavatula, Chandra and
18
- Choi, Yejin and
19
- Ren, Xiang",
20
- booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020",
21
- month = nov,
22
- year = "2020",
23
- address = "Online",
24
- publisher = "Association for Computational Linguistics",
25
- url = "https://www.aclweb.org/anthology/2020.findings-emnlp.165",
26
- doi = "10.18653/v1/2020.findings-emnlp.165",
27
- pages = "1823--1840"
28
- }
29
- """
30
-
31
- _DESCRIPTION = """\
32
- CommonGen is a constrained text generation task, associated with a benchmark dataset,
33
- to explicitly test machines for the ability of generative commonsense reasoning. Given
34
- a set of common concepts; the task is to generate a coherent sentence describing an
35
- everyday scenario using these concepts.
36
-
37
- CommonGen is challenging because it inherently requires 1) relational reasoning using
38
- background commonsense knowledge, and 2) compositional generalization ability to work
39
- on unseen concept combinations. Our dataset, constructed through a combination of
40
- crowd-sourcing from AMT and existing caption corpora, consists of 30k concept-sets and
41
- 50k sentences in total.
42
- """
43
- _URL = "https://storage.googleapis.com/huggingface-nlp/datasets/common_gen/commongen_data.zip"
44
-
45
-
46
- class CommonGen(datasets.GeneratorBasedBuilder):
47
- VERSION = datasets.Version("2020.5.30")
48
-
49
- def _info(self):
50
- features = datasets.Features(
51
- {
52
- "concept_set_idx": datasets.Value("int32"),
53
- "concepts": datasets.Sequence(datasets.Value("string")),
54
- "target": datasets.Value("string"),
55
- }
56
- )
57
- return datasets.DatasetInfo(
58
- description=_DESCRIPTION,
59
- features=features,
60
- supervised_keys=datasets.info.SupervisedKeysData(input="concepts", output="target"),
61
- homepage="https://inklab.usc.edu/CommonGen/index.html",
62
- citation=_CITATION,
63
- )
64
-
65
- def _split_generators(self, dl_manager):
66
- """Returns SplitGenerators."""
67
-
68
- dl_dir = dl_manager.download_and_extract(_URL)
69
-
70
- return [
71
- datasets.SplitGenerator(
72
- name=datasets.Split.TRAIN,
73
- gen_kwargs={"filepath": os.path.join(dl_dir, "commongen.train.jsonl"), "split": "train"},
74
- ),
75
- datasets.SplitGenerator(
76
- name=datasets.Split.VALIDATION,
77
- gen_kwargs={"filepath": os.path.join(dl_dir, "commongen.dev.jsonl"), "split": "dev"},
78
- ),
79
- datasets.SplitGenerator(
80
- name=datasets.Split.TEST,
81
- gen_kwargs={"filepath": os.path.join(dl_dir, "commongen.test_noref.jsonl"), "split": "test"},
82
- ),
83
- ]
84
-
85
- def _generate_examples(self, filepath, split):
86
- """Yields examples."""
87
- with open(filepath, encoding="utf-8") as f:
88
- id_ = 0
89
- for idx, row in enumerate(f):
90
- row = row.replace(", }", "}") # Fix possible JSON format error
91
- data = json.loads(row)
92
-
93
- rand_order = [word for word in data["concept_set"].split("#")]
94
- random.shuffle(rand_order)
95
-
96
- if split == "test":
97
- yield idx, {
98
- "concept_set_idx": idx,
99
- "concepts": rand_order,
100
- "target": "",
101
- }
102
- else:
103
- for scene in data["scene"]:
104
- yield id_, {
105
- "concept_set_idx": idx,
106
- "concepts": rand_order,
107
- "target": scene,
108
- }
109
- id_ += 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dataset_infos.json DELETED
@@ -1 +0,0 @@
1
- {"default": {"description": "CommonGen is a constrained text generation task, associated with a benchmark dataset, \nto explicitly test machines for the ability of generative commonsense reasoning. Given \na set of common concepts; the task is to generate a coherent sentence describing an \neveryday scenario using these concepts.\n\nCommonGen is challenging because it inherently requires 1) relational reasoning using \nbackground commonsense knowledge, and 2) compositional generalization ability to work \non unseen concept combinations. Our dataset, constructed through a combination of \ncrowd-sourcing from AMT and existing caption corpora, consists of 30k concept-sets and \n50k sentences in total.\n", "citation": "@inproceedings{lin-etal-2020-commongen,\n author = {Bill Yuchen Lin and Wangchunshu Zhou and Ming Shen and Pei Zhou and Chandra Bhagavatula and Yejin Choi and Xiang Ren},\n title = {{C}ommon{G}en: A Constrained Text Generation Challenge for Generative Commonsense Reasoning},\n booktitle = {Findings of the Association for Computational Linguistics: EMNLP 2020},\n year = {2020}\n}\n", "homepage": "https://inklab.usc.edu/CommonGen/index.html", "license": "", "features": {"concept_set_idx": {"dtype": "int32", "id": null, "_type": "Value"}, "concepts": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "target": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": {"input": "concepts", "output": "target"}, "builder_name": "common_gen", "config_name": "default", "version": {"version_str": "2020.5.30", "description": null, "datasets_version_to_prepare": null, "major": 2020, "minor": 5, "patch": 30}, "splits": {"train": {"name": "train", "num_bytes": 6724250, "num_examples": 67389, "dataset_name": "common_gen"}, "validation": {"name": "validation", "num_bytes": 408752, "num_examples": 4018, "dataset_name": "common_gen"}, "test": {"name": "test", "num_bytes": 77530, "num_examples": 1497, "dataset_name": "common_gen"}}, "download_checksums": {"https://storage.googleapis.com/huggingface-nlp/datasets/common_gen/commongen_data.zip": {"num_bytes": 1845699, "checksum": "a3f19ca607da4e874fc5f2dd1f53c13a6788a497f883d74cc3f9a1fcda44c594"}}, "download_size": 1845699, "post_processing_size": null, "dataset_size": 7210532, "size_in_bytes": 9056231}}
 
 
default/common_gen-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2c7642eb9ea7d6ea78fb4d834d5e95ecca049aabf69e3f66e3d74da848f292e
3
+ size 31156
default/common_gen-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b249952e1280a48057ef558684dbb4229b85967509715478ff20cf5bac904e8
3
+ size 3232447
default/common_gen-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0560c096ddc90966c4a98587cbb2644bcfe21a648aba1f3f5f9b1e57ba5da5b1
3
+ size 171259