Datasets:
Tasks:
Text Classification
Modalities:
Text
Formats:
parquet
Sub-tasks:
topic-classification
Languages:
English
Size:
100K - 1M
ArXiv:
License:
Commit
•
25b1233
0
Parent(s):
Update files from the datasets library (from 1.2.0)
Browse filesRelease notes: https://github.com/huggingface/datasets/releases/tag/1.2.0
- .gitattributes +27 -0
- README.md +159 -0
- dataset_infos.json +1 -0
- dbpedia_14.py +146 -0
- dummy/dbpedia_14/2.0.0/dummy_data.zip +3 -0
.gitattributes
ADDED
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
5 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
12 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
13 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
14 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
15 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
16 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
17 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
18 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
20 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
21 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
22 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
24 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
25 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
26 |
+
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
27 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,159 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
annotations_creators:
|
3 |
+
- machine-generated
|
4 |
+
language_creators:
|
5 |
+
- crowdsourced
|
6 |
+
languages:
|
7 |
+
- en
|
8 |
+
licenses:
|
9 |
+
- cc-by-sa-3-0
|
10 |
+
multilinguality:
|
11 |
+
- monolingual
|
12 |
+
size_categories:
|
13 |
+
- 100K<n<1M
|
14 |
+
source_datasets:
|
15 |
+
- original
|
16 |
+
task_categories:
|
17 |
+
- text-classification
|
18 |
+
task_ids:
|
19 |
+
- topic-classification
|
20 |
+
---
|
21 |
+
|
22 |
+
# Dataset Card for DBpedia14
|
23 |
+
|
24 |
+
## Table of Contents
|
25 |
+
- [Dataset Description](#dataset-description)
|
26 |
+
- [Dataset Summary](#dataset-summary)
|
27 |
+
- [Supported Tasks](#supported-tasks-and-leaderboards)
|
28 |
+
- [Languages](#languages)
|
29 |
+
- [Dataset Structure](#dataset-structure)
|
30 |
+
- [Data Instances](#data-instances)
|
31 |
+
- [Data Fields](#data-instances)
|
32 |
+
- [Data Splits](#data-instances)
|
33 |
+
- [Dataset Creation](#dataset-creation)
|
34 |
+
- [Curation Rationale](#curation-rationale)
|
35 |
+
- [Source Data](#source-data)
|
36 |
+
- [Annotations](#annotations)
|
37 |
+
- [Personal and Sensitive Information](#personal-and-sensitive-information)
|
38 |
+
- [Considerations for Using the Data](#considerations-for-using-the-data)
|
39 |
+
- [Social Impact of Dataset](#social-impact-of-dataset)
|
40 |
+
- [Discussion of Biases](#discussion-of-biases)
|
41 |
+
- [Other Known Limitations](#other-known-limitations)
|
42 |
+
- [Additional Information](#additional-information)
|
43 |
+
- [Dataset Curators](#dataset-curators)
|
44 |
+
- [Licensing Information](#licensing-information)
|
45 |
+
- [Citation Information](#citation-information)
|
46 |
+
|
47 |
+
## Dataset Description
|
48 |
+
|
49 |
+
- **Homepage:** [DBpedia14 homepage](https://wiki.dbpedia.org/develop/datasets)
|
50 |
+
- **Repository:** [DBpedia14 repository](https://github.com/dbpedia/extraction-framework)
|
51 |
+
- **Paper:** [DBpedia--a large-scale, multilingual knowledge base extracted from Wikipedia](https://content.iospress.com/articles/semantic-web/sw134)
|
52 |
+
- **Point of Contact:** [Xiang Zhang](mailto:[email protected])
|
53 |
+
|
54 |
+
### Dataset Summary
|
55 |
+
|
56 |
+
The DBpedia ontology classification dataset is constructed by picking 14 non-overlapping classes
|
57 |
+
from DBpedia 2014. They are listed in classes.txt. From each of thse 14 ontology classes, we
|
58 |
+
randomly choose 40,000 training samples and 5,000 testing samples. Therefore, the total size
|
59 |
+
of the training dataset is 560,000 and testing dataset 70,000.
|
60 |
+
There are 3 columns in the dataset (same for train and test splits), corresponding to class index
|
61 |
+
(1 to 14), title and content. The title and content are escaped using double quotes ("), and any
|
62 |
+
internal double quote is escaped by 2 double quotes (""). There are no new lines in title or content.
|
63 |
+
|
64 |
+
### Supported Tasks and Leaderboards
|
65 |
+
|
66 |
+
- `text-classification`, `topic-classification`: The dataset is mainly used for text classification: given the content
|
67 |
+
and the title, predict the correct topic.
|
68 |
+
|
69 |
+
### Languages
|
70 |
+
|
71 |
+
Although DBpedia is a multilingual knowledge base, the DBpedia14 extract contains English data mainly, other languages may appear
|
72 |
+
(e.g. a film whose title is origanlly not English).
|
73 |
+
|
74 |
+
## Dataset Structure
|
75 |
+
|
76 |
+
### Data Instances
|
77 |
+
|
78 |
+
A typical data point, comprises of a title, a content and the corresponding label.
|
79 |
+
|
80 |
+
An example from the DBpedia test set looks as follows:
|
81 |
+
```
|
82 |
+
{
|
83 |
+
'title':'',
|
84 |
+
'content':" TY KU /taɪkuː/ is an American alcoholic beverage company that specializes in sake and other spirits. The privately-held company was founded in 2004 and is headquartered in New York City New York. While based in New York TY KU's beverages are made in Japan through a joint venture with two sake breweries. Since 2011 TY KU's growth has extended its products into all 50 states.",
|
85 |
+
'label':0
|
86 |
+
}
|
87 |
+
```
|
88 |
+
|
89 |
+
### Data Fields
|
90 |
+
|
91 |
+
- 'title': a string containing the title of the document - escaped using double quotes (") and any internal double quote is escaped by 2 double quotes ("").
|
92 |
+
- 'content': a string containing the body of the document - escaped using double quotes (") and any internal double quote is escaped by 2 double quotes ("").
|
93 |
+
- 'label': one of the 14 possible topics.
|
94 |
+
|
95 |
+
### Data Splits
|
96 |
+
|
97 |
+
The data is split into a training and test set.
|
98 |
+
For each of the 14 classes we have 40,000 training samples and 5,000 testing samples.
|
99 |
+
Therefore, the total size of the training dataset is 560,000 and testing dataset 70,000.
|
100 |
+
|
101 |
+
## Dataset Creation
|
102 |
+
|
103 |
+
### Curation Rationale
|
104 |
+
|
105 |
+
The DBPedia ontology classification dataset is constructed by Xiang Zhang ([email protected]), licensed under the terms of the Creative Commons Attribution-ShareAlike License and the GNU Free Documentation License. It is used as a text classification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann LeCun. Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015).
|
106 |
+
|
107 |
+
### Source Data
|
108 |
+
|
109 |
+
#### Initial Data Collection and Normalization
|
110 |
+
|
111 |
+
[More Information Needed]
|
112 |
+
|
113 |
+
#### Who are the source language producers?
|
114 |
+
|
115 |
+
[More Information Needed]
|
116 |
+
|
117 |
+
### Annotations
|
118 |
+
|
119 |
+
#### Annotation process
|
120 |
+
|
121 |
+
[More Information Needed]
|
122 |
+
|
123 |
+
#### Who are the annotators?
|
124 |
+
|
125 |
+
[More Information Needed]
|
126 |
+
|
127 |
+
### Personal and Sensitive Information
|
128 |
+
|
129 |
+
[More Information Needed]
|
130 |
+
|
131 |
+
## Considerations for Using the Data
|
132 |
+
|
133 |
+
### Social Impact of Dataset
|
134 |
+
|
135 |
+
[More Information Needed]
|
136 |
+
|
137 |
+
### Discussion of Biases
|
138 |
+
|
139 |
+
[More Information Needed]
|
140 |
+
|
141 |
+
### Other Known Limitations
|
142 |
+
|
143 |
+
[More Information Needed]
|
144 |
+
|
145 |
+
## Additional Information
|
146 |
+
|
147 |
+
### Dataset Curators
|
148 |
+
|
149 |
+
The DBPedia ontology classification dataset is constructed by Xiang Zhang ([email protected]), licensed under the terms of the Creative Commons Attribution-ShareAlike License and the GNU Free Documentation License. It is used as a text classification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann LeCun. Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015).
|
150 |
+
|
151 |
+
### Licensing Information
|
152 |
+
|
153 |
+
The DBPedia ontology classification dataset is licensed under the terms of the Creative Commons Attribution-ShareAlike License and the GNU Free Documentation License.
|
154 |
+
|
155 |
+
### Citation Information
|
156 |
+
|
157 |
+
Xiang Zhang, Junbo Zhao, Yann LeCun. Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015).
|
158 |
+
|
159 |
+
Lehmann, Jens, Robert Isele, Max Jakob, Anja Jentzsch, Dimitris Kontokostas, Pablo N. Mendes, Sebastian Hellmann et al. "DBpedia–a large-scale, multilingual knowledge base extracted from Wikipedia." Semantic web 6, no. 2 (2015): 167-195.
|
dataset_infos.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dbpedia_14": {"description": "The DBpedia ontology classification dataset is constructed by picking 14 non-overlapping classes\nfrom DBpedia 2014. They are listed in classes.txt. From each of thse 14 ontology classes, we\nrandomly choose 40,000 training samples and 5,000 testing samples. Therefore, the total size\nof the training dataset is 560,000 and testing dataset 70,000.\nThere are 3 columns in the dataset (same for train and test splits), corresponding to class index\n(1 to 14), title and content. The title and content are escaped using double quotes (\"), and any\ninternal double quote is escaped by 2 double quotes (\"\"). There are no new lines in title or content.\n", "citation": "@article{lehmann2015dbpedia,\n title={DBpedia--a large-scale, multilingual knowledge base extracted from Wikipedia},\n author={Lehmann, Jens and Isele, Robert and Jakob, Max and Jentzsch, Anja and Kontokostas, \n Dimitris and Mendes, Pablo N and Hellmann, Sebastian and Morsey, Mohamed and Van Kleef, \n Patrick and Auer, S{\"o}ren and others},\n journal={Semantic web},\n volume={6},\n number={2},\n pages={167--195},\n year={2015},\n publisher={IOS Press}\n}\n", "homepage": "https://wiki.dbpedia.org/develop/datasets", "license": "Creative Commons Attribution-ShareAlike 3.0 and the GNU Free Documentation License", "features": {"label": {"num_classes": 14, "names": ["Company", "EducationalInstitution", "Artist", "Athlete", "OfficeHolder", "MeanOfTransportation", "Building", "NaturalPlace", "Village", "Animal", "Plant", "Album", "Film", "WrittenWork"], "names_file": null, "id": null, "_type": "ClassLabel"}, "title": {"dtype": "string", "id": null, "_type": "Value"}, "content": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "d_bpedia14", "config_name": "dbpedia_14", "version": {"version_str": "2.0.0", "description": null, "major": 2, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 178429418, "num_examples": 560000, "dataset_name": "d_bpedia14"}, "test": {"name": "test", "num_bytes": 22310341, "num_examples": 70000, "dataset_name": "d_bpedia14"}}, "download_checksums": {"https://drive.google.com/uc?export=download&id=0Bz8a_Dbh9QhbQ2Vic1kxMmZZQ1k": {"num_bytes": 68341698, "checksum": "cad5773f85d7501bb2783833768bc624641cdddf7056000a06f12bcd0239a310"}}, "download_size": 68341698, "post_processing_size": null, "dataset_size": 200739759, "size_in_bytes": 269081457}}
|
dbpedia_14.py
ADDED
@@ -0,0 +1,146 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# coding=utf-8
|
2 |
+
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
|
3 |
+
#
|
4 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
5 |
+
# you may not use this file except in compliance with the License.
|
6 |
+
# You may obtain a copy of the License at
|
7 |
+
#
|
8 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
9 |
+
#
|
10 |
+
# Unless required by applicable law or agreed to in writing, software
|
11 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
12 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
+
# See the License for the specific language governing permissions and
|
14 |
+
# limitations under the License.
|
15 |
+
"""The DBpedia dataset for text classification."""
|
16 |
+
|
17 |
+
from __future__ import absolute_import, division, print_function
|
18 |
+
|
19 |
+
import csv
|
20 |
+
import os
|
21 |
+
|
22 |
+
import datasets
|
23 |
+
|
24 |
+
|
25 |
+
# TODO: Add BibTeX citation
|
26 |
+
# Find for instance the citation on arxiv or on the dataset repo/website
|
27 |
+
_CITATION = """\
|
28 |
+
@article{lehmann2015dbpedia,
|
29 |
+
title={DBpedia--a large-scale, multilingual knowledge base extracted from Wikipedia},
|
30 |
+
author={Lehmann, Jens and Isele, Robert and Jakob, Max and Jentzsch, Anja and Kontokostas,
|
31 |
+
Dimitris and Mendes, Pablo N and Hellmann, Sebastian and Morsey, Mohamed and Van Kleef,
|
32 |
+
Patrick and Auer, S{\"o}ren and others},
|
33 |
+
journal={Semantic web},
|
34 |
+
volume={6},
|
35 |
+
number={2},
|
36 |
+
pages={167--195},
|
37 |
+
year={2015},
|
38 |
+
publisher={IOS Press}
|
39 |
+
}
|
40 |
+
"""
|
41 |
+
|
42 |
+
_DESCRIPTION = """\
|
43 |
+
The DBpedia ontology classification dataset is constructed by picking 14 non-overlapping classes
|
44 |
+
from DBpedia 2014. They are listed in classes.txt. From each of thse 14 ontology classes, we
|
45 |
+
randomly choose 40,000 training samples and 5,000 testing samples. Therefore, the total size
|
46 |
+
of the training dataset is 560,000 and testing dataset 70,000.
|
47 |
+
There are 3 columns in the dataset (same for train and test splits), corresponding to class index
|
48 |
+
(1 to 14), title and content. The title and content are escaped using double quotes ("), and any
|
49 |
+
internal double quote is escaped by 2 double quotes (""). There are no new lines in title or content.
|
50 |
+
"""
|
51 |
+
|
52 |
+
_HOMEPAGE = "https://wiki.dbpedia.org/develop/datasets"
|
53 |
+
|
54 |
+
_LICENSE = "Creative Commons Attribution-ShareAlike 3.0 and the GNU Free Documentation License"
|
55 |
+
|
56 |
+
_URLs = {
|
57 |
+
"dbpedia_14": "https://drive.google.com/uc?export=download&id=0Bz8a_Dbh9QhbQ2Vic1kxMmZZQ1k",
|
58 |
+
}
|
59 |
+
|
60 |
+
|
61 |
+
class DBpedia14Config(datasets.BuilderConfig):
|
62 |
+
"""BuilderConfig for DBpedia."""
|
63 |
+
|
64 |
+
def __init__(self, **kwargs):
|
65 |
+
"""BuilderConfig for DBpedia.
|
66 |
+
|
67 |
+
Args:
|
68 |
+
**kwargs: keyword arguments forwarded to super.
|
69 |
+
"""
|
70 |
+
super(DBpedia14Config, self).__init__(**kwargs)
|
71 |
+
|
72 |
+
|
73 |
+
class DBpedia14(datasets.GeneratorBasedBuilder):
|
74 |
+
"""DBpedia 2014 Ontology Classification Dataset."""
|
75 |
+
|
76 |
+
VERSION = datasets.Version("2.0.0")
|
77 |
+
|
78 |
+
BUILDER_CONFIGS = [
|
79 |
+
DBpedia14Config(
|
80 |
+
name="dbpedia_14", version=VERSION, description="DBpedia 2014 Ontology Classification Dataset."
|
81 |
+
),
|
82 |
+
]
|
83 |
+
|
84 |
+
def _info(self):
|
85 |
+
features = datasets.Features(
|
86 |
+
{
|
87 |
+
"label": datasets.features.ClassLabel(
|
88 |
+
names=[
|
89 |
+
"Company",
|
90 |
+
"EducationalInstitution",
|
91 |
+
"Artist",
|
92 |
+
"Athlete",
|
93 |
+
"OfficeHolder",
|
94 |
+
"MeanOfTransportation",
|
95 |
+
"Building",
|
96 |
+
"NaturalPlace",
|
97 |
+
"Village",
|
98 |
+
"Animal",
|
99 |
+
"Plant",
|
100 |
+
"Album",
|
101 |
+
"Film",
|
102 |
+
"WrittenWork",
|
103 |
+
]
|
104 |
+
),
|
105 |
+
"title": datasets.Value("string"),
|
106 |
+
"content": datasets.Value("string"),
|
107 |
+
}
|
108 |
+
)
|
109 |
+
return datasets.DatasetInfo(
|
110 |
+
description=_DESCRIPTION,
|
111 |
+
features=features,
|
112 |
+
supervised_keys=None,
|
113 |
+
homepage=_HOMEPAGE,
|
114 |
+
license=_LICENSE,
|
115 |
+
citation=_CITATION,
|
116 |
+
)
|
117 |
+
|
118 |
+
def _split_generators(self, dl_manager):
|
119 |
+
"""Returns SplitGenerators."""
|
120 |
+
my_urls = _URLs[self.config.name]
|
121 |
+
data_dir = dl_manager.download_and_extract(my_urls)
|
122 |
+
return [
|
123 |
+
datasets.SplitGenerator(
|
124 |
+
name=datasets.Split.TRAIN,
|
125 |
+
gen_kwargs={
|
126 |
+
"filepath": os.path.join(data_dir, "dbpedia_csv/train.csv"),
|
127 |
+
"split": "train",
|
128 |
+
},
|
129 |
+
),
|
130 |
+
datasets.SplitGenerator(
|
131 |
+
name=datasets.Split.TEST,
|
132 |
+
gen_kwargs={"filepath": os.path.join(data_dir, "dbpedia_csv/test.csv"), "split": "test"},
|
133 |
+
),
|
134 |
+
]
|
135 |
+
|
136 |
+
def _generate_examples(self, filepath, split):
|
137 |
+
""" Yields examples. """
|
138 |
+
|
139 |
+
with open(filepath, encoding="utf-8") as f:
|
140 |
+
data = csv.reader(f, delimiter=",", quoting=csv.QUOTE_NONNUMERIC)
|
141 |
+
for id_, row in enumerate(data):
|
142 |
+
yield id_, {
|
143 |
+
"title": row[1],
|
144 |
+
"content": row[2],
|
145 |
+
"label": int(row[0]) - 1,
|
146 |
+
}
|
dummy/dbpedia_14/2.0.0/dummy_data.zip
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:98d4ad02edaedc492459d94204bb149b82a55c2eb5bed0ad77b76ccbe387cd97
|
3 |
+
size 3319
|