File size: 2,232 Bytes
f9bd01a
86bc1d9
 
dadfc6c
5b5233f
 
f9bd01a
 
 
 
86bc1d9
 
5b5233f
 
86bc1d9
 
 
 
 
 
 
 
42f294d
86bc1d9
 
42f294d
86bc1d9
42f294d
 
86bc1d9
 
 
 
 
 
 
92ecc52
cd3c2f4
92ecc52
 
 
 
 
b6c46b6
 
 
 
 
 
 
 
 
fd4c3f4
 
 
 
b6c46b6
fd4c3f4
 
cd3c2f4
d492250
 
 
 
 
f3cb635
d492250
 
 
f3cb635
d492250
 
 
 
ea929d1
 
 
dadfc6c
 
 
ea929d1
dadfc6c
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
---
language:
- vi
license: mit
size_categories:
- 10K<n<100K
task_categories:
- text-generation
- question-answering
- text-classification
dataset_info:
  features:
  - name: context
    dtype: string
  - name: question
    dtype: string
  - name: answer
    dtype: string
  - name: answer_start
    dtype: int64
  splits:
  - name: train
    num_bytes: 54478998
    num_examples: 48460
  - name: test
    num_bytes: 6041628
    num_examples: 5385
  download_size: 33267124
  dataset_size: 60520626
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
  - split: test
    path: data/test-*
---
# Dataset Describe

<!-- Provide a quick summary of the dataset. -->

This dataset is collected from internet sources, SQuAD dataset, wiki, etc. It has been translated into Vietnamese using "google translate" and word segmented using VnCoreNLP (https://github.com/vncorenlp/VnCoreNLP).

## Data structure

Dataset includes the following columns:

- `question`: Question related to the content of the text.
- `context`: Paragraph of text.
- `answer`: The answer to the question is based on the content of the text.
- `answer_start`: The starting position of the answer in the text.

### How to use
You can load this dataset using Hugging Face's `datasets` library:
```python
from datasets import load_dataset

dataset = load_dataset("ShynBui/Vietnamese_Reading_Comprehension_Dataset")
```

### Split data
The dataset is divided into train/test sections.
```
DatasetDict({
    train: Dataset({
        features: ['context', 'question', 'answer', 'answer_start'],
        num_rows: 48460
    })
    test: Dataset({
        features: ['context', 'question', 'answer', 'answer_start'],
        num_rows: 5385
    })
})
```

### Task categories
This dataset can be used for the following main tasks
- `question-answering`.
- `reading-comprehension`.
- `natural-language-processing`.


### Contribute
We welcome all contributions to this dataset. If you discover an error or have feedback, please create an Issue or Pull Request on our Hub repository.

### License
This dataset is released under the MIT License.

### Contact
If you have any questions, please contact us via email: [email protected].