|
--- |
|
language: |
|
- en |
|
- zh |
|
- ar |
|
- es |
|
pretty_name: 'MOSLA: Moments of Second Language Acquisition' |
|
annotations_creators: |
|
- expert-generated |
|
- machine-generated |
|
size_categories: |
|
- 100K<n<1M |
|
source_datasets: |
|
- original |
|
tags: |
|
- language-learning |
|
- audio |
|
- voice |
|
- video |
|
configs: |
|
- config_name: default |
|
data_files: merged.jsonl |
|
- config_name: human |
|
data_files: human.jsonl |
|
- config_name: machine |
|
data_files: machine.jsonl |
|
- config_name: annotation_records |
|
data_files: annotation_records.jsonl |
|
arxiv: 2403.17314 |
|
extra_gated_prompt: > |
|
By accessing this dataset, you are agreeing to the terms laid out in the |
|
[license](https://huggingface.co/datasets/octanove/mosla/blob/main/LICENSE.md). |
|
|
|
In particular, you are agreeing to: |
|
1) Only use the data for non-commercial research and educational purposes |
|
2) Not redistribute the data |
|
3) Not attempt to extract personally identifiable information from the data or otherwise misuse it |
|
|
|
extra_gated_fields: |
|
Full Legal Name: text |
|
Affiliation: text |
|
Contact Email: text |
|
license: other |
|
license_name: mosla-license |
|
license_link: LICENSE.md |
|
task_categories: |
|
- automatic-speech-recognition |
|
- audio-classification |
|
- voice-activity-detection |
|
--- |
|
|
|
## Dataset Description |
|
|
|
- **Paper:** [Project MOSLA: Recording Every Moment of Second Language Acquisition](https://arxiv.org/abs/2403.17314) |
|
- **Author/Contact:** [Masato Hagiwara](mailto:[email protected]) |
|
- **Author/Contact:** [Josh Tanner](mailto:[email protected]) |
|
|
|
# Overview |
|
The MOSLA dataset ("MOSLA") is a longitudinal, multimodal, multilingual, and controlled dataset created by inviting participants to learn one |
|
of three target languages (Arabic, Spanish, and Chinese) from scratch over a span of two years, exclusively through online instruction, |
|
and recording every lesson using Zoom. The dataset is semi-automatically annotated with speaker/language IDs and transcripts by both human |
|
annotators and fine-tuned state-of-the-art speech models. Because learners had no previous exposure to the language they are learning in MOSLA |
|
and their only exposure to the language over this time period was these online language lessons, MOSLA provides a complete picture of the |
|
first two years of language acquisition for its participants. |
|
|
|
Concretely, MOSLA includes three types of data: |
|
- Video recordings of all of the language lessons |
|
- Automatically generated speech annotations for all of the language lessons |
|
- Human speech annotation for some five-minute segments in some of the language lessons |
|
|
|
For detailed information please see the [paper](https://arxiv.org/abs/2403.17314), |
|
or for study background and ethical considerations the [datasheet](https://huggingface.co/datasets/octanove/mosla/blob/main/DATASHEET.md). |
|
|
|
The raw data can also be [cloned](https://huggingface.co/datasets/octanove/mosla?clone=true) using git and git-lfs. |
|
If you are using Python we recommend using the 🤗 Datasets library, but we understand that not everyone uses Python. |
|
|
|
## Data Volume |
|
There are approximately 270 hours of video data, 202,000 machine generated |
|
annotation utterances, and 7,700 human annotated utterances. |
|
The entirety of the video data takes up about 71GB. |
|
|
|
# Dataset Structure |
|
|
|
## Configurations / Data Groupings |
|
MOSLA has four configurations, three of which are different mixes of human and/or machine annotations, and one of which |
|
is just a record of what data was presented to humans for annotation. Concretely, these can be seen below: |
|
```python |
|
from datasets import load_dataset |
|
|
|
# the "merged" data is the default configuration. it contains machine annotations for all sections of video EXCEPT |
|
# those annotated by humans, and human annotations for the sections of video annotated by humans |
|
# merged.jsonl |
|
merged = load_dataset('octanove/mosla') |
|
|
|
# the human annotation data, containing only the annotations made by bilingual human annotators |
|
# human.jsonl |
|
human = load_dataset('octanove/mosla', 'human') |
|
|
|
# the machine annotation data, containing machine annotations for ALL sections of video and not human annotations |
|
# machine.jsonl |
|
machine = load_dataset('octanove/mosla', 'machine') |
|
|
|
# annotation record data, where each instance represents a five-minute chunk that was shown to humans for annotation |
|
# this data is useful to understand which sections of the videos were shown to human annotators, but not much else |
|
# annotation_records.jsonl |
|
annotation_records = load_dataset('octanove/mosla', 'annotation_records') |
|
``` |
|
|
|
## Instance Schema |
|
Instances look like this: |
|
```python |
|
{ |
|
"start":1748.4499511719, # beginning of the annotated utterance in the video, in seconds |
|
"end":1758.5999755859, # end of the annotated utterance in the video, in seconds |
|
"video":"20210918.mp4", # name of the annotated video file |
|
"date":1631923200000, # date of the lesson |
|
"lesson_language":"ara", # language being taught in the lesson |
|
"speaker":"s", #the speaker of the utterance |
|
"language":"eng", # the primary language of the utterance |
|
"text":"Oh okay, oh, when you see يلعبوا first, then you don't have to change it?", # utterance content |
|
"memo":"", # annotator memo, if any was provided |
|
"human_annotation":true # true if annotation was generated by a human annotator |
|
} |
|
``` |
|
|
|
Note that annotation records are simpler, and have only the following fields: `("start", "end", "lesson_language", "video")`. |
|
|
|
|
|
## Video Files |
|
The 🤗 dataset only includes the annotations by default, but video files are present in the git repository and can be |
|
downloaded like this: |
|
```python |
|
from huggingface_hub import hf_hub_download |
|
|
|
hf_hub_download(repo_id="octanove/mosla", repo_type="dataset", filename="video/zho/20210205.mp4") |
|
|
|
# or from a specific example |
|
def download_video(ex: dict): |
|
hf_hub_download(repo_id="octanove/mosla", repo_type="dataset", |
|
filename=f"video/{ex['lesson_language']}/{ex['video']}") |
|
``` |
|
|
|
## Test/Train Splits |
|
Right now there are no officially defined test/train splits, but we may add these in the future. |
|
|
|
|
|
# Use Cases / Q&A |
|
|
|
Below we address some common use cases / things that may be unclear just from looking at the data. |
|
|
|
## I just want to analyze text data - what should I use? |
|
|
|
You should probably use the merged machine/human annotations in `merged.jsonl`. This data was used for analysis in the paper, |
|
and contains human data for sections that were human annotated and machine generated annotations for the rest of the data. |
|
|
|
|
|
## I want to train ASR models - what should I use? |
|
|
|
You should probably use the human annotations in `human.jsonl`, unless you want to train on the output of a machine learning model. |
|
|
|
|
|
## What are the annotation records? |
|
|
|
The annotation records contain information about what spans of time were shown to humans for annotation. This can safely be ignored |
|
for most use cases, but becomes important if you want to train or evaluate a system that includes speech detection or diarization, |
|
because this information is necessary to know which periods of time humans _chose not_ to annotate with any utterances. That is, |
|
the contents of `human.jsonl` contain only the utterances that humans chose to annotate, but you need `annotation_records.jsonl` |
|
to know the sections of each video that were shown to humans. For example, with the below annotation record: |
|
```json |
|
{"start":600.0,"end":900.0,"lesson_language":"ara","video":"20210828.mp4"} |
|
``` |
|
From this information, we can say that any period of time between the 600th and 900th seconds in this video for which there is no human |
|
annotation has a gold label/human annotation indicating the absence of speech. |
|
|
|
|
|
|