annotations_creators: []
language_creators:
- crowdsourced
languages:
- ro
- en
licenses:
- cc-by-4-0
multilinguality:
- multilingual
size_categories:
- 10K<n<100K
source_datasets:
- extended|other-sts-b
task_categories:
- conditional-text-generation
task_ids:
- machine-translation
Dataset Card for RO-STS-Parallel
Table of Contents
- Dataset Description
- Dataset Structure
- Dataset Creation
- Considerations for Using the Data
- Additional Information
Dataset Description
- Homepage: GitHub
- Repository: GitHub
- Paper: [Needs More Information]
- Leaderboard: [Needs More Information]
- Point of Contact: [email protected]
Dataset Summary
We present RO-STS-Parallel - a Parallel Romanian-English dataset obtained by translating the STS English dataset dataset into Romanian. It contains 17256 sentences in Romanian and English.
Supported Tasks and Leaderboards
[Needs More Information]
Languages
The text dataset is in Romanian and English (ro
, en
)
Dataset Structure
Data Instances
An example looks like this:
{
'translation': {
'ro': 'Problema e si mai simpla.',
'en': 'The problem is simpler than that.'
}
}
Data Fields
- translation:
- ro: text in Romanian
- en: text in English
Data Splits
The train/validation/test split contain 11,498/3,000/2,758 sentence pairs.
Dataset Creation
Curation Rationale
Source Data
Initial Data Collection and Normalization
*To construct the dataset, we first obtained automatic translations using Google's translation engine. These were then manually checked, corrected, and cross-validated by human volunteers. *
Who are the source language producers?
[Needs More Information]
Annotations
Annotation process
Who are the annotators?
Personal and Sensitive Information
[Needs More Information]
Considerations for Using the Data
Social Impact of Dataset
[Needs More Information]
Discussion of Biases
[Needs More Information]
Other Known Limitations
[Needs More Information]
Additional Information
Dataset Curators
[Needs More Information]
Licensing Information
CC BY-SA 4.0 License
Citation Information
[Needs More Information]
Contributions
Thanks to @lorinczb for adding this dataset.