Datasets:
configs:
- config_name: all
data_files:
- path:
- all.txt.zst
split: train
default: true
- config_name: ar
data_files:
- path:
- ar.txt.zst
split: train
- config_name: az
data_files:
- path:
- az.txt.zst
split: train
- config_name: bg
data_files:
- path:
- bg.txt.zst
split: train
- config_name: bn
data_files:
- path:
- bn.txt.zst
split: train
- config_name: ca
data_files:
- path:
- ca.txt.zst
split: train
- config_name: cs
data_files:
- path:
- cs.txt.zst
split: train
- config_name: da
data_files:
- path:
- da.txt.zst
split: train
- config_name: de
data_files:
- path:
- de.txt.zst
split: train
- config_name: el
data_files:
- path:
- el.txt.zst
split: train
- config_name: en
data_files:
- path:
- en.txt.zst
split: train
- config_name: es
data_files:
- path:
- es.txt.zst
split: train
- config_name: et
data_files:
- path:
- et.txt.zst
split: train
- config_name: fa
data_files:
- path:
- fa.txt.zst
split: train
- config_name: fi
data_files:
- path:
- fi.txt.zst
split: train
- config_name: fr
data_files:
- path:
- fr.txt.zst
split: train
- config_name: he
data_files:
- path:
- he.txt.zst
split: train
- config_name: hi
data_files:
- path:
- hi.txt.zst
split: train
- config_name: hu
data_files:
- path:
- hu.txt.zst
split: train
- config_name: hy
data_files:
- path:
- hy.txt.zst
split: train
- config_name: id
data_files:
- path:
- id.txt.zst
split: train
- config_name: is
data_files:
- path:
- is.txt.zst
split: train
- config_name: it
data_files:
- path:
- it.txt.zst
split: train
- config_name: ja
data_files:
- path:
- ja.txt.zst
split: train
- config_name: ka
data_files:
- path:
- ka.txt.zst
split: train
- config_name: kk
data_files:
- path:
- kk.txt.zst
split: train
- config_name: ko
data_files:
- path:
- ko.txt.zst
split: train
- config_name: lt
data_files:
- path:
- lt.txt.zst
split: train
- config_name: lv
data_files:
- path:
- lv.txt.zst
split: train
- config_name: mk
data_files:
- path:
- mk.txt.zst
split: train
- config_name: ml
data_files:
- path:
- ml.txt.zst
split: train
- config_name: mr
data_files:
- path:
- mr.txt.zst
split: train
- config_name: ne
data_files:
- path:
- ne.txt.zst
split: train
- config_name: nl
data_files:
- path:
- nl.txt.zst
split: train
- config_name: 'no'
data_files:
- path:
- no.txt.zst
split: train
- config_name: pl
data_files:
- path:
- pl.txt.zst
split: train
- config_name: pt
data_files:
- path:
- pt.txt.zst
split: train
- config_name: ro
data_files:
- path:
- ro.txt.zst
split: train
- config_name: ru
data_files:
- path:
- ru.txt.zst
split: train
- config_name: sk
data_files:
- path:
- sk.txt.zst
split: train
- config_name: sl
data_files:
- path:
- sl.txt.zst
split: train
- config_name: sq
data_files:
- path:
- sq.txt.zst
split: train
- config_name: sr
data_files:
- path:
- sr.txt.zst
split: train
- config_name: sv
data_files:
- path:
- sv.txt.zst
split: train
- config_name: ta
data_files:
- path:
- ta.txt.zst
split: train
- config_name: th
data_files:
- path:
- th.txt.zst
split: train
- config_name: tr
data_files:
- path:
- tr.txt.zst
split: train
- config_name: uk
data_files:
- path:
- uk.txt.zst
split: train
- config_name: ur
data_files:
- path:
- ur.txt.zst
split: train
- config_name: vi
data_files:
- path:
- vi.txt.zst
split: train
- config_name: zh
data_files:
- path:
- zh.txt.zst
split: train
language:
- multilingual
- ar
- az
- bg
- bn
- ca
- cs
- da
- de
- el
- en
- es
- et
- fa
- fi
- fr
- he
- hi
- hu
- hy
- id
- is
- it
- ja
- ka
- kk
- ko
- lt
- lv
- mk
- ml
- mr
- ne
- nl
- 'no'
- pl
- pt
- ro
- ru
- sk
- sl
- sq
- sr
- sv
- ta
- th
- tr
- uk
- ur
- vi
- zh
task_categories:
- text-generation
- text-classification
- text-retrieval
size_categories:
- 1M<n<10M
Multilingual Sentences
Dataset contains sentences from 50 languages, grouped by their two-letter ISO 639-1 codes. The "all" configuration includes sentences from all languages.
Dataset Overview
Multilingual Sentence Dataset is a comprehensive collection of high-quality, linguistically diverse sentences. Dataset is designed to support a wide range of natural language processing tasks, including but not limited to language modeling, machine translation, and cross-lingual studies.
Methods
Rigorous methodology consisted of three main stages: text preprocessing, language detection, and dataset processing.
Text Preprocessing
Sophisticated text cleaning pipeline using the textacy library, which included:
- Removal of HTML tags, email addresses, URLs, and emojis
- Unicode and whitespace normalization
- Standardization of punctuation and word formats
Language Detection
Google CLD3 library utilized for accurate language identification:
- Implemented NNetLanguageIdentifier
- Configured for processing texts between 0-1000 bytes
- Included reliability assessment for each language detection
Dataset Processing
Workflow for dataset creation involved the following steps:
- Streamed loading of the LinguaNova multilingual dataset
- Application of the text preprocessing pipeline
- Sentence segmentation using PyICU for accurate boundary detection
- Quality filtering:
- Length constraint (maximum 2048 characters per sentence)
- High-reliability language verification
- Extraction of unique sentences
- Random shuffling for unbiased sampling
- Generation of language-specific files
Technical Details
Libraries and Tools
- textacy: Advanced text preprocessing
- Google CLD3: State-of-the-art language detection
- Hugging Face datasets: Efficient data handling and processing
- SentenceBreaker: Accurate sentence segmentation
Implementation Notes
- Process was executed consistently across all 50 languages to ensure uniformity and high quality in the multilingual dataset preparation.
- Special attention was given to maintaining the integrity of each language's unique characteristics throughout the processing pipeline.
Data Splits
Dataset is organized into the following splits:
- Individual language files: Contains sentences for each of the 50 languages
- "all" configuration: Aggregates sentences from all languages into a single dataset
Limitations and Biases
While extensive efforts were made to ensure dataset quality, users should be aware of potential limitations:
- Language detection accuracy may vary for very short texts or closely related languages
- Dataset may not fully represent all dialects or regional variations within each language
- Potential biases in the original LinguaNova dataset could be carried over
Ethical Considerations
Users of this dataset should be mindful of:
- Potential biases in language representation
- Need for responsible use in AI applications, especially in multilingual contexts
- Privacy considerations, although personal identifiable information should have been removed