Datasets:

ArXiv:
License:
beaye_lexicon / beaye_lexicon.py
holylovenia's picture
Upload beaye_lexicon.py with huggingface_hub
8bbddd9 verified
from pathlib import Path
from typing import Dict, List, Tuple
import datasets
import pandas as pd
from seacrowd.utils.configs import SEACrowdConfig
from seacrowd.utils.constants import Licenses
_CITATION = """\
@misc{lopo2024constructing,
title={Constructing and Expanding Low-Resource and Underrepresented Parallel Datasets for Indonesian Local Languages},
author={Joanito Agili Lopo and Radius Tanone},
year={2024},
eprint={2404.01009},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
"""
_DATASETNAME = "beaye_lexicon"
_DESCRIPTION = """The Beaye Lexicon is a lexicon resource encompassing translations between Indonesian, English, and
Beaye words. Developed through a collaborative effort involving two native Beaye speakers and evaluated by linguistic
experts, this lexicon comprises 984 Beaye vocabularies. The creation of the Beaye Lexicon marks the inaugural effort in
documenting the previously unrecorded Beaye language."""
_HOMEPAGE = "https://github.com/joanitolopo/bhinneka-korpus/tree/main/lexicon"
_LICENSE = Licenses.APACHE_2_0.value
_URLS = "https://raw.githubusercontent.com/joanitolopo/bhinneka-korpus/main/lexicon"
_SUPPORTED_TASKS = []
_SOURCE_VERSION = "1.0.0"
_SEACROWD_VERSION = "2024.06.20"
_LOCAL = False
_LANGUAGES = ["ind", "day", "eng"]
class BeayeLexicon(datasets.GeneratorBasedBuilder):
"""Beaye Lexicon is a lexicon resource encompassing translations between Indonesian, English, and Beaye words"""
SOURCE_VERSION = datasets.Version(_SOURCE_VERSION)
SEACROWD_VERSION = datasets.Version(_SEACROWD_VERSION)
BUILDER_CONFIGS = (
[
SEACrowdConfig(
name=f"{_DATASETNAME}_{lang}_source",
version=datasets.Version(_SOURCE_VERSION),
description=f"beaye lexicon with source schema for {lang} language",
schema="source",
subset_id="beaye_lexicon",
)
for lang in _LANGUAGES if lang != "eng"
]
+ [
SEACrowdConfig(
name=f"{_DATASETNAME}_ext_{lang}_source",
version=datasets.Version(_SOURCE_VERSION),
description=f"beaye lexicon with source schema for extensive definiton of beaye language",
schema="source",
subset_id="beaye_lexicon",
)
for lang in _LANGUAGES if lang != "ind"
]
)
DEFAULT_CONFIG_NAME = f"{_DATASETNAME}_ind_source"
def _info(self) -> datasets.DatasetInfo:
schema = self.config.schema
if schema == "source":
features = datasets.Features({"id": datasets.Value("string"), "word": datasets.Value("string")})
else:
raise NotImplementedError()
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
homepage=_HOMEPAGE,
license=_LICENSE,
citation=_CITATION,
)
def _split_generators(self, dl_manager: datasets.DownloadManager) -> List[datasets.SplitGenerator]:
"""Returns SplitGenerators."""
if "ext" in self.config.name.split("_"):
data_dir = Path(dl_manager.download(_URLS + "/english.xlsx"))
else:
data_dir = Path(dl_manager.download(_URLS + "/lexicon.xlsx"))
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={
"filepath": data_dir,
"split": "train",
}
)
]
def _generate_examples(self, filepath: Path, split: str) -> Tuple[int, Dict]:
"""Yields examples as (key, example) tuples."""
dfs = pd.read_excel(filepath, engine="openpyxl")
if "ext" in self.config.name.split("_"):
lang = self.config.name.split("_")[3]
else:
lang = self.config.name.split("_")[2]
text = dfs[lang]
if self.config.schema == "source":
for idx, word in enumerate(text.values):
row = {"id": str(idx), "word": word}
yield idx, row
else:
raise ValueError(f"Invalid config: {self.config.name}")