Distilled Medium Whisper ASR Model for Thai

Model Description

This is a distilled Automatic Speech Recognition (ASR) model, based on the Whisper architecture. It has been specifically tailored for Thai language speech recognition. The model features 4 decoder layers (vs 24 in teacher model) and has been distilled from a larger teacher model, focusing on enhancing performance and efficiency.

Distillation Details

  • Teacher Model: Medium Whisper ASR model
  • Datasets Used for Distillation:
    • Common Voice v13
    • Gowajee
    • Thai Elderly Speech Corpus
    • Custom Scraped Data
    • Thai-Central Dialect from SLSCU Thai Dialect Corpus

Model Performance

  • DeepCut Tokenized WER on Common Voice 13 Test Set:
    • Distilled Model: 7.58%
    • Teacher Model: 7.42%

Additional datasets for distillation or more decoder layers might improve the WER. More to come soon!

Intended Use

This model is intended for use in applications requiring Thai language speech recognition.

Limitations

  • The model is specifically trained for the Thai language and may not perform well with other languages.
  • Performance might vary across different Thai dialects and accents.
  • As with any ASR system, background noise and speech clarity can impact recognition accuracy.

Acknowledgments

This model was developed using resources and datasets provided by the speech and language technology community. Special thanks to the teams behind Common Voice, Gowajee, SLSCU, and the Thai Elderly Speech Corpus for their valuable datasets.

Framework versions

  • Transformers 4.35.2
  • Pytorch 2.1.2
  • Datasets 2.16.1
  • Tokenizers 0.15.0

Citation

Cite using Bibtex:

@inproceedings{aung-etal-2024-thonburian,
    title = "Thonburian Whisper: Robust Fine-tuned and Distilled Whisper for {T}hai",
    author = "Aung, Zaw Htet  and
      Thavornmongkol, Thanachot  and
      Boribalburephan, Atirut  and
      Tangsriworakan, Vittavas  and
      Pipatsrisawat, Knot  and
      Achakulvisut, Titipat",
    editor = "Abbas, Mourad  and
      Freihat, Abed Alhakim",
    booktitle = "Proceedings of the 7th International Conference on Natural Language and Speech Processing (ICNLSP 2024)",
    month = oct,
    year = "2024",
    address = "Trento",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2024.icnlsp-1.17",
    pages = "149--156",
}

Downloads last month
302
Safetensors
Model size
481M params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for biodatlab/distill-whisper-th-medium

Finetunes
1 model