|
--- |
|
language: |
|
- vi |
|
tags: |
|
- audio |
|
- automatic-speech-recognition |
|
- hf-asr-leaderboard |
|
library_name: transformers.js |
|
pipeline_tag: automatic-speech-recognition |
|
widget: |
|
- example_title: Librispeech sample 1 |
|
src: https://cdn-media.huggingface.co/speech_samples/sample1.flac |
|
- example_title: Librispeech sample 2 |
|
src: https://cdn-media.huggingface.co/speech_samples/sample2.flac |
|
license: wtfpl |
|
--- |
|
https://hf.co/vinai/PhoWhisper-tiny with ONNX weights to be compatible with Transformers.js. |
|
|
|
Please check out this demo using the model: |
|
[![Open in Spaces](https://huggingface.co/datasets/huggingface/badges/resolve/main/open-in-hf-spaces-md-dark.svg)](https://huggingface.co/spaces/huuquyet/PhoWhisper-next) |
|
|
|
|
|
# PhoWhisper: Automatic Speech Recognition for Vietnamese |
|
|
|
|
|
We introduce **PhoWhisper** in five versions for Vietnamese automatic speech recognition. PhoWhisper's robustness is achieved through fine-tuning the multilingual [Whisper](https://github.com/openai/whisper) on an 844-hour dataset that encompasses diverse Vietnamese accents. Our experimental study demonstrates state-of-the-art performances of PhoWhisper on benchmark Vietnamese ASR datasets. Please **cite** our PhoWhisper paper when it is used to help produce published results or is incorporated into other software: |
|
|
|
``` |
|
@inproceedings{PhoWhisper, |
|
title = {{PhoWhisper: Automatic Speech Recognition for Vietnamese}}, |
|
author = {Thanh-Thien Le and Linh The Nguyen and Dat Quoc Nguyen}, |
|
booktitle = {Proceedings of the ICLR 2024 Tiny Papers track}, |
|
year = {2024} |
|
} |
|
``` |
|
|
|
For further information or requests, please go to [PhoWhisper's homepage](https://github.com/VinAIResearch/PhoWhisper)! |
|
|