|
--- |
|
license: apache-2.0 |
|
language: |
|
- en |
|
pipeline_tag: text-generation |
|
inference: false |
|
fine-tuning: true |
|
tags: |
|
- generative error correction |
|
- large language model |
|
- LLaMA |
|
metrics: |
|
- wer |
|
datasets: |
|
- PeacefulData/Robust-HyPoradise |
|
--- |
|
|
|
This repo releases the trained LLaMA-adapter weights in paper "Large Language Models are Efficient Learners of Noise-Robust Speech Recognition." |
|
|
|
**GitHub:** https://github.com/YUCHEN005/RobustGER |
|
|
|
**Data:** https://huggingface.co/datasets/PeacefulData/Robust-HyPoradise |
|
|
|
**Model:** This repo |
|
|
|
If you consider this work would be related or useful for your research, please kindly consider to cite the work in ICLR 2024. Thank you. |
|
|
|
```bib |
|
@inproceedings{hu2024large, |
|
title={Large Language Models are Efficient Learners of Noise-Robust Speech Recognition}, |
|
author={Hu, Yuchen and Chen, Chen and Yang, Chao-Han Huck and Li, Ruizhe and Zhang, Chao and Chen, Pin-Yu and Chng, Eng Siong}, |
|
booktitle={International Conference on Learning Representations}, |
|
year={2024} |
|
} |
|
``` |