--- license: apache-2.0 language: - en pipeline_tag: text-generation inference: false fine-tuning: true tags: - generative error correction - large language model - LLaMA metrics: - wer datasets: - PeacefulData/Robust-HyPoradise --- This repo releases the trained LLaMA-adapter weights in paper "Large Language Models are Efficient Learners of Noise-Robust Speech Recognition." GitHub: https://github.com/YUCHEN005/RobustGER Data: https://huggingface.co/datasets/PeacefulData/Robust-HyPoradise Model: This repo If you consider this work would be related or useful for your research, please kindly consider to cite the work in ICLR 2024. Thank you. ```bib @inproceedings{hu2024large, title={Large Language Models are Efficient Learners of Noise-Robust Speech Recognition}, author={Hu, Yuchen and Chen, Chen and Yang, Chao-Han Huck and Li, Ruizhe and Zhang, Chao and Chen, Pin-Yu and Chng, Eng Siong}, booktitle={International Conference on Learning Representations}, year={2024} } ```