Papers
arxiv:2406.15487

Improving Text-To-Audio Models with Synthetic Captions

Published on Jun 18
Authors:
,
,
,
,
,

Abstract

It is an open challenge to obtain high quality training data, especially captions, for text-to-audio models. Although prior methods have leveraged text-only language models to augment and improve captions, such methods have limitations related to scale and coherence between audio and captions. In this work, we propose an audio captioning pipeline that uses an audio language model to synthesize accurate and diverse captions for audio at scale. We leverage this pipeline to produce a dataset of synthetic captions for AudioSet, named AF-AudioSet, and then evaluate the benefit of pre-training text-to-audio models on these synthetic captions. Through systematic evaluations on AudioCaps and MusicCaps, we find leveraging our pipeline and synthetic captions leads to significant improvements on audio generation quality, achieving a new state-of-the-art.

Community

Paper author

Sharing our latest work on text-to-audio models.

๐Ÿš€ We propose a data labeling pipeline to generate large-scale high-quality synthetic captions for audio.
๐Ÿš€ We introduce AF-AudioSet: a large, diverse, and high-quality synthetic caption dataset produced with our pipeline.
๐Ÿš€ We obtain state-of-the-art models on text-to-audio and text-to-music through pre-training on AF-AudioSet and conduct
a systematic study across various settings.

๐Ÿ”ฅ We will soon release TangoMusic which obtained great results on the MusicBench benchmark.

Sign up or log in to comment

Models citing this paper 2

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2406.15487 in a dataset README.md to link it from this page.

Spaces citing this paper 1

Collections including this paper 1