Text-to-Audio
Inference Endpoints
hollownight commited on
Commit
8c9e37b
·
verified ·
1 Parent(s): 0a3d880

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -0
README.md CHANGED
@@ -44,6 +44,8 @@ TangoFlux: Super Fast and Faithful Text to Audio Generation with Flow Matching a
44
 
45
  </div>
46
 
 
 
47
  ## Model Overview
48
 
49
  TangoFlux consists of FluxTransformer blocks which are Diffusion Transformer (DiT) and Multimodal Diffusion Transformer (MMDiT), conditioned on textual prompt and duration embedding to generate audio at 44.1kHz up to 30 seconds. TangoFlux learns a rectified flow trajectory from audio latent representation encoded by a variational autoencoder (VAE). The TangoFlux training pipeline consists of three stages: pre-training, fine-tuning, and preference optimization. TangoFlux is aligned via CRPO which iteratively generates new synthetic data and constructs preference pairs to perform preference optimization.
@@ -69,6 +71,13 @@ audio = model.generate('Hammer slowly hitting the wooden table', steps=50, durat
69
 
70
  Audio(data=audio, rate=44100)
71
  ```
 
 
 
 
 
 
 
72
  ## Citation
73
 
74
  https://arxiv.org/abs/2412.21037
 
44
 
45
  </div>
46
 
47
+ * Powered by **Stability AI**
48
+
49
  ## Model Overview
50
 
51
  TangoFlux consists of FluxTransformer blocks which are Diffusion Transformer (DiT) and Multimodal Diffusion Transformer (MMDiT), conditioned on textual prompt and duration embedding to generate audio at 44.1kHz up to 30 seconds. TangoFlux learns a rectified flow trajectory from audio latent representation encoded by a variational autoencoder (VAE). The TangoFlux training pipeline consists of three stages: pre-training, fine-tuning, and preference optimization. TangoFlux is aligned via CRPO which iteratively generates new synthetic data and constructs preference pairs to perform preference optimization.
 
71
 
72
  Audio(data=audio, rate=44100)
73
  ```
74
+
75
+ ## License
76
+
77
+ The TangoFlux checkpoints are for non-commercial research use only. They are subject to the [Stable Audio Open’s license](https://huggingface.co/stabilityai/stable-audio-open-1.0/blob/main/LICENSE.md), [WavCap’s license](https://github.com/XinhaoMei/WavCaps?tab=readme-ov-file#license), and the original licenses accompanying each training dataset.
78
+
79
+ This Stability AI Model is licensed under the Stability AI Community License, Copyright © Stability AI Ltd. All Rights Reserved
80
+
81
  ## Citation
82
 
83
  https://arxiv.org/abs/2412.21037