PHBJT commited on
Commit
3dbff6e
·
verified ·
1 Parent(s): f0ab368

Upload fine-tuned XTTS v2 model

Browse files
Files changed (6) hide show
  1. README.md +31 -0
  2. config.json +201 -0
  3. dvae.pth +3 -0
  4. mel_stats.pth +3 -0
  5. model.pth +3 -0
  6. vocab.json +0 -0
README.md ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # XTTS v2 Fine-tuned Model
3
+
4
+ This is a fine-tuned version of the XTTS v2 text-to-speech model.
5
+
6
+ ## Model Details
7
+ - Base Model: XTTS v2
8
+ - Fine-tuning Date: 2024-12-01
9
+ - Repository: PHBJT/replicate_test_01
10
+
11
+ ## Usage
12
+
13
+ ```python
14
+ from TTS.api import TTS
15
+
16
+ # Initialize TTS with the custom model
17
+ tts = TTS(model_path="path_to_model_folder", config_path="path_to_config.json")
18
+
19
+ # Generate speech
20
+ tts.tts_to_file(text="Your text here",
21
+ speaker_wav="path_to_speaker.wav",
22
+ language="en",
23
+ file_path="output.wav")
24
+ ```
25
+
26
+ ## Files
27
+ - model.pth: Main model weights
28
+ - config.json: Model configuration
29
+ - vocab.json: Tokenizer vocabulary
30
+ - mel_stats.pth: Mel-spectrogram statistics
31
+ - dvae.pth: Discrete VAE model
config.json ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "/src/run/training",
3
+ "logger_uri": null,
4
+ "run_name": "GPT_XTTS_v2_FT",
5
+ "project_name": "XTTS_v2_trainer",
6
+ "run_description": "GPT XTTS v2 fine-tuning",
7
+ "print_step": 50,
8
+ "plot_step": 100,
9
+ "model_param_stats": false,
10
+ "wandb_entity": null,
11
+ "dashboard_logger": "tensorboard",
12
+ "save_on_interrupt": true,
13
+ "log_model_step": 100,
14
+ "save_step": 10000,
15
+ "save_n_checkpoints": 1,
16
+ "save_checkpoints": true,
17
+ "save_all_best": false,
18
+ "save_best_after": 0,
19
+ "target_loss": null,
20
+ "print_eval": false,
21
+ "test_delay_epochs": 0,
22
+ "run_eval": true,
23
+ "run_eval_steps": null,
24
+ "distributed_backend": "nccl",
25
+ "distributed_url": "tcp://localhost:54321",
26
+ "mixed_precision": false,
27
+ "precision": "fp16",
28
+ "epochs": 10,
29
+ "batch_size": 4,
30
+ "eval_batch_size": 4,
31
+ "grad_clip": 0.0,
32
+ "scheduler_after_epoch": true,
33
+ "lr": 5e-06,
34
+ "optimizer": "AdamW",
35
+ "optimizer_params": {
36
+ "betas": [
37
+ 0.9,
38
+ 0.96
39
+ ],
40
+ "eps": 1e-08,
41
+ "weight_decay": 0.01
42
+ },
43
+ "lr_scheduler": "MultiStepLR",
44
+ "lr_scheduler_params": {
45
+ "milestones": [
46
+ 900000,
47
+ 2700000,
48
+ 5400000
49
+ ],
50
+ "gamma": 0.5,
51
+ "last_epoch": -1
52
+ },
53
+ "use_grad_scaler": false,
54
+ "allow_tf32": false,
55
+ "cudnn_enable": true,
56
+ "cudnn_deterministic": false,
57
+ "cudnn_benchmark": false,
58
+ "training_seed": 1,
59
+ "model": "xtts",
60
+ "num_loader_workers": 8,
61
+ "num_eval_loader_workers": 0,
62
+ "use_noise_augment": false,
63
+ "audio": {
64
+ "sample_rate": 22050,
65
+ "output_sample_rate": 24000,
66
+ "dvae_sample_rate": 22050
67
+ },
68
+ "use_phonemes": false,
69
+ "phonemizer": null,
70
+ "phoneme_language": null,
71
+ "compute_input_seq_cache": false,
72
+ "text_cleaner": null,
73
+ "enable_eos_bos_chars": false,
74
+ "test_sentences_file": "",
75
+ "phoneme_cache_path": null,
76
+ "characters": null,
77
+ "add_blank": false,
78
+ "batch_group_size": 48,
79
+ "loss_masking": null,
80
+ "min_audio_len": 1,
81
+ "max_audio_len": Infinity,
82
+ "min_text_len": 1,
83
+ "max_text_len": Infinity,
84
+ "compute_f0": false,
85
+ "compute_energy": false,
86
+ "compute_linear_spec": false,
87
+ "precompute_num_workers": 0,
88
+ "start_by_longest": false,
89
+ "shuffle": false,
90
+ "drop_last": false,
91
+ "datasets": [
92
+ {
93
+ "formatter": "",
94
+ "dataset_name": "",
95
+ "path": "",
96
+ "meta_file_train": "",
97
+ "ignored_speakers": null,
98
+ "language": "",
99
+ "phonemizer": "",
100
+ "meta_file_val": "",
101
+ "meta_file_attn_mask": ""
102
+ }
103
+ ],
104
+ "test_sentences": [
105
+ {
106
+ "text": "It took me quite a long time to develop a voice, and now that I have it I'm not going to be silent.",
107
+ "speaker_wav": "/tmp/tmp443oojbjQS_Train_VO_241113__Ivan__ArbreDeVie.wav",
108
+ "language": "fr"
109
+ },
110
+ {
111
+ "text": "This cake is great. It's so delicious and moist.",
112
+ "speaker_wav": "/tmp/tmp443oojbjQS_Train_VO_241113__Ivan__ArbreDeVie.wav",
113
+ "language": "fr"
114
+ }
115
+ ],
116
+ "eval_split_max_size": 32,
117
+ "eval_split_size": 0.01,
118
+ "use_speaker_weighted_sampler": false,
119
+ "speaker_weighted_sampler_alpha": 1.0,
120
+ "use_language_weighted_sampler": false,
121
+ "language_weighted_sampler_alpha": 1.0,
122
+ "use_length_weighted_sampler": false,
123
+ "length_weighted_sampler_alpha": 1.0,
124
+ "model_args": {
125
+ "gpt_batch_size": 1,
126
+ "enable_redaction": false,
127
+ "kv_cache": true,
128
+ "gpt_checkpoint": "",
129
+ "clvp_checkpoint": null,
130
+ "decoder_checkpoint": null,
131
+ "num_chars": 255,
132
+ "tokenizer_file": "/src/run/training/XTTS_v2_original_model_files/vocab.json",
133
+ "gpt_max_audio_tokens": 605,
134
+ "gpt_max_text_tokens": 402,
135
+ "gpt_max_prompt_tokens": 70,
136
+ "gpt_layers": 30,
137
+ "gpt_n_model_channels": 1024,
138
+ "gpt_n_heads": 16,
139
+ "gpt_number_text_tokens": 6681,
140
+ "gpt_start_text_token": 261,
141
+ "gpt_stop_text_token": 0,
142
+ "gpt_num_audio_tokens": 1026,
143
+ "gpt_start_audio_token": 1024,
144
+ "gpt_stop_audio_token": 1025,
145
+ "gpt_code_stride_len": 1024,
146
+ "gpt_use_masking_gt_prompt_approach": true,
147
+ "gpt_use_perceiver_resampler": false,
148
+ "input_sample_rate": 22050,
149
+ "output_sample_rate": 24000,
150
+ "output_hop_length": 256,
151
+ "decoder_input_dim": 1024,
152
+ "d_vector_dim": 512,
153
+ "cond_d_vector_in_each_upsampling_layer": true,
154
+ "duration_const": 102400,
155
+ "min_conditioning_length": 66150,
156
+ "max_conditioning_length": 132300,
157
+ "gpt_loss_text_ce_weight": 0.01,
158
+ "gpt_loss_mel_ce_weight": 1.0,
159
+ "debug_loading_failures": false,
160
+ "max_wav_length": 255995,
161
+ "max_text_length": 200,
162
+ "mel_norm_file": "/src/run/training/XTTS_v2_original_model_files/mel_stats.pth",
163
+ "dvae_checkpoint": "/src/run/training/XTTS_v2_original_model_files/dvae.pth",
164
+ "xtts_checkpoint": "/src/run/training/XTTS_v2_original_model_files/model.pth",
165
+ "vocoder": ""
166
+ },
167
+ "model_dir": null,
168
+ "languages": [
169
+ "en",
170
+ "es",
171
+ "fr",
172
+ "de",
173
+ "it",
174
+ "pt",
175
+ "pl",
176
+ "tr",
177
+ "ru",
178
+ "nl",
179
+ "cs",
180
+ "ar",
181
+ "zh-cn",
182
+ "hu",
183
+ "ko",
184
+ "ja",
185
+ "hi"
186
+ ],
187
+ "temperature": 0.85,
188
+ "length_penalty": 1.0,
189
+ "repetition_penalty": 2.0,
190
+ "top_k": 50,
191
+ "top_p": 0.85,
192
+ "num_gpt_outputs": 1,
193
+ "gpt_cond_len": 12,
194
+ "gpt_cond_chunk_len": 4,
195
+ "max_ref_len": 10,
196
+ "sound_norm_refs": false,
197
+ "optimizer_wd_only_on_weights": true,
198
+ "weighted_loss_attrs": null,
199
+ "weighted_loss_multipliers": null,
200
+ "github_branch": "inside_docker"
201
+ }
dvae.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b29bc227d410d4991e0a8c09b858f77415013eeb9fba9650258e96095557d97a
3
+ size 210514388
mel_stats.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f69422a8a8f344c4fca2f0c6b8d41d2151d6615b7321e48e6bb15ae949b119c
3
+ size 1067
model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb4acb15efaa13659d20af8c295cf190547ae6a09dbd2e423220aeaaa2adc66e
3
+ size 5361163503
vocab.json ADDED
The diff for this file is too large to render. See raw diff