Upload 8 files
Browse files- .gitattributes +3 -0
- loras/ckpt/lora-glados-r128-a128/lora.sft +3 -0
- loras/ckpt/lora-max-r128-a128/lora.sft +3 -0
- loras/ckpt/lora-sam-r128-a128/lora.sft +3 -0
- loras/config.lora.glados.yaml +130 -0
- loras/config.lora.max.yaml +130 -0
- loras/config.lora.sam.yaml +130 -0
- models/ckpt/ar+nar-tts+stt-llama-8/fp32.sft +1 -1
- models/config.llama.tts+stt.yaml +8 -12
.gitattributes
CHANGED
@@ -7,3 +7,6 @@ models/experiments/ckpt/nar-len-llama-9/fp32.pth filter=lfs diff=lfs merge=lfs -
|
|
7 |
models/ckpt/ar+nar-llama-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
|
8 |
models/ckpt/ar+nar-retnet-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
|
9 |
models/ckpt/ar+nar-tts+stt-llama-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
7 |
models/ckpt/ar+nar-llama-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
|
8 |
models/ckpt/ar+nar-retnet-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
|
9 |
models/ckpt/ar+nar-tts+stt-llama-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
|
10 |
+
loras/ckpt/lora-glados-r128-a128/lora.sft filter=lfs diff=lfs merge=lfs -text
|
11 |
+
loras/ckpt/lora-max-r128-a128/lora.sft filter=lfs diff=lfs merge=lfs -text
|
12 |
+
loras/ckpt/lora-sam-r128-a128/lora.sft filter=lfs diff=lfs merge=lfs -text
|
loras/ckpt/lora-glados-r128-a128/lora.sft
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6e9bec9574b8a074c9a87f8997b27706f4f3611efa12d26dc9763cd2bca9eb9e
|
3 |
+
size 66076312
|
loras/ckpt/lora-max-r128-a128/lora.sft
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e8065b17f981dba303af2e5c92275902a45e8e38e9eea578f0a9e4b50690f1e
|
3 |
+
size 66076312
|
loras/ckpt/lora-sam-r128-a128/lora.sft
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:433ada1d6f9cba5ebce7917bb51df09f050f4d221afdc4cdad72964b6624f9fc
|
3 |
+
size 66076312
|
loras/config.lora.glados.yaml
ADDED
@@ -0,0 +1,130 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
sample_rate: 24_000
|
2 |
+
audio_backend: "vocos"
|
3 |
+
|
4 |
+
models:
|
5 |
+
- name: "ar+nar-tts+stt"
|
6 |
+
size: "full"
|
7 |
+
resp_levels: 8
|
8 |
+
prom_levels: 8
|
9 |
+
tasks: 9
|
10 |
+
langs: 2
|
11 |
+
tones: 1
|
12 |
+
arch_type: llama
|
13 |
+
training: False
|
14 |
+
version: 5
|
15 |
+
attention: auto
|
16 |
+
dropout: 0.1
|
17 |
+
#loss_factors:
|
18 |
+
# text: 0.01
|
19 |
+
# prom: 0.5
|
20 |
+
# resp: 1.0
|
21 |
+
capabilities: ["ar", "nar"]
|
22 |
+
experimental:
|
23 |
+
p_rvq_levels: "auto"
|
24 |
+
audio_embedding_sums: True
|
25 |
+
unified_position_ids: False
|
26 |
+
split_classifiers: True
|
27 |
+
#
|
28 |
+
causal_size: 1
|
29 |
+
interleave: False
|
30 |
+
rvq_level_range: []
|
31 |
+
tie_classifier_to_embedding: False
|
32 |
+
|
33 |
+
loras:
|
34 |
+
- name : "lora-glados"
|
35 |
+
rank: 128
|
36 |
+
alpha: 128
|
37 |
+
training: True
|
38 |
+
rvq_levels: []
|
39 |
+
|
40 |
+
hyperparameters:
|
41 |
+
batch_size: 32
|
42 |
+
gradient_accumulation_steps: 8
|
43 |
+
gradient_clipping: 1.0
|
44 |
+
warmup_steps: 10
|
45 |
+
|
46 |
+
optimizer: Prodigy
|
47 |
+
learning_rate: 1.0
|
48 |
+
torch_optimizer: True
|
49 |
+
|
50 |
+
scheduler: "" # ScheduleFree
|
51 |
+
torch_scheduler: True
|
52 |
+
|
53 |
+
evaluation:
|
54 |
+
batch_size: 4
|
55 |
+
frequency: 250
|
56 |
+
size: 4
|
57 |
+
|
58 |
+
steps: 500
|
59 |
+
ar_temperature: 1.0
|
60 |
+
nar_temperature: 0.0
|
61 |
+
|
62 |
+
trainer:
|
63 |
+
iterations: 1_000_000
|
64 |
+
save_frequency: 250
|
65 |
+
keep_last_checkpoints: 4
|
66 |
+
|
67 |
+
resize_modules: True
|
68 |
+
|
69 |
+
check_for_oom: False
|
70 |
+
gradient_checkpointing: True
|
71 |
+
|
72 |
+
weight_dtype: bfloat16
|
73 |
+
amp: True
|
74 |
+
|
75 |
+
backend: deepspeed
|
76 |
+
deepspeed:
|
77 |
+
inferencing: False
|
78 |
+
amp: False
|
79 |
+
|
80 |
+
load_webui: False
|
81 |
+
|
82 |
+
inference:
|
83 |
+
backend: local
|
84 |
+
normalize: False
|
85 |
+
|
86 |
+
weight_dtype: bfloat16
|
87 |
+
amp: True
|
88 |
+
|
89 |
+
optimizations:
|
90 |
+
injects: False
|
91 |
+
replace: True
|
92 |
+
|
93 |
+
linear: False
|
94 |
+
embedding: False
|
95 |
+
optimizers: True
|
96 |
+
|
97 |
+
bitsandbytes: False
|
98 |
+
dadaptation: False
|
99 |
+
bitnet: False
|
100 |
+
fp8: False
|
101 |
+
|
102 |
+
dataset:
|
103 |
+
use_hdf5: True
|
104 |
+
hdf5_flag: r
|
105 |
+
|
106 |
+
use_metadata: True
|
107 |
+
validate: True
|
108 |
+
|
109 |
+
workers: 1
|
110 |
+
cache: True
|
111 |
+
|
112 |
+
duration_range: [3.0, 12.0]
|
113 |
+
|
114 |
+
random_utterance: 1.0
|
115 |
+
max_prompts: 1
|
116 |
+
prompt_duration_range: [3.0, 3.0]
|
117 |
+
|
118 |
+
max_resps: 1
|
119 |
+
p_resp_append: 0.25
|
120 |
+
|
121 |
+
sample_type: path # path # speaker
|
122 |
+
sample_order: duration
|
123 |
+
sample_max_duration_batch: 300
|
124 |
+
sample_shuffle: False
|
125 |
+
|
126 |
+
tasks_list: [ "tts", "stt" ]
|
127 |
+
|
128 |
+
training: []
|
129 |
+
validation: []
|
130 |
+
noise: []
|
loras/config.lora.max.yaml
ADDED
@@ -0,0 +1,130 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
sample_rate: 24_000
|
2 |
+
audio_backend: "vocos"
|
3 |
+
|
4 |
+
models:
|
5 |
+
- name: "ar+nar-tts+stt"
|
6 |
+
size: "full"
|
7 |
+
resp_levels: 8
|
8 |
+
prom_levels: 8
|
9 |
+
tasks: 9
|
10 |
+
langs: 2
|
11 |
+
tones: 1
|
12 |
+
arch_type: llama
|
13 |
+
training: False
|
14 |
+
version: 5
|
15 |
+
attention: auto
|
16 |
+
dropout: 0.1
|
17 |
+
#loss_factors:
|
18 |
+
# text: 0.01
|
19 |
+
# prom: 0.5
|
20 |
+
# resp: 1.0
|
21 |
+
capabilities: ["ar", "nar"]
|
22 |
+
experimental:
|
23 |
+
p_rvq_levels: "auto"
|
24 |
+
audio_embedding_sums: True
|
25 |
+
unified_position_ids: False
|
26 |
+
split_classifiers: True
|
27 |
+
#
|
28 |
+
causal_size: 1
|
29 |
+
interleave: False
|
30 |
+
rvq_level_range: []
|
31 |
+
tie_classifier_to_embedding: False
|
32 |
+
|
33 |
+
loras:
|
34 |
+
- name : "lora-max"
|
35 |
+
rank: 128
|
36 |
+
alpha: 128
|
37 |
+
training: True
|
38 |
+
rvq_levels: []
|
39 |
+
|
40 |
+
hyperparameters:
|
41 |
+
batch_size: 32
|
42 |
+
gradient_accumulation_steps: 8
|
43 |
+
gradient_clipping: 1.0
|
44 |
+
warmup_steps: 10
|
45 |
+
|
46 |
+
optimizer: Prodigy
|
47 |
+
learning_rate: 1.0
|
48 |
+
torch_optimizer: True
|
49 |
+
|
50 |
+
scheduler: "" # ScheduleFree
|
51 |
+
torch_scheduler: True
|
52 |
+
|
53 |
+
evaluation:
|
54 |
+
batch_size: 4
|
55 |
+
frequency: 250
|
56 |
+
size: 4
|
57 |
+
|
58 |
+
steps: 500
|
59 |
+
ar_temperature: 1.0
|
60 |
+
nar_temperature: 0.0
|
61 |
+
|
62 |
+
trainer:
|
63 |
+
iterations: 1_000_000
|
64 |
+
save_frequency: 250
|
65 |
+
keep_last_checkpoints: 4
|
66 |
+
|
67 |
+
resize_modules: True
|
68 |
+
|
69 |
+
check_for_oom: False
|
70 |
+
gradient_checkpointing: True
|
71 |
+
|
72 |
+
weight_dtype: bfloat16
|
73 |
+
amp: True
|
74 |
+
|
75 |
+
backend: deepspeed
|
76 |
+
deepspeed:
|
77 |
+
inferencing: False
|
78 |
+
amp: False
|
79 |
+
|
80 |
+
load_webui: False
|
81 |
+
|
82 |
+
inference:
|
83 |
+
backend: local
|
84 |
+
normalize: False
|
85 |
+
|
86 |
+
weight_dtype: bfloat16
|
87 |
+
amp: True
|
88 |
+
|
89 |
+
optimizations:
|
90 |
+
injects: False
|
91 |
+
replace: True
|
92 |
+
|
93 |
+
linear: False
|
94 |
+
embedding: False
|
95 |
+
optimizers: True
|
96 |
+
|
97 |
+
bitsandbytes: False
|
98 |
+
dadaptation: False
|
99 |
+
bitnet: False
|
100 |
+
fp8: False
|
101 |
+
|
102 |
+
dataset:
|
103 |
+
use_hdf5: True
|
104 |
+
hdf5_flag: r
|
105 |
+
|
106 |
+
use_metadata: True
|
107 |
+
validate: True
|
108 |
+
|
109 |
+
workers: 1
|
110 |
+
cache: True
|
111 |
+
|
112 |
+
duration_range: [3.0, 12.0]
|
113 |
+
|
114 |
+
random_utterance: 1.0
|
115 |
+
max_prompts: 1
|
116 |
+
prompt_duration_range: [3.0, 3.0]
|
117 |
+
|
118 |
+
max_resps: 1
|
119 |
+
p_resp_append: 0.25
|
120 |
+
|
121 |
+
sample_type: path # path # speaker
|
122 |
+
sample_order: duration
|
123 |
+
sample_max_duration_batch: 300
|
124 |
+
sample_shuffle: False
|
125 |
+
|
126 |
+
tasks_list: [ "tts", "stt" ]
|
127 |
+
|
128 |
+
training: []
|
129 |
+
validation: []
|
130 |
+
noise: []
|
loras/config.lora.sam.yaml
ADDED
@@ -0,0 +1,130 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
sample_rate: 24_000
|
2 |
+
audio_backend: "vocos"
|
3 |
+
|
4 |
+
models:
|
5 |
+
- name: "ar+nar-tts+stt"
|
6 |
+
size: "full"
|
7 |
+
resp_levels: 8
|
8 |
+
prom_levels: 8
|
9 |
+
tasks: 9
|
10 |
+
langs: 2
|
11 |
+
tones: 1
|
12 |
+
arch_type: llama
|
13 |
+
training: False
|
14 |
+
version: 5
|
15 |
+
attention: auto
|
16 |
+
dropout: 0.1
|
17 |
+
#loss_factors:
|
18 |
+
# text: 0.01
|
19 |
+
# prom: 0.5
|
20 |
+
# resp: 1.0
|
21 |
+
capabilities: ["ar", "nar"]
|
22 |
+
experimental:
|
23 |
+
p_rvq_levels: "auto"
|
24 |
+
audio_embedding_sums: True
|
25 |
+
unified_position_ids: False
|
26 |
+
split_classifiers: True
|
27 |
+
#
|
28 |
+
causal_size: 1
|
29 |
+
interleave: False
|
30 |
+
rvq_level_range: []
|
31 |
+
tie_classifier_to_embedding: False
|
32 |
+
|
33 |
+
loras:
|
34 |
+
- name : "lora-sam"
|
35 |
+
rank: 128
|
36 |
+
alpha: 128
|
37 |
+
training: True
|
38 |
+
rvq_levels: []
|
39 |
+
|
40 |
+
hyperparameters:
|
41 |
+
batch_size: 32
|
42 |
+
gradient_accumulation_steps: 8
|
43 |
+
gradient_clipping: 1.0
|
44 |
+
warmup_steps: 10
|
45 |
+
|
46 |
+
optimizer: Prodigy
|
47 |
+
learning_rate: 1.0
|
48 |
+
torch_optimizer: True
|
49 |
+
|
50 |
+
scheduler: "" # ScheduleFree
|
51 |
+
torch_scheduler: True
|
52 |
+
|
53 |
+
evaluation:
|
54 |
+
batch_size: 4
|
55 |
+
frequency: 250
|
56 |
+
size: 4
|
57 |
+
|
58 |
+
steps: 500
|
59 |
+
ar_temperature: 1.0
|
60 |
+
nar_temperature: 0.0
|
61 |
+
|
62 |
+
trainer:
|
63 |
+
iterations: 1_000_000
|
64 |
+
save_frequency: 250
|
65 |
+
keep_last_checkpoints: 4
|
66 |
+
|
67 |
+
resize_modules: True
|
68 |
+
|
69 |
+
check_for_oom: False
|
70 |
+
gradient_checkpointing: True
|
71 |
+
|
72 |
+
weight_dtype: bfloat16
|
73 |
+
amp: True
|
74 |
+
|
75 |
+
backend: deepspeed
|
76 |
+
deepspeed:
|
77 |
+
inferencing: False
|
78 |
+
amp: False
|
79 |
+
|
80 |
+
load_webui: False
|
81 |
+
|
82 |
+
inference:
|
83 |
+
backend: local
|
84 |
+
normalize: False
|
85 |
+
|
86 |
+
weight_dtype: bfloat16
|
87 |
+
amp: True
|
88 |
+
|
89 |
+
optimizations:
|
90 |
+
injects: False
|
91 |
+
replace: True
|
92 |
+
|
93 |
+
linear: False
|
94 |
+
embedding: False
|
95 |
+
optimizers: True
|
96 |
+
|
97 |
+
bitsandbytes: False
|
98 |
+
dadaptation: False
|
99 |
+
bitnet: False
|
100 |
+
fp8: False
|
101 |
+
|
102 |
+
dataset:
|
103 |
+
use_hdf5: True
|
104 |
+
hdf5_flag: r
|
105 |
+
|
106 |
+
use_metadata: True
|
107 |
+
validate: True
|
108 |
+
|
109 |
+
workers: 1
|
110 |
+
cache: True
|
111 |
+
|
112 |
+
duration_range: [3.0, 12.0]
|
113 |
+
|
114 |
+
random_utterance: 1.0
|
115 |
+
max_prompts: 1
|
116 |
+
prompt_duration_range: [3.0, 3.0]
|
117 |
+
|
118 |
+
max_resps: 1
|
119 |
+
p_resp_append: 0.25
|
120 |
+
|
121 |
+
sample_type: path # path # speaker
|
122 |
+
sample_order: duration
|
123 |
+
sample_max_duration_batch: 300
|
124 |
+
sample_shuffle: False
|
125 |
+
|
126 |
+
tasks_list: [ "tts", "stt" ]
|
127 |
+
|
128 |
+
training: []
|
129 |
+
validation: []
|
130 |
+
noise: []
|
models/ckpt/ar+nar-tts+stt-llama-8/fp32.sft
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 456272602
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc995fd0c81f7540518cd41a2c0482eb81f03a86c5d8f58dfd4aca4ec5fead3d
|
3 |
size 456272602
|
models/config.llama.tts+stt.yaml
CHANGED
@@ -20,14 +20,15 @@ models:
|
|
20 |
# resp: 1.0
|
21 |
capabilities: ["ar", "nar"]
|
22 |
experimental:
|
23 |
-
|
24 |
-
|
25 |
unified_position_ids: False
|
26 |
-
rvq_level_range: []
|
27 |
split_classifiers: True
|
28 |
-
|
29 |
causal_size: 1
|
30 |
-
|
|
|
|
|
31 |
|
32 |
#loras:
|
33 |
#- name : "lora"
|
@@ -99,13 +100,6 @@ optimizations:
|
|
99 |
fp8: False
|
100 |
|
101 |
dataset:
|
102 |
-
speaker_name_getter: "lambda p: f'{p.parts[-3]}_{p.parts[-2]}'"
|
103 |
-
speaker_group_getter: "lambda p: f'{p.parts[-3]}'"
|
104 |
-
speaker_languages:
|
105 |
-
ja: [
|
106 |
-
"housamo"
|
107 |
-
]
|
108 |
-
|
109 |
use_hdf5: True
|
110 |
hdf5_flag: r
|
111 |
|
@@ -129,6 +123,8 @@ dataset:
|
|
129 |
sample_max_duration_batch: 300
|
130 |
sample_shuffle: False
|
131 |
|
|
|
|
|
132 |
training: []
|
133 |
validation: []
|
134 |
noise: []
|
|
|
20 |
# resp: 1.0
|
21 |
capabilities: ["ar", "nar"]
|
22 |
experimental:
|
23 |
+
p_rvq_levels: "auto"
|
24 |
+
audio_embedding_sums: True
|
25 |
unified_position_ids: False
|
|
|
26 |
split_classifiers: True
|
27 |
+
#
|
28 |
causal_size: 1
|
29 |
+
interleave: False
|
30 |
+
rvq_level_range: []
|
31 |
+
tie_classifier_to_embedding: False
|
32 |
|
33 |
#loras:
|
34 |
#- name : "lora"
|
|
|
100 |
fp8: False
|
101 |
|
102 |
dataset:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
103 |
use_hdf5: True
|
104 |
hdf5_flag: r
|
105 |
|
|
|
123 |
sample_max_duration_batch: 300
|
124 |
sample_shuffle: False
|
125 |
|
126 |
+
tasks_list: [ "tts", "stt" ]
|
127 |
+
|
128 |
training: []
|
129 |
validation: []
|
130 |
noise: []
|