mrq commited on
Commit
f8eb280
1 Parent(s): 6542f44
loras/ckpt/lora-glados-r128-a128/lora.sft DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3395ef66c721a404b6204af6e61f607199b97acf430cfb6151f7d926ca09196e
3
- size 66076304
 
 
 
 
loras/ckpt/lora-max-r128-a128/lora.sft DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e8065b17f981dba303af2e5c92275902a45e8e38e9eea578f0a9e4b50690f1e
3
- size 66076312
 
 
 
 
loras/ckpt/lora-sam-r128-a128/lora.sft DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:433ada1d6f9cba5ebce7917bb51df09f050f4d221afdc4cdad72964b6624f9fc
3
- size 66076312
 
 
 
 
loras/ckpt/lora-shodan-r128-a128/lora.sft DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:206cc3add30585858531744857613e2d6d80ddaa238357ae3f407e4cc15318da
3
- size 66076304
 
 
 
 
loras/config.lora.glados.yaml DELETED
@@ -1,130 +0,0 @@
1
- sample_rate: 24_000
2
- audio_backend: "vocos"
3
-
4
- models:
5
- - name: "ar+nar"
6
- size: "full"
7
- resp_levels: 8
8
- prom_levels: 8
9
- tasks: 9
10
- langs: 4
11
- tones: 1
12
- arch_type: llama
13
- training: True
14
- version: 5
15
- attention: auto
16
- dropout: 0.1
17
- #loss_factors:
18
- # text: 0.01
19
- # prom: 0.5
20
- # resp: 1.0
21
- capabilities: ["ar", "nar"]
22
- experimental:
23
- p_rvq_levels: "auto"
24
- audio_embedding_sums: True
25
- unified_position_ids: False
26
- split_classifiers: True
27
- #
28
- causal_size: 1
29
- interleave: False
30
- rvq_level_range: []
31
- tie_classifier_to_embedding: False
32
-
33
- loras:
34
- - name : "lora-glados"
35
- rank: 128
36
- alpha: 128
37
- training: True
38
- rvq_levels: []
39
-
40
- hyperparameters:
41
- batch_size: 32
42
- gradient_accumulation_steps: 8
43
- gradient_clipping: 1.0
44
- warmup_steps: 10
45
-
46
- optimizer: Prodigy
47
- learning_rate: 1.0
48
- torch_optimizer: True
49
-
50
- scheduler: "" # ScheduleFree
51
- torch_scheduler: True
52
-
53
- evaluation:
54
- batch_size: 4
55
- frequency: 250
56
- size: 4
57
-
58
- steps: 500
59
- ar_temperature: 1.0
60
- nar_temperature: 0.0
61
-
62
- trainer:
63
- iterations: 1_000_000
64
- save_frequency: 250
65
- keep_last_checkpoints: 4
66
-
67
- resize_modules: True
68
-
69
- check_for_oom: False
70
- gradient_checkpointing: True
71
-
72
- weight_dtype: bfloat16
73
- amp: True
74
-
75
- backend: deepspeed
76
- deepspeed:
77
- inferencing: False
78
- amp: False
79
-
80
- load_webui: False
81
-
82
- inference:
83
- backend: local
84
- normalize: False
85
-
86
- weight_dtype: bfloat16
87
- amp: True
88
-
89
- optimizations:
90
- injects: False
91
- replace: True
92
-
93
- linear: False
94
- embedding: False
95
- optimizers: True
96
-
97
- bitsandbytes: False
98
- dadaptation: False
99
- bitnet: False
100
- fp8: False
101
-
102
- dataset:
103
- use_hdf5: True
104
- hdf5_flag: r
105
-
106
- use_metadata: True
107
- validate: True
108
-
109
- workers: 1
110
- cache: True
111
-
112
- duration_range: [3.0, 12.0]
113
-
114
- random_utterance: 1.0
115
- max_prompts: 1
116
- prompt_duration_range: [3.0, 3.0]
117
-
118
- max_resps: 1
119
- p_resp_append: 0.25
120
-
121
- sample_type: path # path # speaker
122
- sample_order: duration
123
- sample_max_duration_batch: 300
124
- sample_shuffle: False
125
-
126
- tasks_list: [ "tts", "stt" ]
127
-
128
- training: []
129
- validation: []
130
- noise: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
loras/config.lora.max.yaml DELETED
@@ -1,130 +0,0 @@
1
- sample_rate: 24_000
2
- audio_backend: "vocos"
3
-
4
- models:
5
- - name: "ar+nar-tts+stt"
6
- size: "full"
7
- resp_levels: 8
8
- prom_levels: 8
9
- tasks: 9
10
- langs: 2
11
- tones: 1
12
- arch_type: llama
13
- training: False
14
- version: 5
15
- attention: auto
16
- dropout: 0.1
17
- #loss_factors:
18
- # text: 0.01
19
- # prom: 0.5
20
- # resp: 1.0
21
- capabilities: ["ar", "nar"]
22
- experimental:
23
- p_rvq_levels: "auto"
24
- audio_embedding_sums: True
25
- unified_position_ids: False
26
- split_classifiers: True
27
- #
28
- causal_size: 1
29
- interleave: False
30
- rvq_level_range: []
31
- tie_classifier_to_embedding: False
32
-
33
- loras:
34
- - name : "lora-max"
35
- rank: 128
36
- alpha: 128
37
- training: True
38
- rvq_levels: []
39
-
40
- hyperparameters:
41
- batch_size: 32
42
- gradient_accumulation_steps: 8
43
- gradient_clipping: 1.0
44
- warmup_steps: 10
45
-
46
- optimizer: Prodigy
47
- learning_rate: 1.0
48
- torch_optimizer: True
49
-
50
- scheduler: "" # ScheduleFree
51
- torch_scheduler: True
52
-
53
- evaluation:
54
- batch_size: 4
55
- frequency: 250
56
- size: 4
57
-
58
- steps: 500
59
- ar_temperature: 1.0
60
- nar_temperature: 0.0
61
-
62
- trainer:
63
- iterations: 1_000_000
64
- save_frequency: 250
65
- keep_last_checkpoints: 4
66
-
67
- resize_modules: True
68
-
69
- check_for_oom: False
70
- gradient_checkpointing: True
71
-
72
- weight_dtype: bfloat16
73
- amp: True
74
-
75
- backend: deepspeed
76
- deepspeed:
77
- inferencing: False
78
- amp: False
79
-
80
- load_webui: False
81
-
82
- inference:
83
- backend: local
84
- normalize: False
85
-
86
- weight_dtype: bfloat16
87
- amp: True
88
-
89
- optimizations:
90
- injects: False
91
- replace: True
92
-
93
- linear: False
94
- embedding: False
95
- optimizers: True
96
-
97
- bitsandbytes: False
98
- dadaptation: False
99
- bitnet: False
100
- fp8: False
101
-
102
- dataset:
103
- use_hdf5: True
104
- hdf5_flag: r
105
-
106
- use_metadata: True
107
- validate: True
108
-
109
- workers: 1
110
- cache: True
111
-
112
- duration_range: [3.0, 12.0]
113
-
114
- random_utterance: 1.0
115
- max_prompts: 1
116
- prompt_duration_range: [3.0, 3.0]
117
-
118
- max_resps: 1
119
- p_resp_append: 0.25
120
-
121
- sample_type: path # path # speaker
122
- sample_order: duration
123
- sample_max_duration_batch: 300
124
- sample_shuffle: False
125
-
126
- tasks_list: [ "tts", "stt" ]
127
-
128
- training: []
129
- validation: []
130
- noise: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
loras/config.lora.sam.yaml DELETED
@@ -1,130 +0,0 @@
1
- sample_rate: 24_000
2
- audio_backend: "vocos"
3
-
4
- models:
5
- - name: "ar+nar-tts+stt"
6
- size: "full"
7
- resp_levels: 8
8
- prom_levels: 8
9
- tasks: 9
10
- langs: 2
11
- tones: 1
12
- arch_type: llama
13
- training: False
14
- version: 5
15
- attention: auto
16
- dropout: 0.1
17
- #loss_factors:
18
- # text: 0.01
19
- # prom: 0.5
20
- # resp: 1.0
21
- capabilities: ["ar", "nar"]
22
- experimental:
23
- p_rvq_levels: "auto"
24
- audio_embedding_sums: True
25
- unified_position_ids: False
26
- split_classifiers: True
27
- #
28
- causal_size: 1
29
- interleave: False
30
- rvq_level_range: []
31
- tie_classifier_to_embedding: False
32
-
33
- loras:
34
- - name : "lora-sam"
35
- rank: 128
36
- alpha: 128
37
- training: True
38
- rvq_levels: []
39
-
40
- hyperparameters:
41
- batch_size: 32
42
- gradient_accumulation_steps: 8
43
- gradient_clipping: 1.0
44
- warmup_steps: 10
45
-
46
- optimizer: Prodigy
47
- learning_rate: 1.0
48
- torch_optimizer: True
49
-
50
- scheduler: "" # ScheduleFree
51
- torch_scheduler: True
52
-
53
- evaluation:
54
- batch_size: 4
55
- frequency: 250
56
- size: 4
57
-
58
- steps: 500
59
- ar_temperature: 1.0
60
- nar_temperature: 0.0
61
-
62
- trainer:
63
- iterations: 1_000_000
64
- save_frequency: 250
65
- keep_last_checkpoints: 4
66
-
67
- resize_modules: True
68
-
69
- check_for_oom: False
70
- gradient_checkpointing: True
71
-
72
- weight_dtype: bfloat16
73
- amp: True
74
-
75
- backend: deepspeed
76
- deepspeed:
77
- inferencing: False
78
- amp: False
79
-
80
- load_webui: False
81
-
82
- inference:
83
- backend: local
84
- normalize: False
85
-
86
- weight_dtype: bfloat16
87
- amp: True
88
-
89
- optimizations:
90
- injects: False
91
- replace: True
92
-
93
- linear: False
94
- embedding: False
95
- optimizers: True
96
-
97
- bitsandbytes: False
98
- dadaptation: False
99
- bitnet: False
100
- fp8: False
101
-
102
- dataset:
103
- use_hdf5: True
104
- hdf5_flag: r
105
-
106
- use_metadata: True
107
- validate: True
108
-
109
- workers: 1
110
- cache: True
111
-
112
- duration_range: [3.0, 12.0]
113
-
114
- random_utterance: 1.0
115
- max_prompts: 1
116
- prompt_duration_range: [3.0, 3.0]
117
-
118
- max_resps: 1
119
- p_resp_append: 0.25
120
-
121
- sample_type: path # path # speaker
122
- sample_order: duration
123
- sample_max_duration_batch: 300
124
- sample_shuffle: False
125
-
126
- tasks_list: [ "tts", "stt" ]
127
-
128
- training: []
129
- validation: []
130
- noise: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
loras/config.lora.shodan.yaml DELETED
@@ -1,130 +0,0 @@
1
- sample_rate: 24_000
2
- audio_backend: "vocos"
3
-
4
- models:
5
- - name: "ar+nar"
6
- size: "full"
7
- resp_levels: 8
8
- prom_levels: 8
9
- tasks: 9
10
- langs: 4
11
- tones: 1
12
- arch_type: llama
13
- training: True
14
- version: 5
15
- attention: auto
16
- dropout: 0.1
17
- #loss_factors:
18
- # text: 0.01
19
- # prom: 0.5
20
- # resp: 1.0
21
- capabilities: ["ar", "nar"]
22
- experimental:
23
- p_rvq_levels: "auto"
24
- audio_embedding_sums: True
25
- unified_position_ids: False
26
- split_classifiers: True
27
- #
28
- causal_size: 1
29
- interleave: False
30
- rvq_level_range: []
31
- tie_classifier_to_embedding: False
32
-
33
- loras:
34
- - name : "lora-shodan"
35
- rank: 128
36
- alpha: 128
37
- training: True
38
- rvq_levels: []
39
-
40
- hyperparameters:
41
- batch_size: 32
42
- gradient_accumulation_steps: 8
43
- gradient_clipping: 1.0
44
- warmup_steps: 10
45
-
46
- optimizer: Prodigy
47
- learning_rate: 1.0
48
- torch_optimizer: True
49
-
50
- scheduler: "" # ScheduleFree
51
- torch_scheduler: True
52
-
53
- evaluation:
54
- batch_size: 4
55
- frequency: 250
56
- size: 4
57
-
58
- steps: 500
59
- ar_temperature: 1.0
60
- nar_temperature: 0.0
61
-
62
- trainer:
63
- iterations: 1_000_000
64
- save_frequency: 250
65
- keep_last_checkpoints: 4
66
-
67
- resize_modules: True
68
-
69
- check_for_oom: False
70
- gradient_checkpointing: True
71
-
72
- weight_dtype: bfloat16
73
- amp: True
74
-
75
- backend: deepspeed
76
- deepspeed:
77
- inferencing: False
78
- amp: False
79
-
80
- load_webui: False
81
-
82
- inference:
83
- backend: local
84
- normalize: False
85
-
86
- weight_dtype: bfloat16
87
- amp: True
88
-
89
- optimizations:
90
- injects: False
91
- replace: True
92
-
93
- linear: False
94
- embedding: False
95
- optimizers: True
96
-
97
- bitsandbytes: False
98
- dadaptation: False
99
- bitnet: False
100
- fp8: False
101
-
102
- dataset:
103
- use_hdf5: True
104
- hdf5_flag: r
105
-
106
- use_metadata: True
107
- validate: True
108
-
109
- workers: 1
110
- cache: True
111
-
112
- duration_range: [3.0, 12.0]
113
-
114
- random_utterance: 1.0
115
- max_prompts: 1
116
- prompt_duration_range: [3.0, 3.0]
117
-
118
- max_resps: 1
119
- p_resp_append: 0.25
120
-
121
- sample_type: path # path # speaker
122
- sample_order: duration
123
- sample_max_duration_batch: 300
124
- sample_shuffle: False
125
-
126
- tasks_list: [ "tts", "stt" ]
127
-
128
- training: []
129
- validation: []
130
- noise: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
loras/config.lora[delamain].yaml ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ models:
2
+ - name: "ar+nar"
3
+ size: "full"
4
+ resp_levels: 8
5
+ tasks: 9
6
+ langs: 4
7
+ tones: 1
8
+ arch_type: llama
9
+ attention: auto
10
+ version: 5
11
+
12
+ capabilities: ["ar", "nar"]
13
+ experimental:
14
+ split_classifiers: True
15
+ audio_embedding_sums: True
16
+ unified_position_ids: False
17
+ rvq_levels_p: [
18
+ 0, 0, 0, 0, 0, 0, 0,
19
+ 1, 2, 3, 4, 5, 6, 7
20
+ ]
21
+
22
+ loras:
23
+ - name : "lora-cyberpunk-delamain"
24
+ rank: 128
25
+ alpha: 128
26
+ rvq_levels: []
loras/config.lora[glados].yaml ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ models:
2
+ - name: "ar+nar"
3
+ size: "full"
4
+ resp_levels: 8
5
+ tasks: 9
6
+ langs: 4
7
+ tones: 1
8
+ arch_type: llama
9
+ attention: auto
10
+ version: 5
11
+
12
+ capabilities: ["ar", "nar"]
13
+ experimental:
14
+ split_classifiers: True
15
+ audio_embedding_sums: True
16
+ unified_position_ids: False
17
+ rvq_levels_p: [
18
+ 0, 0, 0, 0, 0, 0, 0,
19
+ 1, 2, 3, 4, 5, 6, 7
20
+ ]
21
+
22
+ loras:
23
+ - name : "lora-portal-glados"
24
+ rank: 128
25
+ alpha: 128
26
+ rvq_levels: []
loras/config.lora[johnny].yaml ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ models:
2
+ - name: "ar+nar"
3
+ size: "full"
4
+ resp_levels: 8
5
+ tasks: 9
6
+ langs: 4
7
+ tones: 1
8
+ arch_type: llama
9
+ attention: auto
10
+ version: 5
11
+
12
+ capabilities: ["ar", "nar"]
13
+ experimental:
14
+ split_classifiers: True
15
+ audio_embedding_sums: True
16
+ unified_position_ids: False
17
+ rvq_levels_p: [
18
+ 0, 0, 0, 0, 0, 0, 0,
19
+ 1, 2, 3, 4, 5, 6, 7
20
+ ]
21
+
22
+ loras:
23
+ - name : "lora-cyberpunk-silverhand"
24
+ rank: 128
25
+ alpha: 128
26
+ rvq_levels: []
loras/config.lora[sam].yaml ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ models:
2
+ - name: "ar+nar"
3
+ size: "full"
4
+ resp_levels: 8
5
+ tasks: 9
6
+ langs: 4
7
+ tones: 1
8
+ arch_type: llama
9
+ attention: auto
10
+ version: 5
11
+
12
+ capabilities: ["ar", "nar"]
13
+ experimental:
14
+ split_classifiers: True
15
+ audio_embedding_sums: True
16
+ unified_position_ids: False
17
+ rvq_levels_p: [
18
+ 0, 0, 0, 0, 0, 0, 0,
19
+ 1, 2, 3, 4, 5, 6, 7
20
+ ]
21
+
22
+ loras:
23
+ - name : "lora-samandmax-sam"
24
+ rank: 128
25
+ alpha: 128
26
+ rvq_levels: []