Akjava commited on
Commit
ec91abc
1 Parent(s): 6ea41c2

Upload folder using huggingface_hub

Browse files
runs/2024-09-14_04-08-31/.hydra/config.yaml ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ task_name: train
2
+ run_name: merged_ja004_ja005
3
+ tags:
4
+ - merged_ja004_ja005
5
+ train: true
6
+ test: true
7
+ ckpt_path: datas/merged_ja004_ja005/checkpoint_epoch=2669.ckpt
8
+ seed: 1234
9
+ data:
10
+ _target_: matcha.data.text_mel_datamodule.TextMelDataModule
11
+ name: ja006
12
+ train_filelist_path: datas/merged_ja004_ja005/train.txt
13
+ valid_filelist_path: datas/merged_ja004_ja005/valid.txt
14
+ batch_size: 80
15
+ num_workers: 1
16
+ pin_memory: true
17
+ cleaners:
18
+ - basic_cleaners2
19
+ add_blank: true
20
+ n_spks: 1
21
+ n_fft: 1024
22
+ n_feats: 80
23
+ sample_rate: 22050
24
+ hop_length: 256
25
+ win_length: 1024
26
+ f_min: 0
27
+ f_max: 8000
28
+ data_statistics:
29
+ mel_mean: -5.794878959655762
30
+ mel_std: 2.2488205432891846
31
+ seed: 3000
32
+ model:
33
+ _target_: matcha.models.matcha_tts.MatchaTTS
34
+ n_vocab: 178
35
+ n_spks: ${data.n_spks}
36
+ spk_emb_dim: 64
37
+ n_feats: 80
38
+ data_statistics: ${data.data_statistics}
39
+ out_size: null
40
+ prior_loss: true
41
+ encoder:
42
+ encoder_type: RoPE Encoder
43
+ encoder_params:
44
+ n_feats: ${model.n_feats}
45
+ n_channels: 192
46
+ filter_channels: 768
47
+ filter_channels_dp: 256
48
+ n_heads: 2
49
+ n_layers: 6
50
+ kernel_size: 3
51
+ p_dropout: 0.1
52
+ spk_emb_dim: 64
53
+ n_spks: 1
54
+ prenet: true
55
+ duration_predictor_params:
56
+ filter_channels_dp: ${model.encoder.encoder_params.filter_channels_dp}
57
+ kernel_size: 3
58
+ p_dropout: ${model.encoder.encoder_params.p_dropout}
59
+ decoder:
60
+ channels:
61
+ - 256
62
+ - 256
63
+ dropout: 0.05
64
+ attention_head_dim: 64
65
+ n_blocks: 1
66
+ num_mid_blocks: 2
67
+ num_heads: 2
68
+ act_fn: snakebeta
69
+ cfm:
70
+ name: CFM
71
+ solver: euler
72
+ sigma_min: 0.0001
73
+ optimizer:
74
+ _target_: torch.optim.Adam
75
+ _partial_: true
76
+ lr: 0.0001
77
+ weight_decay: 0.0
78
+ callbacks:
79
+ model_checkpoint:
80
+ _target_: lightning.pytorch.callbacks.ModelCheckpoint
81
+ dirpath: ${paths.output_dir}/checkpoints
82
+ filename: checkpoint_{epoch:03d}
83
+ monitor: epoch
84
+ verbose: false
85
+ save_last: true
86
+ save_top_k: 25
87
+ mode: max
88
+ auto_insert_metric_name: true
89
+ save_weights_only: false
90
+ every_n_train_steps: null
91
+ train_time_interval: null
92
+ every_n_epochs: 25
93
+ save_on_train_epoch_end: null
94
+ model_summary:
95
+ _target_: lightning.pytorch.callbacks.RichModelSummary
96
+ max_depth: 3
97
+ rich_progress_bar:
98
+ _target_: lightning.pytorch.callbacks.RichProgressBar
99
+ logger:
100
+ tensorboard:
101
+ _target_: lightning.pytorch.loggers.tensorboard.TensorBoardLogger
102
+ save_dir: ${paths.output_dir}/tensorboard/
103
+ name: null
104
+ log_graph: false
105
+ default_hp_metric: true
106
+ prefix: ''
107
+ trainer:
108
+ _target_: lightning.pytorch.trainer.Trainer
109
+ default_root_dir: ${paths.output_dir}
110
+ max_epochs: -1
111
+ accelerator: gpu
112
+ devices:
113
+ - 0
114
+ precision: 16-mixed
115
+ check_val_every_n_epoch: 1
116
+ deterministic: false
117
+ gradient_clip_val: 5.0
118
+ paths:
119
+ root_dir: ${oc.env:PROJECT_ROOT}
120
+ data_dir: ${paths.root_dir}/data/
121
+ log_dir: ${paths.root_dir}/logs/
122
+ output_dir: ${hydra:runtime.output_dir}
123
+ work_dir: ${hydra:runtime.cwd}
124
+ extras:
125
+ ignore_warnings: false
126
+ enforce_tags: true
127
+ print_config: true
runs/2024-09-14_04-08-31/.hydra/hydra.yaml ADDED
@@ -0,0 +1,182 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: ${paths.log_dir}/${task_name}/${run_name}/runs/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: ${paths.log_dir}/${task_name}/${run_name}/multiruns/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ colorlog:
72
+ (): colorlog.ColoredFormatter
73
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
74
+ handlers:
75
+ console:
76
+ class: logging.StreamHandler
77
+ formatter: colorlog
78
+ stream: ext://sys.stdout
79
+ root:
80
+ level: INFO
81
+ handlers:
82
+ - console
83
+ disable_existing_loggers: false
84
+ job_logging:
85
+ version: 1
86
+ formatters:
87
+ simple:
88
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
89
+ colorlog:
90
+ (): colorlog.ColoredFormatter
91
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
92
+ - %(message)s'
93
+ log_colors:
94
+ DEBUG: purple
95
+ INFO: green
96
+ WARNING: yellow
97
+ ERROR: red
98
+ CRITICAL: red
99
+ handlers:
100
+ console:
101
+ class: logging.StreamHandler
102
+ formatter: colorlog
103
+ stream: ext://sys.stdout
104
+ file:
105
+ class: logging.FileHandler
106
+ formatter: simple
107
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
108
+ root:
109
+ level: INFO
110
+ handlers:
111
+ - console
112
+ - file
113
+ disable_existing_loggers: false
114
+ env: {}
115
+ mode: RUN
116
+ searchpath: []
117
+ callbacks: {}
118
+ output_subdir: .hydra
119
+ overrides:
120
+ hydra:
121
+ - hydra.mode=RUN
122
+ task:
123
+ - experiment=merged_ja004_ja005
124
+ job:
125
+ name: train
126
+ chdir: null
127
+ override_dirname: experiment=merged_ja004_ja005
128
+ id: ???
129
+ num: ???
130
+ config_name: train.yaml
131
+ env_set: {}
132
+ env_copy: []
133
+ config:
134
+ override_dirname:
135
+ kv_sep: '='
136
+ item_sep: ','
137
+ exclude_keys: []
138
+ runtime:
139
+ version: 1.3.2
140
+ version_base: '1.3'
141
+ cwd: /notebooks/Matcha-TTS
142
+ config_sources:
143
+ - path: hydra.conf
144
+ schema: pkg
145
+ provider: hydra
146
+ - path: /notebooks/Matcha-TTS/configs
147
+ schema: file
148
+ provider: main
149
+ - path: hydra_plugins.hydra_colorlog.conf
150
+ schema: pkg
151
+ provider: hydra-colorlog
152
+ - path: ''
153
+ schema: structured
154
+ provider: schema
155
+ output_dir: /notebooks/Matcha-TTS/logs/train/merged_ja004_ja005/runs/2024-09-14_04-08-31
156
+ choices:
157
+ debug: null
158
+ local: default
159
+ hparams_search: null
160
+ experiment: merged_ja004_ja005
161
+ hydra: default
162
+ extras: default
163
+ paths: default
164
+ trainer: default
165
+ logger: tensorboard
166
+ callbacks: default
167
+ model: matcha
168
+ model/optimizer: adam.yaml
169
+ model/cfm: default.yaml
170
+ model/decoder: default.yaml
171
+ model/encoder: default.yaml
172
+ data: merged_ja004_ja005.yaml
173
+ hydra/env: default
174
+ hydra/callbacks: null
175
+ hydra/job_logging: colorlog
176
+ hydra/hydra_logging: colorlog
177
+ hydra/hydra_help: default
178
+ hydra/help: default
179
+ hydra/sweeper: basic
180
+ hydra/launcher: basic
181
+ hydra/output: default
182
+ verbose: false
runs/2024-09-14_04-08-31/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - experiment=merged_ja004_ja005
runs/2024-09-14_04-08-31/checkpoints/checkpoint_epoch=2674.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12a2217eac3d9e3581629e0209fa88d059941df5b46f2f2000738fcbbc11b05f
3
+ size 218839178
runs/2024-09-14_04-08-31/checkpoints/checkpoint_epoch=2699.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5034fe39ee79e7be8e7a59bd2761828d17fb9b1fc4ccd12dbaf2b272ed0fcdd
3
+ size 218839561
runs/2024-09-14_04-08-31/checkpoints/checkpoint_epoch=2724.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6751ea9b98d47cd7bed9ddc68791c7e63fe260f18c011c70c76defd91d020d3
3
+ size 218839944
runs/2024-09-14_04-08-31/checkpoints/checkpoint_epoch=2749.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:288f895e4a60c860ca1917d53fee963bb4e2d67ed2ec5ee73182c73e989b9d20
3
+ size 218840327
runs/2024-09-14_04-08-31/checkpoints/checkpoint_epoch=2774.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d109474fe96df9e64539fcbb063aac9b7f1d90e7dc2af70a84b4a2aa9b41185
3
+ size 218840710
runs/2024-09-14_04-08-31/checkpoints/checkpoint_epoch=2799.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad2bab9aceb545ca2b78ce379dc3494fedad71d3d41aa487329d753c46c99440
3
+ size 218841093
runs/2024-09-14_04-08-31/checkpoints/last.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad2bab9aceb545ca2b78ce379dc3494fedad71d3d41aa487329d753c46c99440
3
+ size 218841093
runs/2024-09-14_04-08-31/config_tree.log ADDED
@@ -0,0 +1,143 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ CONFIG
2
+ ├── data
3
+ │ └── _target_: matcha.data.text_mel_datamodule.TextMelDataModule
4
+ │ name: ja006
5
+ │ train_filelist_path: datas/merged_ja004_ja005/train.txt
6
+ │ valid_filelist_path: datas/merged_ja004_ja005/valid.txt
7
+ │ batch_size: 80
8
+ │ num_workers: 1
9
+ │ pin_memory: true
10
+ │ cleaners:
11
+ │ - basic_cleaners2
12
+ │ add_blank: true
13
+ │ n_spks: 1
14
+ │ n_fft: 1024
15
+ │ n_feats: 80
16
+ │ sample_rate: 22050
17
+ │ hop_length: 256
18
+ │ win_length: 1024
19
+ │ f_min: 0
20
+ │ f_max: 8000
21
+ │ data_statistics:
22
+ │ mel_mean: -5.794878959655762
23
+ │ mel_std: 2.2488205432891846
24
+ │ seed: 3000
25
+
26
+ ├── model
27
+ │ └── _target_: matcha.models.matcha_tts.MatchaTTS
28
+ │ n_vocab: 178
29
+ │ n_spks: 1
30
+ │ spk_emb_dim: 64
31
+ │ n_feats: 80
32
+ │ data_statistics:
33
+ │ mel_mean: -5.794878959655762
34
+ │ mel_std: 2.2488205432891846
35
+ │ out_size: null
36
+ │ prior_loss: true
37
+ │ encoder:
38
+ │ encoder_type: RoPE Encoder
39
+ │ encoder_params:
40
+ │ n_feats: 80
41
+ │ n_channels: 192
42
+ │ filter_channels: 768
43
+ │ filter_channels_dp: 256
44
+ │ n_heads: 2
45
+ │ n_layers: 6
46
+ │ kernel_size: 3
47
+ │ p_dropout: 0.1
48
+ │ spk_emb_dim: 64
49
+ │ n_spks: 1
50
+ │ prenet: true
51
+ │ duration_predictor_params:
52
+ │ filter_channels_dp: 256
53
+ │ kernel_size: 3
54
+ │ p_dropout: 0.1
55
+ │ decoder:
56
+ │ channels:
57
+ │ - 256
58
+ │ - 256
59
+ │ dropout: 0.05
60
+ │ attention_head_dim: 64
61
+ │ n_blocks: 1
62
+ │ num_mid_blocks: 2
63
+ │ num_heads: 2
64
+ │ act_fn: snakebeta
65
+ │ cfm:
66
+ │ name: CFM
67
+ │ solver: euler
68
+ │ sigma_min: 0.0001
69
+ │ optimizer:
70
+ │ _target_: torch.optim.Adam
71
+ │ _partial_: true
72
+ │ lr: 0.0001
73
+ │ weight_decay: 0.0
74
+
75
+ ├── callbacks
76
+ │ └── model_checkpoint:
77
+ │ _target_: lightning.pytorch.callbacks.ModelCheckpoint
78
+ │ dirpath: /notebooks/Matcha-TTS/logs/train/merged_ja004_ja005/runs/2024-09-14_04-08-31/checkpoints
79
+ │ filename: checkpoint_{epoch:03d}
80
+ │ monitor: epoch
81
+ │ verbose: false
82
+ │ save_last: true
83
+ │ save_top_k: 25
84
+ │ mode: max
85
+ │ auto_insert_metric_name: true
86
+ │ save_weights_only: false
87
+ │ every_n_train_steps: null
88
+ │ train_time_interval: null
89
+ │ every_n_epochs: 25
90
+ │ save_on_train_epoch_end: null
91
+ │ model_summary:
92
+ │ _target_: lightning.pytorch.callbacks.RichModelSummary
93
+ │ max_depth: 3
94
+ │ rich_progress_bar:
95
+ │ _target_: lightning.pytorch.callbacks.RichProgressBar
96
+
97
+ ├── logger
98
+ │ └── tensorboard:
99
+ │ _target_: lightning.pytorch.loggers.tensorboard.TensorBoardLogger
100
+ │ save_dir: /notebooks/Matcha-TTS/logs/train/merged_ja004_ja005/runs/2024-09-14_04-08-31/tensorboard/
101
+ │ name: null
102
+ │ log_graph: false
103
+ │ default_hp_metric: true
104
+ │ prefix: ''
105
+
106
+ ├── trainer
107
+ │ └── _target_: lightning.pytorch.trainer.Trainer
108
+ │ default_root_dir: /notebooks/Matcha-TTS/logs/train/merged_ja004_ja005/runs/2024-09-14_04-08-31
109
+ │ max_epochs: -1
110
+ │ accelerator: gpu
111
+ │ devices:
112
+ │ - 0
113
+ │ precision: 16-mixed
114
+ │ check_val_every_n_epoch: 1
115
+ │ deterministic: false
116
+ │ gradient_clip_val: 5.0
117
+
118
+ ├── paths
119
+ │ └── root_dir: /notebooks/Matcha-TTS
120
+ │ data_dir: /notebooks/Matcha-TTS/data/
121
+ │ log_dir: /notebooks/Matcha-TTS/logs/
122
+ │ output_dir: /notebooks/Matcha-TTS/logs/train/merged_ja004_ja005/runs/2024-09-14_04-08-31
123
+ │ work_dir: /notebooks/Matcha-TTS
124
+
125
+ ├── extras
126
+ │ └── ignore_warnings: false
127
+ │ enforce_tags: true
128
+ │ print_config: true
129
+
130
+ ├── task_name
131
+ │ └── train
132
+ ├── run_name
133
+ │ └── merged_ja004_ja005
134
+ ├── tags
135
+ │ └── ['merged_ja004_ja005']
136
+ ├── train
137
+ │ └── True
138
+ ├── test
139
+ │ └── True
140
+ ├── ckpt_path
141
+ │ └── datas/merged_ja004_ja005/checkpoint_epoch=2669.ckpt
142
+ └── seed
143
+ └── 1234
runs/2024-09-14_04-08-31/tags.log ADDED
@@ -0,0 +1 @@
 
 
1
+ ['merged_ja004_ja005']
runs/2024-09-14_04-08-31/tensorboard/version_0/events.out.tfevents.1726286913.nz86mdl5mv.3125.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a304038f1c64af05d910bfde06855e4456141ccda1d84365a9d38404534c6a6c
3
+ size 43863552
runs/2024-09-14_04-08-31/tensorboard/version_0/hparams.yaml ADDED
@@ -0,0 +1,113 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ _target_: matcha.models.matcha_tts.MatchaTTS
3
+ n_vocab: 178
4
+ n_spks: ${data.n_spks}
5
+ spk_emb_dim: 64
6
+ n_feats: 80
7
+ data_statistics: ${data.data_statistics}
8
+ out_size: null
9
+ prior_loss: true
10
+ encoder:
11
+ encoder_type: RoPE Encoder
12
+ encoder_params:
13
+ n_feats: ${model.n_feats}
14
+ n_channels: 192
15
+ filter_channels: 768
16
+ filter_channels_dp: 256
17
+ n_heads: 2
18
+ n_layers: 6
19
+ kernel_size: 3
20
+ p_dropout: 0.1
21
+ spk_emb_dim: 64
22
+ n_spks: 1
23
+ prenet: true
24
+ duration_predictor_params:
25
+ filter_channels_dp: ${model.encoder.encoder_params.filter_channels_dp}
26
+ kernel_size: 3
27
+ p_dropout: ${model.encoder.encoder_params.p_dropout}
28
+ decoder:
29
+ channels:
30
+ - 256
31
+ - 256
32
+ dropout: 0.05
33
+ attention_head_dim: 64
34
+ n_blocks: 1
35
+ num_mid_blocks: 2
36
+ num_heads: 2
37
+ act_fn: snakebeta
38
+ cfm:
39
+ name: CFM
40
+ solver: euler
41
+ sigma_min: 0.0001
42
+ optimizer:
43
+ _target_: torch.optim.Adam
44
+ _partial_: true
45
+ lr: 0.0001
46
+ weight_decay: 0.0
47
+ model/params/total: 18204193
48
+ model/params/trainable: 18204193
49
+ model/params/non_trainable: 0
50
+ data:
51
+ _target_: matcha.data.text_mel_datamodule.TextMelDataModule
52
+ name: ja006
53
+ train_filelist_path: datas/merged_ja004_ja005/train.txt
54
+ valid_filelist_path: datas/merged_ja004_ja005/valid.txt
55
+ batch_size: 80
56
+ num_workers: 1
57
+ pin_memory: true
58
+ cleaners:
59
+ - basic_cleaners2
60
+ add_blank: true
61
+ n_spks: 1
62
+ n_fft: 1024
63
+ n_feats: 80
64
+ sample_rate: 22050
65
+ hop_length: 256
66
+ win_length: 1024
67
+ f_min: 0
68
+ f_max: 8000
69
+ data_statistics:
70
+ mel_mean: -5.794878959655762
71
+ mel_std: 2.2488205432891846
72
+ seed: 3000
73
+ trainer:
74
+ _target_: lightning.pytorch.trainer.Trainer
75
+ default_root_dir: ${paths.output_dir}
76
+ max_epochs: -1
77
+ accelerator: gpu
78
+ devices:
79
+ - 0
80
+ precision: 16-mixed
81
+ check_val_every_n_epoch: 1
82
+ deterministic: false
83
+ gradient_clip_val: 5.0
84
+ callbacks:
85
+ model_checkpoint:
86
+ _target_: lightning.pytorch.callbacks.ModelCheckpoint
87
+ dirpath: ${paths.output_dir}/checkpoints
88
+ filename: checkpoint_{epoch:03d}
89
+ monitor: epoch
90
+ verbose: false
91
+ save_last: true
92
+ save_top_k: 25
93
+ mode: max
94
+ auto_insert_metric_name: true
95
+ save_weights_only: false
96
+ every_n_train_steps: null
97
+ train_time_interval: null
98
+ every_n_epochs: 25
99
+ save_on_train_epoch_end: null
100
+ model_summary:
101
+ _target_: lightning.pytorch.callbacks.RichModelSummary
102
+ max_depth: 3
103
+ rich_progress_bar:
104
+ _target_: lightning.pytorch.callbacks.RichProgressBar
105
+ extras:
106
+ ignore_warnings: false
107
+ enforce_tags: true
108
+ print_config: true
109
+ task_name: train
110
+ tags:
111
+ - merged_ja004_ja005
112
+ ckpt_path: datas/merged_ja004_ja005/checkpoint_epoch=2669.ckpt
113
+ seed: 1234
runs/2024-09-14_04-08-31/train.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2024-09-14 04:08:31,694][matcha.utils.utils][INFO] - Enforcing tags! <cfg.extras.enforce_tags=True>
2
+ [2024-09-14 04:08:31,698][matcha.utils.utils][INFO] - Printing config tree with Rich! <cfg.extras.print_config=True>
3
+ [2024-09-14 04:08:31,754][__main__][INFO] - Instantiating datamodule <matcha.data.text_mel_datamodule.TextMelDataModule>
4
+ [2024-09-14 04:08:32,661][__main__][INFO] - Instantiating model <matcha.models.matcha_tts.MatchaTTS>
5
+ [2024-09-14 04:08:33,427][__main__][INFO] - Instantiating callbacks...
6
+ [2024-09-14 04:08:33,428][matcha.utils.instantiators][INFO] - Instantiating callback <lightning.pytorch.callbacks.ModelCheckpoint>
7
+ [2024-09-14 04:08:33,432][matcha.utils.instantiators][INFO] - Instantiating callback <lightning.pytorch.callbacks.RichModelSummary>
8
+ [2024-09-14 04:08:33,433][matcha.utils.instantiators][INFO] - Instantiating callback <lightning.pytorch.callbacks.RichProgressBar>
9
+ [2024-09-14 04:08:33,433][__main__][INFO] - Instantiating loggers...
10
+ [2024-09-14 04:08:33,434][matcha.utils.instantiators][INFO] - Instantiating logger <lightning.pytorch.loggers.tensorboard.TensorBoardLogger>
11
+ [2024-09-14 04:08:33,436][__main__][INFO] - Instantiating trainer <lightning.pytorch.trainer.Trainer>
12
+ [2024-09-14 04:08:33,452][__main__][INFO] - Logging hyperparameters!
13
+ [2024-09-14 04:08:33,512][__main__][INFO] - Starting training!
14
+ [2024-09-14 09:36:59,593][matcha.utils.utils][INFO] - Output dir: /notebooks/Matcha-TTS/logs/train/merged_ja004_ja005/runs/2024-09-14_04-08-31