imenLa commited on
Commit
9cac002
·
verified ·
1 Parent(s): 16b5b44

Upload 4 files

Browse files
train_tun_parallel_wavegan.v3/checkpoint-400000steps.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64ea64092be605397bbadc778c570640616eec0a83201a81795bdd19b72a1166
3
+ size 225511291
train_tun_parallel_wavegan.v3/checkpoint-560000steps.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53258aa40300a21ebbadea1b1fbc496cad8a41688edb03372e9d32d1864c8b4c
3
+ size 225511291
train_tun_parallel_wavegan.v3/config.yml ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ allow_cache: true
2
+ batch_max_steps: 8192
3
+ batch_size: 16
4
+ config: conf/parallel_wavegan.v3.yaml
5
+ dev_dumpdir: dump/dev/norm
6
+ dev_feats_scp: null
7
+ dev_segments: null
8
+ dev_wav_scp: null
9
+ discriminator_grad_norm: 1
10
+ discriminator_optimizer_params:
11
+ eps: 1.0e-06
12
+ lr: 5.0e-05
13
+ weight_decay: 0.0
14
+ discriminator_params:
15
+ channels: 16
16
+ downsample_pooling: AvgPool1d
17
+ downsample_pooling_params:
18
+ count_include_pad: false
19
+ kernel_size: 4
20
+ padding: 1
21
+ stride: 2
22
+ downsample_scales:
23
+ - 4
24
+ - 4
25
+ - 4
26
+ - 4
27
+ in_channels: 1
28
+ kernel_sizes:
29
+ - 5
30
+ - 3
31
+ max_downsample_channels: 1024
32
+ nonlinear_activation: LeakyReLU
33
+ nonlinear_activation_params:
34
+ negative_slope: 0.2
35
+ out_channels: 1
36
+ scales: 3
37
+ use_weight_norm: true
38
+ discriminator_scheduler_params:
39
+ gamma: 0.5
40
+ step_size: 3000000
41
+ discriminator_train_start_steps: 100000
42
+ discriminator_type: MelGANMultiScaleDiscriminator
43
+ distributed: false
44
+ eval_interval_steps: 1000
45
+ fft_size: 1024
46
+ fmax: 7600
47
+ fmin: 80
48
+ format: hdf5
49
+ generator_grad_norm: 10
50
+ generator_optimizer_params:
51
+ eps: 1.0e-06
52
+ lr: 0.0001
53
+ weight_decay: 0.0
54
+ generator_params:
55
+ aux_channels: 80
56
+ aux_context_window: 2
57
+ dropout: 0.0
58
+ gate_channels: 128
59
+ in_channels: 1
60
+ kernel_size: 5
61
+ layers: 30
62
+ out_channels: 1
63
+ residual_channels: 64
64
+ skip_channels: 64
65
+ stacks: 3
66
+ upsample_net: ConvInUpsampleNetwork
67
+ upsample_params:
68
+ upsample_scales:
69
+ - 4
70
+ - 4
71
+ - 4
72
+ - 4
73
+ use_weight_norm: true
74
+ generator_scheduler_params:
75
+ gamma: 0.5
76
+ step_size: 3000000
77
+ global_gain_scale: 1.0
78
+ hop_size: 256
79
+ lambda_adv: 4.0
80
+ lambda_feat_match: 25.0
81
+ log_interval_steps: 100
82
+ num_mels: 80
83
+ num_save_intermediate_results: 4
84
+ num_workers: 2
85
+ outdir: exp/train_tun_parallel_wavegan.v3
86
+ pin_memory: true
87
+ pretrain: ''
88
+ rank: 0
89
+ remove_short_samples: true
90
+ resume: exp/train_tun_parallel_wavegan.v3/checkpoint-502000steps.pkl
91
+ sampling_rate: 22050
92
+ save_interval_steps: 2000
93
+ stft_loss_params:
94
+ fft_sizes:
95
+ - 1024
96
+ - 2048
97
+ - 512
98
+ hop_sizes:
99
+ - 120
100
+ - 240
101
+ - 50
102
+ win_lengths:
103
+ - 600
104
+ - 1200
105
+ - 240
106
+ window: hann_window
107
+ train_dumpdir: dump/train/norm
108
+ train_feats_scp: null
109
+ train_max_steps: 3000000
110
+ train_segments: null
111
+ train_wav_scp: null
112
+ trim_frame_size: 2048
113
+ trim_hop_size: 512
114
+ trim_silence: true
115
+ trim_threshold_in_db: 60
116
+ use_feat_match_loss: true
117
+ verbose: 1
118
+ version: 0.6.0
119
+ win_length: null
120
+ window: hann
train_tun_parallel_wavegan.v3/stats.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4c3fcaef240a21f8045ba9bd6630280850f79466951971c847e26971e002894
3
+ size 4736