zhniu commited on
Commit
7cd7712
·
verified ·
1 Parent(s): 3325999

Upload LvcWhisperForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +24 -120
  2. generation_config.json +1 -1
  3. model.safetensors +3 -0
config.json CHANGED
@@ -1,144 +1,48 @@
1
  {
2
- "_name_or_path": "openai/whisper-tiny",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
 
5
  "architectures": [
6
- "WhisperForConditionalGeneration"
7
  ],
8
  "attention_dropout": 0.0,
9
- "begin_suppress_tokens": [
10
- 220,
11
- 50257
12
- ],
13
  "bos_token_id": 50257,
14
- "d_model": 384,
15
- "decoder_attention_heads": 6,
16
- "decoder_ffn_dim": 1536,
 
17
  "decoder_layerdrop": 0.0,
18
- "decoder_layers": 4,
19
  "decoder_start_token_id": 50258,
20
  "dropout": 0.0,
21
- "encoder_attention_heads": 6,
22
- "encoder_ffn_dim": 1536,
23
  "encoder_layerdrop": 0.0,
24
- "encoder_layers": 4,
25
  "eos_token_id": 50257,
26
- "forced_decoder_ids": [
27
- [
28
- 1,
29
- 50259
30
- ],
31
- [
32
- 2,
33
- 50359
34
- ],
35
- [
36
- 3,
37
- 50363
38
- ]
39
- ],
40
  "init_std": 0.02,
41
  "is_encoder_decoder": true,
42
- "max_length": 448,
 
 
 
 
 
 
43
  "max_source_positions": 1500,
44
  "max_target_positions": 448,
 
45
  "model_type": "whisper",
46
- "num_hidden_layers": 4,
47
  "num_mel_bins": 80,
48
  "pad_token_id": 50257,
49
  "scale_embedding": false,
50
- "suppress_tokens": [
51
- 1,
52
- 2,
53
- 7,
54
- 8,
55
- 9,
56
- 10,
57
- 14,
58
- 25,
59
- 26,
60
- 27,
61
- 28,
62
- 29,
63
- 31,
64
- 58,
65
- 59,
66
- 60,
67
- 61,
68
- 62,
69
- 63,
70
- 90,
71
- 91,
72
- 92,
73
- 93,
74
- 359,
75
- 503,
76
- 522,
77
- 542,
78
- 873,
79
- 893,
80
- 902,
81
- 918,
82
- 922,
83
- 931,
84
- 1350,
85
- 1853,
86
- 1982,
87
- 2460,
88
- 2627,
89
- 3246,
90
- 3253,
91
- 3268,
92
- 3536,
93
- 3846,
94
- 3961,
95
- 4183,
96
- 4667,
97
- 6585,
98
- 6647,
99
- 7273,
100
- 9061,
101
- 9383,
102
- 10428,
103
- 10929,
104
- 11938,
105
- 12033,
106
- 12331,
107
- 12562,
108
- 13793,
109
- 14157,
110
- 14635,
111
- 15265,
112
- 15618,
113
- 16553,
114
- 16604,
115
- 18362,
116
- 18956,
117
- 20075,
118
- 21675,
119
- 22520,
120
- 26130,
121
- 26161,
122
- 26435,
123
- 28279,
124
- 29464,
125
- 31650,
126
- 32302,
127
- 32470,
128
- 36865,
129
- 42863,
130
- 47425,
131
- 49870,
132
- 50254,
133
- 50258,
134
- 50358,
135
- 50359,
136
- 50360,
137
- 50361,
138
- 50362
139
- ],
140
  "torch_dtype": "float32",
141
- "transformers_version": "4.27.0.dev0",
142
  "use_cache": true,
 
143
  "vocab_size": 51865
144
  }
 
1
  {
2
+ "_name_or_path": "output_model/",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
+ "apply_spec_augment": false,
6
  "architectures": [
7
+ "LvcWhisperForConditionalGeneration"
8
  ],
9
  "attention_dropout": 0.0,
10
+ "begin_suppress_tokens": null,
 
 
 
11
  "bos_token_id": 50257,
12
+ "classifier_proj_size": 256,
13
+ "d_model": 1024,
14
+ "decoder_attention_heads": 16,
15
+ "decoder_ffn_dim": 4096,
16
  "decoder_layerdrop": 0.0,
17
+ "decoder_layers": 24,
18
  "decoder_start_token_id": 50258,
19
  "dropout": 0.0,
20
+ "encoder_attention_heads": 16,
21
+ "encoder_ffn_dim": 4096,
22
  "encoder_layerdrop": 0.0,
23
+ "encoder_layers": 24,
24
  "eos_token_id": 50257,
25
+ "forced_decoder_ids": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  "init_std": 0.02,
27
  "is_encoder_decoder": true,
28
+ "mask_feature_length": 10,
29
+ "mask_feature_min_masks": 0,
30
+ "mask_feature_prob": 0.0,
31
+ "mask_time_length": 10,
32
+ "mask_time_min_masks": 2,
33
+ "mask_time_prob": 0.05,
34
+ "max_length": null,
35
  "max_source_positions": 1500,
36
  "max_target_positions": 448,
37
+ "median_filter_width": 7,
38
  "model_type": "whisper",
39
+ "num_hidden_layers": 24,
40
  "num_mel_bins": 80,
41
  "pad_token_id": 50257,
42
  "scale_embedding": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
  "torch_dtype": "float32",
44
+ "transformers_version": "4.48.2",
45
  "use_cache": true,
46
+ "use_weighted_layer_sum": false,
47
  "vocab_size": 51865
48
  }
generation_config.json CHANGED
@@ -244,5 +244,5 @@
244
  "transcribe": 50359,
245
  "translate": 50358
246
  },
247
- "transformers_version": "4.31.0.dev0"
248
  }
 
244
  "transcribe": 50359,
245
  "translate": 50358
246
  },
247
+ "transformers_version": "4.48.2"
248
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fec1ca47f253fc99ceee2880d76c9fe75641e97d43c6640feb9c5426687500c
3
+ size 3055544304