csikasote commited on
Commit
91b55e7
·
verified ·
1 Parent(s): 0143ebc

End of training

Browse files
README.md CHANGED
@@ -4,11 +4,23 @@ license: apache-2.0
4
  base_model: openai/whisper-medium
5
  tags:
6
  - generated_from_trainer
 
 
7
  metrics:
8
  - wer
9
  model-index:
10
  - name: whisper-medium-toigen-balanced-model
11
- results: []
 
 
 
 
 
 
 
 
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -16,10 +28,10 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # whisper-medium-toigen-balanced-model
18
 
19
- This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.8748
22
- - Wer: 0.4604
23
 
24
  ## Model description
25
 
 
4
  base_model: openai/whisper-medium
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - toigen
9
  metrics:
10
  - wer
11
  model-index:
12
  - name: whisper-medium-toigen-balanced-model
13
+ results:
14
+ - task:
15
+ name: Automatic Speech Recognition
16
+ type: automatic-speech-recognition
17
+ dataset:
18
+ name: toigen
19
+ type: toigen
20
+ metrics:
21
+ - name: Wer
22
+ type: wer
23
+ value: 0.48498635122838946
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
28
 
29
  # whisper-medium-toigen-balanced-model
30
 
31
+ This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the toigen dataset.
32
  It achieves the following results on the evaluation set:
33
+ - Loss: 0.7784
34
+ - Wer: 0.4850
35
 
36
  ## Model description
37
 
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.766884531590414,
3
+ "eval_loss": 0.7783711552619934,
4
+ "eval_runtime": 130.1477,
5
+ "eval_samples": 215,
6
+ "eval_samples_per_second": 1.652,
7
+ "eval_steps_per_second": 0.83,
8
+ "eval_wer": 0.48498635122838946,
9
+ "total_flos": 8.19749861720064e+18,
10
+ "train_loss": 2.2032353465557097,
11
+ "train_runtime": 2667.1173,
12
+ "train_samples": 918,
13
+ "train_samples_per_second": 14.997,
14
+ "train_steps_per_second": 1.875
15
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.766884531590414,
3
+ "eval_loss": 0.7783711552619934,
4
+ "eval_runtime": 130.1477,
5
+ "eval_samples": 215,
6
+ "eval_samples_per_second": 1.652,
7
+ "eval_steps_per_second": 0.83,
8
+ "eval_wer": 0.48498635122838946
9
+ }
runs/Jan05_15-12-44_srvrocgpu011.uct.ac.za/events.out.tfevents.1736085866.srvrocgpu011.uct.ac.za ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a0042f46ac94c50db88f8cbbb16753c1ee493389162e7f7c13356409130c612
3
+ size 40
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.766884531590414,
3
+ "total_flos": 8.19749861720064e+18,
4
+ "train_loss": 2.2032353465557097,
5
+ "train_runtime": 2667.1173,
6
+ "train_samples": 918,
7
+ "train_samples_per_second": 14.997,
8
+ "train_steps_per_second": 1.875
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,376 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7783711552619934,
3
+ "best_model_checkpoint": "/scratch/skscla001/speech/results/whisper-medium-toigen-balanced-model/checkpoint-400",
4
+ "epoch": 8.766884531590414,
5
+ "eval_steps": 200,
6
+ "global_step": 1000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.2178649237472767,
13
+ "grad_norm": 130.9698028564453,
14
+ "learning_rate": 4.0000000000000003e-07,
15
+ "loss": 13.9102,
16
+ "step": 25
17
+ },
18
+ {
19
+ "epoch": 0.4357298474945534,
20
+ "grad_norm": 102.07112884521484,
21
+ "learning_rate": 9.000000000000001e-07,
22
+ "loss": 11.1467,
23
+ "step": 50
24
+ },
25
+ {
26
+ "epoch": 0.6535947712418301,
27
+ "grad_norm": 79.69112396240234,
28
+ "learning_rate": 1.4000000000000001e-06,
29
+ "loss": 8.6162,
30
+ "step": 75
31
+ },
32
+ {
33
+ "epoch": 0.8714596949891068,
34
+ "grad_norm": 67.27130126953125,
35
+ "learning_rate": 1.9000000000000002e-06,
36
+ "loss": 6.3881,
37
+ "step": 100
38
+ },
39
+ {
40
+ "epoch": 1.0958605664488017,
41
+ "grad_norm": 61.87217712402344,
42
+ "learning_rate": 2.4000000000000003e-06,
43
+ "loss": 5.4679,
44
+ "step": 125
45
+ },
46
+ {
47
+ "epoch": 1.3137254901960784,
48
+ "grad_norm": 70.62369537353516,
49
+ "learning_rate": 2.9e-06,
50
+ "loss": 4.4849,
51
+ "step": 150
52
+ },
53
+ {
54
+ "epoch": 1.5315904139433552,
55
+ "grad_norm": 72.37535858154297,
56
+ "learning_rate": 3.4000000000000005e-06,
57
+ "loss": 4.2698,
58
+ "step": 175
59
+ },
60
+ {
61
+ "epoch": 1.7494553376906317,
62
+ "grad_norm": 78.36324310302734,
63
+ "learning_rate": 3.900000000000001e-06,
64
+ "loss": 3.415,
65
+ "step": 200
66
+ },
67
+ {
68
+ "epoch": 1.7494553376906317,
69
+ "eval_loss": 0.8852776885032654,
70
+ "eval_runtime": 133.304,
71
+ "eval_samples_per_second": 1.613,
72
+ "eval_steps_per_second": 0.81,
73
+ "eval_wer": 0.6050955414012739,
74
+ "step": 200
75
+ },
76
+ {
77
+ "epoch": 1.9673202614379086,
78
+ "grad_norm": 52.00922775268555,
79
+ "learning_rate": 4.4e-06,
80
+ "loss": 3.5546,
81
+ "step": 225
82
+ },
83
+ {
84
+ "epoch": 2.1917211328976034,
85
+ "grad_norm": 59.29832077026367,
86
+ "learning_rate": 4.9000000000000005e-06,
87
+ "loss": 2.9609,
88
+ "step": 250
89
+ },
90
+ {
91
+ "epoch": 2.4095860566448803,
92
+ "grad_norm": 43.57518768310547,
93
+ "learning_rate": 5.400000000000001e-06,
94
+ "loss": 2.3833,
95
+ "step": 275
96
+ },
97
+ {
98
+ "epoch": 2.627450980392157,
99
+ "grad_norm": 54.65665054321289,
100
+ "learning_rate": 5.9e-06,
101
+ "loss": 2.3348,
102
+ "step": 300
103
+ },
104
+ {
105
+ "epoch": 2.845315904139434,
106
+ "grad_norm": 68.74827575683594,
107
+ "learning_rate": 6.4000000000000006e-06,
108
+ "loss": 2.5705,
109
+ "step": 325
110
+ },
111
+ {
112
+ "epoch": 3.0697167755991286,
113
+ "grad_norm": 34.880619049072266,
114
+ "learning_rate": 6.9e-06,
115
+ "loss": 2.0538,
116
+ "step": 350
117
+ },
118
+ {
119
+ "epoch": 3.287581699346405,
120
+ "grad_norm": 47.74470901489258,
121
+ "learning_rate": 7.4e-06,
122
+ "loss": 1.3222,
123
+ "step": 375
124
+ },
125
+ {
126
+ "epoch": 3.505446623093682,
127
+ "grad_norm": 53.504234313964844,
128
+ "learning_rate": 7.9e-06,
129
+ "loss": 1.4267,
130
+ "step": 400
131
+ },
132
+ {
133
+ "epoch": 3.505446623093682,
134
+ "eval_loss": 0.7783711552619934,
135
+ "eval_runtime": 133.3648,
136
+ "eval_samples_per_second": 1.612,
137
+ "eval_steps_per_second": 0.81,
138
+ "eval_wer": 0.48498635122838946,
139
+ "step": 400
140
+ },
141
+ {
142
+ "epoch": 3.7233115468409586,
143
+ "grad_norm": 36.2320671081543,
144
+ "learning_rate": 8.400000000000001e-06,
145
+ "loss": 1.561,
146
+ "step": 425
147
+ },
148
+ {
149
+ "epoch": 3.9411764705882355,
150
+ "grad_norm": 37.68352508544922,
151
+ "learning_rate": 8.900000000000001e-06,
152
+ "loss": 1.4199,
153
+ "step": 450
154
+ },
155
+ {
156
+ "epoch": 4.16557734204793,
157
+ "grad_norm": 30.85091209411621,
158
+ "learning_rate": 9.4e-06,
159
+ "loss": 1.078,
160
+ "step": 475
161
+ },
162
+ {
163
+ "epoch": 4.383442265795207,
164
+ "grad_norm": 31.413185119628906,
165
+ "learning_rate": 9.9e-06,
166
+ "loss": 0.7431,
167
+ "step": 500
168
+ },
169
+ {
170
+ "epoch": 4.601307189542483,
171
+ "grad_norm": 29.73971939086914,
172
+ "learning_rate": 9.955555555555556e-06,
173
+ "loss": 0.8453,
174
+ "step": 525
175
+ },
176
+ {
177
+ "epoch": 4.819172113289761,
178
+ "grad_norm": 32.07612609863281,
179
+ "learning_rate": 9.9e-06,
180
+ "loss": 0.8495,
181
+ "step": 550
182
+ },
183
+ {
184
+ "epoch": 5.0435729847494555,
185
+ "grad_norm": 35.00574493408203,
186
+ "learning_rate": 9.844444444444446e-06,
187
+ "loss": 0.8732,
188
+ "step": 575
189
+ },
190
+ {
191
+ "epoch": 5.261437908496732,
192
+ "grad_norm": 24.1536922454834,
193
+ "learning_rate": 9.78888888888889e-06,
194
+ "loss": 0.4437,
195
+ "step": 600
196
+ },
197
+ {
198
+ "epoch": 5.261437908496732,
199
+ "eval_loss": 0.783809244632721,
200
+ "eval_runtime": 142.8442,
201
+ "eval_samples_per_second": 1.505,
202
+ "eval_steps_per_second": 0.756,
203
+ "eval_wer": 0.5104640582347588,
204
+ "step": 600
205
+ },
206
+ {
207
+ "epoch": 5.4793028322440085,
208
+ "grad_norm": 24.725513458251953,
209
+ "learning_rate": 9.733333333333334e-06,
210
+ "loss": 0.4725,
211
+ "step": 625
212
+ },
213
+ {
214
+ "epoch": 5.697167755991286,
215
+ "grad_norm": 17.23900604248047,
216
+ "learning_rate": 9.677777777777778e-06,
217
+ "loss": 0.5129,
218
+ "step": 650
219
+ },
220
+ {
221
+ "epoch": 5.915032679738562,
222
+ "grad_norm": 21.22838592529297,
223
+ "learning_rate": 9.622222222222222e-06,
224
+ "loss": 0.4442,
225
+ "step": 675
226
+ },
227
+ {
228
+ "epoch": 6.139433551198257,
229
+ "grad_norm": 20.201953887939453,
230
+ "learning_rate": 9.566666666666668e-06,
231
+ "loss": 0.3515,
232
+ "step": 700
233
+ },
234
+ {
235
+ "epoch": 6.357298474945534,
236
+ "grad_norm": 28.23876190185547,
237
+ "learning_rate": 9.511111111111112e-06,
238
+ "loss": 0.2276,
239
+ "step": 725
240
+ },
241
+ {
242
+ "epoch": 6.57516339869281,
243
+ "grad_norm": 35.3404541015625,
244
+ "learning_rate": 9.455555555555557e-06,
245
+ "loss": 0.2374,
246
+ "step": 750
247
+ },
248
+ {
249
+ "epoch": 6.793028322440087,
250
+ "grad_norm": 21.424623489379883,
251
+ "learning_rate": 9.4e-06,
252
+ "loss": 0.2962,
253
+ "step": 775
254
+ },
255
+ {
256
+ "epoch": 7.017429193899782,
257
+ "grad_norm": 13.183309555053711,
258
+ "learning_rate": 9.344444444444446e-06,
259
+ "loss": 0.2779,
260
+ "step": 800
261
+ },
262
+ {
263
+ "epoch": 7.017429193899782,
264
+ "eval_loss": 0.8308559060096741,
265
+ "eval_runtime": 133.876,
266
+ "eval_samples_per_second": 1.606,
267
+ "eval_steps_per_second": 0.807,
268
+ "eval_wer": 0.4572338489535942,
269
+ "step": 800
270
+ },
271
+ {
272
+ "epoch": 7.235294117647059,
273
+ "grad_norm": 13.984848022460938,
274
+ "learning_rate": 9.28888888888889e-06,
275
+ "loss": 0.1554,
276
+ "step": 825
277
+ },
278
+ {
279
+ "epoch": 7.453159041394335,
280
+ "grad_norm": 18.76070785522461,
281
+ "learning_rate": 9.233333333333334e-06,
282
+ "loss": 0.1674,
283
+ "step": 850
284
+ },
285
+ {
286
+ "epoch": 7.671023965141612,
287
+ "grad_norm": 25.312210083007812,
288
+ "learning_rate": 9.17777777777778e-06,
289
+ "loss": 0.1438,
290
+ "step": 875
291
+ },
292
+ {
293
+ "epoch": 7.888888888888889,
294
+ "grad_norm": 19.22443962097168,
295
+ "learning_rate": 9.122222222222223e-06,
296
+ "loss": 0.1863,
297
+ "step": 900
298
+ },
299
+ {
300
+ "epoch": 8.113289760348584,
301
+ "grad_norm": 7.898106575012207,
302
+ "learning_rate": 9.066666666666667e-06,
303
+ "loss": 0.1604,
304
+ "step": 925
305
+ },
306
+ {
307
+ "epoch": 8.33115468409586,
308
+ "grad_norm": 19.490093231201172,
309
+ "learning_rate": 9.011111111111111e-06,
310
+ "loss": 0.1181,
311
+ "step": 950
312
+ },
313
+ {
314
+ "epoch": 8.549019607843137,
315
+ "grad_norm": 10.587133407592773,
316
+ "learning_rate": 8.955555555555555e-06,
317
+ "loss": 0.1209,
318
+ "step": 975
319
+ },
320
+ {
321
+ "epoch": 8.766884531590414,
322
+ "grad_norm": 22.309471130371094,
323
+ "learning_rate": 8.900000000000001e-06,
324
+ "loss": 0.1379,
325
+ "step": 1000
326
+ },
327
+ {
328
+ "epoch": 8.766884531590414,
329
+ "eval_loss": 0.8748428821563721,
330
+ "eval_runtime": 132.6338,
331
+ "eval_samples_per_second": 1.621,
332
+ "eval_steps_per_second": 0.814,
333
+ "eval_wer": 0.4604185623293904,
334
+ "step": 1000
335
+ },
336
+ {
337
+ "epoch": 8.766884531590414,
338
+ "step": 1000,
339
+ "total_flos": 8.19749861720064e+18,
340
+ "train_loss": 2.2032353465557097,
341
+ "train_runtime": 2667.1173,
342
+ "train_samples_per_second": 14.997,
343
+ "train_steps_per_second": 1.875
344
+ }
345
+ ],
346
+ "logging_steps": 25,
347
+ "max_steps": 5000,
348
+ "num_input_tokens_seen": 0,
349
+ "num_train_epochs": 44,
350
+ "save_steps": 200,
351
+ "stateful_callbacks": {
352
+ "EarlyStoppingCallback": {
353
+ "args": {
354
+ "early_stopping_patience": 3,
355
+ "early_stopping_threshold": 0.0
356
+ },
357
+ "attributes": {
358
+ "early_stopping_patience_counter": 3
359
+ }
360
+ },
361
+ "TrainerControl": {
362
+ "args": {
363
+ "should_epoch_stop": false,
364
+ "should_evaluate": false,
365
+ "should_log": false,
366
+ "should_save": true,
367
+ "should_training_stop": true
368
+ },
369
+ "attributes": {}
370
+ }
371
+ },
372
+ "total_flos": 8.19749861720064e+18,
373
+ "train_batch_size": 2,
374
+ "trial_name": null,
375
+ "trial_params": null
376
+ }