Aditya3107 commited on
Commit
f8fa2aa
·
verified ·
1 Parent(s): 8714407

End of training

Browse files
Files changed (5) hide show
  1. README.md +4 -3
  2. all_results.json +16 -0
  3. eval_results.json +10 -0
  4. train_results.json +9 -0
  5. trainer_state.json +388 -0
README.md CHANGED
@@ -3,6 +3,7 @@ library_name: transformers
3
  license: apache-2.0
4
  base_model: facebook/wav2vec2-large-xlsr-53
5
  tags:
 
6
  - generated_from_trainer
7
  metrics:
8
  - wer
@@ -16,11 +17,11 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # wav2vec2-MYST-demo-dist
18
 
19
- This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 22.9483
22
  - Wer: 1.0
23
- - Cer: 1.1343
24
 
25
  ## Model description
26
 
 
3
  license: apache-2.0
4
  base_model: facebook/wav2vec2-large-xlsr-53
5
  tags:
6
+ - automatic-speech-recognition
7
  - generated_from_trainer
8
  metrics:
9
  - wer
 
17
 
18
  # wav2vec2-MYST-demo-dist
19
 
20
+ This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the MYST dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 19.7114
23
  - Wer: 1.0
24
+ - Cer: 0.9951
25
 
26
  ## Model description
27
 
all_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "eval_cer": 0.99505515405097,
4
+ "eval_loss": 19.71142578125,
5
+ "eval_runtime": 1.7764,
6
+ "eval_samples": 55,
7
+ "eval_samples_per_second": 30.962,
8
+ "eval_steps_per_second": 2.252,
9
+ "eval_wer": 1.0,
10
+ "total_flos": 7.281788547445555e+16,
11
+ "train_loss": 16.014588832855225,
12
+ "train_runtime": 78.9646,
13
+ "train_samples": 192,
14
+ "train_samples_per_second": 4.863,
15
+ "train_steps_per_second": 0.608
16
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "eval_cer": 0.99505515405097,
4
+ "eval_loss": 19.71142578125,
5
+ "eval_runtime": 1.7764,
6
+ "eval_samples": 55,
7
+ "eval_samples_per_second": 30.962,
8
+ "eval_steps_per_second": 2.252,
9
+ "eval_wer": 1.0
10
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "total_flos": 7.281788547445555e+16,
4
+ "train_loss": 16.014588832855225,
5
+ "train_runtime": 78.9646,
6
+ "train_samples": 192,
7
+ "train_samples_per_second": 4.863,
8
+ "train_steps_per_second": 0.608
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,388 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.0,
5
+ "eval_steps": 40,
6
+ "global_step": 48,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.041666666666666664,
13
+ "grad_norm": Infinity,
14
+ "learning_rate": 0.0,
15
+ "loss": 11.0247,
16
+ "step": 1
17
+ },
18
+ {
19
+ "epoch": 0.08333333333333333,
20
+ "grad_norm": Infinity,
21
+ "learning_rate": 0.0,
22
+ "loss": 12.7514,
23
+ "step": 2
24
+ },
25
+ {
26
+ "epoch": 0.125,
27
+ "grad_norm": 8.806501388549805,
28
+ "learning_rate": 6e-07,
29
+ "loss": 14.0529,
30
+ "step": 3
31
+ },
32
+ {
33
+ "epoch": 0.16666666666666666,
34
+ "grad_norm": Infinity,
35
+ "learning_rate": 6e-07,
36
+ "loss": 15.7234,
37
+ "step": 4
38
+ },
39
+ {
40
+ "epoch": 0.20833333333333334,
41
+ "grad_norm": Infinity,
42
+ "learning_rate": 6e-07,
43
+ "loss": 26.8783,
44
+ "step": 5
45
+ },
46
+ {
47
+ "epoch": 0.25,
48
+ "grad_norm": 20.22827911376953,
49
+ "learning_rate": 1.2e-06,
50
+ "loss": 24.4708,
51
+ "step": 6
52
+ },
53
+ {
54
+ "epoch": 0.2916666666666667,
55
+ "grad_norm": 7.415353775024414,
56
+ "learning_rate": 1.8e-06,
57
+ "loss": 11.0688,
58
+ "step": 7
59
+ },
60
+ {
61
+ "epoch": 0.3333333333333333,
62
+ "grad_norm": 7.202558517456055,
63
+ "learning_rate": 2.4e-06,
64
+ "loss": 11.3372,
65
+ "step": 8
66
+ },
67
+ {
68
+ "epoch": 0.375,
69
+ "grad_norm": 7.113807201385498,
70
+ "learning_rate": 2.9999999999999997e-06,
71
+ "loss": 11.8504,
72
+ "step": 9
73
+ },
74
+ {
75
+ "epoch": 0.4166666666666667,
76
+ "grad_norm": 7.7839155197143555,
77
+ "learning_rate": 3.6e-06,
78
+ "loss": 14.2548,
79
+ "step": 10
80
+ },
81
+ {
82
+ "epoch": 0.4583333333333333,
83
+ "grad_norm": 15.441122055053711,
84
+ "learning_rate": 4.2e-06,
85
+ "loss": 30.0114,
86
+ "step": 11
87
+ },
88
+ {
89
+ "epoch": 0.5,
90
+ "grad_norm": 10.832345008850098,
91
+ "learning_rate": 4.8e-06,
92
+ "loss": 18.9673,
93
+ "step": 12
94
+ },
95
+ {
96
+ "epoch": 0.5416666666666666,
97
+ "grad_norm": 8.285051345825195,
98
+ "learning_rate": 5.399999999999999e-06,
99
+ "loss": 11.0917,
100
+ "step": 13
101
+ },
102
+ {
103
+ "epoch": 0.5833333333333334,
104
+ "grad_norm": 8.244933128356934,
105
+ "learning_rate": 5.999999999999999e-06,
106
+ "loss": 12.1921,
107
+ "step": 14
108
+ },
109
+ {
110
+ "epoch": 0.625,
111
+ "grad_norm": 8.572517395019531,
112
+ "learning_rate": 6.599999999999999e-06,
113
+ "loss": 12.7475,
114
+ "step": 15
115
+ },
116
+ {
117
+ "epoch": 0.6666666666666666,
118
+ "grad_norm": 9.099493026733398,
119
+ "learning_rate": 7.2e-06,
120
+ "loss": 13.8911,
121
+ "step": 16
122
+ },
123
+ {
124
+ "epoch": 0.7083333333333334,
125
+ "grad_norm": 14.261921882629395,
126
+ "learning_rate": 7.799999999999998e-06,
127
+ "loss": 24.8105,
128
+ "step": 17
129
+ },
130
+ {
131
+ "epoch": 0.75,
132
+ "grad_norm": 7.782045841217041,
133
+ "learning_rate": 8.4e-06,
134
+ "loss": 15.1234,
135
+ "step": 18
136
+ },
137
+ {
138
+ "epoch": 0.7916666666666666,
139
+ "grad_norm": 9.227529525756836,
140
+ "learning_rate": 8.999999999999999e-06,
141
+ "loss": 13.6529,
142
+ "step": 19
143
+ },
144
+ {
145
+ "epoch": 0.8333333333333334,
146
+ "grad_norm": 7.739823818206787,
147
+ "learning_rate": 9.6e-06,
148
+ "loss": 11.3012,
149
+ "step": 20
150
+ },
151
+ {
152
+ "epoch": 0.875,
153
+ "grad_norm": 14.564947128295898,
154
+ "learning_rate": 1.02e-05,
155
+ "loss": 23.4074,
156
+ "step": 21
157
+ },
158
+ {
159
+ "epoch": 0.9166666666666666,
160
+ "grad_norm": 9.717215538024902,
161
+ "learning_rate": 1.0799999999999998e-05,
162
+ "loss": 16.9068,
163
+ "step": 22
164
+ },
165
+ {
166
+ "epoch": 0.9583333333333334,
167
+ "grad_norm": 9.182290077209473,
168
+ "learning_rate": 1.14e-05,
169
+ "loss": 17.5192,
170
+ "step": 23
171
+ },
172
+ {
173
+ "epoch": 1.0,
174
+ "grad_norm": 15.464149475097656,
175
+ "learning_rate": 1.1999999999999999e-05,
176
+ "loss": 20.7789,
177
+ "step": 24
178
+ },
179
+ {
180
+ "epoch": 1.0416666666666667,
181
+ "grad_norm": 7.396650791168213,
182
+ "learning_rate": 1.26e-05,
183
+ "loss": 10.7157,
184
+ "step": 25
185
+ },
186
+ {
187
+ "epoch": 1.0833333333333333,
188
+ "grad_norm": 9.421314239501953,
189
+ "learning_rate": 1.3199999999999997e-05,
190
+ "loss": 14.0375,
191
+ "step": 26
192
+ },
193
+ {
194
+ "epoch": 1.125,
195
+ "grad_norm": 15.237442016601562,
196
+ "learning_rate": 1.3799999999999998e-05,
197
+ "loss": 24.223,
198
+ "step": 27
199
+ },
200
+ {
201
+ "epoch": 1.1666666666666667,
202
+ "grad_norm": 9.531441688537598,
203
+ "learning_rate": 1.44e-05,
204
+ "loss": 15.9401,
205
+ "step": 28
206
+ },
207
+ {
208
+ "epoch": 1.2083333333333333,
209
+ "grad_norm": 12.435009002685547,
210
+ "learning_rate": 1.4999999999999999e-05,
211
+ "loss": 23.032,
212
+ "step": 29
213
+ },
214
+ {
215
+ "epoch": 1.25,
216
+ "grad_norm": 14.035198211669922,
217
+ "learning_rate": 1.5599999999999996e-05,
218
+ "loss": 16.6483,
219
+ "step": 30
220
+ },
221
+ {
222
+ "epoch": 1.2916666666666667,
223
+ "grad_norm": 8.37588882446289,
224
+ "learning_rate": 1.6199999999999997e-05,
225
+ "loss": 12.122,
226
+ "step": 31
227
+ },
228
+ {
229
+ "epoch": 1.3333333333333333,
230
+ "grad_norm": 7.98301362991333,
231
+ "learning_rate": 1.68e-05,
232
+ "loss": 12.153,
233
+ "step": 32
234
+ },
235
+ {
236
+ "epoch": 1.375,
237
+ "grad_norm": 9.900574684143066,
238
+ "learning_rate": 1.74e-05,
239
+ "loss": 16.227,
240
+ "step": 33
241
+ },
242
+ {
243
+ "epoch": 1.4166666666666667,
244
+ "grad_norm": 6.607911109924316,
245
+ "learning_rate": 1.7999999999999997e-05,
246
+ "loss": 11.2944,
247
+ "step": 34
248
+ },
249
+ {
250
+ "epoch": 1.4583333333333333,
251
+ "grad_norm": 21.229183197021484,
252
+ "learning_rate": 1.8599999999999998e-05,
253
+ "loss": 27.8815,
254
+ "step": 35
255
+ },
256
+ {
257
+ "epoch": 1.5,
258
+ "grad_norm": 17.414100646972656,
259
+ "learning_rate": 1.92e-05,
260
+ "loss": 18.9613,
261
+ "step": 36
262
+ },
263
+ {
264
+ "epoch": 1.5416666666666665,
265
+ "grad_norm": 7.7176361083984375,
266
+ "learning_rate": 1.98e-05,
267
+ "loss": 11.2252,
268
+ "step": 37
269
+ },
270
+ {
271
+ "epoch": 1.5833333333333335,
272
+ "grad_norm": 6.465058326721191,
273
+ "learning_rate": 2.04e-05,
274
+ "loss": 9.7911,
275
+ "step": 38
276
+ },
277
+ {
278
+ "epoch": 1.625,
279
+ "grad_norm": 10.792200088500977,
280
+ "learning_rate": 2.1e-05,
281
+ "loss": 16.627,
282
+ "step": 39
283
+ },
284
+ {
285
+ "epoch": 1.6666666666666665,
286
+ "grad_norm": 5.8488054275512695,
287
+ "learning_rate": 2.1599999999999996e-05,
288
+ "loss": 10.3385,
289
+ "step": 40
290
+ },
291
+ {
292
+ "epoch": 1.6666666666666665,
293
+ "eval_cer": 1.1342715861544312,
294
+ "eval_loss": 22.94832420349121,
295
+ "eval_runtime": 1.8395,
296
+ "eval_samples_per_second": 29.899,
297
+ "eval_steps_per_second": 2.174,
298
+ "eval_wer": 1.0,
299
+ "step": 40
300
+ },
301
+ {
302
+ "epoch": 1.7083333333333335,
303
+ "grad_norm": 21.320125579833984,
304
+ "learning_rate": 2.2199999999999998e-05,
305
+ "loss": 24.4931,
306
+ "step": 41
307
+ },
308
+ {
309
+ "epoch": 1.75,
310
+ "grad_norm": 21.40454864501953,
311
+ "learning_rate": 2.28e-05,
312
+ "loss": 19.2591,
313
+ "step": 42
314
+ },
315
+ {
316
+ "epoch": 1.7916666666666665,
317
+ "grad_norm": 6.241701126098633,
318
+ "learning_rate": 2.34e-05,
319
+ "loss": 10.151,
320
+ "step": 43
321
+ },
322
+ {
323
+ "epoch": 1.8333333333333335,
324
+ "grad_norm": 10.232379913330078,
325
+ "learning_rate": 2.3999999999999997e-05,
326
+ "loss": 11.456,
327
+ "step": 44
328
+ },
329
+ {
330
+ "epoch": 1.875,
331
+ "grad_norm": 10.987034797668457,
332
+ "learning_rate": 2.4599999999999998e-05,
333
+ "loss": 10.9538,
334
+ "step": 45
335
+ },
336
+ {
337
+ "epoch": 1.9166666666666665,
338
+ "grad_norm": 24.182756423950195,
339
+ "learning_rate": 2.52e-05,
340
+ "loss": 16.68,
341
+ "step": 46
342
+ },
343
+ {
344
+ "epoch": 1.9583333333333335,
345
+ "grad_norm": 15.300764083862305,
346
+ "learning_rate": 2.5799999999999997e-05,
347
+ "loss": 13.9462,
348
+ "step": 47
349
+ },
350
+ {
351
+ "epoch": 2.0,
352
+ "grad_norm": 19.339834213256836,
353
+ "learning_rate": 2.6399999999999995e-05,
354
+ "loss": 14.7293,
355
+ "step": 48
356
+ },
357
+ {
358
+ "epoch": 2.0,
359
+ "step": 48,
360
+ "total_flos": 7.281788547445555e+16,
361
+ "train_loss": 16.014588832855225,
362
+ "train_runtime": 78.9646,
363
+ "train_samples_per_second": 4.863,
364
+ "train_steps_per_second": 0.608
365
+ }
366
+ ],
367
+ "logging_steps": 1.0,
368
+ "max_steps": 48,
369
+ "num_input_tokens_seen": 0,
370
+ "num_train_epochs": 2,
371
+ "save_steps": 400,
372
+ "stateful_callbacks": {
373
+ "TrainerControl": {
374
+ "args": {
375
+ "should_epoch_stop": false,
376
+ "should_evaluate": false,
377
+ "should_log": false,
378
+ "should_save": true,
379
+ "should_training_stop": true
380
+ },
381
+ "attributes": {}
382
+ }
383
+ },
384
+ "total_flos": 7.281788547445555e+16,
385
+ "train_batch_size": 4,
386
+ "trial_name": null,
387
+ "trial_params": null
388
+ }