warmestman commited on
Commit
565148a
·
verified ·
1 Parent(s): a8d188b

Training in progress, step 9000, checkpoint

Browse files
last-checkpoint/model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da65298fcbe65ec90ace7a34200b691786bbbf741cb0bcefde53bae793c0c150
3
  size 4993448880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13b0b8605d056f8ee7ce244a61451dfe6635de98c1b69d84b5289c60f0ab8eaa
3
  size 4993448880
last-checkpoint/model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:05b4684e020b85bbfca68cd478e00f24ed336b20da011cf724b3610bbdaf0ebb
3
  size 1180663192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e6087bf4500e0ab3b2cab5f97f24befceaa6ebbfdae315b4bd2490c05a5d511
3
  size 1180663192
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bab032ab8a6d161826bec34f3047be0921deb634160832f7b92503b7128c5ea4
3
  size 3095446256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13e72aba5acacb0929adffd5014b3aaa0ab0fa8f3ab6b8a64e401afdf7596199
3
  size 3095446256
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2de0b7a38bec3fb98aa32f323336a214d83264d9703f7d612b7245ed18581e0
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33ddc40084ed0bf1ba89d8a115875f6b967da1b3460fbfda40f8da959e6708a6
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c52f6b6405132e6ee658c09ea95ff4a1e46ef9dee4259ea08581b6ffffd433f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1b89bc8aed79bb7d8a2beccf0e2fe565be37ff238354d0b75a836cc8219e4fe
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 41.91738628238271,
3
  "best_model_checkpoint": "warmestman/whisper-large-v3-mn-cv-fleurs/checkpoint-6000",
4
- "epoch": 47.90419161676647,
5
  "eval_steps": 1000,
6
- "global_step": 8000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1999,6 +1999,255 @@
1999
  "eval_steps_per_second": 0.089,
2000
  "eval_wer": 42.22540014300644,
2001
  "step": 8000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2002
  }
2003
  ],
2004
  "logging_steps": 25,
@@ -2006,7 +2255,7 @@
2006
  "num_input_tokens_seen": 0,
2007
  "num_train_epochs": 120,
2008
  "save_steps": 1000,
2009
- "total_flos": 4.3440073597845504e+20,
2010
  "train_batch_size": 16,
2011
  "trial_name": null,
2012
  "trial_params": null
 
1
  {
2
  "best_metric": 41.91738628238271,
3
  "best_model_checkpoint": "warmestman/whisper-large-v3-mn-cv-fleurs/checkpoint-6000",
4
+ "epoch": 53.89221556886228,
5
  "eval_steps": 1000,
6
+ "global_step": 9000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1999
  "eval_steps_per_second": 0.089,
2000
  "eval_wer": 42.22540014300644,
2001
  "step": 8000
2002
+ },
2003
+ {
2004
+ "epoch": 48.05,
2005
+ "learning_rate": 6.142051282051281e-07,
2006
+ "loss": 0.0005,
2007
+ "step": 8025
2008
+ },
2009
+ {
2010
+ "epoch": 48.2,
2011
+ "learning_rate": 6.129230769230769e-07,
2012
+ "loss": 0.0004,
2013
+ "step": 8050
2014
+ },
2015
+ {
2016
+ "epoch": 48.35,
2017
+ "learning_rate": 6.116410256410256e-07,
2018
+ "loss": 0.0004,
2019
+ "step": 8075
2020
+ },
2021
+ {
2022
+ "epoch": 48.5,
2023
+ "learning_rate": 6.103589743589743e-07,
2024
+ "loss": 0.0004,
2025
+ "step": 8100
2026
+ },
2027
+ {
2028
+ "epoch": 48.65,
2029
+ "learning_rate": 6.09076923076923e-07,
2030
+ "loss": 0.0004,
2031
+ "step": 8125
2032
+ },
2033
+ {
2034
+ "epoch": 48.8,
2035
+ "learning_rate": 6.077948717948718e-07,
2036
+ "loss": 0.0004,
2037
+ "step": 8150
2038
+ },
2039
+ {
2040
+ "epoch": 48.95,
2041
+ "learning_rate": 6.065128205128205e-07,
2042
+ "loss": 0.0005,
2043
+ "step": 8175
2044
+ },
2045
+ {
2046
+ "epoch": 49.1,
2047
+ "learning_rate": 6.052307692307692e-07,
2048
+ "loss": 0.0004,
2049
+ "step": 8200
2050
+ },
2051
+ {
2052
+ "epoch": 49.25,
2053
+ "learning_rate": 6.039487179487179e-07,
2054
+ "loss": 0.0004,
2055
+ "step": 8225
2056
+ },
2057
+ {
2058
+ "epoch": 49.4,
2059
+ "learning_rate": 6.026666666666667e-07,
2060
+ "loss": 0.0004,
2061
+ "step": 8250
2062
+ },
2063
+ {
2064
+ "epoch": 49.55,
2065
+ "learning_rate": 6.013846153846154e-07,
2066
+ "loss": 0.0004,
2067
+ "step": 8275
2068
+ },
2069
+ {
2070
+ "epoch": 49.7,
2071
+ "learning_rate": 6.001025641025641e-07,
2072
+ "loss": 0.0004,
2073
+ "step": 8300
2074
+ },
2075
+ {
2076
+ "epoch": 49.85,
2077
+ "learning_rate": 5.988205128205128e-07,
2078
+ "loss": 0.0004,
2079
+ "step": 8325
2080
+ },
2081
+ {
2082
+ "epoch": 50.0,
2083
+ "learning_rate": 5.975384615384616e-07,
2084
+ "loss": 0.0004,
2085
+ "step": 8350
2086
+ },
2087
+ {
2088
+ "epoch": 50.15,
2089
+ "learning_rate": 5.962564102564103e-07,
2090
+ "loss": 0.0003,
2091
+ "step": 8375
2092
+ },
2093
+ {
2094
+ "epoch": 50.3,
2095
+ "learning_rate": 5.94974358974359e-07,
2096
+ "loss": 0.0004,
2097
+ "step": 8400
2098
+ },
2099
+ {
2100
+ "epoch": 50.45,
2101
+ "learning_rate": 5.936923076923077e-07,
2102
+ "loss": 0.0003,
2103
+ "step": 8425
2104
+ },
2105
+ {
2106
+ "epoch": 50.6,
2107
+ "learning_rate": 5.924102564102565e-07,
2108
+ "loss": 0.0003,
2109
+ "step": 8450
2110
+ },
2111
+ {
2112
+ "epoch": 50.75,
2113
+ "learning_rate": 5.911282051282051e-07,
2114
+ "loss": 0.0003,
2115
+ "step": 8475
2116
+ },
2117
+ {
2118
+ "epoch": 50.9,
2119
+ "learning_rate": 5.898461538461538e-07,
2120
+ "loss": 0.0003,
2121
+ "step": 8500
2122
+ },
2123
+ {
2124
+ "epoch": 51.05,
2125
+ "learning_rate": 5.885641025641025e-07,
2126
+ "loss": 0.0003,
2127
+ "step": 8525
2128
+ },
2129
+ {
2130
+ "epoch": 51.2,
2131
+ "learning_rate": 5.872820512820513e-07,
2132
+ "loss": 0.0003,
2133
+ "step": 8550
2134
+ },
2135
+ {
2136
+ "epoch": 51.35,
2137
+ "learning_rate": 5.86e-07,
2138
+ "loss": 0.0003,
2139
+ "step": 8575
2140
+ },
2141
+ {
2142
+ "epoch": 51.5,
2143
+ "learning_rate": 5.847179487179487e-07,
2144
+ "loss": 0.0003,
2145
+ "step": 8600
2146
+ },
2147
+ {
2148
+ "epoch": 51.65,
2149
+ "learning_rate": 5.834358974358974e-07,
2150
+ "loss": 0.0003,
2151
+ "step": 8625
2152
+ },
2153
+ {
2154
+ "epoch": 51.8,
2155
+ "learning_rate": 5.821538461538462e-07,
2156
+ "loss": 0.0003,
2157
+ "step": 8650
2158
+ },
2159
+ {
2160
+ "epoch": 51.95,
2161
+ "learning_rate": 5.808717948717949e-07,
2162
+ "loss": 0.0003,
2163
+ "step": 8675
2164
+ },
2165
+ {
2166
+ "epoch": 52.1,
2167
+ "learning_rate": 5.795897435897436e-07,
2168
+ "loss": 0.0003,
2169
+ "step": 8700
2170
+ },
2171
+ {
2172
+ "epoch": 52.25,
2173
+ "learning_rate": 5.783076923076922e-07,
2174
+ "loss": 0.0003,
2175
+ "step": 8725
2176
+ },
2177
+ {
2178
+ "epoch": 52.4,
2179
+ "learning_rate": 5.77025641025641e-07,
2180
+ "loss": 0.0003,
2181
+ "step": 8750
2182
+ },
2183
+ {
2184
+ "epoch": 52.54,
2185
+ "learning_rate": 5.757435897435897e-07,
2186
+ "loss": 0.0003,
2187
+ "step": 8775
2188
+ },
2189
+ {
2190
+ "epoch": 52.69,
2191
+ "learning_rate": 5.744615384615384e-07,
2192
+ "loss": 0.0003,
2193
+ "step": 8800
2194
+ },
2195
+ {
2196
+ "epoch": 52.84,
2197
+ "learning_rate": 5.731794871794871e-07,
2198
+ "loss": 0.0003,
2199
+ "step": 8825
2200
+ },
2201
+ {
2202
+ "epoch": 52.99,
2203
+ "learning_rate": 5.718974358974358e-07,
2204
+ "loss": 0.0003,
2205
+ "step": 8850
2206
+ },
2207
+ {
2208
+ "epoch": 53.14,
2209
+ "learning_rate": 5.706153846153846e-07,
2210
+ "loss": 0.0003,
2211
+ "step": 8875
2212
+ },
2213
+ {
2214
+ "epoch": 53.29,
2215
+ "learning_rate": 5.693333333333333e-07,
2216
+ "loss": 0.0003,
2217
+ "step": 8900
2218
+ },
2219
+ {
2220
+ "epoch": 53.44,
2221
+ "learning_rate": 5.68051282051282e-07,
2222
+ "loss": 0.0003,
2223
+ "step": 8925
2224
+ },
2225
+ {
2226
+ "epoch": 53.59,
2227
+ "learning_rate": 5.667692307692307e-07,
2228
+ "loss": 0.0003,
2229
+ "step": 8950
2230
+ },
2231
+ {
2232
+ "epoch": 53.74,
2233
+ "learning_rate": 5.654871794871795e-07,
2234
+ "loss": 0.0003,
2235
+ "step": 8975
2236
+ },
2237
+ {
2238
+ "epoch": 53.89,
2239
+ "learning_rate": 5.642051282051282e-07,
2240
+ "loss": 0.0003,
2241
+ "step": 9000
2242
+ },
2243
+ {
2244
+ "epoch": 53.89,
2245
+ "eval_loss": 0.7420342564582825,
2246
+ "eval_runtime": 596.5217,
2247
+ "eval_samples_per_second": 0.702,
2248
+ "eval_steps_per_second": 0.089,
2249
+ "eval_wer": 42.20889940047302,
2250
+ "step": 9000
2251
  }
2252
  ],
2253
  "logging_steps": 25,
 
2255
  "num_input_tokens_seen": 0,
2256
  "num_train_epochs": 120,
2257
  "save_steps": 1000,
2258
+ "total_flos": 4.8869955391389696e+20,
2259
  "train_batch_size": 16,
2260
  "trial_name": null,
2261
  "trial_params": null