abdiharyadi
commited on
Training in progress, epoch 13
Browse files- last-checkpoint/model.safetensors +1 -1
- last-checkpoint/optimizer.pt +1 -1
- last-checkpoint/rng_state.pth +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +5 -1173
- model.safetensors +1 -1
- run-2024-10-28T15:22:08+00:00.log +0 -0
- val_outputs/val_generated_predictions_23953367.txt +0 -0
last-checkpoint/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1575259780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b35a3bc2ac180ed070b42029c9a9dd327a1a9559e81df276f329c07eb21d04fc
|
3 |
size 1575259780
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3150397656
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba79276529acf6b7fcac21ab5be5fc5756c900e697415161ce71f7759f9fa8e0
|
3 |
size 3150397656
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e52ca2f5b1048c2984d9cb01ff8bc5c06ec7e6e1ac850eb54ef8fe7147dcf65
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df05e968bcb6cb9f8c607bdaf90fbac1131121f0efd29f2e5e7bc42c79c2d577
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"best_metric": 12.
|
3 |
-
"best_model_checkpoint": "/kaggle/tmp/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted/checkpoint-
|
4 |
-
"epoch": 11.
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -12891,1174 +12891,6 @@
|
|
12891 |
"eval_samples_per_second": 2.165,
|
12892 |
"eval_steps_per_second": 1.083,
|
12893 |
"step": 42562
|
12894 |
-
},
|
12895 |
-
{
|
12896 |
-
"epoch": 11.0054270577594,
|
12897 |
-
"learning_rate": 1.3131725657980034e-07,
|
12898 |
-
"loss": 2.4441,
|
12899 |
-
"step": 42580
|
12900 |
-
},
|
12901 |
-
{
|
12902 |
-
"epoch": 11.010595684196925,
|
12903 |
-
"learning_rate": 1.31284843770258e-07,
|
12904 |
-
"loss": 2.5027,
|
12905 |
-
"step": 42600
|
12906 |
-
},
|
12907 |
-
{
|
12908 |
-
"epoch": 11.015764310634449,
|
12909 |
-
"learning_rate": 1.3125243096071566e-07,
|
12910 |
-
"loss": 2.4709,
|
12911 |
-
"step": 42620
|
12912 |
-
},
|
12913 |
-
{
|
12914 |
-
"epoch": 11.020932937071974,
|
12915 |
-
"learning_rate": 1.3122001815117335e-07,
|
12916 |
-
"loss": 2.4576,
|
12917 |
-
"step": 42640
|
12918 |
-
},
|
12919 |
-
{
|
12920 |
-
"epoch": 11.026101563509497,
|
12921 |
-
"learning_rate": 1.31187605341631e-07,
|
12922 |
-
"loss": 2.4395,
|
12923 |
-
"step": 42660
|
12924 |
-
},
|
12925 |
-
{
|
12926 |
-
"epoch": 11.031270189947021,
|
12927 |
-
"learning_rate": 1.3115519253208867e-07,
|
12928 |
-
"loss": 2.504,
|
12929 |
-
"step": 42680
|
12930 |
-
},
|
12931 |
-
{
|
12932 |
-
"epoch": 11.036438816384546,
|
12933 |
-
"learning_rate": 1.3112277972254634e-07,
|
12934 |
-
"loss": 2.5521,
|
12935 |
-
"step": 42700
|
12936 |
-
},
|
12937 |
-
{
|
12938 |
-
"epoch": 11.04160744282207,
|
12939 |
-
"learning_rate": 1.31090366913004e-07,
|
12940 |
-
"loss": 2.5074,
|
12941 |
-
"step": 42720
|
12942 |
-
},
|
12943 |
-
{
|
12944 |
-
"epoch": 11.046776069259595,
|
12945 |
-
"learning_rate": 1.3105795410346169e-07,
|
12946 |
-
"loss": 2.5079,
|
12947 |
-
"step": 42740
|
12948 |
-
},
|
12949 |
-
{
|
12950 |
-
"epoch": 11.051944695697118,
|
12951 |
-
"learning_rate": 1.3102554129391935e-07,
|
12952 |
-
"loss": 2.5337,
|
12953 |
-
"step": 42760
|
12954 |
-
},
|
12955 |
-
{
|
12956 |
-
"epoch": 11.057113322134644,
|
12957 |
-
"learning_rate": 1.30993128484377e-07,
|
12958 |
-
"loss": 2.4815,
|
12959 |
-
"step": 42780
|
12960 |
-
},
|
12961 |
-
{
|
12962 |
-
"epoch": 11.062281948572167,
|
12963 |
-
"learning_rate": 1.309607156748347e-07,
|
12964 |
-
"loss": 2.4897,
|
12965 |
-
"step": 42800
|
12966 |
-
},
|
12967 |
-
{
|
12968 |
-
"epoch": 11.06745057500969,
|
12969 |
-
"learning_rate": 1.3092830286529236e-07,
|
12970 |
-
"loss": 2.5184,
|
12971 |
-
"step": 42820
|
12972 |
-
},
|
12973 |
-
{
|
12974 |
-
"epoch": 11.072619201447216,
|
12975 |
-
"learning_rate": 1.3089589005575002e-07,
|
12976 |
-
"loss": 2.459,
|
12977 |
-
"step": 42840
|
12978 |
-
},
|
12979 |
-
{
|
12980 |
-
"epoch": 11.07778782788474,
|
12981 |
-
"learning_rate": 1.308634772462077e-07,
|
12982 |
-
"loss": 2.4623,
|
12983 |
-
"step": 42860
|
12984 |
-
},
|
12985 |
-
{
|
12986 |
-
"epoch": 11.082956454322265,
|
12987 |
-
"learning_rate": 1.3083106443666535e-07,
|
12988 |
-
"loss": 2.4422,
|
12989 |
-
"step": 42880
|
12990 |
-
},
|
12991 |
-
{
|
12992 |
-
"epoch": 11.088125080759788,
|
12993 |
-
"learning_rate": 1.3079865162712303e-07,
|
12994 |
-
"loss": 2.5212,
|
12995 |
-
"step": 42900
|
12996 |
-
},
|
12997 |
-
{
|
12998 |
-
"epoch": 11.093293707197311,
|
12999 |
-
"learning_rate": 1.307662388175807e-07,
|
13000 |
-
"loss": 2.4923,
|
13001 |
-
"step": 42920
|
13002 |
-
},
|
13003 |
-
{
|
13004 |
-
"epoch": 11.098462333634837,
|
13005 |
-
"learning_rate": 1.3073382600803836e-07,
|
13006 |
-
"loss": 2.5076,
|
13007 |
-
"step": 42940
|
13008 |
-
},
|
13009 |
-
{
|
13010 |
-
"epoch": 11.10363096007236,
|
13011 |
-
"learning_rate": 1.3070141319849605e-07,
|
13012 |
-
"loss": 2.5291,
|
13013 |
-
"step": 42960
|
13014 |
-
},
|
13015 |
-
{
|
13016 |
-
"epoch": 11.108799586509885,
|
13017 |
-
"learning_rate": 1.306690003889537e-07,
|
13018 |
-
"loss": 2.4459,
|
13019 |
-
"step": 42980
|
13020 |
-
},
|
13021 |
-
{
|
13022 |
-
"epoch": 11.113968212947409,
|
13023 |
-
"learning_rate": 1.3063658757941137e-07,
|
13024 |
-
"loss": 2.4522,
|
13025 |
-
"step": 43000
|
13026 |
-
},
|
13027 |
-
{
|
13028 |
-
"epoch": 11.119136839384934,
|
13029 |
-
"learning_rate": 1.3060417476986906e-07,
|
13030 |
-
"loss": 2.4752,
|
13031 |
-
"step": 43020
|
13032 |
-
},
|
13033 |
-
{
|
13034 |
-
"epoch": 11.124305465822458,
|
13035 |
-
"learning_rate": 1.3057176196032672e-07,
|
13036 |
-
"loss": 2.4877,
|
13037 |
-
"step": 43040
|
13038 |
-
},
|
13039 |
-
{
|
13040 |
-
"epoch": 11.129474092259981,
|
13041 |
-
"learning_rate": 1.3053934915078438e-07,
|
13042 |
-
"loss": 2.4969,
|
13043 |
-
"step": 43060
|
13044 |
-
},
|
13045 |
-
{
|
13046 |
-
"epoch": 11.134642718697506,
|
13047 |
-
"learning_rate": 1.3050693634124207e-07,
|
13048 |
-
"loss": 2.5331,
|
13049 |
-
"step": 43080
|
13050 |
-
},
|
13051 |
-
{
|
13052 |
-
"epoch": 11.13981134513503,
|
13053 |
-
"learning_rate": 1.304745235316997e-07,
|
13054 |
-
"loss": 2.4214,
|
13055 |
-
"step": 43100
|
13056 |
-
},
|
13057 |
-
{
|
13058 |
-
"epoch": 11.144979971572555,
|
13059 |
-
"learning_rate": 1.304421107221574e-07,
|
13060 |
-
"loss": 2.4795,
|
13061 |
-
"step": 43120
|
13062 |
-
},
|
13063 |
-
{
|
13064 |
-
"epoch": 11.150148598010079,
|
13065 |
-
"learning_rate": 1.3040969791261506e-07,
|
13066 |
-
"loss": 2.491,
|
13067 |
-
"step": 43140
|
13068 |
-
},
|
13069 |
-
{
|
13070 |
-
"epoch": 11.155317224447604,
|
13071 |
-
"learning_rate": 1.3037728510307272e-07,
|
13072 |
-
"loss": 2.4823,
|
13073 |
-
"step": 43160
|
13074 |
-
},
|
13075 |
-
{
|
13076 |
-
"epoch": 11.160485850885127,
|
13077 |
-
"learning_rate": 1.303448722935304e-07,
|
13078 |
-
"loss": 2.5129,
|
13079 |
-
"step": 43180
|
13080 |
-
},
|
13081 |
-
{
|
13082 |
-
"epoch": 11.16565447732265,
|
13083 |
-
"learning_rate": 1.3031245948398807e-07,
|
13084 |
-
"loss": 2.4543,
|
13085 |
-
"step": 43200
|
13086 |
-
},
|
13087 |
-
{
|
13088 |
-
"epoch": 11.170823103760176,
|
13089 |
-
"learning_rate": 1.3028004667444573e-07,
|
13090 |
-
"loss": 2.4735,
|
13091 |
-
"step": 43220
|
13092 |
-
},
|
13093 |
-
{
|
13094 |
-
"epoch": 11.1759917301977,
|
13095 |
-
"learning_rate": 1.3024763386490342e-07,
|
13096 |
-
"loss": 2.5127,
|
13097 |
-
"step": 43240
|
13098 |
-
},
|
13099 |
-
{
|
13100 |
-
"epoch": 11.181160356635225,
|
13101 |
-
"learning_rate": 1.3021522105536105e-07,
|
13102 |
-
"loss": 2.4045,
|
13103 |
-
"step": 43260
|
13104 |
-
},
|
13105 |
-
{
|
13106 |
-
"epoch": 11.186328983072748,
|
13107 |
-
"learning_rate": 1.3018280824581874e-07,
|
13108 |
-
"loss": 2.4599,
|
13109 |
-
"step": 43280
|
13110 |
-
},
|
13111 |
-
{
|
13112 |
-
"epoch": 11.191497609510273,
|
13113 |
-
"learning_rate": 1.301503954362764e-07,
|
13114 |
-
"loss": 2.4875,
|
13115 |
-
"step": 43300
|
13116 |
-
},
|
13117 |
-
{
|
13118 |
-
"epoch": 11.196666235947797,
|
13119 |
-
"learning_rate": 1.3011798262673407e-07,
|
13120 |
-
"loss": 2.4523,
|
13121 |
-
"step": 43320
|
13122 |
-
},
|
13123 |
-
{
|
13124 |
-
"epoch": 11.20183486238532,
|
13125 |
-
"learning_rate": 1.3008556981719175e-07,
|
13126 |
-
"loss": 2.47,
|
13127 |
-
"step": 43340
|
13128 |
-
},
|
13129 |
-
{
|
13130 |
-
"epoch": 11.207003488822846,
|
13131 |
-
"learning_rate": 1.3005315700764942e-07,
|
13132 |
-
"loss": 2.464,
|
13133 |
-
"step": 43360
|
13134 |
-
},
|
13135 |
-
{
|
13136 |
-
"epoch": 11.212172115260369,
|
13137 |
-
"learning_rate": 1.3002074419810708e-07,
|
13138 |
-
"loss": 2.5223,
|
13139 |
-
"step": 43380
|
13140 |
-
},
|
13141 |
-
{
|
13142 |
-
"epoch": 11.217340741697894,
|
13143 |
-
"learning_rate": 1.2998833138856477e-07,
|
13144 |
-
"loss": 2.4749,
|
13145 |
-
"step": 43400
|
13146 |
-
},
|
13147 |
-
{
|
13148 |
-
"epoch": 11.222509368135418,
|
13149 |
-
"learning_rate": 1.2995591857902243e-07,
|
13150 |
-
"loss": 2.5268,
|
13151 |
-
"step": 43420
|
13152 |
-
},
|
13153 |
-
{
|
13154 |
-
"epoch": 11.227677994572943,
|
13155 |
-
"learning_rate": 1.299235057694801e-07,
|
13156 |
-
"loss": 2.4837,
|
13157 |
-
"step": 43440
|
13158 |
-
},
|
13159 |
-
{
|
13160 |
-
"epoch": 11.232846621010466,
|
13161 |
-
"learning_rate": 1.2989109295993778e-07,
|
13162 |
-
"loss": 2.4285,
|
13163 |
-
"step": 43460
|
13164 |
-
},
|
13165 |
-
{
|
13166 |
-
"epoch": 11.23801524744799,
|
13167 |
-
"learning_rate": 1.2985868015039541e-07,
|
13168 |
-
"loss": 2.4918,
|
13169 |
-
"step": 43480
|
13170 |
-
},
|
13171 |
-
{
|
13172 |
-
"epoch": 11.243183873885515,
|
13173 |
-
"learning_rate": 1.298262673408531e-07,
|
13174 |
-
"loss": 2.4585,
|
13175 |
-
"step": 43500
|
13176 |
-
},
|
13177 |
-
{
|
13178 |
-
"epoch": 11.248352500323039,
|
13179 |
-
"learning_rate": 1.2979385453131076e-07,
|
13180 |
-
"loss": 2.5156,
|
13181 |
-
"step": 43520
|
13182 |
-
},
|
13183 |
-
{
|
13184 |
-
"epoch": 11.253521126760564,
|
13185 |
-
"learning_rate": 1.2976144172176843e-07,
|
13186 |
-
"loss": 2.5031,
|
13187 |
-
"step": 43540
|
13188 |
-
},
|
13189 |
-
{
|
13190 |
-
"epoch": 11.258689753198087,
|
13191 |
-
"learning_rate": 1.2972902891222611e-07,
|
13192 |
-
"loss": 2.456,
|
13193 |
-
"step": 43560
|
13194 |
-
},
|
13195 |
-
{
|
13196 |
-
"epoch": 11.263858379635613,
|
13197 |
-
"learning_rate": 1.2969661610268378e-07,
|
13198 |
-
"loss": 2.5083,
|
13199 |
-
"step": 43580
|
13200 |
-
},
|
13201 |
-
{
|
13202 |
-
"epoch": 11.269027006073136,
|
13203 |
-
"learning_rate": 1.2966420329314144e-07,
|
13204 |
-
"loss": 2.5241,
|
13205 |
-
"step": 43600
|
13206 |
-
},
|
13207 |
-
{
|
13208 |
-
"epoch": 11.27419563251066,
|
13209 |
-
"learning_rate": 1.2963179048359913e-07,
|
13210 |
-
"loss": 2.4751,
|
13211 |
-
"step": 43620
|
13212 |
-
},
|
13213 |
-
{
|
13214 |
-
"epoch": 11.279364258948185,
|
13215 |
-
"learning_rate": 1.295993776740568e-07,
|
13216 |
-
"loss": 2.4695,
|
13217 |
-
"step": 43640
|
13218 |
-
},
|
13219 |
-
{
|
13220 |
-
"epoch": 11.284532885385708,
|
13221 |
-
"learning_rate": 1.2956696486451445e-07,
|
13222 |
-
"loss": 2.5578,
|
13223 |
-
"step": 43660
|
13224 |
-
},
|
13225 |
-
{
|
13226 |
-
"epoch": 11.289701511823234,
|
13227 |
-
"learning_rate": 1.2953455205497214e-07,
|
13228 |
-
"loss": 2.4649,
|
13229 |
-
"step": 43680
|
13230 |
-
},
|
13231 |
-
{
|
13232 |
-
"epoch": 11.294870138260757,
|
13233 |
-
"learning_rate": 1.2950213924542977e-07,
|
13234 |
-
"loss": 2.4747,
|
13235 |
-
"step": 43700
|
13236 |
-
},
|
13237 |
-
{
|
13238 |
-
"epoch": 11.300038764698282,
|
13239 |
-
"learning_rate": 1.2946972643588746e-07,
|
13240 |
-
"loss": 2.4679,
|
13241 |
-
"step": 43720
|
13242 |
-
},
|
13243 |
-
{
|
13244 |
-
"epoch": 11.305207391135806,
|
13245 |
-
"learning_rate": 1.2943731362634512e-07,
|
13246 |
-
"loss": 2.4683,
|
13247 |
-
"step": 43740
|
13248 |
-
},
|
13249 |
-
{
|
13250 |
-
"epoch": 11.31037601757333,
|
13251 |
-
"learning_rate": 1.2940490081680279e-07,
|
13252 |
-
"loss": 2.4857,
|
13253 |
-
"step": 43760
|
13254 |
-
},
|
13255 |
-
{
|
13256 |
-
"epoch": 11.315544644010854,
|
13257 |
-
"learning_rate": 1.2937248800726047e-07,
|
13258 |
-
"loss": 2.426,
|
13259 |
-
"step": 43780
|
13260 |
-
},
|
13261 |
-
{
|
13262 |
-
"epoch": 11.320713270448378,
|
13263 |
-
"learning_rate": 1.2934007519771814e-07,
|
13264 |
-
"loss": 2.4733,
|
13265 |
-
"step": 43800
|
13266 |
-
},
|
13267 |
-
{
|
13268 |
-
"epoch": 11.325881896885903,
|
13269 |
-
"learning_rate": 1.293076623881758e-07,
|
13270 |
-
"loss": 2.4812,
|
13271 |
-
"step": 43820
|
13272 |
-
},
|
13273 |
-
{
|
13274 |
-
"epoch": 11.331050523323427,
|
13275 |
-
"learning_rate": 1.2927524957863349e-07,
|
13276 |
-
"loss": 2.4528,
|
13277 |
-
"step": 43840
|
13278 |
-
},
|
13279 |
-
{
|
13280 |
-
"epoch": 11.33621914976095,
|
13281 |
-
"learning_rate": 1.2924283676909112e-07,
|
13282 |
-
"loss": 2.494,
|
13283 |
-
"step": 43860
|
13284 |
-
},
|
13285 |
-
{
|
13286 |
-
"epoch": 11.341387776198475,
|
13287 |
-
"learning_rate": 1.292104239595488e-07,
|
13288 |
-
"loss": 2.4795,
|
13289 |
-
"step": 43880
|
13290 |
-
},
|
13291 |
-
{
|
13292 |
-
"epoch": 11.346556402635999,
|
13293 |
-
"learning_rate": 1.2917801115000647e-07,
|
13294 |
-
"loss": 2.4816,
|
13295 |
-
"step": 43900
|
13296 |
-
},
|
13297 |
-
{
|
13298 |
-
"epoch": 11.351725029073524,
|
13299 |
-
"learning_rate": 1.2914559834046413e-07,
|
13300 |
-
"loss": 2.5282,
|
13301 |
-
"step": 43920
|
13302 |
-
},
|
13303 |
-
{
|
13304 |
-
"epoch": 11.356893655511048,
|
13305 |
-
"learning_rate": 1.2911318553092182e-07,
|
13306 |
-
"loss": 2.4849,
|
13307 |
-
"step": 43940
|
13308 |
-
},
|
13309 |
-
{
|
13310 |
-
"epoch": 11.362062281948573,
|
13311 |
-
"learning_rate": 1.2908077272137948e-07,
|
13312 |
-
"loss": 2.4246,
|
13313 |
-
"step": 43960
|
13314 |
-
},
|
13315 |
-
{
|
13316 |
-
"epoch": 11.367230908386096,
|
13317 |
-
"learning_rate": 1.2904835991183715e-07,
|
13318 |
-
"loss": 2.5491,
|
13319 |
-
"step": 43980
|
13320 |
-
},
|
13321 |
-
{
|
13322 |
-
"epoch": 11.372399534823622,
|
13323 |
-
"learning_rate": 1.2901594710229483e-07,
|
13324 |
-
"loss": 2.4584,
|
13325 |
-
"step": 44000
|
13326 |
-
},
|
13327 |
-
{
|
13328 |
-
"epoch": 11.377568161261145,
|
13329 |
-
"learning_rate": 1.289835342927525e-07,
|
13330 |
-
"loss": 2.5351,
|
13331 |
-
"step": 44020
|
13332 |
-
},
|
13333 |
-
{
|
13334 |
-
"epoch": 11.382736787698668,
|
13335 |
-
"learning_rate": 1.2895112148321016e-07,
|
13336 |
-
"loss": 2.5108,
|
13337 |
-
"step": 44040
|
13338 |
-
},
|
13339 |
-
{
|
13340 |
-
"epoch": 11.387905414136194,
|
13341 |
-
"learning_rate": 1.2891870867366782e-07,
|
13342 |
-
"loss": 2.5328,
|
13343 |
-
"step": 44060
|
13344 |
-
},
|
13345 |
-
{
|
13346 |
-
"epoch": 11.393074040573717,
|
13347 |
-
"learning_rate": 1.2888629586412548e-07,
|
13348 |
-
"loss": 2.4501,
|
13349 |
-
"step": 44080
|
13350 |
-
},
|
13351 |
-
{
|
13352 |
-
"epoch": 11.398242667011242,
|
13353 |
-
"learning_rate": 1.2885388305458317e-07,
|
13354 |
-
"loss": 2.434,
|
13355 |
-
"step": 44100
|
13356 |
-
},
|
13357 |
-
{
|
13358 |
-
"epoch": 11.403411293448766,
|
13359 |
-
"learning_rate": 1.2882147024504083e-07,
|
13360 |
-
"loss": 2.4713,
|
13361 |
-
"step": 44120
|
13362 |
-
},
|
13363 |
-
{
|
13364 |
-
"epoch": 11.40857991988629,
|
13365 |
-
"learning_rate": 1.287890574354985e-07,
|
13366 |
-
"loss": 2.5093,
|
13367 |
-
"step": 44140
|
13368 |
-
},
|
13369 |
-
{
|
13370 |
-
"epoch": 11.413748546323815,
|
13371 |
-
"learning_rate": 1.2875664462595618e-07,
|
13372 |
-
"loss": 2.5031,
|
13373 |
-
"step": 44160
|
13374 |
-
},
|
13375 |
-
{
|
13376 |
-
"epoch": 11.418917172761338,
|
13377 |
-
"learning_rate": 1.2872423181641384e-07,
|
13378 |
-
"loss": 2.4712,
|
13379 |
-
"step": 44180
|
13380 |
-
},
|
13381 |
-
{
|
13382 |
-
"epoch": 11.424085799198863,
|
13383 |
-
"learning_rate": 1.286918190068715e-07,
|
13384 |
-
"loss": 2.4963,
|
13385 |
-
"step": 44200
|
13386 |
-
},
|
13387 |
-
{
|
13388 |
-
"epoch": 11.429254425636387,
|
13389 |
-
"learning_rate": 1.2865940619732917e-07,
|
13390 |
-
"loss": 2.4866,
|
13391 |
-
"step": 44220
|
13392 |
-
},
|
13393 |
-
{
|
13394 |
-
"epoch": 11.434423052073912,
|
13395 |
-
"learning_rate": 1.2862699338778686e-07,
|
13396 |
-
"loss": 2.4665,
|
13397 |
-
"step": 44240
|
13398 |
-
},
|
13399 |
-
{
|
13400 |
-
"epoch": 11.439591678511436,
|
13401 |
-
"learning_rate": 1.2859458057824452e-07,
|
13402 |
-
"loss": 2.4525,
|
13403 |
-
"step": 44260
|
13404 |
-
},
|
13405 |
-
{
|
13406 |
-
"epoch": 11.444760304948959,
|
13407 |
-
"learning_rate": 1.2856216776870218e-07,
|
13408 |
-
"loss": 2.5271,
|
13409 |
-
"step": 44280
|
13410 |
-
},
|
13411 |
-
{
|
13412 |
-
"epoch": 11.449928931386484,
|
13413 |
-
"learning_rate": 1.2852975495915984e-07,
|
13414 |
-
"loss": 2.4676,
|
13415 |
-
"step": 44300
|
13416 |
-
},
|
13417 |
-
{
|
13418 |
-
"epoch": 11.455097557824008,
|
13419 |
-
"learning_rate": 1.2849734214961753e-07,
|
13420 |
-
"loss": 2.4996,
|
13421 |
-
"step": 44320
|
13422 |
-
},
|
13423 |
-
{
|
13424 |
-
"epoch": 11.460266184261533,
|
13425 |
-
"learning_rate": 1.284649293400752e-07,
|
13426 |
-
"loss": 2.4901,
|
13427 |
-
"step": 44340
|
13428 |
-
},
|
13429 |
-
{
|
13430 |
-
"epoch": 11.465434810699056,
|
13431 |
-
"learning_rate": 1.2843251653053285e-07,
|
13432 |
-
"loss": 2.411,
|
13433 |
-
"step": 44360
|
13434 |
-
},
|
13435 |
-
{
|
13436 |
-
"epoch": 11.470603437136582,
|
13437 |
-
"learning_rate": 1.2840010372099054e-07,
|
13438 |
-
"loss": 2.4688,
|
13439 |
-
"step": 44380
|
13440 |
-
},
|
13441 |
-
{
|
13442 |
-
"epoch": 11.475772063574105,
|
13443 |
-
"learning_rate": 1.283676909114482e-07,
|
13444 |
-
"loss": 2.4458,
|
13445 |
-
"step": 44400
|
13446 |
-
},
|
13447 |
-
{
|
13448 |
-
"epoch": 11.480940690011629,
|
13449 |
-
"learning_rate": 1.2833527810190587e-07,
|
13450 |
-
"loss": 2.4871,
|
13451 |
-
"step": 44420
|
13452 |
-
},
|
13453 |
-
{
|
13454 |
-
"epoch": 11.486109316449154,
|
13455 |
-
"learning_rate": 1.2830286529236353e-07,
|
13456 |
-
"loss": 2.4476,
|
13457 |
-
"step": 44440
|
13458 |
-
},
|
13459 |
-
{
|
13460 |
-
"epoch": 11.491277942886677,
|
13461 |
-
"learning_rate": 1.282704524828212e-07,
|
13462 |
-
"loss": 2.4647,
|
13463 |
-
"step": 44460
|
13464 |
-
},
|
13465 |
-
{
|
13466 |
-
"epoch": 11.496446569324203,
|
13467 |
-
"learning_rate": 1.2823803967327888e-07,
|
13468 |
-
"loss": 2.4887,
|
13469 |
-
"step": 44480
|
13470 |
-
},
|
13471 |
-
{
|
13472 |
-
"epoch": 11.501615195761726,
|
13473 |
-
"learning_rate": 1.2820562686373654e-07,
|
13474 |
-
"loss": 2.4835,
|
13475 |
-
"step": 44500
|
13476 |
-
},
|
13477 |
-
{
|
13478 |
-
"epoch": 11.506783822199251,
|
13479 |
-
"learning_rate": 1.281732140541942e-07,
|
13480 |
-
"loss": 2.4435,
|
13481 |
-
"step": 44520
|
13482 |
-
},
|
13483 |
-
{
|
13484 |
-
"epoch": 11.511952448636775,
|
13485 |
-
"learning_rate": 1.281408012446519e-07,
|
13486 |
-
"loss": 2.4716,
|
13487 |
-
"step": 44540
|
13488 |
-
},
|
13489 |
-
{
|
13490 |
-
"epoch": 11.517121075074298,
|
13491 |
-
"learning_rate": 1.2810838843510955e-07,
|
13492 |
-
"loss": 2.457,
|
13493 |
-
"step": 44560
|
13494 |
-
},
|
13495 |
-
{
|
13496 |
-
"epoch": 11.522289701511824,
|
13497 |
-
"learning_rate": 1.2807597562556721e-07,
|
13498 |
-
"loss": 2.4465,
|
13499 |
-
"step": 44580
|
13500 |
-
},
|
13501 |
-
{
|
13502 |
-
"epoch": 11.527458327949347,
|
13503 |
-
"learning_rate": 1.2804356281602488e-07,
|
13504 |
-
"loss": 2.4804,
|
13505 |
-
"step": 44600
|
13506 |
-
},
|
13507 |
-
{
|
13508 |
-
"epoch": 11.532626954386872,
|
13509 |
-
"learning_rate": 1.2801115000648256e-07,
|
13510 |
-
"loss": 2.5168,
|
13511 |
-
"step": 44620
|
13512 |
-
},
|
13513 |
-
{
|
13514 |
-
"epoch": 11.537795580824396,
|
13515 |
-
"learning_rate": 1.2797873719694023e-07,
|
13516 |
-
"loss": 2.4759,
|
13517 |
-
"step": 44640
|
13518 |
-
},
|
13519 |
-
{
|
13520 |
-
"epoch": 11.542964207261921,
|
13521 |
-
"learning_rate": 1.279463243873979e-07,
|
13522 |
-
"loss": 2.4929,
|
13523 |
-
"step": 44660
|
13524 |
-
},
|
13525 |
-
{
|
13526 |
-
"epoch": 11.548132833699444,
|
13527 |
-
"learning_rate": 1.2791391157785555e-07,
|
13528 |
-
"loss": 2.4136,
|
13529 |
-
"step": 44680
|
13530 |
-
},
|
13531 |
-
{
|
13532 |
-
"epoch": 11.553301460136968,
|
13533 |
-
"learning_rate": 1.2788149876831324e-07,
|
13534 |
-
"loss": 2.4497,
|
13535 |
-
"step": 44700
|
13536 |
-
},
|
13537 |
-
{
|
13538 |
-
"epoch": 11.558470086574493,
|
13539 |
-
"learning_rate": 1.278490859587709e-07,
|
13540 |
-
"loss": 2.4912,
|
13541 |
-
"step": 44720
|
13542 |
-
},
|
13543 |
-
{
|
13544 |
-
"epoch": 11.563638713012017,
|
13545 |
-
"learning_rate": 1.2781667314922856e-07,
|
13546 |
-
"loss": 2.5089,
|
13547 |
-
"step": 44740
|
13548 |
-
},
|
13549 |
-
{
|
13550 |
-
"epoch": 11.568807339449542,
|
13551 |
-
"learning_rate": 1.2778426033968622e-07,
|
13552 |
-
"loss": 2.4516,
|
13553 |
-
"step": 44760
|
13554 |
-
},
|
13555 |
-
{
|
13556 |
-
"epoch": 11.573975965887065,
|
13557 |
-
"learning_rate": 1.277518475301439e-07,
|
13558 |
-
"loss": 2.5023,
|
13559 |
-
"step": 44780
|
13560 |
-
},
|
13561 |
-
{
|
13562 |
-
"epoch": 11.579144592324589,
|
13563 |
-
"learning_rate": 1.2771943472060157e-07,
|
13564 |
-
"loss": 2.4883,
|
13565 |
-
"step": 44800
|
13566 |
-
},
|
13567 |
-
{
|
13568 |
-
"epoch": 11.584313218762114,
|
13569 |
-
"learning_rate": 1.2768702191105924e-07,
|
13570 |
-
"loss": 2.4885,
|
13571 |
-
"step": 44820
|
13572 |
-
},
|
13573 |
-
{
|
13574 |
-
"epoch": 11.589481845199638,
|
13575 |
-
"learning_rate": 1.2765460910151692e-07,
|
13576 |
-
"loss": 2.5421,
|
13577 |
-
"step": 44840
|
13578 |
-
},
|
13579 |
-
{
|
13580 |
-
"epoch": 11.594650471637163,
|
13581 |
-
"learning_rate": 1.2762219629197459e-07,
|
13582 |
-
"loss": 2.4938,
|
13583 |
-
"step": 44860
|
13584 |
-
},
|
13585 |
-
{
|
13586 |
-
"epoch": 11.599819098074686,
|
13587 |
-
"learning_rate": 1.2758978348243225e-07,
|
13588 |
-
"loss": 2.4712,
|
13589 |
-
"step": 44880
|
13590 |
-
},
|
13591 |
-
{
|
13592 |
-
"epoch": 11.604987724512211,
|
13593 |
-
"learning_rate": 1.275573706728899e-07,
|
13594 |
-
"loss": 2.5227,
|
13595 |
-
"step": 44900
|
13596 |
-
},
|
13597 |
-
{
|
13598 |
-
"epoch": 11.610156350949735,
|
13599 |
-
"learning_rate": 1.275249578633476e-07,
|
13600 |
-
"loss": 2.4463,
|
13601 |
-
"step": 44920
|
13602 |
-
},
|
13603 |
-
{
|
13604 |
-
"epoch": 11.61532497738726,
|
13605 |
-
"learning_rate": 1.2749254505380526e-07,
|
13606 |
-
"loss": 2.4526,
|
13607 |
-
"step": 44940
|
13608 |
-
},
|
13609 |
-
{
|
13610 |
-
"epoch": 11.620493603824784,
|
13611 |
-
"learning_rate": 1.2746013224426292e-07,
|
13612 |
-
"loss": 2.4763,
|
13613 |
-
"step": 44960
|
13614 |
-
},
|
13615 |
-
{
|
13616 |
-
"epoch": 11.625662230262307,
|
13617 |
-
"learning_rate": 1.2742771943472058e-07,
|
13618 |
-
"loss": 2.5575,
|
13619 |
-
"step": 44980
|
13620 |
-
},
|
13621 |
-
{
|
13622 |
-
"epoch": 11.630830856699832,
|
13623 |
-
"learning_rate": 1.2739530662517827e-07,
|
13624 |
-
"loss": 2.5079,
|
13625 |
-
"step": 45000
|
13626 |
-
},
|
13627 |
-
{
|
13628 |
-
"epoch": 11.635999483137356,
|
13629 |
-
"learning_rate": 1.2736289381563593e-07,
|
13630 |
-
"loss": 2.4868,
|
13631 |
-
"step": 45020
|
13632 |
-
},
|
13633 |
-
{
|
13634 |
-
"epoch": 11.641168109574881,
|
13635 |
-
"learning_rate": 1.273304810060936e-07,
|
13636 |
-
"loss": 2.4547,
|
13637 |
-
"step": 45040
|
13638 |
-
},
|
13639 |
-
{
|
13640 |
-
"epoch": 11.646336736012405,
|
13641 |
-
"learning_rate": 1.2729806819655126e-07,
|
13642 |
-
"loss": 2.4594,
|
13643 |
-
"step": 45060
|
13644 |
-
},
|
13645 |
-
{
|
13646 |
-
"epoch": 11.651505362449928,
|
13647 |
-
"learning_rate": 1.2726565538700895e-07,
|
13648 |
-
"loss": 2.495,
|
13649 |
-
"step": 45080
|
13650 |
-
},
|
13651 |
-
{
|
13652 |
-
"epoch": 11.656673988887453,
|
13653 |
-
"learning_rate": 1.272332425774666e-07,
|
13654 |
-
"loss": 2.5174,
|
13655 |
-
"step": 45100
|
13656 |
-
},
|
13657 |
-
{
|
13658 |
-
"epoch": 11.661842615324977,
|
13659 |
-
"learning_rate": 1.2720082976792427e-07,
|
13660 |
-
"loss": 2.4223,
|
13661 |
-
"step": 45120
|
13662 |
-
},
|
13663 |
-
{
|
13664 |
-
"epoch": 11.667011241762502,
|
13665 |
-
"learning_rate": 1.2716841695838196e-07,
|
13666 |
-
"loss": 2.4523,
|
13667 |
-
"step": 45140
|
13668 |
-
},
|
13669 |
-
{
|
13670 |
-
"epoch": 11.672179868200026,
|
13671 |
-
"learning_rate": 1.2713600414883962e-07,
|
13672 |
-
"loss": 2.4271,
|
13673 |
-
"step": 45160
|
13674 |
-
},
|
13675 |
-
{
|
13676 |
-
"epoch": 11.67734849463755,
|
13677 |
-
"learning_rate": 1.2710359133929728e-07,
|
13678 |
-
"loss": 2.5235,
|
13679 |
-
"step": 45180
|
13680 |
-
},
|
13681 |
-
{
|
13682 |
-
"epoch": 11.682517121075074,
|
13683 |
-
"learning_rate": 1.2707117852975494e-07,
|
13684 |
-
"loss": 2.4666,
|
13685 |
-
"step": 45200
|
13686 |
-
},
|
13687 |
-
{
|
13688 |
-
"epoch": 11.687685747512598,
|
13689 |
-
"learning_rate": 1.2703876572021263e-07,
|
13690 |
-
"loss": 2.4413,
|
13691 |
-
"step": 45220
|
13692 |
-
},
|
13693 |
-
{
|
13694 |
-
"epoch": 11.692854373950123,
|
13695 |
-
"learning_rate": 1.270063529106703e-07,
|
13696 |
-
"loss": 2.422,
|
13697 |
-
"step": 45240
|
13698 |
-
},
|
13699 |
-
{
|
13700 |
-
"epoch": 11.698023000387646,
|
13701 |
-
"learning_rate": 1.2697394010112796e-07,
|
13702 |
-
"loss": 2.5057,
|
13703 |
-
"step": 45260
|
13704 |
-
},
|
13705 |
-
{
|
13706 |
-
"epoch": 11.703191626825172,
|
13707 |
-
"learning_rate": 1.2694152729158562e-07,
|
13708 |
-
"loss": 2.5153,
|
13709 |
-
"step": 45280
|
13710 |
-
},
|
13711 |
-
{
|
13712 |
-
"epoch": 11.708360253262695,
|
13713 |
-
"learning_rate": 1.269091144820433e-07,
|
13714 |
-
"loss": 2.4894,
|
13715 |
-
"step": 45300
|
13716 |
-
},
|
13717 |
-
{
|
13718 |
-
"epoch": 11.71352887970022,
|
13719 |
-
"learning_rate": 1.2687670167250097e-07,
|
13720 |
-
"loss": 2.5211,
|
13721 |
-
"step": 45320
|
13722 |
-
},
|
13723 |
-
{
|
13724 |
-
"epoch": 11.718697506137744,
|
13725 |
-
"learning_rate": 1.2684428886295863e-07,
|
13726 |
-
"loss": 2.4449,
|
13727 |
-
"step": 45340
|
13728 |
-
},
|
13729 |
-
{
|
13730 |
-
"epoch": 11.723866132575267,
|
13731 |
-
"learning_rate": 1.268118760534163e-07,
|
13732 |
-
"loss": 2.4859,
|
13733 |
-
"step": 45360
|
13734 |
-
},
|
13735 |
-
{
|
13736 |
-
"epoch": 11.729034759012793,
|
13737 |
-
"learning_rate": 1.2677946324387398e-07,
|
13738 |
-
"loss": 2.4578,
|
13739 |
-
"step": 45380
|
13740 |
-
},
|
13741 |
-
{
|
13742 |
-
"epoch": 11.734203385450316,
|
13743 |
-
"learning_rate": 1.2674705043433164e-07,
|
13744 |
-
"loss": 2.4715,
|
13745 |
-
"step": 45400
|
13746 |
-
},
|
13747 |
-
{
|
13748 |
-
"epoch": 11.739372011887841,
|
13749 |
-
"learning_rate": 1.267146376247893e-07,
|
13750 |
-
"loss": 2.4258,
|
13751 |
-
"step": 45420
|
13752 |
-
},
|
13753 |
-
{
|
13754 |
-
"epoch": 11.744540638325365,
|
13755 |
-
"learning_rate": 1.26682224815247e-07,
|
13756 |
-
"loss": 2.4635,
|
13757 |
-
"step": 45440
|
13758 |
-
},
|
13759 |
-
{
|
13760 |
-
"epoch": 11.74970926476289,
|
13761 |
-
"learning_rate": 1.2664981200570465e-07,
|
13762 |
-
"loss": 2.4803,
|
13763 |
-
"step": 45460
|
13764 |
-
},
|
13765 |
-
{
|
13766 |
-
"epoch": 11.754877891200413,
|
13767 |
-
"learning_rate": 1.2661739919616232e-07,
|
13768 |
-
"loss": 2.4333,
|
13769 |
-
"step": 45480
|
13770 |
-
},
|
13771 |
-
{
|
13772 |
-
"epoch": 11.760046517637937,
|
13773 |
-
"learning_rate": 1.2658498638661998e-07,
|
13774 |
-
"loss": 2.5486,
|
13775 |
-
"step": 45500
|
13776 |
-
},
|
13777 |
-
{
|
13778 |
-
"epoch": 11.765215144075462,
|
13779 |
-
"learning_rate": 1.2655257357707767e-07,
|
13780 |
-
"loss": 2.4803,
|
13781 |
-
"step": 45520
|
13782 |
-
},
|
13783 |
-
{
|
13784 |
-
"epoch": 11.770383770512986,
|
13785 |
-
"learning_rate": 1.2652016076753533e-07,
|
13786 |
-
"loss": 2.4958,
|
13787 |
-
"step": 45540
|
13788 |
-
},
|
13789 |
-
{
|
13790 |
-
"epoch": 11.775552396950511,
|
13791 |
-
"learning_rate": 1.26487747957993e-07,
|
13792 |
-
"loss": 2.4918,
|
13793 |
-
"step": 45560
|
13794 |
-
},
|
13795 |
-
{
|
13796 |
-
"epoch": 11.780721023388034,
|
13797 |
-
"learning_rate": 1.2645533514845065e-07,
|
13798 |
-
"loss": 2.513,
|
13799 |
-
"step": 45580
|
13800 |
-
},
|
13801 |
-
{
|
13802 |
-
"epoch": 11.78588964982556,
|
13803 |
-
"learning_rate": 1.2642292233890834e-07,
|
13804 |
-
"loss": 2.4897,
|
13805 |
-
"step": 45600
|
13806 |
-
},
|
13807 |
-
{
|
13808 |
-
"epoch": 11.791058276263083,
|
13809 |
-
"learning_rate": 1.26390509529366e-07,
|
13810 |
-
"loss": 2.4468,
|
13811 |
-
"step": 45620
|
13812 |
-
},
|
13813 |
-
{
|
13814 |
-
"epoch": 11.796226902700607,
|
13815 |
-
"learning_rate": 1.2635809671982366e-07,
|
13816 |
-
"loss": 2.4928,
|
13817 |
-
"step": 45640
|
13818 |
-
},
|
13819 |
-
{
|
13820 |
-
"epoch": 11.801395529138132,
|
13821 |
-
"learning_rate": 1.2632568391028132e-07,
|
13822 |
-
"loss": 2.4793,
|
13823 |
-
"step": 45660
|
13824 |
-
},
|
13825 |
-
{
|
13826 |
-
"epoch": 11.806564155575655,
|
13827 |
-
"learning_rate": 1.26293271100739e-07,
|
13828 |
-
"loss": 2.4788,
|
13829 |
-
"step": 45680
|
13830 |
-
},
|
13831 |
-
{
|
13832 |
-
"epoch": 11.81173278201318,
|
13833 |
-
"learning_rate": 1.2626085829119668e-07,
|
13834 |
-
"loss": 2.4913,
|
13835 |
-
"step": 45700
|
13836 |
-
},
|
13837 |
-
{
|
13838 |
-
"epoch": 11.816901408450704,
|
13839 |
-
"learning_rate": 1.2622844548165434e-07,
|
13840 |
-
"loss": 2.4966,
|
13841 |
-
"step": 45720
|
13842 |
-
},
|
13843 |
-
{
|
13844 |
-
"epoch": 11.82207003488823,
|
13845 |
-
"learning_rate": 1.2619603267211203e-07,
|
13846 |
-
"loss": 2.4954,
|
13847 |
-
"step": 45740
|
13848 |
-
},
|
13849 |
-
{
|
13850 |
-
"epoch": 11.827238661325753,
|
13851 |
-
"learning_rate": 1.261636198625697e-07,
|
13852 |
-
"loss": 2.4608,
|
13853 |
-
"step": 45760
|
13854 |
-
},
|
13855 |
-
{
|
13856 |
-
"epoch": 11.832407287763276,
|
13857 |
-
"learning_rate": 1.2613120705302735e-07,
|
13858 |
-
"loss": 2.4901,
|
13859 |
-
"step": 45780
|
13860 |
-
},
|
13861 |
-
{
|
13862 |
-
"epoch": 11.837575914200801,
|
13863 |
-
"learning_rate": 1.26098794243485e-07,
|
13864 |
-
"loss": 2.5462,
|
13865 |
-
"step": 45800
|
13866 |
-
},
|
13867 |
-
{
|
13868 |
-
"epoch": 11.842744540638325,
|
13869 |
-
"learning_rate": 1.260663814339427e-07,
|
13870 |
-
"loss": 2.5217,
|
13871 |
-
"step": 45820
|
13872 |
-
},
|
13873 |
-
{
|
13874 |
-
"epoch": 11.84791316707585,
|
13875 |
-
"learning_rate": 1.2603396862440036e-07,
|
13876 |
-
"loss": 2.4889,
|
13877 |
-
"step": 45840
|
13878 |
-
},
|
13879 |
-
{
|
13880 |
-
"epoch": 11.853081793513374,
|
13881 |
-
"learning_rate": 1.2600155581485802e-07,
|
13882 |
-
"loss": 2.5021,
|
13883 |
-
"step": 45860
|
13884 |
-
},
|
13885 |
-
{
|
13886 |
-
"epoch": 11.858250419950899,
|
13887 |
-
"learning_rate": 1.2596914300531568e-07,
|
13888 |
-
"loss": 2.5027,
|
13889 |
-
"step": 45880
|
13890 |
-
},
|
13891 |
-
{
|
13892 |
-
"epoch": 11.863419046388422,
|
13893 |
-
"learning_rate": 1.2593673019577337e-07,
|
13894 |
-
"loss": 2.4681,
|
13895 |
-
"step": 45900
|
13896 |
-
},
|
13897 |
-
{
|
13898 |
-
"epoch": 11.868587672825946,
|
13899 |
-
"learning_rate": 1.2590431738623104e-07,
|
13900 |
-
"loss": 2.409,
|
13901 |
-
"step": 45920
|
13902 |
-
},
|
13903 |
-
{
|
13904 |
-
"epoch": 11.873756299263471,
|
13905 |
-
"learning_rate": 1.258719045766887e-07,
|
13906 |
-
"loss": 2.4426,
|
13907 |
-
"step": 45940
|
13908 |
-
},
|
13909 |
-
{
|
13910 |
-
"epoch": 11.878924925700995,
|
13911 |
-
"learning_rate": 1.2583949176714636e-07,
|
13912 |
-
"loss": 2.4285,
|
13913 |
-
"step": 45960
|
13914 |
-
},
|
13915 |
-
{
|
13916 |
-
"epoch": 11.88409355213852,
|
13917 |
-
"learning_rate": 1.2580707895760405e-07,
|
13918 |
-
"loss": 2.4452,
|
13919 |
-
"step": 45980
|
13920 |
-
},
|
13921 |
-
{
|
13922 |
-
"epoch": 11.889262178576043,
|
13923 |
-
"learning_rate": 1.257746661480617e-07,
|
13924 |
-
"loss": 2.4816,
|
13925 |
-
"step": 46000
|
13926 |
-
},
|
13927 |
-
{
|
13928 |
-
"epoch": 11.894430805013567,
|
13929 |
-
"learning_rate": 1.2574225333851937e-07,
|
13930 |
-
"loss": 2.4585,
|
13931 |
-
"step": 46020
|
13932 |
-
},
|
13933 |
-
{
|
13934 |
-
"epoch": 11.899599431451092,
|
13935 |
-
"learning_rate": 1.2570984052897706e-07,
|
13936 |
-
"loss": 2.4171,
|
13937 |
-
"step": 46040
|
13938 |
-
},
|
13939 |
-
{
|
13940 |
-
"epoch": 11.904768057888615,
|
13941 |
-
"learning_rate": 1.2567742771943472e-07,
|
13942 |
-
"loss": 2.4597,
|
13943 |
-
"step": 46060
|
13944 |
-
},
|
13945 |
-
{
|
13946 |
-
"epoch": 11.90993668432614,
|
13947 |
-
"learning_rate": 1.2564501490989238e-07,
|
13948 |
-
"loss": 2.4195,
|
13949 |
-
"step": 46080
|
13950 |
-
},
|
13951 |
-
{
|
13952 |
-
"epoch": 11.915105310763664,
|
13953 |
-
"learning_rate": 1.2561260210035004e-07,
|
13954 |
-
"loss": 2.5034,
|
13955 |
-
"step": 46100
|
13956 |
-
},
|
13957 |
-
{
|
13958 |
-
"epoch": 11.92027393720119,
|
13959 |
-
"learning_rate": 1.2558018929080773e-07,
|
13960 |
-
"loss": 2.4407,
|
13961 |
-
"step": 46120
|
13962 |
-
},
|
13963 |
-
{
|
13964 |
-
"epoch": 11.925442563638713,
|
13965 |
-
"learning_rate": 1.255477764812654e-07,
|
13966 |
-
"loss": 2.4813,
|
13967 |
-
"step": 46140
|
13968 |
-
},
|
13969 |
-
{
|
13970 |
-
"epoch": 11.930611190076238,
|
13971 |
-
"learning_rate": 1.2551536367172306e-07,
|
13972 |
-
"loss": 2.4599,
|
13973 |
-
"step": 46160
|
13974 |
-
},
|
13975 |
-
{
|
13976 |
-
"epoch": 11.935779816513762,
|
13977 |
-
"learning_rate": 1.2548295086218072e-07,
|
13978 |
-
"loss": 2.4641,
|
13979 |
-
"step": 46180
|
13980 |
-
},
|
13981 |
-
{
|
13982 |
-
"epoch": 11.940948442951285,
|
13983 |
-
"learning_rate": 1.254505380526384e-07,
|
13984 |
-
"loss": 2.5262,
|
13985 |
-
"step": 46200
|
13986 |
-
},
|
13987 |
-
{
|
13988 |
-
"epoch": 11.94611706938881,
|
13989 |
-
"learning_rate": 1.2541812524309607e-07,
|
13990 |
-
"loss": 2.4518,
|
13991 |
-
"step": 46220
|
13992 |
-
},
|
13993 |
-
{
|
13994 |
-
"epoch": 11.951285695826334,
|
13995 |
-
"learning_rate": 1.2538571243355373e-07,
|
13996 |
-
"loss": 2.5018,
|
13997 |
-
"step": 46240
|
13998 |
-
},
|
13999 |
-
{
|
14000 |
-
"epoch": 11.956454322263859,
|
14001 |
-
"learning_rate": 1.253532996240114e-07,
|
14002 |
-
"loss": 2.4634,
|
14003 |
-
"step": 46260
|
14004 |
-
},
|
14005 |
-
{
|
14006 |
-
"epoch": 11.961622948701383,
|
14007 |
-
"learning_rate": 1.2532088681446908e-07,
|
14008 |
-
"loss": 2.4889,
|
14009 |
-
"step": 46280
|
14010 |
-
},
|
14011 |
-
{
|
14012 |
-
"epoch": 11.966791575138906,
|
14013 |
-
"learning_rate": 1.2528847400492674e-07,
|
14014 |
-
"loss": 2.4577,
|
14015 |
-
"step": 46300
|
14016 |
-
},
|
14017 |
-
{
|
14018 |
-
"epoch": 11.971960201576431,
|
14019 |
-
"learning_rate": 1.252560611953844e-07,
|
14020 |
-
"loss": 2.4812,
|
14021 |
-
"step": 46320
|
14022 |
-
},
|
14023 |
-
{
|
14024 |
-
"epoch": 11.977128828013955,
|
14025 |
-
"learning_rate": 1.252236483858421e-07,
|
14026 |
-
"loss": 2.4448,
|
14027 |
-
"step": 46340
|
14028 |
-
},
|
14029 |
-
{
|
14030 |
-
"epoch": 11.98229745445148,
|
14031 |
-
"learning_rate": 1.2519123557629975e-07,
|
14032 |
-
"loss": 2.4594,
|
14033 |
-
"step": 46360
|
14034 |
-
},
|
14035 |
-
{
|
14036 |
-
"epoch": 11.987466080889003,
|
14037 |
-
"learning_rate": 1.2515882276675742e-07,
|
14038 |
-
"loss": 2.4473,
|
14039 |
-
"step": 46380
|
14040 |
-
},
|
14041 |
-
{
|
14042 |
-
"epoch": 11.992634707326529,
|
14043 |
-
"learning_rate": 1.2512640995721508e-07,
|
14044 |
-
"loss": 2.4831,
|
14045 |
-
"step": 46400
|
14046 |
-
},
|
14047 |
-
{
|
14048 |
-
"epoch": 11.997803333764052,
|
14049 |
-
"learning_rate": 1.2509399714767277e-07,
|
14050 |
-
"loss": 2.4706,
|
14051 |
-
"step": 46420
|
14052 |
-
},
|
14053 |
-
{
|
14054 |
-
"epoch": 11.999870784339063,
|
14055 |
-
"eval_bleu": 12.51,
|
14056 |
-
"eval_gen_len": 37.4146,
|
14057 |
-
"eval_loss": 2.4478776454925537,
|
14058 |
-
"eval_runtime": 807.7318,
|
14059 |
-
"eval_samples_per_second": 2.132,
|
14060 |
-
"eval_steps_per_second": 1.066,
|
14061 |
-
"step": 46428
|
14062 |
}
|
14063 |
],
|
14064 |
"logging_steps": 20,
|
@@ -14078,7 +12910,7 @@
|
|
14078 |
"attributes": {}
|
14079 |
}
|
14080 |
},
|
14081 |
-
"total_flos":
|
14082 |
"train_batch_size": 2,
|
14083 |
"trial_name": null,
|
14084 |
"trial_params": null
|
|
|
1 |
{
|
2 |
+
"best_metric": 12.1716,
|
3 |
+
"best_model_checkpoint": "/kaggle/tmp/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted/checkpoint-42562",
|
4 |
+
"epoch": 11.0,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 42562,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
12891 |
"eval_samples_per_second": 2.165,
|
12892 |
"eval_steps_per_second": 1.083,
|
12893 |
"step": 42562
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12894 |
}
|
12895 |
],
|
12896 |
"logging_steps": 20,
|
|
|
12910 |
"attributes": {}
|
12911 |
}
|
12912 |
},
|
12913 |
+
"total_flos": 2.793935715793797e+17,
|
12914 |
"train_batch_size": 2,
|
12915 |
"trial_name": null,
|
12916 |
"trial_params": null
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1575259780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5f63231d876ab3ad240434253a3c1d9b5e552e066e216bb8da212e99bc37119e
|
3 |
size 1575259780
|
run-2024-10-28T15:22:08+00:00.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
val_outputs/val_generated_predictions_23953367.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|