sagarpatelcompstak's picture
Upload folder using huggingface_hub
2c85d63 verified
{
"best_metric": 0.3821594715118408,
"best_model_checkpoint": "classify-google-augment-3/checkpoint-4950",
"epoch": 3.0,
"eval_steps": 500,
"global_step": 4950,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.015151515151515152,
"grad_norm": 12.085535049438477,
"learning_rate": 2.3232323232323234e-06,
"loss": 2.4513,
"step": 25
},
{
"epoch": 0.030303030303030304,
"grad_norm": 10.15959358215332,
"learning_rate": 4.848484848484849e-06,
"loss": 2.234,
"step": 50
},
{
"epoch": 0.045454545454545456,
"grad_norm": 12.21349048614502,
"learning_rate": 7.3737373737373745e-06,
"loss": 1.9342,
"step": 75
},
{
"epoch": 0.06060606060606061,
"grad_norm": 13.986493110656738,
"learning_rate": 9.898989898989899e-06,
"loss": 1.7956,
"step": 100
},
{
"epoch": 0.07575757575757576,
"grad_norm": 13.179245948791504,
"learning_rate": 1.2424242424242424e-05,
"loss": 1.7165,
"step": 125
},
{
"epoch": 0.09090909090909091,
"grad_norm": 15.372575759887695,
"learning_rate": 1.494949494949495e-05,
"loss": 1.519,
"step": 150
},
{
"epoch": 0.10606060606060606,
"grad_norm": 9.555817604064941,
"learning_rate": 1.7474747474747475e-05,
"loss": 1.4716,
"step": 175
},
{
"epoch": 0.12121212121212122,
"grad_norm": 14.15468692779541,
"learning_rate": 2e-05,
"loss": 1.5611,
"step": 200
},
{
"epoch": 0.13636363636363635,
"grad_norm": 9.72482681274414,
"learning_rate": 2.2525252525252528e-05,
"loss": 1.3692,
"step": 225
},
{
"epoch": 0.15151515151515152,
"grad_norm": 11.219098091125488,
"learning_rate": 2.505050505050505e-05,
"loss": 1.3761,
"step": 250
},
{
"epoch": 0.16666666666666666,
"grad_norm": 14.293442726135254,
"learning_rate": 2.7575757575757578e-05,
"loss": 1.5677,
"step": 275
},
{
"epoch": 0.18181818181818182,
"grad_norm": 12.68210506439209,
"learning_rate": 3.01010101010101e-05,
"loss": 1.4212,
"step": 300
},
{
"epoch": 0.19696969696969696,
"grad_norm": 10.856423377990723,
"learning_rate": 3.2626262626262624e-05,
"loss": 1.3204,
"step": 325
},
{
"epoch": 0.21212121212121213,
"grad_norm": 11.644044876098633,
"learning_rate": 3.515151515151515e-05,
"loss": 1.3228,
"step": 350
},
{
"epoch": 0.22727272727272727,
"grad_norm": 7.3117356300354,
"learning_rate": 3.767676767676768e-05,
"loss": 1.4005,
"step": 375
},
{
"epoch": 0.24242424242424243,
"grad_norm": 12.309542655944824,
"learning_rate": 4.0202020202020204e-05,
"loss": 1.2739,
"step": 400
},
{
"epoch": 0.25757575757575757,
"grad_norm": 11.981978416442871,
"learning_rate": 4.2727272727272724e-05,
"loss": 1.3082,
"step": 425
},
{
"epoch": 0.2727272727272727,
"grad_norm": 8.469183921813965,
"learning_rate": 4.525252525252526e-05,
"loss": 1.1932,
"step": 450
},
{
"epoch": 0.2878787878787879,
"grad_norm": 13.018028259277344,
"learning_rate": 4.7777777777777784e-05,
"loss": 1.2612,
"step": 475
},
{
"epoch": 0.30303030303030304,
"grad_norm": 11.944835662841797,
"learning_rate": 4.9966329966329964e-05,
"loss": 1.3266,
"step": 500
},
{
"epoch": 0.3181818181818182,
"grad_norm": 11.489705085754395,
"learning_rate": 4.968574635241302e-05,
"loss": 1.3537,
"step": 525
},
{
"epoch": 0.3333333333333333,
"grad_norm": 14.265945434570312,
"learning_rate": 4.940516273849607e-05,
"loss": 1.4162,
"step": 550
},
{
"epoch": 0.3484848484848485,
"grad_norm": 10.913326263427734,
"learning_rate": 4.912457912457913e-05,
"loss": 1.0754,
"step": 575
},
{
"epoch": 0.36363636363636365,
"grad_norm": 7.5248236656188965,
"learning_rate": 4.884399551066218e-05,
"loss": 1.3724,
"step": 600
},
{
"epoch": 0.3787878787878788,
"grad_norm": 9.98138427734375,
"learning_rate": 4.856341189674523e-05,
"loss": 1.184,
"step": 625
},
{
"epoch": 0.3939393939393939,
"grad_norm": 9.07705307006836,
"learning_rate": 4.828282828282829e-05,
"loss": 1.0418,
"step": 650
},
{
"epoch": 0.4090909090909091,
"grad_norm": 8.512435913085938,
"learning_rate": 4.800224466891134e-05,
"loss": 1.1425,
"step": 675
},
{
"epoch": 0.42424242424242425,
"grad_norm": 13.559457778930664,
"learning_rate": 4.7721661054994394e-05,
"loss": 1.4285,
"step": 700
},
{
"epoch": 0.4393939393939394,
"grad_norm": 7.820692539215088,
"learning_rate": 4.7441077441077445e-05,
"loss": 1.1221,
"step": 725
},
{
"epoch": 0.45454545454545453,
"grad_norm": 6.074573040008545,
"learning_rate": 4.7160493827160495e-05,
"loss": 0.9862,
"step": 750
},
{
"epoch": 0.4696969696969697,
"grad_norm": 6.816392421722412,
"learning_rate": 4.687991021324355e-05,
"loss": 1.1894,
"step": 775
},
{
"epoch": 0.48484848484848486,
"grad_norm": 9.136533737182617,
"learning_rate": 4.65993265993266e-05,
"loss": 1.0674,
"step": 800
},
{
"epoch": 0.5,
"grad_norm": 8.841058731079102,
"learning_rate": 4.631874298540965e-05,
"loss": 1.0798,
"step": 825
},
{
"epoch": 0.5151515151515151,
"grad_norm": 10.076774597167969,
"learning_rate": 4.60381593714927e-05,
"loss": 1.0903,
"step": 850
},
{
"epoch": 0.5303030303030303,
"grad_norm": 13.101428985595703,
"learning_rate": 4.575757575757576e-05,
"loss": 1.0672,
"step": 875
},
{
"epoch": 0.5454545454545454,
"grad_norm": 10.879769325256348,
"learning_rate": 4.547699214365882e-05,
"loss": 1.005,
"step": 900
},
{
"epoch": 0.5606060606060606,
"grad_norm": 5.972232341766357,
"learning_rate": 4.519640852974186e-05,
"loss": 1.026,
"step": 925
},
{
"epoch": 0.5757575757575758,
"grad_norm": 12.992452621459961,
"learning_rate": 4.491582491582492e-05,
"loss": 1.0436,
"step": 950
},
{
"epoch": 0.5909090909090909,
"grad_norm": 11.431863784790039,
"learning_rate": 4.463524130190797e-05,
"loss": 1.006,
"step": 975
},
{
"epoch": 0.6060606060606061,
"grad_norm": 13.05385971069336,
"learning_rate": 4.4354657687991025e-05,
"loss": 1.1255,
"step": 1000
},
{
"epoch": 0.6212121212121212,
"grad_norm": 9.93174934387207,
"learning_rate": 4.4074074074074076e-05,
"loss": 1.0457,
"step": 1025
},
{
"epoch": 0.6363636363636364,
"grad_norm": 14.331412315368652,
"learning_rate": 4.3793490460157126e-05,
"loss": 0.997,
"step": 1050
},
{
"epoch": 0.6515151515151515,
"grad_norm": 11.081893920898438,
"learning_rate": 4.351290684624018e-05,
"loss": 0.999,
"step": 1075
},
{
"epoch": 0.6666666666666666,
"grad_norm": 8.453628540039062,
"learning_rate": 4.3232323232323234e-05,
"loss": 1.0695,
"step": 1100
},
{
"epoch": 0.6818181818181818,
"grad_norm": 8.907005310058594,
"learning_rate": 4.295173961840629e-05,
"loss": 0.967,
"step": 1125
},
{
"epoch": 0.696969696969697,
"grad_norm": 10.75546646118164,
"learning_rate": 4.267115600448934e-05,
"loss": 1.0517,
"step": 1150
},
{
"epoch": 0.7121212121212122,
"grad_norm": 9.127320289611816,
"learning_rate": 4.239057239057239e-05,
"loss": 1.0049,
"step": 1175
},
{
"epoch": 0.7272727272727273,
"grad_norm": 10.124237060546875,
"learning_rate": 4.210998877665545e-05,
"loss": 0.969,
"step": 1200
},
{
"epoch": 0.7424242424242424,
"grad_norm": 5.617876052856445,
"learning_rate": 4.18294051627385e-05,
"loss": 1.0031,
"step": 1225
},
{
"epoch": 0.7575757575757576,
"grad_norm": 11.184277534484863,
"learning_rate": 4.154882154882155e-05,
"loss": 1.0209,
"step": 1250
},
{
"epoch": 0.7727272727272727,
"grad_norm": 6.545035362243652,
"learning_rate": 4.12682379349046e-05,
"loss": 0.8416,
"step": 1275
},
{
"epoch": 0.7878787878787878,
"grad_norm": 4.944152355194092,
"learning_rate": 4.0987654320987657e-05,
"loss": 1.0029,
"step": 1300
},
{
"epoch": 0.803030303030303,
"grad_norm": 8.916969299316406,
"learning_rate": 4.070707070707071e-05,
"loss": 0.8808,
"step": 1325
},
{
"epoch": 0.8181818181818182,
"grad_norm": 4.9258713722229,
"learning_rate": 4.0426487093153764e-05,
"loss": 0.9536,
"step": 1350
},
{
"epoch": 0.8333333333333334,
"grad_norm": 11.556329727172852,
"learning_rate": 4.0145903479236814e-05,
"loss": 0.937,
"step": 1375
},
{
"epoch": 0.8484848484848485,
"grad_norm": 10.06991958618164,
"learning_rate": 3.9865319865319865e-05,
"loss": 0.8424,
"step": 1400
},
{
"epoch": 0.8636363636363636,
"grad_norm": 8.390497207641602,
"learning_rate": 3.958473625140292e-05,
"loss": 0.9453,
"step": 1425
},
{
"epoch": 0.8787878787878788,
"grad_norm": 8.02001667022705,
"learning_rate": 3.930415263748597e-05,
"loss": 0.8522,
"step": 1450
},
{
"epoch": 0.8939393939393939,
"grad_norm": 8.541698455810547,
"learning_rate": 3.902356902356902e-05,
"loss": 0.9519,
"step": 1475
},
{
"epoch": 0.9090909090909091,
"grad_norm": 7.404212951660156,
"learning_rate": 3.874298540965208e-05,
"loss": 0.9692,
"step": 1500
},
{
"epoch": 0.9242424242424242,
"grad_norm": 10.684733390808105,
"learning_rate": 3.846240179573513e-05,
"loss": 1.0417,
"step": 1525
},
{
"epoch": 0.9393939393939394,
"grad_norm": 6.182531833648682,
"learning_rate": 3.818181818181819e-05,
"loss": 0.8632,
"step": 1550
},
{
"epoch": 0.9545454545454546,
"grad_norm": 8.63337516784668,
"learning_rate": 3.790123456790123e-05,
"loss": 0.9318,
"step": 1575
},
{
"epoch": 0.9696969696969697,
"grad_norm": 11.64743709564209,
"learning_rate": 3.762065095398429e-05,
"loss": 0.8535,
"step": 1600
},
{
"epoch": 0.9848484848484849,
"grad_norm": 11.15438461303711,
"learning_rate": 3.7340067340067345e-05,
"loss": 0.9558,
"step": 1625
},
{
"epoch": 1.0,
"grad_norm": 7.817443370819092,
"learning_rate": 3.7059483726150395e-05,
"loss": 0.9244,
"step": 1650
},
{
"epoch": 1.0,
"eval_accuracy": 0.7645454545454545,
"eval_f1_macro": 0.6989492792384068,
"eval_f1_micro": 0.7645454545454545,
"eval_f1_weighted": 0.7496986753676582,
"eval_loss": 0.752944827079773,
"eval_precision_macro": 0.8012284685903541,
"eval_precision_micro": 0.7645454545454545,
"eval_precision_weighted": 0.817408067065806,
"eval_recall_macro": 0.7035,
"eval_recall_micro": 0.7645454545454545,
"eval_recall_weighted": 0.7645454545454545,
"eval_runtime": 18.9801,
"eval_samples_per_second": 173.866,
"eval_steps_per_second": 10.906,
"step": 1650
},
{
"epoch": 1.0151515151515151,
"grad_norm": 10.560444831848145,
"learning_rate": 3.677890011223345e-05,
"loss": 0.7833,
"step": 1675
},
{
"epoch": 1.0303030303030303,
"grad_norm": 4.251949310302734,
"learning_rate": 3.6498316498316496e-05,
"loss": 0.7379,
"step": 1700
},
{
"epoch": 1.0454545454545454,
"grad_norm": 10.605439186096191,
"learning_rate": 3.621773288439955e-05,
"loss": 0.7779,
"step": 1725
},
{
"epoch": 1.0606060606060606,
"grad_norm": 7.418851375579834,
"learning_rate": 3.59371492704826e-05,
"loss": 0.7347,
"step": 1750
},
{
"epoch": 1.0757575757575757,
"grad_norm": 10.209125518798828,
"learning_rate": 3.565656565656566e-05,
"loss": 0.8636,
"step": 1775
},
{
"epoch": 1.0909090909090908,
"grad_norm": 5.001626491546631,
"learning_rate": 3.537598204264871e-05,
"loss": 0.8716,
"step": 1800
},
{
"epoch": 1.106060606060606,
"grad_norm": 5.035640239715576,
"learning_rate": 3.509539842873176e-05,
"loss": 0.7529,
"step": 1825
},
{
"epoch": 1.121212121212121,
"grad_norm": 7.653929710388184,
"learning_rate": 3.481481481481482e-05,
"loss": 0.8541,
"step": 1850
},
{
"epoch": 1.1363636363636362,
"grad_norm": 11.623481750488281,
"learning_rate": 3.453423120089787e-05,
"loss": 0.8159,
"step": 1875
},
{
"epoch": 1.1515151515151516,
"grad_norm": 9.897392272949219,
"learning_rate": 3.425364758698092e-05,
"loss": 0.5642,
"step": 1900
},
{
"epoch": 1.1666666666666667,
"grad_norm": 8.24569320678711,
"learning_rate": 3.3973063973063976e-05,
"loss": 1.0238,
"step": 1925
},
{
"epoch": 1.1818181818181819,
"grad_norm": 6.044914245605469,
"learning_rate": 3.3692480359147026e-05,
"loss": 0.8917,
"step": 1950
},
{
"epoch": 1.196969696969697,
"grad_norm": 5.3622002601623535,
"learning_rate": 3.3411896745230084e-05,
"loss": 0.7951,
"step": 1975
},
{
"epoch": 1.2121212121212122,
"grad_norm": 8.479634284973145,
"learning_rate": 3.3131313131313134e-05,
"loss": 0.9219,
"step": 2000
},
{
"epoch": 1.2272727272727273,
"grad_norm": 10.730965614318848,
"learning_rate": 3.2850729517396184e-05,
"loss": 0.6391,
"step": 2025
},
{
"epoch": 1.2424242424242424,
"grad_norm": 2.901815891265869,
"learning_rate": 3.2570145903479235e-05,
"loss": 0.7946,
"step": 2050
},
{
"epoch": 1.2575757575757576,
"grad_norm": 7.174593448638916,
"learning_rate": 3.228956228956229e-05,
"loss": 0.7458,
"step": 2075
},
{
"epoch": 1.2727272727272727,
"grad_norm": 5.743284702301025,
"learning_rate": 3.200897867564535e-05,
"loss": 0.8514,
"step": 2100
},
{
"epoch": 1.2878787878787878,
"grad_norm": 5.6034464836120605,
"learning_rate": 3.172839506172839e-05,
"loss": 0.5711,
"step": 2125
},
{
"epoch": 1.303030303030303,
"grad_norm": 6.2251996994018555,
"learning_rate": 3.144781144781145e-05,
"loss": 0.8115,
"step": 2150
},
{
"epoch": 1.3181818181818181,
"grad_norm": 6.7410078048706055,
"learning_rate": 3.11672278338945e-05,
"loss": 0.7262,
"step": 2175
},
{
"epoch": 1.3333333333333333,
"grad_norm": 5.232451438903809,
"learning_rate": 3.088664421997756e-05,
"loss": 0.7573,
"step": 2200
},
{
"epoch": 1.3484848484848486,
"grad_norm": 4.0043439865112305,
"learning_rate": 3.060606060606061e-05,
"loss": 0.9982,
"step": 2225
},
{
"epoch": 1.3636363636363638,
"grad_norm": 8.409883499145508,
"learning_rate": 3.0325476992143658e-05,
"loss": 0.7902,
"step": 2250
},
{
"epoch": 1.378787878787879,
"grad_norm": 8.475225448608398,
"learning_rate": 3.004489337822671e-05,
"loss": 0.8343,
"step": 2275
},
{
"epoch": 1.393939393939394,
"grad_norm": 4.989223480224609,
"learning_rate": 2.976430976430977e-05,
"loss": 0.8126,
"step": 2300
},
{
"epoch": 1.4090909090909092,
"grad_norm": 5.526180267333984,
"learning_rate": 2.9483726150392822e-05,
"loss": 0.6589,
"step": 2325
},
{
"epoch": 1.4242424242424243,
"grad_norm": 7.382469177246094,
"learning_rate": 2.920314253647587e-05,
"loss": 0.7384,
"step": 2350
},
{
"epoch": 1.4393939393939394,
"grad_norm": 5.158092498779297,
"learning_rate": 2.8922558922558923e-05,
"loss": 0.6578,
"step": 2375
},
{
"epoch": 1.4545454545454546,
"grad_norm": 4.976596832275391,
"learning_rate": 2.8641975308641977e-05,
"loss": 0.8378,
"step": 2400
},
{
"epoch": 1.4696969696969697,
"grad_norm": 9.176383018493652,
"learning_rate": 2.836139169472503e-05,
"loss": 0.7935,
"step": 2425
},
{
"epoch": 1.4848484848484849,
"grad_norm": 10.877906799316406,
"learning_rate": 2.808080808080808e-05,
"loss": 0.8241,
"step": 2450
},
{
"epoch": 1.5,
"grad_norm": 12.3035249710083,
"learning_rate": 2.7800224466891134e-05,
"loss": 0.8298,
"step": 2475
},
{
"epoch": 1.5151515151515151,
"grad_norm": 7.093743324279785,
"learning_rate": 2.7519640852974188e-05,
"loss": 0.7239,
"step": 2500
},
{
"epoch": 1.5303030303030303,
"grad_norm": 14.64828872680664,
"learning_rate": 2.7239057239057242e-05,
"loss": 0.7086,
"step": 2525
},
{
"epoch": 1.5454545454545454,
"grad_norm": 5.189187526702881,
"learning_rate": 2.6958473625140296e-05,
"loss": 0.7458,
"step": 2550
},
{
"epoch": 1.5606060606060606,
"grad_norm": 7.915388107299805,
"learning_rate": 2.6677890011223346e-05,
"loss": 0.765,
"step": 2575
},
{
"epoch": 1.5757575757575757,
"grad_norm": 5.491411209106445,
"learning_rate": 2.63973063973064e-05,
"loss": 0.8382,
"step": 2600
},
{
"epoch": 1.5909090909090908,
"grad_norm": 10.231388092041016,
"learning_rate": 2.6116722783389453e-05,
"loss": 0.6905,
"step": 2625
},
{
"epoch": 1.606060606060606,
"grad_norm": 4.599724769592285,
"learning_rate": 2.5836139169472507e-05,
"loss": 0.7544,
"step": 2650
},
{
"epoch": 1.621212121212121,
"grad_norm": 9.038063049316406,
"learning_rate": 2.5555555555555554e-05,
"loss": 0.5843,
"step": 2675
},
{
"epoch": 1.6363636363636362,
"grad_norm": 7.875892639160156,
"learning_rate": 2.5274971941638608e-05,
"loss": 0.7318,
"step": 2700
},
{
"epoch": 1.6515151515151514,
"grad_norm": 7.997163772583008,
"learning_rate": 2.499438832772166e-05,
"loss": 0.6789,
"step": 2725
},
{
"epoch": 1.6666666666666665,
"grad_norm": 12.051643371582031,
"learning_rate": 2.4725028058361395e-05,
"loss": 0.8427,
"step": 2750
},
{
"epoch": 1.6818181818181817,
"grad_norm": 18.78903579711914,
"learning_rate": 2.4444444444444445e-05,
"loss": 0.7387,
"step": 2775
},
{
"epoch": 1.696969696969697,
"grad_norm": 5.628903865814209,
"learning_rate": 2.41638608305275e-05,
"loss": 0.6316,
"step": 2800
},
{
"epoch": 1.7121212121212122,
"grad_norm": 6.309241771697998,
"learning_rate": 2.388327721661055e-05,
"loss": 0.6701,
"step": 2825
},
{
"epoch": 1.7272727272727273,
"grad_norm": 6.3270392417907715,
"learning_rate": 2.3602693602693603e-05,
"loss": 0.5524,
"step": 2850
},
{
"epoch": 1.7424242424242424,
"grad_norm": 3.400372266769409,
"learning_rate": 2.3322109988776656e-05,
"loss": 0.6322,
"step": 2875
},
{
"epoch": 1.7575757575757576,
"grad_norm": 10.937955856323242,
"learning_rate": 2.304152637485971e-05,
"loss": 0.5497,
"step": 2900
},
{
"epoch": 1.7727272727272727,
"grad_norm": 9.74146842956543,
"learning_rate": 2.2760942760942764e-05,
"loss": 0.7272,
"step": 2925
},
{
"epoch": 1.7878787878787878,
"grad_norm": 24.714502334594727,
"learning_rate": 2.2480359147025814e-05,
"loss": 0.7961,
"step": 2950
},
{
"epoch": 1.803030303030303,
"grad_norm": 3.8761136531829834,
"learning_rate": 2.2199775533108868e-05,
"loss": 0.6801,
"step": 2975
},
{
"epoch": 1.8181818181818183,
"grad_norm": 5.883018493652344,
"learning_rate": 2.191919191919192e-05,
"loss": 0.7082,
"step": 3000
},
{
"epoch": 1.8333333333333335,
"grad_norm": 7.082830905914307,
"learning_rate": 2.1638608305274975e-05,
"loss": 0.5892,
"step": 3025
},
{
"epoch": 1.8484848484848486,
"grad_norm": 10.057491302490234,
"learning_rate": 2.1358024691358026e-05,
"loss": 0.7226,
"step": 3050
},
{
"epoch": 1.8636363636363638,
"grad_norm": 3.2809817790985107,
"learning_rate": 2.107744107744108e-05,
"loss": 0.546,
"step": 3075
},
{
"epoch": 1.878787878787879,
"grad_norm": 8.243097305297852,
"learning_rate": 2.079685746352413e-05,
"loss": 0.6747,
"step": 3100
},
{
"epoch": 1.893939393939394,
"grad_norm": 7.2149457931518555,
"learning_rate": 2.0516273849607184e-05,
"loss": 0.6502,
"step": 3125
},
{
"epoch": 1.9090909090909092,
"grad_norm": 5.218500137329102,
"learning_rate": 2.0235690235690234e-05,
"loss": 0.5866,
"step": 3150
},
{
"epoch": 1.9242424242424243,
"grad_norm": 8.202606201171875,
"learning_rate": 1.995510662177329e-05,
"loss": 0.7029,
"step": 3175
},
{
"epoch": 1.9393939393939394,
"grad_norm": 10.254151344299316,
"learning_rate": 1.967452300785634e-05,
"loss": 0.6972,
"step": 3200
},
{
"epoch": 1.9545454545454546,
"grad_norm": 3.413534641265869,
"learning_rate": 1.9393939393939395e-05,
"loss": 0.6599,
"step": 3225
},
{
"epoch": 1.9696969696969697,
"grad_norm": 5.466010570526123,
"learning_rate": 1.911335578002245e-05,
"loss": 0.6938,
"step": 3250
},
{
"epoch": 1.9848484848484849,
"grad_norm": 4.317994594573975,
"learning_rate": 1.88327721661055e-05,
"loss": 0.7309,
"step": 3275
},
{
"epoch": 2.0,
"grad_norm": 7.49802303314209,
"learning_rate": 1.8552188552188553e-05,
"loss": 0.5147,
"step": 3300
},
{
"epoch": 2.0,
"eval_accuracy": 0.8475757575757575,
"eval_f1_macro": 0.8117394542384239,
"eval_f1_micro": 0.8475757575757575,
"eval_f1_weighted": 0.840866356559439,
"eval_loss": 0.5004541873931885,
"eval_precision_macro": 0.8356258037544531,
"eval_precision_micro": 0.8475757575757575,
"eval_precision_weighted": 0.8505208734646063,
"eval_recall_macro": 0.8084166666666667,
"eval_recall_micro": 0.8475757575757575,
"eval_recall_weighted": 0.8475757575757575,
"eval_runtime": 18.9039,
"eval_samples_per_second": 174.567,
"eval_steps_per_second": 10.95,
"step": 3300
},
{
"epoch": 2.015151515151515,
"grad_norm": 4.281924247741699,
"learning_rate": 1.8271604938271607e-05,
"loss": 0.4413,
"step": 3325
},
{
"epoch": 2.0303030303030303,
"grad_norm": 11.078526496887207,
"learning_rate": 1.799102132435466e-05,
"loss": 0.516,
"step": 3350
},
{
"epoch": 2.0454545454545454,
"grad_norm": 12.010851860046387,
"learning_rate": 1.771043771043771e-05,
"loss": 0.6546,
"step": 3375
},
{
"epoch": 2.0606060606060606,
"grad_norm": 6.438499450683594,
"learning_rate": 1.7429854096520764e-05,
"loss": 0.6509,
"step": 3400
},
{
"epoch": 2.0757575757575757,
"grad_norm": 2.6832661628723145,
"learning_rate": 1.7149270482603815e-05,
"loss": 0.6495,
"step": 3425
},
{
"epoch": 2.090909090909091,
"grad_norm": 6.37797737121582,
"learning_rate": 1.686868686868687e-05,
"loss": 0.6738,
"step": 3450
},
{
"epoch": 2.106060606060606,
"grad_norm": 5.069665908813477,
"learning_rate": 1.6588103254769922e-05,
"loss": 0.5145,
"step": 3475
},
{
"epoch": 2.121212121212121,
"grad_norm": 12.002437591552734,
"learning_rate": 1.6307519640852976e-05,
"loss": 0.7411,
"step": 3500
},
{
"epoch": 2.1363636363636362,
"grad_norm": 4.099394798278809,
"learning_rate": 1.6026936026936026e-05,
"loss": 0.6736,
"step": 3525
},
{
"epoch": 2.1515151515151514,
"grad_norm": 11.855905532836914,
"learning_rate": 1.574635241301908e-05,
"loss": 0.4738,
"step": 3550
},
{
"epoch": 2.1666666666666665,
"grad_norm": 9.715377807617188,
"learning_rate": 1.5465768799102134e-05,
"loss": 0.5321,
"step": 3575
},
{
"epoch": 2.1818181818181817,
"grad_norm": 10.646961212158203,
"learning_rate": 1.5185185185185186e-05,
"loss": 0.5856,
"step": 3600
},
{
"epoch": 2.196969696969697,
"grad_norm": 9.710843086242676,
"learning_rate": 1.490460157126824e-05,
"loss": 0.6343,
"step": 3625
},
{
"epoch": 2.212121212121212,
"grad_norm": 5.35206937789917,
"learning_rate": 1.4624017957351292e-05,
"loss": 0.7256,
"step": 3650
},
{
"epoch": 2.227272727272727,
"grad_norm": 6.582408428192139,
"learning_rate": 1.4343434343434345e-05,
"loss": 0.5843,
"step": 3675
},
{
"epoch": 2.242424242424242,
"grad_norm": 11.637269973754883,
"learning_rate": 1.4062850729517396e-05,
"loss": 0.5725,
"step": 3700
},
{
"epoch": 2.257575757575758,
"grad_norm": 11.256560325622559,
"learning_rate": 1.378226711560045e-05,
"loss": 0.5129,
"step": 3725
},
{
"epoch": 2.2727272727272725,
"grad_norm": 4.118545055389404,
"learning_rate": 1.3501683501683501e-05,
"loss": 0.6303,
"step": 3750
},
{
"epoch": 2.287878787878788,
"grad_norm": 8.819905281066895,
"learning_rate": 1.3221099887766555e-05,
"loss": 0.4861,
"step": 3775
},
{
"epoch": 2.303030303030303,
"grad_norm": 11.89466667175293,
"learning_rate": 1.2940516273849607e-05,
"loss": 0.6803,
"step": 3800
},
{
"epoch": 2.3181818181818183,
"grad_norm": 14.192046165466309,
"learning_rate": 1.2659932659932661e-05,
"loss": 0.6194,
"step": 3825
},
{
"epoch": 2.3333333333333335,
"grad_norm": 8.376368522644043,
"learning_rate": 1.2379349046015713e-05,
"loss": 0.5394,
"step": 3850
},
{
"epoch": 2.3484848484848486,
"grad_norm": 2.375666856765747,
"learning_rate": 1.2098765432098767e-05,
"loss": 0.5348,
"step": 3875
},
{
"epoch": 2.3636363636363638,
"grad_norm": 6.248928070068359,
"learning_rate": 1.1818181818181819e-05,
"loss": 0.6926,
"step": 3900
},
{
"epoch": 2.378787878787879,
"grad_norm": 3.3347110748291016,
"learning_rate": 1.153759820426487e-05,
"loss": 0.5196,
"step": 3925
},
{
"epoch": 2.393939393939394,
"grad_norm": 7.418989658355713,
"learning_rate": 1.1257014590347924e-05,
"loss": 0.5975,
"step": 3950
},
{
"epoch": 2.409090909090909,
"grad_norm": 17.091835021972656,
"learning_rate": 1.0976430976430976e-05,
"loss": 0.5818,
"step": 3975
},
{
"epoch": 2.4242424242424243,
"grad_norm": 4.748253345489502,
"learning_rate": 1.0695847362514029e-05,
"loss": 0.5008,
"step": 4000
},
{
"epoch": 2.4393939393939394,
"grad_norm": 2.7156593799591064,
"learning_rate": 1.0415263748597082e-05,
"loss": 0.5977,
"step": 4025
},
{
"epoch": 2.4545454545454546,
"grad_norm": 3.704885482788086,
"learning_rate": 1.0134680134680136e-05,
"loss": 0.428,
"step": 4050
},
{
"epoch": 2.4696969696969697,
"grad_norm": 6.289621829986572,
"learning_rate": 9.854096520763188e-06,
"loss": 0.6108,
"step": 4075
},
{
"epoch": 2.484848484848485,
"grad_norm": 11.862101554870605,
"learning_rate": 9.573512906846242e-06,
"loss": 0.6182,
"step": 4100
},
{
"epoch": 2.5,
"grad_norm": 2.4363837242126465,
"learning_rate": 9.292929292929294e-06,
"loss": 0.6187,
"step": 4125
},
{
"epoch": 2.515151515151515,
"grad_norm": 4.281114101409912,
"learning_rate": 9.012345679012346e-06,
"loss": 0.5247,
"step": 4150
},
{
"epoch": 2.5303030303030303,
"grad_norm": 7.008238315582275,
"learning_rate": 8.7317620650954e-06,
"loss": 0.4063,
"step": 4175
},
{
"epoch": 2.5454545454545454,
"grad_norm": 10.460261344909668,
"learning_rate": 8.451178451178452e-06,
"loss": 0.4699,
"step": 4200
},
{
"epoch": 2.5606060606060606,
"grad_norm": 6.779576301574707,
"learning_rate": 8.170594837261504e-06,
"loss": 0.632,
"step": 4225
},
{
"epoch": 2.5757575757575757,
"grad_norm": 12.081551551818848,
"learning_rate": 7.890011223344557e-06,
"loss": 0.5803,
"step": 4250
},
{
"epoch": 2.590909090909091,
"grad_norm": 14.813763618469238,
"learning_rate": 7.609427609427609e-06,
"loss": 0.5825,
"step": 4275
},
{
"epoch": 2.606060606060606,
"grad_norm": 3.7761809825897217,
"learning_rate": 7.328843995510662e-06,
"loss": 0.4587,
"step": 4300
},
{
"epoch": 2.621212121212121,
"grad_norm": 2.71801495552063,
"learning_rate": 7.048260381593716e-06,
"loss": 0.5465,
"step": 4325
},
{
"epoch": 2.6363636363636362,
"grad_norm": 5.783569812774658,
"learning_rate": 6.767676767676769e-06,
"loss": 0.5003,
"step": 4350
},
{
"epoch": 2.6515151515151514,
"grad_norm": 12.182541847229004,
"learning_rate": 6.487093153759821e-06,
"loss": 0.5783,
"step": 4375
},
{
"epoch": 2.6666666666666665,
"grad_norm": 12.15306282043457,
"learning_rate": 6.206509539842874e-06,
"loss": 0.5159,
"step": 4400
},
{
"epoch": 2.6818181818181817,
"grad_norm": 8.337636947631836,
"learning_rate": 5.925925925925927e-06,
"loss": 0.4119,
"step": 4425
},
{
"epoch": 2.6969696969696972,
"grad_norm": 3.530006170272827,
"learning_rate": 5.645342312008979e-06,
"loss": 0.4734,
"step": 4450
},
{
"epoch": 2.712121212121212,
"grad_norm": 7.290157794952393,
"learning_rate": 5.364758698092032e-06,
"loss": 0.5549,
"step": 4475
},
{
"epoch": 2.7272727272727275,
"grad_norm": 6.58420467376709,
"learning_rate": 5.0841750841750845e-06,
"loss": 0.4441,
"step": 4500
},
{
"epoch": 2.742424242424242,
"grad_norm": 4.345494747161865,
"learning_rate": 4.803591470258137e-06,
"loss": 0.48,
"step": 4525
},
{
"epoch": 2.757575757575758,
"grad_norm": 10.148066520690918,
"learning_rate": 4.523007856341189e-06,
"loss": 0.4423,
"step": 4550
},
{
"epoch": 2.7727272727272725,
"grad_norm": 6.80890417098999,
"learning_rate": 4.242424242424243e-06,
"loss": 0.4497,
"step": 4575
},
{
"epoch": 2.787878787878788,
"grad_norm": 9.686090469360352,
"learning_rate": 3.961840628507295e-06,
"loss": 0.3885,
"step": 4600
},
{
"epoch": 2.8030303030303028,
"grad_norm": 6.605715274810791,
"learning_rate": 3.681257014590348e-06,
"loss": 0.4989,
"step": 4625
},
{
"epoch": 2.8181818181818183,
"grad_norm": 5.217257499694824,
"learning_rate": 3.400673400673401e-06,
"loss": 0.5379,
"step": 4650
},
{
"epoch": 2.8333333333333335,
"grad_norm": 8.507195472717285,
"learning_rate": 3.1200897867564538e-06,
"loss": 0.6228,
"step": 4675
},
{
"epoch": 2.8484848484848486,
"grad_norm": 5.713125228881836,
"learning_rate": 2.8395061728395062e-06,
"loss": 0.4659,
"step": 4700
},
{
"epoch": 2.8636363636363638,
"grad_norm": 13.924797058105469,
"learning_rate": 2.558922558922559e-06,
"loss": 0.5267,
"step": 4725
},
{
"epoch": 2.878787878787879,
"grad_norm": 9.560832023620605,
"learning_rate": 2.2783389450056116e-06,
"loss": 0.5839,
"step": 4750
},
{
"epoch": 2.893939393939394,
"grad_norm": 4.680112361907959,
"learning_rate": 1.9977553310886645e-06,
"loss": 0.3282,
"step": 4775
},
{
"epoch": 2.909090909090909,
"grad_norm": 3.320807695388794,
"learning_rate": 1.7171717171717171e-06,
"loss": 0.4741,
"step": 4800
},
{
"epoch": 2.9242424242424243,
"grad_norm": 6.167539119720459,
"learning_rate": 1.43658810325477e-06,
"loss": 0.5005,
"step": 4825
},
{
"epoch": 2.9393939393939394,
"grad_norm": 6.241283893585205,
"learning_rate": 1.156004489337823e-06,
"loss": 0.4634,
"step": 4850
},
{
"epoch": 2.9545454545454546,
"grad_norm": 2.0378332138061523,
"learning_rate": 8.754208754208755e-07,
"loss": 0.4892,
"step": 4875
},
{
"epoch": 2.9696969696969697,
"grad_norm": 8.896339416503906,
"learning_rate": 5.948372615039282e-07,
"loss": 0.5611,
"step": 4900
},
{
"epoch": 2.984848484848485,
"grad_norm": 2.1297545433044434,
"learning_rate": 3.1425364758698096e-07,
"loss": 0.516,
"step": 4925
},
{
"epoch": 3.0,
"grad_norm": 10.862909317016602,
"learning_rate": 3.367003367003367e-08,
"loss": 0.4453,
"step": 4950
},
{
"epoch": 3.0,
"eval_accuracy": 0.886969696969697,
"eval_f1_macro": 0.8644630522360383,
"eval_f1_micro": 0.886969696969697,
"eval_f1_weighted": 0.8837489529217776,
"eval_loss": 0.3821594715118408,
"eval_precision_macro": 0.8700338902181693,
"eval_precision_micro": 0.886969696969697,
"eval_precision_weighted": 0.8838390180385471,
"eval_recall_macro": 0.8628333333333335,
"eval_recall_micro": 0.886969696969697,
"eval_recall_weighted": 0.886969696969697,
"eval_runtime": 18.9146,
"eval_samples_per_second": 174.469,
"eval_steps_per_second": 10.944,
"step": 4950
}
],
"logging_steps": 25,
"max_steps": 4950,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 5,
"early_stopping_threshold": 0.01
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.084816011852841e+19,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}