gpt2-xl-lora-multi-6 / all_results.json
MHGanainy's picture
MHGanainy/gpt2-xl-lora-multi-6
4edd52d verified
raw
history blame
387 Bytes
{
"epoch": 1.0,
"eval_loss": 2.0126099586486816,
"eval_runtime": 4087.7012,
"eval_samples_per_second": 7.809,
"eval_steps_per_second": 0.488,
"perplexity": 7.482821738733412,
"total_flos": 1.3078189256852111e+18,
"train_loss": 2.2911263094506222,
"train_runtime": 2843.733,
"train_samples_per_second": 50.495,
"train_steps_per_second": 3.156
}