gpt2-xl-lora-multi-6 / train_results.json
MHGanainy's picture
MHGanainy/gpt2-xl-lora-multi-6
4edd52d verified
raw
history blame
208 Bytes
{
"epoch": 1.0,
"total_flos": 1.3078189256852111e+18,
"train_loss": 2.2911263094506222,
"train_runtime": 2843.733,
"train_samples_per_second": 50.495,
"train_steps_per_second": 3.156
}