gpt2-xl-lora-multi-3 / all_results.json
MHGanainy's picture
MHGanainy/gpt2-xl-lora-multi-3
64c3d2a verified
raw
history blame
384 Bytes
{
"epoch": 1.0,
"eval_loss": 2.447016954421997,
"eval_runtime": 255.0441,
"eval_samples_per_second": 7.795,
"eval_steps_per_second": 0.49,
"perplexity": 11.553829661360664,
"total_flos": 1.0085141819823227e+18,
"train_loss": 2.6225888340860033,
"train_runtime": 2201.4938,
"train_samples_per_second": 50.3,
"train_steps_per_second": 3.144
}