gpt2-xl-lora-multi-3 / train_results.json
MHGanainy's picture
MHGanainy/gpt2-xl-lora-multi-3
64c3d2a verified
raw
history blame
207 Bytes
{
"epoch": 1.0,
"total_flos": 1.0085141819823227e+18,
"train_loss": 2.6225888340860033,
"train_runtime": 2201.4938,
"train_samples_per_second": 50.3,
"train_steps_per_second": 3.144
}