File size: 345 Bytes
f67282b
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
{
    "epoch": 3.0,
    "eval_loss": 0.9407615661621094,
    "eval_runtime": 26.6788,
    "eval_samples_per_second": 39.357,
    "eval_steps_per_second": 0.637,
    "total_flos": 195769978060800.0,
    "train_loss": 0.9480493904178978,
    "train_runtime": 4080.0307,
    "train_samples_per_second": 14.655,
    "train_steps_per_second": 0.029
}