linshoufan's picture
Training in progress, step 500, checkpoint
e7eddaa verified
raw
history blame
3.91 kB
{
"best_metric": 55.83411121482864,
"best_model_checkpoint": "./linshoufanfork-whisper-small-nan-tw/checkpoint-500",
"epoch": 0.1607200257152041,
"eval_steps": 500,
"global_step": 500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01,
"grad_norm": 48.035465240478516,
"learning_rate": 2.5e-06,
"loss": 6.8285,
"step": 25
},
{
"epoch": 0.02,
"grad_norm": 142.98660278320312,
"learning_rate": 5e-06,
"loss": 3.2493,
"step": 50
},
{
"epoch": 0.02,
"grad_norm": 27.529985427856445,
"learning_rate": 7.500000000000001e-06,
"loss": 2.2763,
"step": 75
},
{
"epoch": 0.03,
"grad_norm": 21.324289321899414,
"learning_rate": 1e-05,
"loss": 1.7193,
"step": 100
},
{
"epoch": 0.04,
"grad_norm": 13.650047302246094,
"learning_rate": 9.91697110594487e-06,
"loss": 1.3177,
"step": 125
},
{
"epoch": 0.05,
"grad_norm": 14.939111709594727,
"learning_rate": 9.83394221188974e-06,
"loss": 1.2486,
"step": 150
},
{
"epoch": 0.06,
"grad_norm": 14.85805606842041,
"learning_rate": 9.750913317834608e-06,
"loss": 1.1991,
"step": 175
},
{
"epoch": 0.06,
"grad_norm": 15.52128791809082,
"learning_rate": 9.667884423779476e-06,
"loss": 1.1126,
"step": 200
},
{
"epoch": 0.07,
"grad_norm": 22.642026901245117,
"learning_rate": 9.584855529724345e-06,
"loss": 1.0437,
"step": 225
},
{
"epoch": 0.08,
"grad_norm": 16.712217330932617,
"learning_rate": 9.501826635669213e-06,
"loss": 0.9986,
"step": 250
},
{
"epoch": 0.09,
"grad_norm": 13.973222732543945,
"learning_rate": 9.418797741614083e-06,
"loss": 0.9188,
"step": 275
},
{
"epoch": 0.1,
"grad_norm": 13.404074668884277,
"learning_rate": 9.335768847558952e-06,
"loss": 0.9244,
"step": 300
},
{
"epoch": 0.1,
"grad_norm": 12.21960163116455,
"learning_rate": 9.25273995350382e-06,
"loss": 0.9018,
"step": 325
},
{
"epoch": 0.11,
"grad_norm": 14.460400581359863,
"learning_rate": 9.169711059448689e-06,
"loss": 0.8398,
"step": 350
},
{
"epoch": 0.12,
"grad_norm": 14.076154708862305,
"learning_rate": 9.086682165393557e-06,
"loss": 0.8828,
"step": 375
},
{
"epoch": 0.13,
"grad_norm": 13.204269409179688,
"learning_rate": 9.003653271338426e-06,
"loss": 0.8503,
"step": 400
},
{
"epoch": 0.14,
"grad_norm": 10.978958129882812,
"learning_rate": 8.920624377283296e-06,
"loss": 0.8198,
"step": 425
},
{
"epoch": 0.14,
"grad_norm": 13.963995933532715,
"learning_rate": 8.837595483228164e-06,
"loss": 0.8202,
"step": 450
},
{
"epoch": 0.15,
"grad_norm": 13.337563514709473,
"learning_rate": 8.754566589173033e-06,
"loss": 0.7536,
"step": 475
},
{
"epoch": 0.16,
"grad_norm": 12.711252212524414,
"learning_rate": 8.671537695117903e-06,
"loss": 0.7938,
"step": 500
},
{
"epoch": 0.16,
"eval_cer": 55.83411121482864,
"eval_loss": 0.7767874002456665,
"eval_runtime": 1802.202,
"eval_samples_per_second": 2.461,
"eval_steps_per_second": 0.308,
"step": 500
}
],
"logging_steps": 25,
"max_steps": 3111,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"total_flos": 2.30868320256e+18,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}