|
{ |
|
"best_metric": 0.9648304120191213, |
|
"best_model_checkpoint": "resnet-50-finetuned-eurosat/checkpoint-2745", |
|
"epoch": 4.997724169321803, |
|
"eval_steps": 500, |
|
"global_step": 2745, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.818181818181818e-06, |
|
"loss": 3.6402, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 3.636363636363636e-06, |
|
"loss": 3.6399, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 5.4545454545454545e-06, |
|
"loss": 3.6392, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 7.272727272727272e-06, |
|
"loss": 3.637, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.090909090909091e-06, |
|
"loss": 3.6383, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.0909090909090909e-05, |
|
"loss": 3.6362, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.2727272727272727e-05, |
|
"loss": 3.6312, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.4545454545454545e-05, |
|
"loss": 3.6275, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.6363636363636366e-05, |
|
"loss": 3.6237, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8181818181818182e-05, |
|
"loss": 3.6269, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 2e-05, |
|
"loss": 3.6201, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 2.1818181818181818e-05, |
|
"loss": 3.6147, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 2.3636363636363637e-05, |
|
"loss": 3.6125, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 2.5454545454545454e-05, |
|
"loss": 3.6079, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 2.7272727272727273e-05, |
|
"loss": 3.6001, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 2.909090909090909e-05, |
|
"loss": 3.5919, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 3.090909090909091e-05, |
|
"loss": 3.5876, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 3.272727272727273e-05, |
|
"loss": 3.5793, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 3.454545454545455e-05, |
|
"loss": 3.5733, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 3.6363636363636364e-05, |
|
"loss": 3.5642, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 3.818181818181819e-05, |
|
"loss": 3.556, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4e-05, |
|
"loss": 3.5382, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.181818181818182e-05, |
|
"loss": 3.5308, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.3636363636363636e-05, |
|
"loss": 3.5158, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.545454545454546e-05, |
|
"loss": 3.4985, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.7272727272727275e-05, |
|
"loss": 3.4826, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.909090909090909e-05, |
|
"loss": 3.4614, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.989878542510122e-05, |
|
"loss": 3.4349, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.9696356275303645e-05, |
|
"loss": 3.4079, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.9493927125506076e-05, |
|
"loss": 3.3765, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.9291497975708506e-05, |
|
"loss": 3.3423, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.908906882591093e-05, |
|
"loss": 3.2997, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.888663967611336e-05, |
|
"loss": 3.2531, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 4.868421052631579e-05, |
|
"loss": 3.2099, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.848178137651822e-05, |
|
"loss": 3.1511, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 4.8279352226720646e-05, |
|
"loss": 3.1053, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.8076923076923084e-05, |
|
"loss": 3.03, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.787449392712551e-05, |
|
"loss": 2.9668, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.767206477732794e-05, |
|
"loss": 2.8766, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.746963562753037e-05, |
|
"loss": 2.7964, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.726720647773279e-05, |
|
"loss": 2.7181, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.7064777327935223e-05, |
|
"loss": 2.6812, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.6862348178137654e-05, |
|
"loss": 2.5486, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.6659919028340085e-05, |
|
"loss": 2.4858, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.6457489878542516e-05, |
|
"loss": 2.3997, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.625506072874494e-05, |
|
"loss": 2.3232, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.605263157894737e-05, |
|
"loss": 2.2581, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.58502024291498e-05, |
|
"loss": 2.1588, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 4.564777327935223e-05, |
|
"loss": 2.1098, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.5445344129554655e-05, |
|
"loss": 2.0018, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 4.524291497975709e-05, |
|
"loss": 1.9727, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.504048582995952e-05, |
|
"loss": 1.8964, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.483805668016194e-05, |
|
"loss": 1.7941, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.463562753036438e-05, |
|
"loss": 1.7502, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.795356248577282, |
|
"eval_loss": 1.328504204750061, |
|
"eval_runtime": 106.9422, |
|
"eval_samples_per_second": 164.313, |
|
"eval_steps_per_second": 5.143, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.44331983805668e-05, |
|
"loss": 1.7101, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 4.423076923076923e-05, |
|
"loss": 1.6815, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.402834008097166e-05, |
|
"loss": 1.5733, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 4.3825910931174094e-05, |
|
"loss": 1.5261, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 4.362348178137652e-05, |
|
"loss": 1.479, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 4.342105263157895e-05, |
|
"loss": 1.4515, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.321862348178138e-05, |
|
"loss": 1.3921, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 4.30161943319838e-05, |
|
"loss": 1.3221, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 4.281376518218624e-05, |
|
"loss": 1.332, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 4.2611336032388664e-05, |
|
"loss": 1.2475, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 4.2408906882591095e-05, |
|
"loss": 1.2658, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 4.2206477732793526e-05, |
|
"loss": 1.2328, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.200404858299595e-05, |
|
"loss": 1.1655, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 4.180161943319838e-05, |
|
"loss": 1.1272, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 4.159919028340081e-05, |
|
"loss": 1.0837, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 4.139676113360324e-05, |
|
"loss": 1.1115, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 4.1194331983805666e-05, |
|
"loss": 1.078, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 4.09919028340081e-05, |
|
"loss": 1.0328, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 4.078947368421053e-05, |
|
"loss": 1.0057, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 4.058704453441296e-05, |
|
"loss": 0.9995, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 4.038461538461539e-05, |
|
"loss": 0.975, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 4.018218623481781e-05, |
|
"loss": 0.9552, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 3.997975708502025e-05, |
|
"loss": 0.9118, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 3.9777327935222674e-05, |
|
"loss": 0.9089, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 3.9574898785425104e-05, |
|
"loss": 0.8996, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 3.9372469635627535e-05, |
|
"loss": 0.8773, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 3.917004048582996e-05, |
|
"loss": 0.8736, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 3.896761133603239e-05, |
|
"loss": 0.8889, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 3.876518218623482e-05, |
|
"loss": 0.8332, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 3.856275303643725e-05, |
|
"loss": 0.7772, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 3.8360323886639675e-05, |
|
"loss": 0.7924, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 3.815789473684211e-05, |
|
"loss": 0.8023, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 3.7955465587044536e-05, |
|
"loss": 0.7885, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.775303643724697e-05, |
|
"loss": 0.753, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.75506072874494e-05, |
|
"loss": 0.7889, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.734817813765182e-05, |
|
"loss": 0.7486, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.714574898785425e-05, |
|
"loss": 0.7106, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.694331983805668e-05, |
|
"loss": 0.7279, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.6740890688259113e-05, |
|
"loss": 0.727, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 3.653846153846154e-05, |
|
"loss": 0.7469, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.633603238866397e-05, |
|
"loss": 0.6788, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.61336032388664e-05, |
|
"loss": 0.6962, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.593117408906882e-05, |
|
"loss": 0.6872, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.572874493927126e-05, |
|
"loss": 0.6464, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.5526315789473684e-05, |
|
"loss": 0.6389, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.5323886639676115e-05, |
|
"loss": 0.6092, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3.5121457489878545e-05, |
|
"loss": 0.6159, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.491902834008097e-05, |
|
"loss": 0.6301, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 3.47165991902834e-05, |
|
"loss": 0.6138, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 3.451417004048583e-05, |
|
"loss": 0.6002, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 3.431174089068826e-05, |
|
"loss": 0.6109, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 3.410931174089069e-05, |
|
"loss": 0.6136, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 3.390688259109312e-05, |
|
"loss": 0.5631, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 3.3704453441295546e-05, |
|
"loss": 0.6395, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 3.350202429149798e-05, |
|
"loss": 0.6126, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9301160937855679, |
|
"eval_loss": 0.32050490379333496, |
|
"eval_runtime": 103.9225, |
|
"eval_samples_per_second": 169.088, |
|
"eval_steps_per_second": 5.292, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 3.329959514170041e-05, |
|
"loss": 0.5663, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 3.309716599190283e-05, |
|
"loss": 0.5179, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 3.289473684210527e-05, |
|
"loss": 0.5657, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 3.269230769230769e-05, |
|
"loss": 0.5623, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 3.2489878542510124e-05, |
|
"loss": 0.581, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 3.2287449392712554e-05, |
|
"loss": 0.5664, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 3.208502024291498e-05, |
|
"loss": 0.5485, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 3.188259109311741e-05, |
|
"loss": 0.5518, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.168016194331984e-05, |
|
"loss": 0.5452, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.147773279352227e-05, |
|
"loss": 0.5394, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.1275303643724694e-05, |
|
"loss": 0.5483, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.107287449392713e-05, |
|
"loss": 0.5503, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.0870445344129556e-05, |
|
"loss": 0.5164, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.0668016194331986e-05, |
|
"loss": 0.4868, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.0465587044534417e-05, |
|
"loss": 0.4483, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 3.0263157894736844e-05, |
|
"loss": 0.5105, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 3.006072874493927e-05, |
|
"loss": 0.4933, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 2.9858299595141702e-05, |
|
"loss": 0.4973, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 2.965587044534413e-05, |
|
"loss": 0.4855, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 2.9453441295546557e-05, |
|
"loss": 0.4828, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 2.925101214574899e-05, |
|
"loss": 0.4806, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 2.9048582995951418e-05, |
|
"loss": 0.5003, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 2.8846153846153845e-05, |
|
"loss": 0.5143, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 2.8643724696356276e-05, |
|
"loss": 0.4758, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 2.8441295546558703e-05, |
|
"loss": 0.4782, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.823886639676113e-05, |
|
"loss": 0.4825, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.8036437246963565e-05, |
|
"loss": 0.4637, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 2.7834008097165992e-05, |
|
"loss": 0.4574, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.7631578947368426e-05, |
|
"loss": 0.443, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.7429149797570853e-05, |
|
"loss": 0.4742, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 2.722672064777328e-05, |
|
"loss": 0.4719, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.702429149797571e-05, |
|
"loss": 0.4539, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.682186234817814e-05, |
|
"loss": 0.4393, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.6619433198380566e-05, |
|
"loss": 0.4418, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.6417004048583e-05, |
|
"loss": 0.4801, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.6214574898785427e-05, |
|
"loss": 0.4524, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.6012145748987855e-05, |
|
"loss": 0.4154, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.5809716599190285e-05, |
|
"loss": 0.4374, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.5607287449392713e-05, |
|
"loss": 0.3937, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.540485829959514e-05, |
|
"loss": 0.4008, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.5202429149797574e-05, |
|
"loss": 0.4617, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.4357, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 2.4797570850202432e-05, |
|
"loss": 0.4234, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 2.459514170040486e-05, |
|
"loss": 0.4479, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 2.439271255060729e-05, |
|
"loss": 0.3867, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 2.4190283400809717e-05, |
|
"loss": 0.4096, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.3987854251012144e-05, |
|
"loss": 0.4076, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.3785425101214575e-05, |
|
"loss": 0.4417, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 2.3582995951417006e-05, |
|
"loss": 0.4439, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 2.3380566801619436e-05, |
|
"loss": 0.4154, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 2.3178137651821864e-05, |
|
"loss": 0.4498, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.2975708502024294e-05, |
|
"loss": 0.4068, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 2.2773279352226722e-05, |
|
"loss": 0.3884, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.257085020242915e-05, |
|
"loss": 0.3952, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.236842105263158e-05, |
|
"loss": 0.4063, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.9550990211700433, |
|
"eval_loss": 0.1893182247877121, |
|
"eval_runtime": 104.3895, |
|
"eval_samples_per_second": 168.331, |
|
"eval_steps_per_second": 5.269, |
|
"step": 1647 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 2.216599190283401e-05, |
|
"loss": 0.4239, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 2.1963562753036438e-05, |
|
"loss": 0.4083, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 2.1761133603238868e-05, |
|
"loss": 0.4241, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 2.15587044534413e-05, |
|
"loss": 0.407, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 2.1356275303643726e-05, |
|
"loss": 0.3937, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 2.1153846153846154e-05, |
|
"loss": 0.3753, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 2.0951417004048584e-05, |
|
"loss": 0.3942, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 2.074898785425101e-05, |
|
"loss": 0.4204, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 2.0546558704453442e-05, |
|
"loss": 0.4329, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 2.0344129554655873e-05, |
|
"loss": 0.3618, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 2.0141700404858304e-05, |
|
"loss": 0.3692, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 1.9939271255060727e-05, |
|
"loss": 0.3712, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 1.9736842105263158e-05, |
|
"loss": 0.419, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 1.953441295546559e-05, |
|
"loss": 0.3825, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 1.9331983805668016e-05, |
|
"loss": 0.3896, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 1.9129554655870447e-05, |
|
"loss": 0.3531, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 1.8927125506072877e-05, |
|
"loss": 0.3676, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 1.8724696356275305e-05, |
|
"loss": 0.3713, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 1.8522267206477732e-05, |
|
"loss": 0.3826, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 1.8319838056680163e-05, |
|
"loss": 0.3865, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 1.811740890688259e-05, |
|
"loss": 0.362, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 1.791497975708502e-05, |
|
"loss": 0.3906, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.771255060728745e-05, |
|
"loss": 0.3623, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.751012145748988e-05, |
|
"loss": 0.3954, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 1.730769230769231e-05, |
|
"loss": 0.373, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 1.7105263157894737e-05, |
|
"loss": 0.3897, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 1.6902834008097167e-05, |
|
"loss": 0.3627, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 1.6700404858299595e-05, |
|
"loss": 0.3843, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.6497975708502025e-05, |
|
"loss": 0.3501, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 1.6295546558704456e-05, |
|
"loss": 0.3681, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.6093117408906883e-05, |
|
"loss": 0.359, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 1.5890688259109314e-05, |
|
"loss": 0.3595, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 1.568825910931174e-05, |
|
"loss": 0.3684, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 1.548582995951417e-05, |
|
"loss": 0.377, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 1.52834008097166e-05, |
|
"loss": 0.3604, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 1.508097165991903e-05, |
|
"loss": 0.3432, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 1.4878542510121457e-05, |
|
"loss": 0.3641, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 1.4676113360323888e-05, |
|
"loss": 0.3588, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 1.4473684210526317e-05, |
|
"loss": 0.3579, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 1.4271255060728744e-05, |
|
"loss": 0.326, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 1.4068825910931175e-05, |
|
"loss": 0.3464, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 1.3866396761133604e-05, |
|
"loss": 0.3513, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 1.3663967611336034e-05, |
|
"loss": 0.4124, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 1.3461538461538462e-05, |
|
"loss": 0.3532, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 1.3259109311740892e-05, |
|
"loss": 0.3482, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 1.3056680161943321e-05, |
|
"loss": 0.3489, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 1.2854251012145749e-05, |
|
"loss": 0.3168, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.265182186234818e-05, |
|
"loss": 0.3575, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.2449392712550607e-05, |
|
"loss": 0.3506, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.2246963562753037e-05, |
|
"loss": 0.3338, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.2044534412955466e-05, |
|
"loss": 0.3388, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 1.1842105263157895e-05, |
|
"loss": 0.3355, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 1.1639676113360324e-05, |
|
"loss": 0.3238, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 1.1437246963562753e-05, |
|
"loss": 0.3328, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 1.1234817813765184e-05, |
|
"loss": 0.3333, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.9624402458456636, |
|
"eval_loss": 0.15150564908981323, |
|
"eval_runtime": 102.6282, |
|
"eval_samples_per_second": 171.22, |
|
"eval_steps_per_second": 5.359, |
|
"step": 2197 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 1.1032388663967611e-05, |
|
"loss": 0.3703, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 1.082995951417004e-05, |
|
"loss": 0.3558, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 1.062753036437247e-05, |
|
"loss": 0.3665, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 1.04251012145749e-05, |
|
"loss": 0.3178, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 1.0222672064777327e-05, |
|
"loss": 0.3195, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 1.0020242914979758e-05, |
|
"loss": 0.3246, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 9.817813765182187e-06, |
|
"loss": 0.3535, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 9.615384615384616e-06, |
|
"loss": 0.3413, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 9.412955465587045e-06, |
|
"loss": 0.3725, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 9.210526315789474e-06, |
|
"loss": 0.347, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 9.008097165991904e-06, |
|
"loss": 0.3258, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 8.805668016194332e-06, |
|
"loss": 0.3556, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 8.60323886639676e-06, |
|
"loss": 0.3161, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 8.400809716599191e-06, |
|
"loss": 0.3404, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 8.19838056680162e-06, |
|
"loss": 0.2928, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 7.99595141700405e-06, |
|
"loss": 0.3363, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 7.793522267206478e-06, |
|
"loss": 0.2995, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 7.591093117408906e-06, |
|
"loss": 0.3457, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 7.388663967611337e-06, |
|
"loss": 0.3192, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 7.186234817813765e-06, |
|
"loss": 0.3464, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 6.983805668016195e-06, |
|
"loss": 0.3351, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 6.781376518218624e-06, |
|
"loss": 0.3412, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 6.578947368421053e-06, |
|
"loss": 0.3266, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 6.376518218623483e-06, |
|
"loss": 0.349, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 6.174089068825911e-06, |
|
"loss": 0.3293, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 5.971659919028341e-06, |
|
"loss": 0.34, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 5.76923076923077e-06, |
|
"loss": 0.3109, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 5.566801619433199e-06, |
|
"loss": 0.3153, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 5.364372469635628e-06, |
|
"loss": 0.3318, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 5.161943319838057e-06, |
|
"loss": 0.3565, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 4.9595141700404865e-06, |
|
"loss": 0.3399, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 4.757085020242915e-06, |
|
"loss": 0.3014, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 4.5546558704453445e-06, |
|
"loss": 0.3368, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 4.3522267206477735e-06, |
|
"loss": 0.3252, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 4.149797570850203e-06, |
|
"loss": 0.313, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 3.9473684210526315e-06, |
|
"loss": 0.3495, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 3.744939271255061e-06, |
|
"loss": 0.3318, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 3.5425101214574903e-06, |
|
"loss": 0.3463, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 3.3400809716599197e-06, |
|
"loss": 0.3212, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 3.1376518218623482e-06, |
|
"loss": 0.2976, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 2.9352226720647772e-06, |
|
"loss": 0.3084, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 2.7327935222672066e-06, |
|
"loss": 0.3233, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 2.5303643724696356e-06, |
|
"loss": 0.3123, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 2.327935222672065e-06, |
|
"loss": 0.3201, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 2.125506072874494e-06, |
|
"loss": 0.378, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 1.9230769230769234e-06, |
|
"loss": 0.3351, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 1.7206477732793522e-06, |
|
"loss": 0.3003, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 1.5182186234817814e-06, |
|
"loss": 0.3341, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 1.3157894736842106e-06, |
|
"loss": 0.3253, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 1.1133603238866398e-06, |
|
"loss": 0.306, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 9.109311740890688e-07, |
|
"loss": 0.2901, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 7.08502024291498e-07, |
|
"loss": 0.3485, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 5.060728744939271e-07, |
|
"loss": 0.328, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 3.036437246963563e-07, |
|
"loss": 0.3459, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 1.0121457489878543e-07, |
|
"loss": 0.3365, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.9648304120191213, |
|
"eval_loss": 0.13630270957946777, |
|
"eval_runtime": 102.6364, |
|
"eval_samples_per_second": 171.206, |
|
"eval_steps_per_second": 5.359, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"step": 2745, |
|
"total_flos": 7.483912529986142e+18, |
|
"train_loss": 1.0410587748977438, |
|
"train_runtime": 5056.7157, |
|
"train_samples_per_second": 69.507, |
|
"train_steps_per_second": 0.543 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2745, |
|
"num_train_epochs": 5, |
|
"save_steps": 500, |
|
"total_flos": 7.483912529986142e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|