|
{ |
|
"best_metric": 0.3599020838737488, |
|
"best_model_checkpoint": "./vit-finetune-scrap/checkpoint-1000", |
|
"epoch": 4.0, |
|
"eval_steps": 1000, |
|
"global_step": 1244, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.7885215282440186, |
|
"learning_rate": 0.00019839228295819936, |
|
"loss": 0.3445, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.6006595492362976, |
|
"learning_rate": 0.00019678456591639874, |
|
"loss": 0.0458, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.0847816988825798, |
|
"learning_rate": 0.00019517684887459809, |
|
"loss": 0.1535, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 10.684067726135254, |
|
"learning_rate": 0.00019356913183279743, |
|
"loss": 0.1208, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.08578234165906906, |
|
"learning_rate": 0.00019196141479099678, |
|
"loss": 0.0208, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.06457830220460892, |
|
"learning_rate": 0.00019035369774919616, |
|
"loss": 0.0838, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 12.919880867004395, |
|
"learning_rate": 0.0001887459807073955, |
|
"loss": 0.0491, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 10.229437828063965, |
|
"learning_rate": 0.00018713826366559486, |
|
"loss": 0.1614, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.03680131584405899, |
|
"learning_rate": 0.0001855305466237942, |
|
"loss": 0.2226, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 1.575692892074585, |
|
"learning_rate": 0.0001839228295819936, |
|
"loss": 0.3464, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 11.554421424865723, |
|
"learning_rate": 0.00018231511254019294, |
|
"loss": 0.267, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 11.113668441772461, |
|
"learning_rate": 0.00018070739549839229, |
|
"loss": 0.3107, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 5.346070766448975, |
|
"learning_rate": 0.00017909967845659166, |
|
"loss": 0.6457, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 31.744802474975586, |
|
"learning_rate": 0.000177491961414791, |
|
"loss": 0.4913, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 6.202216148376465, |
|
"learning_rate": 0.00017588424437299036, |
|
"loss": 0.6627, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 12.808435440063477, |
|
"learning_rate": 0.0001742765273311897, |
|
"loss": 0.4071, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.17637468874454498, |
|
"learning_rate": 0.0001726688102893891, |
|
"loss": 0.2312, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 9.431242942810059, |
|
"learning_rate": 0.00017106109324758844, |
|
"loss": 0.1206, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.0686856061220169, |
|
"learning_rate": 0.0001694533762057878, |
|
"loss": 0.2657, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.035174936056137085, |
|
"learning_rate": 0.00016784565916398716, |
|
"loss": 0.2447, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.20221814513206482, |
|
"learning_rate": 0.0001662379421221865, |
|
"loss": 0.0932, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 22.348093032836914, |
|
"learning_rate": 0.00016463022508038586, |
|
"loss": 0.5622, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.09363432973623276, |
|
"learning_rate": 0.0001630225080385852, |
|
"loss": 0.1989, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.10725115239620209, |
|
"learning_rate": 0.0001614147909967846, |
|
"loss": 0.1174, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 2.2332615852355957, |
|
"learning_rate": 0.00015980707395498394, |
|
"loss": 0.3852, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.08053239434957504, |
|
"learning_rate": 0.0001581993569131833, |
|
"loss": 0.2247, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.21441006660461426, |
|
"learning_rate": 0.00015659163987138264, |
|
"loss": 0.2879, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.0631105974316597, |
|
"learning_rate": 0.00015498392282958201, |
|
"loss": 0.2922, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 20.96392822265625, |
|
"learning_rate": 0.00015337620578778136, |
|
"loss": 0.0413, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.4054552912712097, |
|
"learning_rate": 0.0001517684887459807, |
|
"loss": 0.1037, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.8209251165390015, |
|
"learning_rate": 0.0001501607717041801, |
|
"loss": 0.1802, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 11.979750633239746, |
|
"learning_rate": 0.00014855305466237944, |
|
"loss": 0.1788, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.10045678168535233, |
|
"learning_rate": 0.0001469453376205788, |
|
"loss": 0.4287, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.051606420427560806, |
|
"learning_rate": 0.00014533762057877814, |
|
"loss": 0.1055, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.20983509719371796, |
|
"learning_rate": 0.00014372990353697752, |
|
"loss": 0.1667, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.15858705341815948, |
|
"learning_rate": 0.00014212218649517686, |
|
"loss": 0.1135, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 1.420276165008545, |
|
"learning_rate": 0.00014051446945337621, |
|
"loss": 0.2112, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.02672198973596096, |
|
"learning_rate": 0.0001389067524115756, |
|
"loss": 0.0313, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.02562582679092884, |
|
"learning_rate": 0.00013729903536977494, |
|
"loss": 0.1069, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 4.197677135467529, |
|
"learning_rate": 0.0001356913183279743, |
|
"loss": 0.2342, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.04306723549962044, |
|
"learning_rate": 0.00013408360128617364, |
|
"loss": 0.1313, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 4.383702278137207, |
|
"learning_rate": 0.00013247588424437302, |
|
"loss": 0.2145, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.20137301087379456, |
|
"learning_rate": 0.00013086816720257237, |
|
"loss": 0.1514, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.020689483731985092, |
|
"learning_rate": 0.00012926045016077172, |
|
"loss": 0.0712, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.018355347216129303, |
|
"learning_rate": 0.00012765273311897106, |
|
"loss": 0.0923, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 17.85742950439453, |
|
"learning_rate": 0.00012604501607717044, |
|
"loss": 0.0779, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.05268234387040138, |
|
"learning_rate": 0.0001244372990353698, |
|
"loss": 0.0096, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.09329644590616226, |
|
"learning_rate": 0.00012282958199356914, |
|
"loss": 0.0392, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 8.043782234191895, |
|
"learning_rate": 0.0001212218649517685, |
|
"loss": 0.1322, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.016368461772799492, |
|
"learning_rate": 0.00011961414790996785, |
|
"loss": 0.0843, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 13.989496231079102, |
|
"learning_rate": 0.0001180064308681672, |
|
"loss": 0.2665, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 14.703727722167969, |
|
"learning_rate": 0.00011639871382636655, |
|
"loss": 0.1551, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.13277527689933777, |
|
"learning_rate": 0.00011479099678456593, |
|
"loss": 0.1346, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.04265744984149933, |
|
"learning_rate": 0.00011318327974276528, |
|
"loss": 0.1725, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.04861054942011833, |
|
"learning_rate": 0.00011157556270096463, |
|
"loss": 0.0696, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.015978263691067696, |
|
"learning_rate": 0.00010996784565916398, |
|
"loss": 0.0583, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.21797218918800354, |
|
"learning_rate": 0.00010836012861736335, |
|
"loss": 0.1746, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 16.11418342590332, |
|
"learning_rate": 0.0001067524115755627, |
|
"loss": 0.2571, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.025191914290189743, |
|
"learning_rate": 0.00010514469453376205, |
|
"loss": 0.1326, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.03328488767147064, |
|
"learning_rate": 0.00010353697749196143, |
|
"loss": 0.1601, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.017355144023895264, |
|
"learning_rate": 0.00010192926045016078, |
|
"loss": 0.1607, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.15774357318878174, |
|
"learning_rate": 0.00010032154340836013, |
|
"loss": 0.0182, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 0.039228443056344986, |
|
"learning_rate": 9.871382636655949e-05, |
|
"loss": 0.0898, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 27.718074798583984, |
|
"learning_rate": 9.710610932475884e-05, |
|
"loss": 0.1895, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 0.10635074228048325, |
|
"learning_rate": 9.54983922829582e-05, |
|
"loss": 0.0805, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 0.01466443482786417, |
|
"learning_rate": 9.389067524115757e-05, |
|
"loss": 0.0697, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 0.037685755640268326, |
|
"learning_rate": 9.228295819935692e-05, |
|
"loss": 0.0031, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 0.009445942007005215, |
|
"learning_rate": 9.067524115755628e-05, |
|
"loss": 0.0636, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 0.011714440770447254, |
|
"learning_rate": 8.906752411575563e-05, |
|
"loss": 0.0554, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 0.018049364909529686, |
|
"learning_rate": 8.7459807073955e-05, |
|
"loss": 0.0073, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 0.013684015721082687, |
|
"learning_rate": 8.585209003215434e-05, |
|
"loss": 0.004, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.0740511417388916, |
|
"learning_rate": 8.42443729903537e-05, |
|
"loss": 0.0724, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 0.0066555822268128395, |
|
"learning_rate": 8.263665594855306e-05, |
|
"loss": 0.0029, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 0.00826901663094759, |
|
"learning_rate": 8.102893890675242e-05, |
|
"loss": 0.0745, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 0.009980367496609688, |
|
"learning_rate": 7.942122186495177e-05, |
|
"loss": 0.0026, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 7.373602390289307, |
|
"learning_rate": 7.781350482315113e-05, |
|
"loss": 0.1221, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 0.011884603649377823, |
|
"learning_rate": 7.62057877813505e-05, |
|
"loss": 0.0404, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 28.61383628845215, |
|
"learning_rate": 7.459807073954984e-05, |
|
"loss": 0.0483, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 0.006681604776531458, |
|
"learning_rate": 7.299035369774921e-05, |
|
"loss": 0.0729, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 0.02536676451563835, |
|
"learning_rate": 7.138263665594856e-05, |
|
"loss": 0.0023, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 0.016948092728853226, |
|
"learning_rate": 6.977491961414792e-05, |
|
"loss": 0.0426, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 0.021912137046456337, |
|
"learning_rate": 6.816720257234727e-05, |
|
"loss": 0.064, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.014551441185176373, |
|
"learning_rate": 6.655948553054663e-05, |
|
"loss": 0.065, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 0.3249826431274414, |
|
"learning_rate": 6.495176848874598e-05, |
|
"loss": 0.1171, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 17.302183151245117, |
|
"learning_rate": 6.334405144694535e-05, |
|
"loss": 0.066, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 0.03599601984024048, |
|
"learning_rate": 6.173633440514471e-05, |
|
"loss": 0.0203, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 0.011334868147969246, |
|
"learning_rate": 6.012861736334405e-05, |
|
"loss": 0.0188, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 0.012830100953578949, |
|
"learning_rate": 5.8520900321543414e-05, |
|
"loss": 0.0028, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 0.09628697484731674, |
|
"learning_rate": 5.6913183279742764e-05, |
|
"loss": 0.0857, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 0.008330035023391247, |
|
"learning_rate": 5.530546623794213e-05, |
|
"loss": 0.0018, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 0.007389857899397612, |
|
"learning_rate": 5.369774919614148e-05, |
|
"loss": 0.0019, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.011146631091833115, |
|
"learning_rate": 5.209003215434084e-05, |
|
"loss": 0.104, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 0.021591678261756897, |
|
"learning_rate": 5.048231511254019e-05, |
|
"loss": 0.0016, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 0.09255488216876984, |
|
"learning_rate": 4.887459807073955e-05, |
|
"loss": 0.1024, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"grad_norm": 0.01300421915948391, |
|
"learning_rate": 4.726688102893891e-05, |
|
"loss": 0.0019, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"grad_norm": 0.009576304815709591, |
|
"learning_rate": 4.5659163987138265e-05, |
|
"loss": 0.0036, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"grad_norm": 0.006058481056243181, |
|
"learning_rate": 4.405144694533762e-05, |
|
"loss": 0.0333, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"grad_norm": 0.00888640247285366, |
|
"learning_rate": 4.244372990353698e-05, |
|
"loss": 0.0026, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"grad_norm": 0.008608637377619743, |
|
"learning_rate": 4.083601286173634e-05, |
|
"loss": 0.002, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"grad_norm": 0.017658203840255737, |
|
"learning_rate": 3.92282958199357e-05, |
|
"loss": 0.0021, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"eval_accuracy": 0.9260450160771704, |
|
"eval_loss": 0.3599020838737488, |
|
"eval_runtime": 19.6251, |
|
"eval_samples_per_second": 31.694, |
|
"eval_steps_per_second": 3.975, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"grad_norm": 0.015870286151766777, |
|
"learning_rate": 3.7620578778135054e-05, |
|
"loss": 0.002, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"grad_norm": 0.006019009742885828, |
|
"learning_rate": 3.601286173633441e-05, |
|
"loss": 0.0021, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"grad_norm": 0.02200642041862011, |
|
"learning_rate": 3.4405144694533766e-05, |
|
"loss": 0.0484, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"grad_norm": 0.00513369170948863, |
|
"learning_rate": 3.279742765273312e-05, |
|
"loss": 0.0021, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"grad_norm": 0.0066593969240784645, |
|
"learning_rate": 3.118971061093248e-05, |
|
"loss": 0.0017, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"grad_norm": 0.010374795645475388, |
|
"learning_rate": 2.9581993569131832e-05, |
|
"loss": 0.0015, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"grad_norm": 0.006329710595309734, |
|
"learning_rate": 2.7974276527331188e-05, |
|
"loss": 0.0727, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"grad_norm": 0.012291769497096539, |
|
"learning_rate": 2.6366559485530545e-05, |
|
"loss": 0.0022, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 0.009208714589476585, |
|
"learning_rate": 2.4758842443729904e-05, |
|
"loss": 0.0017, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"grad_norm": 0.008606062270700932, |
|
"learning_rate": 2.315112540192926e-05, |
|
"loss": 0.0355, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"grad_norm": 0.005816610064357519, |
|
"learning_rate": 2.154340836012862e-05, |
|
"loss": 0.0024, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"grad_norm": 0.016983453184366226, |
|
"learning_rate": 1.9935691318327977e-05, |
|
"loss": 0.0022, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"grad_norm": 0.013002891093492508, |
|
"learning_rate": 1.8327974276527333e-05, |
|
"loss": 0.0027, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"grad_norm": 0.004919551312923431, |
|
"learning_rate": 1.672025723472669e-05, |
|
"loss": 0.0038, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"grad_norm": 0.01140748430043459, |
|
"learning_rate": 1.5112540192926044e-05, |
|
"loss": 0.0018, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"grad_norm": 0.0061045498587191105, |
|
"learning_rate": 1.3504823151125404e-05, |
|
"loss": 0.0015, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"grad_norm": 0.01204584538936615, |
|
"learning_rate": 1.189710610932476e-05, |
|
"loss": 0.0019, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"grad_norm": 0.005698794033378363, |
|
"learning_rate": 1.0289389067524116e-05, |
|
"loss": 0.0015, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"grad_norm": 0.008717156946659088, |
|
"learning_rate": 8.681672025723474e-06, |
|
"loss": 0.0018, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"grad_norm": 0.03146166726946831, |
|
"learning_rate": 7.07395498392283e-06, |
|
"loss": 0.0305, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"grad_norm": 0.0084818284958601, |
|
"learning_rate": 5.466237942122187e-06, |
|
"loss": 0.0046, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"grad_norm": 0.026195811107754707, |
|
"learning_rate": 3.858520900321544e-06, |
|
"loss": 0.0018, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"grad_norm": 0.01475218404084444, |
|
"learning_rate": 2.2508038585209006e-06, |
|
"loss": 0.0019, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"grad_norm": 0.006672242656350136, |
|
"learning_rate": 6.430868167202573e-07, |
|
"loss": 0.0516, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"step": 1244, |
|
"total_flos": 7.703325099767808e+17, |
|
"train_loss": 0.11289434264701495, |
|
"train_runtime": 3405.5271, |
|
"train_samples_per_second": 2.919, |
|
"train_steps_per_second": 0.365 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1244, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 1000, |
|
"total_flos": 7.703325099767808e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|