|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9956108266276518, |
|
"eval_steps": 500, |
|
"global_step": 1023, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.029261155815654718, |
|
"grad_norm": 1.4015444467822036, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7996, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.058522311631309436, |
|
"grad_norm": 1.2657884873720076, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7383, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.08778346744696415, |
|
"grad_norm": 1.2853443307857617, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7065, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.11704462326261887, |
|
"grad_norm": 1.092609029300215, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6961, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.14630577907827358, |
|
"grad_norm": 1.1461752520371735, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6866, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.1755669348939283, |
|
"grad_norm": 0.9006655574575179, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6763, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.20482809070958302, |
|
"grad_norm": 0.5991449701861865, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6697, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.23408924652523774, |
|
"grad_norm": 0.42048613921733907, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6646, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.26335040234089246, |
|
"grad_norm": 0.430405040916768, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6618, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.29261155815654716, |
|
"grad_norm": 0.4594025425656112, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6542, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3218727139722019, |
|
"grad_norm": 0.3994331184639397, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6631, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.3511338697878566, |
|
"grad_norm": 0.37951451802607, |
|
"learning_rate": 5e-06, |
|
"loss": 0.668, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.38039502560351135, |
|
"grad_norm": 0.37938408152223546, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6518, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.40965618141916604, |
|
"grad_norm": 0.3682764375063777, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6525, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.4389173372348208, |
|
"grad_norm": 0.3506753294754763, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6479, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.4681784930504755, |
|
"grad_norm": 0.38739237025714635, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6498, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.49743964886613024, |
|
"grad_norm": 0.3641990199038121, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6549, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.5267008046817849, |
|
"grad_norm": 0.3387262148596558, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6545, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.5559619604974396, |
|
"grad_norm": 0.35583454417742927, |
|
"learning_rate": 5e-06, |
|
"loss": 0.643, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.5852231163130943, |
|
"grad_norm": 0.3439858964156482, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6443, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.6144842721287491, |
|
"grad_norm": 0.339501405550569, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6508, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.6437454279444038, |
|
"grad_norm": 0.354200530410764, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6388, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.6730065837600585, |
|
"grad_norm": 0.36478465091491635, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6387, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.7022677395757132, |
|
"grad_norm": 0.3350250502378, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6456, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.731528895391368, |
|
"grad_norm": 0.3434477503899696, |
|
"learning_rate": 5e-06, |
|
"loss": 0.644, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.7607900512070227, |
|
"grad_norm": 0.341252472448644, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6466, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.7900512070226774, |
|
"grad_norm": 0.37592505438246726, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6489, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.8193123628383321, |
|
"grad_norm": 0.3419586714412142, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6476, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.8485735186539868, |
|
"grad_norm": 0.33854473212092273, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6399, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.8778346744696416, |
|
"grad_norm": 0.3551824692503222, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6318, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.9070958302852963, |
|
"grad_norm": 0.35666494542849303, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6449, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.936356986100951, |
|
"grad_norm": 0.33636053152251066, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6369, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.9656181419166057, |
|
"grad_norm": 0.3803771266748499, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6401, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.9948792977322605, |
|
"grad_norm": 0.35250821963156515, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6361, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.9978054133138259, |
|
"eval_loss": 0.640618622303009, |
|
"eval_runtime": 345.7926, |
|
"eval_samples_per_second": 26.632, |
|
"eval_steps_per_second": 0.416, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.025237746891002, |
|
"grad_norm": 0.3721731893063918, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6512, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.054498902706657, |
|
"grad_norm": 0.3492849036972559, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6055, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.0837600585223117, |
|
"grad_norm": 0.4115188835261346, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6088, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.1130212143379663, |
|
"grad_norm": 0.3443680714255078, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6062, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.142282370153621, |
|
"grad_norm": 0.39688431726948026, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6115, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.1715435259692757, |
|
"grad_norm": 0.35729977553611475, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6129, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.2008046817849305, |
|
"grad_norm": 0.3587917316338556, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6127, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.2300658376005853, |
|
"grad_norm": 0.3978792090841637, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6008, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.2593269934162399, |
|
"grad_norm": 0.36180899559468804, |
|
"learning_rate": 5e-06, |
|
"loss": 0.611, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.2885881492318947, |
|
"grad_norm": 0.3442675552815694, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6036, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.3178493050475493, |
|
"grad_norm": 0.3717062017395999, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6074, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.347110460863204, |
|
"grad_norm": 0.3796535325867537, |
|
"learning_rate": 5e-06, |
|
"loss": 0.605, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.3763716166788589, |
|
"grad_norm": 0.3273244882099893, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6024, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.4056327724945135, |
|
"grad_norm": 0.3727174937863657, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6033, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.4348939283101683, |
|
"grad_norm": 0.34694786322227344, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6096, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.464155084125823, |
|
"grad_norm": 0.33919862390827415, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5979, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.4934162399414777, |
|
"grad_norm": 0.34508821585413224, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6059, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.5226773957571325, |
|
"grad_norm": 0.38883434839268116, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6083, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.5519385515727873, |
|
"grad_norm": 0.35107933222108867, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6069, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.5811997073884418, |
|
"grad_norm": 0.3691057135251824, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6135, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.6104608632040964, |
|
"grad_norm": 0.35119977655218887, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6045, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.6397220190197512, |
|
"grad_norm": 0.3513289449211162, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6062, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.668983174835406, |
|
"grad_norm": 0.3249818476095366, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5968, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.6982443306510606, |
|
"grad_norm": 0.3517967607582011, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6062, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.7275054864667154, |
|
"grad_norm": 0.4030178347069804, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6064, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.7567666422823702, |
|
"grad_norm": 0.34207358553045886, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6033, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.7860277980980248, |
|
"grad_norm": 0.3457052497930087, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6058, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.8152889539136796, |
|
"grad_norm": 0.32108048224719427, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6061, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.8445501097293344, |
|
"grad_norm": 0.3567570067525494, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6052, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.873811265544989, |
|
"grad_norm": 0.3281186537149481, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6001, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.9030724213606436, |
|
"grad_norm": 0.36018189397498207, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6058, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.9323335771762986, |
|
"grad_norm": 0.3502425747015856, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6063, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.9615947329919532, |
|
"grad_norm": 0.33148897222711265, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5952, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.9908558888076078, |
|
"grad_norm": 0.3580118082608882, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6096, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.9967081199707388, |
|
"eval_loss": 0.6332002878189087, |
|
"eval_runtime": 346.5687, |
|
"eval_samples_per_second": 26.572, |
|
"eval_steps_per_second": 0.416, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 2.0212143379663496, |
|
"grad_norm": 0.39037400649762577, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6196, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.050475493782004, |
|
"grad_norm": 0.4434429381232548, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5673, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.0797366495976592, |
|
"grad_norm": 0.3497622377865716, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5693, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.108997805413314, |
|
"grad_norm": 0.3773856223516478, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5704, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.1382589612289684, |
|
"grad_norm": 0.3522556629706183, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5722, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.1675201170446234, |
|
"grad_norm": 0.3506219317591605, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5775, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.196781272860278, |
|
"grad_norm": 0.3147645395001558, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5761, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.2260424286759326, |
|
"grad_norm": 0.371033572143663, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5736, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.255303584491587, |
|
"grad_norm": 0.33096335665510035, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5724, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.284564740307242, |
|
"grad_norm": 0.3367058473599436, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5715, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.313825896122897, |
|
"grad_norm": 0.326433950142472, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5698, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.3430870519385514, |
|
"grad_norm": 0.3603804154225191, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5776, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.3723482077542064, |
|
"grad_norm": 0.3388463555889941, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5728, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.401609363569861, |
|
"grad_norm": 0.33023802442966366, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5791, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.4308705193855156, |
|
"grad_norm": 0.31991052482202714, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5713, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.4601316752011706, |
|
"grad_norm": 0.35182055756947245, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5718, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.489392831016825, |
|
"grad_norm": 0.3436348657090354, |
|
"learning_rate": 5e-06, |
|
"loss": 0.574, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.5186539868324798, |
|
"grad_norm": 0.3558300443060959, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5671, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.547915142648135, |
|
"grad_norm": 0.39457846380123907, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5695, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.5771762984637894, |
|
"grad_norm": 0.3540156892398613, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5737, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.606437454279444, |
|
"grad_norm": 0.3611060327105372, |
|
"learning_rate": 5e-06, |
|
"loss": 0.573, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.6356986100950985, |
|
"grad_norm": 0.33749542282346184, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5776, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.6649597659107536, |
|
"grad_norm": 0.36124194708101637, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5778, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.694220921726408, |
|
"grad_norm": 0.35710662173949104, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5721, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.723482077542063, |
|
"grad_norm": 0.3566062701794353, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5758, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.7527432333577178, |
|
"grad_norm": 0.34964975557486105, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5747, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 2.7820043891733723, |
|
"grad_norm": 0.34002257310230194, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5735, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.811265544989027, |
|
"grad_norm": 0.37468803583644683, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5773, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 2.840526700804682, |
|
"grad_norm": 0.34310179408919084, |
|
"learning_rate": 5e-06, |
|
"loss": 0.573, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 2.8697878566203365, |
|
"grad_norm": 0.38021125890295365, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5729, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 2.899049012435991, |
|
"grad_norm": 0.3239775742376659, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5658, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 2.928310168251646, |
|
"grad_norm": 0.3262461836425964, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5736, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.9575713240673007, |
|
"grad_norm": 0.34333695957920457, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5654, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 2.9868324798829553, |
|
"grad_norm": 0.3358081690176178, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5771, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 2.9956108266276518, |
|
"eval_loss": 0.6331179141998291, |
|
"eval_runtime": 344.5727, |
|
"eval_samples_per_second": 26.726, |
|
"eval_steps_per_second": 0.418, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 2.9956108266276518, |
|
"step": 1023, |
|
"total_flos": 2144987064041472.0, |
|
"train_loss": 0.6140813032786051, |
|
"train_runtime": 55371.2357, |
|
"train_samples_per_second": 9.479, |
|
"train_steps_per_second": 0.018 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1023, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2144987064041472.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|