|
{ |
|
"best_metric": 0.9225352112676056, |
|
"best_model_checkpoint": "/scratch/camembertv2/runs/results/flue-PAWS-X/camembertv2-base-bf16-p2-17000/max_seq_length-148-gradient_accumulation_steps-2-precision-fp32-learning_rate-3e-05-epochs-6-lr_scheduler-linear-warmup_steps-0/SEED-1/checkpoint-15437", |
|
"epoch": 5.999028340080971, |
|
"eval_steps": 500, |
|
"global_step": 18522, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.032388663967611336, |
|
"grad_norm": 1.3617459535598755, |
|
"learning_rate": 2.983803045027535e-05, |
|
"loss": 0.6917, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.06477732793522267, |
|
"grad_norm": 1.1722190380096436, |
|
"learning_rate": 2.96760609005507e-05, |
|
"loss": 0.6863, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.09716599190283401, |
|
"grad_norm": 5.13236665725708, |
|
"learning_rate": 2.9514091350826045e-05, |
|
"loss": 0.5495, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.12955465587044535, |
|
"grad_norm": 11.99015998840332, |
|
"learning_rate": 2.9352121801101394e-05, |
|
"loss": 0.4849, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.16194331983805668, |
|
"grad_norm": 7.261778831481934, |
|
"learning_rate": 2.9190152251376742e-05, |
|
"loss": 0.4368, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.19433198380566802, |
|
"grad_norm": 8.82419490814209, |
|
"learning_rate": 2.9028182701652093e-05, |
|
"loss": 0.4046, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.22672064777327935, |
|
"grad_norm": 7.460156440734863, |
|
"learning_rate": 2.8866213151927438e-05, |
|
"loss": 0.3746, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.2591093117408907, |
|
"grad_norm": 17.467742919921875, |
|
"learning_rate": 2.8704243602202786e-05, |
|
"loss": 0.3493, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.291497975708502, |
|
"grad_norm": 16.750350952148438, |
|
"learning_rate": 2.8542274052478135e-05, |
|
"loss": 0.3583, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.32388663967611336, |
|
"grad_norm": 8.11487102508545, |
|
"learning_rate": 2.8380304502753486e-05, |
|
"loss": 0.3469, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.3562753036437247, |
|
"grad_norm": 13.264874458312988, |
|
"learning_rate": 2.821833495302883e-05, |
|
"loss": 0.3359, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.38866396761133604, |
|
"grad_norm": 38.216304779052734, |
|
"learning_rate": 2.805636540330418e-05, |
|
"loss": 0.32, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.42105263157894735, |
|
"grad_norm": 3.897590398788452, |
|
"learning_rate": 2.7894395853579527e-05, |
|
"loss": 0.3134, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.4534412955465587, |
|
"grad_norm": 19.601438522338867, |
|
"learning_rate": 2.773242630385488e-05, |
|
"loss": 0.3247, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.48582995951417, |
|
"grad_norm": 6.229040145874023, |
|
"learning_rate": 2.7570456754130224e-05, |
|
"loss": 0.3283, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.5182186234817814, |
|
"grad_norm": 8.433351516723633, |
|
"learning_rate": 2.7408487204405572e-05, |
|
"loss": 0.2851, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.5506072874493927, |
|
"grad_norm": 29.696481704711914, |
|
"learning_rate": 2.724651765468092e-05, |
|
"loss": 0.3067, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.582995951417004, |
|
"grad_norm": 9.64140796661377, |
|
"learning_rate": 2.7084548104956272e-05, |
|
"loss": 0.3072, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.6153846153846154, |
|
"grad_norm": 2.7816696166992188, |
|
"learning_rate": 2.6922578555231617e-05, |
|
"loss": 0.2838, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.6477732793522267, |
|
"grad_norm": 11.457731246948242, |
|
"learning_rate": 2.6760609005506965e-05, |
|
"loss": 0.2692, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.680161943319838, |
|
"grad_norm": 10.894031524658203, |
|
"learning_rate": 2.6598639455782313e-05, |
|
"loss": 0.2969, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.7125506072874493, |
|
"grad_norm": 8.756068229675293, |
|
"learning_rate": 2.6436669906057665e-05, |
|
"loss": 0.2982, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.7449392712550608, |
|
"grad_norm": 6.366727352142334, |
|
"learning_rate": 2.627470035633301e-05, |
|
"loss": 0.2743, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.7773279352226721, |
|
"grad_norm": 3.6022021770477295, |
|
"learning_rate": 2.6112730806608358e-05, |
|
"loss": 0.2497, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.8097165991902834, |
|
"grad_norm": 9.339791297912598, |
|
"learning_rate": 2.5950761256883706e-05, |
|
"loss": 0.2617, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.8421052631578947, |
|
"grad_norm": 0.46193593740463257, |
|
"learning_rate": 2.5788791707159058e-05, |
|
"loss": 0.2591, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.8744939271255061, |
|
"grad_norm": 8.35823917388916, |
|
"learning_rate": 2.5626822157434402e-05, |
|
"loss": 0.2619, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.9068825910931174, |
|
"grad_norm": 11.017343521118164, |
|
"learning_rate": 2.546485260770975e-05, |
|
"loss": 0.2571, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.9392712550607287, |
|
"grad_norm": 12.998709678649902, |
|
"learning_rate": 2.53028830579851e-05, |
|
"loss": 0.2922, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.97165991902834, |
|
"grad_norm": 10.282382011413574, |
|
"learning_rate": 2.514091350826045e-05, |
|
"loss": 0.2421, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.9998380566801619, |
|
"eval_accuracy": 0.8958752515090543, |
|
"eval_loss": 0.3331478238105774, |
|
"eval_runtime": 4.4897, |
|
"eval_samples_per_second": 442.795, |
|
"eval_steps_per_second": 55.461, |
|
"step": 3087 |
|
}, |
|
{ |
|
"epoch": 1.0040485829959513, |
|
"grad_norm": 1.8964722156524658, |
|
"learning_rate": 2.4978943958535795e-05, |
|
"loss": 0.2629, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.0364372469635628, |
|
"grad_norm": 20.548351287841797, |
|
"learning_rate": 2.4816974408811143e-05, |
|
"loss": 0.251, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.0688259109311742, |
|
"grad_norm": 9.91510009765625, |
|
"learning_rate": 2.465500485908649e-05, |
|
"loss": 0.205, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.1012145748987854, |
|
"grad_norm": 15.848530769348145, |
|
"learning_rate": 2.4493035309361843e-05, |
|
"loss": 0.1904, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.1336032388663968, |
|
"grad_norm": 2.94976544380188, |
|
"learning_rate": 2.4331065759637188e-05, |
|
"loss": 0.2101, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.165991902834008, |
|
"grad_norm": 2.3269777297973633, |
|
"learning_rate": 2.4169096209912536e-05, |
|
"loss": 0.1924, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.1983805668016194, |
|
"grad_norm": 9.416550636291504, |
|
"learning_rate": 2.4007126660187884e-05, |
|
"loss": 0.2243, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.2307692307692308, |
|
"grad_norm": 18.740257263183594, |
|
"learning_rate": 2.3845157110463236e-05, |
|
"loss": 0.204, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.263157894736842, |
|
"grad_norm": 11.511466026306152, |
|
"learning_rate": 2.368318756073858e-05, |
|
"loss": 0.2211, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.2955465587044535, |
|
"grad_norm": 48.617271423339844, |
|
"learning_rate": 2.352121801101393e-05, |
|
"loss": 0.2006, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.3279352226720649, |
|
"grad_norm": 1.9806264638900757, |
|
"learning_rate": 2.3359248461289277e-05, |
|
"loss": 0.2331, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.360323886639676, |
|
"grad_norm": 11.272978782653809, |
|
"learning_rate": 2.319727891156463e-05, |
|
"loss": 0.228, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.3927125506072875, |
|
"grad_norm": 13.344253540039062, |
|
"learning_rate": 2.3035309361839974e-05, |
|
"loss": 0.2166, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.425101214574899, |
|
"grad_norm": 6.339032173156738, |
|
"learning_rate": 2.2873339812115322e-05, |
|
"loss": 0.1978, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.45748987854251, |
|
"grad_norm": 12.384753227233887, |
|
"learning_rate": 2.271137026239067e-05, |
|
"loss": 0.2137, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.4898785425101215, |
|
"grad_norm": 2.5586955547332764, |
|
"learning_rate": 2.254940071266602e-05, |
|
"loss": 0.1965, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.522267206477733, |
|
"grad_norm": 30.300275802612305, |
|
"learning_rate": 2.2387431162941366e-05, |
|
"loss": 0.2013, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.5546558704453441, |
|
"grad_norm": 0.9059699773788452, |
|
"learning_rate": 2.2225461613216715e-05, |
|
"loss": 0.215, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.5870445344129553, |
|
"grad_norm": 24.580528259277344, |
|
"learning_rate": 2.2063492063492063e-05, |
|
"loss": 0.2025, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.6194331983805668, |
|
"grad_norm": 46.1649169921875, |
|
"learning_rate": 2.1901522513767414e-05, |
|
"loss": 0.1775, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.6518218623481782, |
|
"grad_norm": 9.169315338134766, |
|
"learning_rate": 2.173955296404276e-05, |
|
"loss": 0.2034, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.6842105263157894, |
|
"grad_norm": 22.808425903320312, |
|
"learning_rate": 2.1577583414318108e-05, |
|
"loss": 0.2338, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.7165991902834008, |
|
"grad_norm": 0.972823977470398, |
|
"learning_rate": 2.1415613864593456e-05, |
|
"loss": 0.2079, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.7489878542510122, |
|
"grad_norm": 12.798443794250488, |
|
"learning_rate": 2.1253644314868807e-05, |
|
"loss": 0.2155, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.7813765182186234, |
|
"grad_norm": 14.221925735473633, |
|
"learning_rate": 2.1091674765144152e-05, |
|
"loss": 0.2016, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.8137651821862348, |
|
"grad_norm": 16.415285110473633, |
|
"learning_rate": 2.09297052154195e-05, |
|
"loss": 0.187, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.8461538461538463, |
|
"grad_norm": 3.618945598602295, |
|
"learning_rate": 2.0767735665694852e-05, |
|
"loss": 0.2213, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.8785425101214575, |
|
"grad_norm": 39.05452346801758, |
|
"learning_rate": 2.06057661159702e-05, |
|
"loss": 0.1785, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 1.9109311740890689, |
|
"grad_norm": 19.066164016723633, |
|
"learning_rate": 2.0443796566245545e-05, |
|
"loss": 0.2173, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 1.9433198380566803, |
|
"grad_norm": 48.79111862182617, |
|
"learning_rate": 2.0281827016520893e-05, |
|
"loss": 0.1846, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.9757085020242915, |
|
"grad_norm": 0.30500343441963196, |
|
"learning_rate": 2.0119857466796245e-05, |
|
"loss": 0.2054, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9094567404426559, |
|
"eval_loss": 0.30133694410324097, |
|
"eval_runtime": 4.0681, |
|
"eval_samples_per_second": 488.676, |
|
"eval_steps_per_second": 61.207, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 2.0080971659919027, |
|
"grad_norm": 5.170982837677002, |
|
"learning_rate": 1.9957887917071593e-05, |
|
"loss": 0.1975, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.0404858299595143, |
|
"grad_norm": 17.76828956604004, |
|
"learning_rate": 1.9795918367346938e-05, |
|
"loss": 0.1424, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.0728744939271255, |
|
"grad_norm": 0.5401751399040222, |
|
"learning_rate": 1.9633948817622286e-05, |
|
"loss": 0.1564, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.1052631578947367, |
|
"grad_norm": 0.1927350014448166, |
|
"learning_rate": 1.9471979267897638e-05, |
|
"loss": 0.1559, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.1376518218623484, |
|
"grad_norm": 6.872354984283447, |
|
"learning_rate": 1.9310009718172986e-05, |
|
"loss": 0.1577, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.1700404858299596, |
|
"grad_norm": 0.9195345640182495, |
|
"learning_rate": 1.914804016844833e-05, |
|
"loss": 0.1655, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 2.2024291497975708, |
|
"grad_norm": 0.5888263583183289, |
|
"learning_rate": 1.898607061872368e-05, |
|
"loss": 0.1404, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.234817813765182, |
|
"grad_norm": 20.80263900756836, |
|
"learning_rate": 1.882410106899903e-05, |
|
"loss": 0.1828, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.2672064777327936, |
|
"grad_norm": 11.595118522644043, |
|
"learning_rate": 1.866213151927438e-05, |
|
"loss": 0.1768, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.299595141700405, |
|
"grad_norm": 4.5267333984375, |
|
"learning_rate": 1.8500161969549723e-05, |
|
"loss": 0.1546, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.331983805668016, |
|
"grad_norm": 0.5598276257514954, |
|
"learning_rate": 1.833819241982507e-05, |
|
"loss": 0.1605, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.3643724696356276, |
|
"grad_norm": 1.0213172435760498, |
|
"learning_rate": 1.8176222870100423e-05, |
|
"loss": 0.1715, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.396761133603239, |
|
"grad_norm": 4.324398040771484, |
|
"learning_rate": 1.801425332037577e-05, |
|
"loss": 0.1597, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.42914979757085, |
|
"grad_norm": 0.6324844360351562, |
|
"learning_rate": 1.7852283770651116e-05, |
|
"loss": 0.1659, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.4615384615384617, |
|
"grad_norm": 0.5160787105560303, |
|
"learning_rate": 1.7690314220926464e-05, |
|
"loss": 0.1799, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.493927125506073, |
|
"grad_norm": 0.26999789476394653, |
|
"learning_rate": 1.7528344671201816e-05, |
|
"loss": 0.1347, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.526315789473684, |
|
"grad_norm": 4.179189682006836, |
|
"learning_rate": 1.7366375121477164e-05, |
|
"loss": 0.1737, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.5587044534412957, |
|
"grad_norm": 14.371667861938477, |
|
"learning_rate": 1.720440557175251e-05, |
|
"loss": 0.1825, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.591093117408907, |
|
"grad_norm": 0.13881312310695648, |
|
"learning_rate": 1.7042436022027857e-05, |
|
"loss": 0.1757, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.623481781376518, |
|
"grad_norm": 10.17100715637207, |
|
"learning_rate": 1.688046647230321e-05, |
|
"loss": 0.1501, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 2.6558704453441297, |
|
"grad_norm": 43.74755859375, |
|
"learning_rate": 1.6718496922578557e-05, |
|
"loss": 0.1845, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 2.688259109311741, |
|
"grad_norm": 17.117637634277344, |
|
"learning_rate": 1.6556527372853902e-05, |
|
"loss": 0.1745, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 2.720647773279352, |
|
"grad_norm": 31.1845645904541, |
|
"learning_rate": 1.639455782312925e-05, |
|
"loss": 0.1664, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 2.753036437246964, |
|
"grad_norm": 9.80735969543457, |
|
"learning_rate": 1.6232588273404602e-05, |
|
"loss": 0.1632, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 2.785425101214575, |
|
"grad_norm": 9.704376220703125, |
|
"learning_rate": 1.607061872367995e-05, |
|
"loss": 0.1443, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 2.817813765182186, |
|
"grad_norm": 4.445658206939697, |
|
"learning_rate": 1.5908649173955295e-05, |
|
"loss": 0.1775, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 2.850202429149798, |
|
"grad_norm": 1.7219343185424805, |
|
"learning_rate": 1.5746679624230643e-05, |
|
"loss": 0.1581, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 2.882591093117409, |
|
"grad_norm": 8.307001113891602, |
|
"learning_rate": 1.5584710074505995e-05, |
|
"loss": 0.1536, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 2.91497975708502, |
|
"grad_norm": 9.363420486450195, |
|
"learning_rate": 1.5422740524781343e-05, |
|
"loss": 0.1862, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 2.9473684210526314, |
|
"grad_norm": 20.355863571166992, |
|
"learning_rate": 1.5260770975056688e-05, |
|
"loss": 0.1679, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 2.979757085020243, |
|
"grad_norm": 16.128374099731445, |
|
"learning_rate": 1.5098801425332037e-05, |
|
"loss": 0.1513, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 2.999838056680162, |
|
"eval_accuracy": 0.9054325955734407, |
|
"eval_loss": 0.40321749448776245, |
|
"eval_runtime": 4.0611, |
|
"eval_samples_per_second": 489.52, |
|
"eval_steps_per_second": 61.313, |
|
"step": 9262 |
|
}, |
|
{ |
|
"epoch": 3.0121457489878543, |
|
"grad_norm": 2.3643200397491455, |
|
"learning_rate": 1.4936831875607386e-05, |
|
"loss": 0.1294, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 3.0445344129554655, |
|
"grad_norm": 0.2633157968521118, |
|
"learning_rate": 1.4774862325882734e-05, |
|
"loss": 0.1198, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 3.076923076923077, |
|
"grad_norm": 116.33301544189453, |
|
"learning_rate": 1.4612892776158082e-05, |
|
"loss": 0.1199, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 3.1093117408906883, |
|
"grad_norm": 110.17882537841797, |
|
"learning_rate": 1.445092322643343e-05, |
|
"loss": 0.1101, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 3.1417004048582995, |
|
"grad_norm": 66.73577117919922, |
|
"learning_rate": 1.4288953676708779e-05, |
|
"loss": 0.1634, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 3.174089068825911, |
|
"grad_norm": 0.10323189944028854, |
|
"learning_rate": 1.4126984126984127e-05, |
|
"loss": 0.1326, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 3.2064777327935223, |
|
"grad_norm": 0.39017683267593384, |
|
"learning_rate": 1.3965014577259475e-05, |
|
"loss": 0.1305, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 3.2388663967611335, |
|
"grad_norm": 13.028061866760254, |
|
"learning_rate": 1.3803045027534823e-05, |
|
"loss": 0.1254, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 3.2712550607287447, |
|
"grad_norm": 3.545095205307007, |
|
"learning_rate": 1.3641075477810171e-05, |
|
"loss": 0.1227, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 3.3036437246963564, |
|
"grad_norm": 0.2068515121936798, |
|
"learning_rate": 1.347910592808552e-05, |
|
"loss": 0.1107, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 3.3360323886639676, |
|
"grad_norm": 0.7598180174827576, |
|
"learning_rate": 1.3317136378360868e-05, |
|
"loss": 0.1431, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 3.3684210526315788, |
|
"grad_norm": 0.3546907603740692, |
|
"learning_rate": 1.3155166828636216e-05, |
|
"loss": 0.128, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 3.4008097165991904, |
|
"grad_norm": 1.108842372894287, |
|
"learning_rate": 1.2993197278911564e-05, |
|
"loss": 0.1247, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 3.4331983805668016, |
|
"grad_norm": 31.11785316467285, |
|
"learning_rate": 1.2831227729186914e-05, |
|
"loss": 0.1293, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 3.465587044534413, |
|
"grad_norm": 0.4772971272468567, |
|
"learning_rate": 1.266925817946226e-05, |
|
"loss": 0.136, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 3.4979757085020244, |
|
"grad_norm": 3.384209632873535, |
|
"learning_rate": 1.250728862973761e-05, |
|
"loss": 0.1197, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 3.5303643724696356, |
|
"grad_norm": 13.939515113830566, |
|
"learning_rate": 1.2345319080012957e-05, |
|
"loss": 0.1414, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 3.562753036437247, |
|
"grad_norm": 0.4108864963054657, |
|
"learning_rate": 1.2183349530288307e-05, |
|
"loss": 0.107, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 3.5951417004048585, |
|
"grad_norm": 12.29635238647461, |
|
"learning_rate": 1.2021379980563655e-05, |
|
"loss": 0.1357, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 3.6275303643724697, |
|
"grad_norm": 49.79674530029297, |
|
"learning_rate": 1.1859410430839003e-05, |
|
"loss": 0.1333, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 3.659919028340081, |
|
"grad_norm": 0.1119026318192482, |
|
"learning_rate": 1.1697440881114352e-05, |
|
"loss": 0.1311, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 3.6923076923076925, |
|
"grad_norm": 0.2087603062391281, |
|
"learning_rate": 1.15354713313897e-05, |
|
"loss": 0.145, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 3.7246963562753037, |
|
"grad_norm": 0.3955753743648529, |
|
"learning_rate": 1.1373501781665048e-05, |
|
"loss": 0.1225, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 3.757085020242915, |
|
"grad_norm": 0.05121416971087456, |
|
"learning_rate": 1.1211532231940396e-05, |
|
"loss": 0.0975, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 3.7894736842105265, |
|
"grad_norm": 21.439380645751953, |
|
"learning_rate": 1.1049562682215744e-05, |
|
"loss": 0.1398, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 3.8218623481781377, |
|
"grad_norm": 1.1940436363220215, |
|
"learning_rate": 1.0887593132491093e-05, |
|
"loss": 0.1231, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 3.854251012145749, |
|
"grad_norm": 12.916104316711426, |
|
"learning_rate": 1.072562358276644e-05, |
|
"loss": 0.1392, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 3.8866396761133606, |
|
"grad_norm": 0.13800889253616333, |
|
"learning_rate": 1.0563654033041789e-05, |
|
"loss": 0.142, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 3.919028340080972, |
|
"grad_norm": 0.10773531347513199, |
|
"learning_rate": 1.0401684483317137e-05, |
|
"loss": 0.1444, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 3.951417004048583, |
|
"grad_norm": 4.032077312469482, |
|
"learning_rate": 1.0239714933592485e-05, |
|
"loss": 0.127, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 3.983805668016194, |
|
"grad_norm": 0.21410343050956726, |
|
"learning_rate": 1.0077745383867834e-05, |
|
"loss": 0.1475, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.9079476861167002, |
|
"eval_loss": 0.35845422744750977, |
|
"eval_runtime": 4.0822, |
|
"eval_samples_per_second": 486.989, |
|
"eval_steps_per_second": 60.996, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 4.016194331983805, |
|
"grad_norm": 108.23912048339844, |
|
"learning_rate": 9.915775834143182e-06, |
|
"loss": 0.1174, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 4.048582995951417, |
|
"grad_norm": 0.086638443171978, |
|
"learning_rate": 9.75380628441853e-06, |
|
"loss": 0.1094, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 4.080971659919029, |
|
"grad_norm": 31.371986389160156, |
|
"learning_rate": 9.591836734693878e-06, |
|
"loss": 0.1086, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 4.113360323886639, |
|
"grad_norm": 0.3369753658771515, |
|
"learning_rate": 9.429867184969226e-06, |
|
"loss": 0.0997, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 4.145748987854251, |
|
"grad_norm": 4.937185287475586, |
|
"learning_rate": 9.267897635244575e-06, |
|
"loss": 0.1187, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 4.178137651821863, |
|
"grad_norm": 0.08810489624738693, |
|
"learning_rate": 9.105928085519923e-06, |
|
"loss": 0.0714, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 4.2105263157894735, |
|
"grad_norm": 2.339200496673584, |
|
"learning_rate": 8.943958535795271e-06, |
|
"loss": 0.1018, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 4.242914979757085, |
|
"grad_norm": 0.8731828331947327, |
|
"learning_rate": 8.78198898607062e-06, |
|
"loss": 0.0783, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 4.275303643724697, |
|
"grad_norm": 0.03349796682596207, |
|
"learning_rate": 8.620019436345967e-06, |
|
"loss": 0.076, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 4.3076923076923075, |
|
"grad_norm": 0.3485426902770996, |
|
"learning_rate": 8.458049886621316e-06, |
|
"loss": 0.0891, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 4.340080971659919, |
|
"grad_norm": 0.18010343611240387, |
|
"learning_rate": 8.296080336896664e-06, |
|
"loss": 0.1094, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 4.372469635627531, |
|
"grad_norm": 0.1857542097568512, |
|
"learning_rate": 8.134110787172012e-06, |
|
"loss": 0.0976, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 4.4048582995951415, |
|
"grad_norm": 0.5226219296455383, |
|
"learning_rate": 7.97214123744736e-06, |
|
"loss": 0.1342, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 4.437246963562753, |
|
"grad_norm": 0.23230472207069397, |
|
"learning_rate": 7.810171687722709e-06, |
|
"loss": 0.0997, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 4.469635627530364, |
|
"grad_norm": 0.15814034640789032, |
|
"learning_rate": 7.648202137998057e-06, |
|
"loss": 0.0924, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 4.502024291497976, |
|
"grad_norm": 6.45848274230957, |
|
"learning_rate": 7.486232588273405e-06, |
|
"loss": 0.0867, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 4.534412955465587, |
|
"grad_norm": 0.31769734621047974, |
|
"learning_rate": 7.324263038548753e-06, |
|
"loss": 0.1186, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 4.566801619433198, |
|
"grad_norm": 5.155035495758057, |
|
"learning_rate": 7.162293488824101e-06, |
|
"loss": 0.0912, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 4.59919028340081, |
|
"grad_norm": 43.147640228271484, |
|
"learning_rate": 7.0003239390994495e-06, |
|
"loss": 0.0967, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 4.631578947368421, |
|
"grad_norm": 0.5134268999099731, |
|
"learning_rate": 6.838354389374798e-06, |
|
"loss": 0.0945, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 4.663967611336032, |
|
"grad_norm": 0.2741609811782837, |
|
"learning_rate": 6.676384839650146e-06, |
|
"loss": 0.1012, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 4.696356275303644, |
|
"grad_norm": 0.4370046854019165, |
|
"learning_rate": 6.514415289925494e-06, |
|
"loss": 0.123, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 4.728744939271255, |
|
"grad_norm": 4.210660457611084, |
|
"learning_rate": 6.352445740200842e-06, |
|
"loss": 0.1047, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 4.761133603238866, |
|
"grad_norm": 6.13052225112915, |
|
"learning_rate": 6.190476190476191e-06, |
|
"loss": 0.0966, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 4.793522267206478, |
|
"grad_norm": 0.41901007294654846, |
|
"learning_rate": 6.028506640751539e-06, |
|
"loss": 0.1014, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 4.825910931174089, |
|
"grad_norm": 0.28807559609413147, |
|
"learning_rate": 5.866537091026887e-06, |
|
"loss": 0.1042, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 4.8582995951417, |
|
"grad_norm": 28.3045654296875, |
|
"learning_rate": 5.704567541302235e-06, |
|
"loss": 0.1027, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 4.890688259109312, |
|
"grad_norm": 148.7666473388672, |
|
"learning_rate": 5.542597991577583e-06, |
|
"loss": 0.0804, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 4.923076923076923, |
|
"grad_norm": 1.260237216949463, |
|
"learning_rate": 5.380628441852932e-06, |
|
"loss": 0.1201, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 4.955465587044534, |
|
"grad_norm": 0.15356577932834625, |
|
"learning_rate": 5.21865889212828e-06, |
|
"loss": 0.1002, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 4.987854251012146, |
|
"grad_norm": 0.7852919697761536, |
|
"learning_rate": 5.056689342403628e-06, |
|
"loss": 0.1202, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 4.9998380566801615, |
|
"eval_accuracy": 0.9225352112676056, |
|
"eval_loss": 0.3642682433128357, |
|
"eval_runtime": 4.0522, |
|
"eval_samples_per_second": 490.597, |
|
"eval_steps_per_second": 61.448, |
|
"step": 15437 |
|
}, |
|
{ |
|
"epoch": 5.020242914979757, |
|
"grad_norm": 0.10688459873199463, |
|
"learning_rate": 4.894719792678976e-06, |
|
"loss": 0.1028, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 5.052631578947368, |
|
"grad_norm": 20.821151733398438, |
|
"learning_rate": 4.7327502429543244e-06, |
|
"loss": 0.0861, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 5.08502024291498, |
|
"grad_norm": 0.06313851475715637, |
|
"learning_rate": 4.5707806932296735e-06, |
|
"loss": 0.0749, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 5.117408906882591, |
|
"grad_norm": 5.4235734939575195, |
|
"learning_rate": 4.408811143505022e-06, |
|
"loss": 0.0756, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 5.149797570850202, |
|
"grad_norm": 0.07610571384429932, |
|
"learning_rate": 4.24684159378037e-06, |
|
"loss": 0.0821, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 5.182186234817814, |
|
"grad_norm": 189.49761962890625, |
|
"learning_rate": 4.084872044055718e-06, |
|
"loss": 0.0694, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 5.2145748987854255, |
|
"grad_norm": 0.05370008572936058, |
|
"learning_rate": 3.922902494331066e-06, |
|
"loss": 0.0767, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 5.246963562753036, |
|
"grad_norm": 0.05699535831809044, |
|
"learning_rate": 3.7609329446064145e-06, |
|
"loss": 0.0777, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 5.279352226720648, |
|
"grad_norm": 0.04399504140019417, |
|
"learning_rate": 3.5989633948817623e-06, |
|
"loss": 0.0753, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 5.3117408906882595, |
|
"grad_norm": 0.09438109397888184, |
|
"learning_rate": 3.4369938451571105e-06, |
|
"loss": 0.0759, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 5.34412955465587, |
|
"grad_norm": 0.10693158209323883, |
|
"learning_rate": 3.2750242954324587e-06, |
|
"loss": 0.0627, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 5.376518218623482, |
|
"grad_norm": 0.16652892529964447, |
|
"learning_rate": 3.113054745707807e-06, |
|
"loss": 0.0738, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 5.4089068825910935, |
|
"grad_norm": 17.245641708374023, |
|
"learning_rate": 2.951085195983155e-06, |
|
"loss": 0.0806, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 5.441295546558704, |
|
"grad_norm": 0.10547757893800735, |
|
"learning_rate": 2.7891156462585034e-06, |
|
"loss": 0.0671, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 5.473684210526316, |
|
"grad_norm": 1.494895100593567, |
|
"learning_rate": 2.6271460965338516e-06, |
|
"loss": 0.0733, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 5.506072874493928, |
|
"grad_norm": 0.5026708841323853, |
|
"learning_rate": 2.4651765468091998e-06, |
|
"loss": 0.0579, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 5.538461538461538, |
|
"grad_norm": 0.033295173197984695, |
|
"learning_rate": 2.303206997084548e-06, |
|
"loss": 0.0665, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 5.57085020242915, |
|
"grad_norm": 38.09762954711914, |
|
"learning_rate": 2.141237447359896e-06, |
|
"loss": 0.076, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 5.603238866396762, |
|
"grad_norm": 6.565536022186279, |
|
"learning_rate": 1.9792678976352444e-06, |
|
"loss": 0.0695, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 5.635627530364372, |
|
"grad_norm": 0.03646261617541313, |
|
"learning_rate": 1.817298347910593e-06, |
|
"loss": 0.0941, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 5.668016194331984, |
|
"grad_norm": 2.2503933906555176, |
|
"learning_rate": 1.6553287981859412e-06, |
|
"loss": 0.0992, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 5.700404858299595, |
|
"grad_norm": 0.26883581280708313, |
|
"learning_rate": 1.4933592484612894e-06, |
|
"loss": 0.0849, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 5.732793522267206, |
|
"grad_norm": 0.1083364486694336, |
|
"learning_rate": 1.3313896987366376e-06, |
|
"loss": 0.0798, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 5.765182186234818, |
|
"grad_norm": 1.5160281658172607, |
|
"learning_rate": 1.1694201490119858e-06, |
|
"loss": 0.0535, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 5.797570850202429, |
|
"grad_norm": 0.054747115820646286, |
|
"learning_rate": 1.007450599287334e-06, |
|
"loss": 0.0828, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 5.82995951417004, |
|
"grad_norm": 0.062284424901008606, |
|
"learning_rate": 8.454810495626823e-07, |
|
"loss": 0.0816, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 5.862348178137652, |
|
"grad_norm": 0.07957690209150314, |
|
"learning_rate": 6.835114998380305e-07, |
|
"loss": 0.0873, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 5.894736842105263, |
|
"grad_norm": 18.025510787963867, |
|
"learning_rate": 5.215419501133787e-07, |
|
"loss": 0.0786, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 5.9271255060728745, |
|
"grad_norm": 3.286792755126953, |
|
"learning_rate": 3.5957240038872693e-07, |
|
"loss": 0.0761, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 5.959514170040486, |
|
"grad_norm": 0.04595565423369408, |
|
"learning_rate": 1.9760285066407517e-07, |
|
"loss": 0.0551, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 5.991902834008097, |
|
"grad_norm": 0.07384829223155975, |
|
"learning_rate": 3.563330093942339e-08, |
|
"loss": 0.092, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 5.999028340080971, |
|
"eval_accuracy": 0.9164989939637826, |
|
"eval_loss": 0.4001348912715912, |
|
"eval_runtime": 4.0447, |
|
"eval_samples_per_second": 491.512, |
|
"eval_steps_per_second": 61.563, |
|
"step": 18522 |
|
}, |
|
{ |
|
"epoch": 5.999028340080971, |
|
"step": 18522, |
|
"total_flos": 1.33712370278538e+16, |
|
"train_loss": 0.17084743083006051, |
|
"train_runtime": 2225.7449, |
|
"train_samples_per_second": 133.166, |
|
"train_steps_per_second": 8.322 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 18522, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 6, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.33712370278538e+16, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|