|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 1578, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.019011406844106463, |
|
"grad_norm": 9.1989966206838, |
|
"learning_rate": 6.329113924050634e-07, |
|
"loss": 0.8874, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03802281368821293, |
|
"grad_norm": 4.161122432720083, |
|
"learning_rate": 1.2658227848101267e-06, |
|
"loss": 0.791, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.057034220532319393, |
|
"grad_norm": 1.6740419759238627, |
|
"learning_rate": 1.8987341772151901e-06, |
|
"loss": 0.7354, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07604562737642585, |
|
"grad_norm": 1.472773729260504, |
|
"learning_rate": 2.5316455696202535e-06, |
|
"loss": 0.6972, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09505703422053231, |
|
"grad_norm": 1.6632418888769052, |
|
"learning_rate": 3.164556962025317e-06, |
|
"loss": 0.698, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.11406844106463879, |
|
"grad_norm": 1.2299395567832287, |
|
"learning_rate": 3.7974683544303802e-06, |
|
"loss": 0.6778, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.13307984790874525, |
|
"grad_norm": 1.5380428593757542, |
|
"learning_rate": 4.430379746835443e-06, |
|
"loss": 0.6751, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.1520912547528517, |
|
"grad_norm": 1.3704564329134865, |
|
"learning_rate": 4.999995058613287e-06, |
|
"loss": 0.6713, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.17110266159695817, |
|
"grad_norm": 1.9490540655863884, |
|
"learning_rate": 4.999402118469357e-06, |
|
"loss": 0.6671, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.19011406844106463, |
|
"grad_norm": 1.5630755002882903, |
|
"learning_rate": 4.997821199394829e-06, |
|
"loss": 0.6676, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.20912547528517111, |
|
"grad_norm": 2.133710896815452, |
|
"learning_rate": 4.995252995758543e-06, |
|
"loss": 0.6695, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.22813688212927757, |
|
"grad_norm": 1.5943526267612576, |
|
"learning_rate": 4.99169863556294e-06, |
|
"loss": 0.6575, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.24714828897338403, |
|
"grad_norm": 1.832479949115814, |
|
"learning_rate": 4.987159679948624e-06, |
|
"loss": 0.6612, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.2661596958174905, |
|
"grad_norm": 1.4945020437443977, |
|
"learning_rate": 4.981638122508684e-06, |
|
"loss": 0.658, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.28517110266159695, |
|
"grad_norm": 1.619599628436461, |
|
"learning_rate": 4.975136388413065e-06, |
|
"loss": 0.6656, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.3041825095057034, |
|
"grad_norm": 1.5360935621329557, |
|
"learning_rate": 4.967657333343394e-06, |
|
"loss": 0.6511, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.3231939163498099, |
|
"grad_norm": 1.4510475916499845, |
|
"learning_rate": 4.959204242238707e-06, |
|
"loss": 0.6573, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.34220532319391633, |
|
"grad_norm": 1.532931462135121, |
|
"learning_rate": 4.949780827852648e-06, |
|
"loss": 0.6507, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.3612167300380228, |
|
"grad_norm": 1.1719507987669484, |
|
"learning_rate": 4.939391229122757e-06, |
|
"loss": 0.6523, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.38022813688212925, |
|
"grad_norm": 1.2585592582289158, |
|
"learning_rate": 4.928040009352568e-06, |
|
"loss": 0.6524, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.39923954372623577, |
|
"grad_norm": 1.7296360428483535, |
|
"learning_rate": 4.91573215420733e-06, |
|
"loss": 0.6614, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.41825095057034223, |
|
"grad_norm": 1.5691004380249787, |
|
"learning_rate": 4.902473069524204e-06, |
|
"loss": 0.6484, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.4372623574144487, |
|
"grad_norm": 1.247775759644616, |
|
"learning_rate": 4.888268578937923e-06, |
|
"loss": 0.6508, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.45627376425855515, |
|
"grad_norm": 1.3594034920568745, |
|
"learning_rate": 4.873124921322945e-06, |
|
"loss": 0.6469, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.4752851711026616, |
|
"grad_norm": 1.2309510240137682, |
|
"learning_rate": 4.8570487480532196e-06, |
|
"loss": 0.6401, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.49429657794676807, |
|
"grad_norm": 1.3299272061400234, |
|
"learning_rate": 4.840047120080787e-06, |
|
"loss": 0.6416, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.5133079847908745, |
|
"grad_norm": 1.3129636158199345, |
|
"learning_rate": 4.822127504834472e-06, |
|
"loss": 0.6438, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.532319391634981, |
|
"grad_norm": 1.4703016714267874, |
|
"learning_rate": 4.8032977729400585e-06, |
|
"loss": 0.6506, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.5513307984790875, |
|
"grad_norm": 1.3282370638504304, |
|
"learning_rate": 4.783566194763359e-06, |
|
"loss": 0.6521, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.5703422053231939, |
|
"grad_norm": 1.187364149838104, |
|
"learning_rate": 4.762941436777721e-06, |
|
"loss": 0.6371, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.5893536121673004, |
|
"grad_norm": 1.2164074784365948, |
|
"learning_rate": 4.7414325577575484e-06, |
|
"loss": 0.6407, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.6083650190114068, |
|
"grad_norm": 1.4208300435315424, |
|
"learning_rate": 4.719049004799525e-06, |
|
"loss": 0.6328, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.6273764258555133, |
|
"grad_norm": 1.324349326714873, |
|
"learning_rate": 4.695800609173274e-06, |
|
"loss": 0.6526, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.6463878326996197, |
|
"grad_norm": 1.1238559883921406, |
|
"learning_rate": 4.671697582003279e-06, |
|
"loss": 0.6371, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.6653992395437263, |
|
"grad_norm": 1.3772583883883136, |
|
"learning_rate": 4.646750509783975e-06, |
|
"loss": 0.6413, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.6844106463878327, |
|
"grad_norm": 1.3311507053384555, |
|
"learning_rate": 4.620970349729961e-06, |
|
"loss": 0.638, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.7034220532319392, |
|
"grad_norm": 1.1084009519269555, |
|
"learning_rate": 4.594368424963392e-06, |
|
"loss": 0.6402, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.7224334600760456, |
|
"grad_norm": 1.272517045012054, |
|
"learning_rate": 4.56695641954065e-06, |
|
"loss": 0.6332, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.7414448669201521, |
|
"grad_norm": 1.068763291195587, |
|
"learning_rate": 4.538746373320499e-06, |
|
"loss": 0.6451, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.7604562737642585, |
|
"grad_norm": 1.17609263311229, |
|
"learning_rate": 4.5097506766759465e-06, |
|
"loss": 0.6398, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.779467680608365, |
|
"grad_norm": 1.2993920513145316, |
|
"learning_rate": 4.479982065052171e-06, |
|
"loss": 0.6333, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.7984790874524715, |
|
"grad_norm": 1.4521328868048515, |
|
"learning_rate": 4.4494536133728754e-06, |
|
"loss": 0.6389, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.8174904942965779, |
|
"grad_norm": 1.134966447023594, |
|
"learning_rate": 4.418178730297542e-06, |
|
"loss": 0.6369, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.8365019011406845, |
|
"grad_norm": 1.1802112020601854, |
|
"learning_rate": 4.3861711523321e-06, |
|
"loss": 0.6355, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.8555133079847909, |
|
"grad_norm": 1.2438925294191414, |
|
"learning_rate": 4.353444937795595e-06, |
|
"loss": 0.6337, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.8745247148288974, |
|
"grad_norm": 1.169622958547474, |
|
"learning_rate": 4.320014460645523e-06, |
|
"loss": 0.6357, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.8935361216730038, |
|
"grad_norm": 1.0360592829770976, |
|
"learning_rate": 4.2858944041645196e-06, |
|
"loss": 0.6308, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.9125475285171103, |
|
"grad_norm": 1.133882806727274, |
|
"learning_rate": 4.251099754511189e-06, |
|
"loss": 0.633, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.9315589353612167, |
|
"grad_norm": 1.0230579665487711, |
|
"learning_rate": 4.2156457941379095e-06, |
|
"loss": 0.637, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.9505703422053232, |
|
"grad_norm": 1.0897950528236497, |
|
"learning_rate": 4.179548095078498e-06, |
|
"loss": 0.6315, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.9695817490494296, |
|
"grad_norm": 1.4631268078629502, |
|
"learning_rate": 4.142822512108683e-06, |
|
"loss": 0.632, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.9885931558935361, |
|
"grad_norm": 1.2859460507714569, |
|
"learning_rate": 4.105485175782396e-06, |
|
"loss": 0.6226, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.6353716254234314, |
|
"eval_runtime": 49.2084, |
|
"eval_samples_per_second": 287.816, |
|
"eval_steps_per_second": 1.138, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.0076045627376427, |
|
"grad_norm": 2.6186402518104304, |
|
"learning_rate": 4.067552485346939e-06, |
|
"loss": 0.5901, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.026615969581749, |
|
"grad_norm": 1.4429614147480494, |
|
"learning_rate": 4.029041101540122e-06, |
|
"loss": 0.546, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.0456273764258555, |
|
"grad_norm": 1.1901165511323117, |
|
"learning_rate": 3.989967939272569e-06, |
|
"loss": 0.5434, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.064638783269962, |
|
"grad_norm": 1.133670318062506, |
|
"learning_rate": 3.950350160198373e-06, |
|
"loss": 0.5365, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.0836501901140685, |
|
"grad_norm": 1.1569131629508171, |
|
"learning_rate": 3.91020516517738e-06, |
|
"loss": 0.5394, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.102661596958175, |
|
"grad_norm": 1.1578867626444747, |
|
"learning_rate": 3.869550586632413e-06, |
|
"loss": 0.5454, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.1216730038022813, |
|
"grad_norm": 1.1710585399638043, |
|
"learning_rate": 3.828404280804782e-06, |
|
"loss": 0.5437, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.1406844106463878, |
|
"grad_norm": 1.2968456937546469, |
|
"learning_rate": 3.7867843199114996e-06, |
|
"loss": 0.545, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.1596958174904943, |
|
"grad_norm": 1.173788116358311, |
|
"learning_rate": 3.7447089842076238e-06, |
|
"loss": 0.5401, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.1787072243346008, |
|
"grad_norm": 1.1736747473104372, |
|
"learning_rate": 3.7021967539572343e-06, |
|
"loss": 0.5451, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.1977186311787071, |
|
"grad_norm": 1.1583258876929678, |
|
"learning_rate": 3.6592663013165636e-06, |
|
"loss": 0.5413, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.2167300380228137, |
|
"grad_norm": 1.1187865805782111, |
|
"learning_rate": 3.6159364821328325e-06, |
|
"loss": 0.5531, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.2357414448669202, |
|
"grad_norm": 1.2500694936887193, |
|
"learning_rate": 3.572226327662423e-06, |
|
"loss": 0.5481, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.2547528517110267, |
|
"grad_norm": 1.2649269376886392, |
|
"learning_rate": 3.528155036211995e-06, |
|
"loss": 0.5402, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.2737642585551332, |
|
"grad_norm": 1.2215327457350609, |
|
"learning_rate": 3.4837419647062344e-06, |
|
"loss": 0.5452, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.2927756653992395, |
|
"grad_norm": 1.2145466936928535, |
|
"learning_rate": 3.4390066201859355e-06, |
|
"loss": 0.5376, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.311787072243346, |
|
"grad_norm": 1.16823342480946, |
|
"learning_rate": 3.393968651240153e-06, |
|
"loss": 0.5442, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.3307984790874525, |
|
"grad_norm": 1.1720513921800486, |
|
"learning_rate": 3.3486478393761707e-06, |
|
"loss": 0.5449, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.3498098859315588, |
|
"grad_norm": 1.32127207540034, |
|
"learning_rate": 3.303064090331106e-06, |
|
"loss": 0.537, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.3688212927756653, |
|
"grad_norm": 1.170488361503433, |
|
"learning_rate": 3.2572374253289385e-06, |
|
"loss": 0.5525, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.3878326996197718, |
|
"grad_norm": 1.2540643595664949, |
|
"learning_rate": 3.2111879722868204e-06, |
|
"loss": 0.5513, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.4068441064638784, |
|
"grad_norm": 1.1125376049437519, |
|
"learning_rate": 3.164935956974522e-06, |
|
"loss": 0.5428, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.4258555133079849, |
|
"grad_norm": 1.2222128167327286, |
|
"learning_rate": 3.1185016941309033e-06, |
|
"loss": 0.5443, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.4448669201520912, |
|
"grad_norm": 1.1189014403528423, |
|
"learning_rate": 3.071905578541304e-06, |
|
"loss": 0.55, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.4638783269961977, |
|
"grad_norm": 1.106630875988617, |
|
"learning_rate": 3.0251680760797737e-06, |
|
"loss": 0.5512, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.4828897338403042, |
|
"grad_norm": 1.238247599173364, |
|
"learning_rate": 2.978309714720092e-06, |
|
"loss": 0.5457, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.5019011406844105, |
|
"grad_norm": 1.1877938363765261, |
|
"learning_rate": 2.931351075519494e-06, |
|
"loss": 0.5472, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.5209125475285172, |
|
"grad_norm": 1.0765509288814237, |
|
"learning_rate": 2.8843127835790922e-06, |
|
"loss": 0.5467, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.5399239543726235, |
|
"grad_norm": 1.1769551690370097, |
|
"learning_rate": 2.837215498984956e-06, |
|
"loss": 0.5423, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.55893536121673, |
|
"grad_norm": 1.0525687613765247, |
|
"learning_rate": 2.7900799077338137e-06, |
|
"loss": 0.5377, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.5779467680608366, |
|
"grad_norm": 1.1872738975990165, |
|
"learning_rate": 2.742926712647389e-06, |
|
"loss": 0.5422, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.5969581749049429, |
|
"grad_norm": 1.2556831499142438, |
|
"learning_rate": 2.6957766242793286e-06, |
|
"loss": 0.5548, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.6159695817490496, |
|
"grad_norm": 1.0530666465416862, |
|
"learning_rate": 2.648650351818758e-06, |
|
"loss": 0.5396, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.6349809885931559, |
|
"grad_norm": 1.0626505034866265, |
|
"learning_rate": 2.6015685939944113e-06, |
|
"loss": 0.5443, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.6539923954372624, |
|
"grad_norm": 1.02476310680603, |
|
"learning_rate": 2.554552029983375e-06, |
|
"loss": 0.5404, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.673003802281369, |
|
"grad_norm": 1.0885196976387683, |
|
"learning_rate": 2.5076213103284107e-06, |
|
"loss": 0.5443, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.6920152091254752, |
|
"grad_norm": 1.1420366693644837, |
|
"learning_rate": 2.4607970478678515e-06, |
|
"loss": 0.5464, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.7110266159695817, |
|
"grad_norm": 1.1255789898459938, |
|
"learning_rate": 2.4140998086820665e-06, |
|
"loss": 0.5447, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.7300380228136882, |
|
"grad_norm": 1.3189103855357065, |
|
"learning_rate": 2.367550103060459e-06, |
|
"loss": 0.5475, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.7490494296577945, |
|
"grad_norm": 1.0465751177563847, |
|
"learning_rate": 2.3211683764929664e-06, |
|
"loss": 0.5484, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.7680608365019013, |
|
"grad_norm": 1.1704397396223116, |
|
"learning_rate": 2.2749750006900263e-06, |
|
"loss": 0.5431, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.7870722433460076, |
|
"grad_norm": 1.053213291008937, |
|
"learning_rate": 2.2289902646349423e-06, |
|
"loss": 0.5425, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.806083650190114, |
|
"grad_norm": 1.2515940528618232, |
|
"learning_rate": 2.1832343656725918e-06, |
|
"loss": 0.5459, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.8250950570342206, |
|
"grad_norm": 1.099127233957744, |
|
"learning_rate": 2.1377274006383737e-06, |
|
"loss": 0.5394, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.8441064638783269, |
|
"grad_norm": 1.12063559389163, |
|
"learning_rate": 2.092489357031315e-06, |
|
"loss": 0.5447, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.8631178707224336, |
|
"grad_norm": 1.102123448980619, |
|
"learning_rate": 2.0475401042351843e-06, |
|
"loss": 0.5305, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.88212927756654, |
|
"grad_norm": 1.1254624333919672, |
|
"learning_rate": 2.0028993847914966e-06, |
|
"loss": 0.5395, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.9011406844106464, |
|
"grad_norm": 1.0910835517042778, |
|
"learning_rate": 1.9585868057282256e-06, |
|
"loss": 0.5406, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.920152091254753, |
|
"grad_norm": 1.3820944438954628, |
|
"learning_rate": 1.9146218299480303e-06, |
|
"loss": 0.5476, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.9391634980988592, |
|
"grad_norm": 1.2180866671022355, |
|
"learning_rate": 1.8710237676797923e-06, |
|
"loss": 0.5429, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.9581749049429658, |
|
"grad_norm": 1.1380236136327964, |
|
"learning_rate": 1.827811767997207e-06, |
|
"loss": 0.5419, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.9771863117870723, |
|
"grad_norm": 1.0602009694403736, |
|
"learning_rate": 1.7850048104081555e-06, |
|
"loss": 0.5369, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.9961977186311786, |
|
"grad_norm": 1.0393675864746768, |
|
"learning_rate": 1.742621696518555e-06, |
|
"loss": 0.5371, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.6316379308700562, |
|
"eval_runtime": 54.2509, |
|
"eval_samples_per_second": 261.065, |
|
"eval_steps_per_second": 1.032, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 2.0152091254752853, |
|
"grad_norm": 1.7644212761083442, |
|
"learning_rate": 1.7006810417743453e-06, |
|
"loss": 0.4716, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 2.0342205323193916, |
|
"grad_norm": 1.3479439652442078, |
|
"learning_rate": 1.6592012672852462e-06, |
|
"loss": 0.4592, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 2.053231939163498, |
|
"grad_norm": 1.1555574696681938, |
|
"learning_rate": 1.618200591733858e-06, |
|
"loss": 0.4568, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 2.0722433460076046, |
|
"grad_norm": 1.1655620913425222, |
|
"learning_rate": 1.5776970233736877e-06, |
|
"loss": 0.4649, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 2.091254752851711, |
|
"grad_norm": 1.246797649914118, |
|
"learning_rate": 1.5377083521195849e-06, |
|
"loss": 0.4584, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.1102661596958177, |
|
"grad_norm": 1.20981005283122, |
|
"learning_rate": 1.4982521417340881e-06, |
|
"loss": 0.457, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 2.129277566539924, |
|
"grad_norm": 1.25314570539992, |
|
"learning_rate": 1.459345722113095e-06, |
|
"loss": 0.4613, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 2.1482889733840302, |
|
"grad_norm": 1.2661139969932325, |
|
"learning_rate": 1.4210061816742487e-06, |
|
"loss": 0.4557, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 2.167300380228137, |
|
"grad_norm": 1.22533837512988, |
|
"learning_rate": 1.3832503598513939e-06, |
|
"loss": 0.4638, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 2.1863117870722433, |
|
"grad_norm": 1.153296199976961, |
|
"learning_rate": 1.3460948396983833e-06, |
|
"loss": 0.4539, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 2.20532319391635, |
|
"grad_norm": 1.2554665358203887, |
|
"learning_rate": 1.3095559406054964e-06, |
|
"loss": 0.4574, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 2.2243346007604563, |
|
"grad_norm": 1.216327809593985, |
|
"learning_rate": 1.2736497111316607e-06, |
|
"loss": 0.4595, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 2.2433460076045626, |
|
"grad_norm": 1.1472080286252149, |
|
"learning_rate": 1.238391921955631e-06, |
|
"loss": 0.4565, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 2.2623574144486693, |
|
"grad_norm": 1.280760900457142, |
|
"learning_rate": 1.2037980589492116e-06, |
|
"loss": 0.4623, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 2.2813688212927756, |
|
"grad_norm": 1.1694311311925278, |
|
"learning_rate": 1.1698833163755793e-06, |
|
"loss": 0.4531, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 2.3003802281368824, |
|
"grad_norm": 1.1799865250234591, |
|
"learning_rate": 1.1366625902156814e-06, |
|
"loss": 0.4555, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 2.3193916349809887, |
|
"grad_norm": 1.1721763967059378, |
|
"learning_rate": 1.1041504716256456e-06, |
|
"loss": 0.4601, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 2.338403041825095, |
|
"grad_norm": 1.1882316774789523, |
|
"learning_rate": 1.0723612405280762e-06, |
|
"loss": 0.4669, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 2.3574144486692017, |
|
"grad_norm": 1.1388831508699453, |
|
"learning_rate": 1.0413088593400464e-06, |
|
"loss": 0.4632, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 2.376425855513308, |
|
"grad_norm": 1.143656084527738, |
|
"learning_rate": 1.0110069668405487e-06, |
|
"loss": 0.4564, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 2.3954372623574143, |
|
"grad_norm": 1.210491576650488, |
|
"learning_rate": 9.8146887218009e-07, |
|
"loss": 0.4639, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 2.414448669201521, |
|
"grad_norm": 1.1273169371405412, |
|
"learning_rate": 9.527075490350695e-07, |
|
"loss": 0.4615, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 2.4334600760456273, |
|
"grad_norm": 1.215505001410675, |
|
"learning_rate": 9.247356299094966e-07, |
|
"loss": 0.4528, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 2.4524714828897336, |
|
"grad_norm": 1.1389739744110048, |
|
"learning_rate": 8.975654005865671e-07, |
|
"loss": 0.4535, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 2.4714828897338403, |
|
"grad_norm": 1.159288850094368, |
|
"learning_rate": 8.712087947325155e-07, |
|
"loss": 0.4514, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 2.4904942965779466, |
|
"grad_norm": 1.2097686536353103, |
|
"learning_rate": 8.456773886551317e-07, |
|
"loss": 0.456, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 2.5095057034220534, |
|
"grad_norm": 1.125891025891593, |
|
"learning_rate": 8.20982396219231e-07, |
|
"loss": 0.4584, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 2.5285171102661597, |
|
"grad_norm": 1.165727380934779, |
|
"learning_rate": 7.971346639213192e-07, |
|
"loss": 0.4576, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 2.5475285171102664, |
|
"grad_norm": 1.2196983806300021, |
|
"learning_rate": 7.74144666125611e-07, |
|
"loss": 0.4585, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 2.5665399239543727, |
|
"grad_norm": 1.1416376267908626, |
|
"learning_rate": 7.520225004634946e-07, |
|
"loss": 0.4617, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 2.585551330798479, |
|
"grad_norm": 1.1374660560345522, |
|
"learning_rate": 7.307778833984666e-07, |
|
"loss": 0.4525, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 2.6045627376425857, |
|
"grad_norm": 1.1600478528127594, |
|
"learning_rate": 7.104201459584814e-07, |
|
"loss": 0.4554, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 2.623574144486692, |
|
"grad_norm": 1.1613018240717523, |
|
"learning_rate": 6.909582296375923e-07, |
|
"loss": 0.4571, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 2.6425855513307983, |
|
"grad_norm": 1.1582568171040142, |
|
"learning_rate": 6.724006824686814e-07, |
|
"loss": 0.458, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 2.661596958174905, |
|
"grad_norm": 1.2297405479958352, |
|
"learning_rate": 6.547556552690069e-07, |
|
"loss": 0.4576, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.6806083650190113, |
|
"grad_norm": 1.1545365546618969, |
|
"learning_rate": 6.380308980602119e-07, |
|
"loss": 0.4559, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 2.6996197718631176, |
|
"grad_norm": 1.1501358792075553, |
|
"learning_rate": 6.222337566643733e-07, |
|
"loss": 0.458, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 2.7186311787072244, |
|
"grad_norm": 1.1596268149561126, |
|
"learning_rate": 6.073711694775805e-07, |
|
"loss": 0.4523, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 2.7376425855513307, |
|
"grad_norm": 1.176536310467956, |
|
"learning_rate": 5.934496644224642e-07, |
|
"loss": 0.4592, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 2.7566539923954374, |
|
"grad_norm": 1.1601201945564672, |
|
"learning_rate": 5.804753560810128e-07, |
|
"loss": 0.4587, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.7756653992395437, |
|
"grad_norm": 1.1503887080909905, |
|
"learning_rate": 5.684539430089329e-07, |
|
"loss": 0.4558, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 2.7946768060836504, |
|
"grad_norm": 1.148286520523666, |
|
"learning_rate": 5.573907052327413e-07, |
|
"loss": 0.4585, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 2.8136882129277567, |
|
"grad_norm": 1.1477000696419715, |
|
"learning_rate": 5.472905019306775e-07, |
|
"loss": 0.4555, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 2.832699619771863, |
|
"grad_norm": 1.1636191591442357, |
|
"learning_rate": 5.38157769298465e-07, |
|
"loss": 0.4529, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 2.8517110266159698, |
|
"grad_norm": 1.226786969124158, |
|
"learning_rate": 5.2999651860085e-07, |
|
"loss": 0.4628, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.870722433460076, |
|
"grad_norm": 1.2111320995578168, |
|
"learning_rate": 5.228103344097825e-07, |
|
"loss": 0.4642, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 2.8897338403041823, |
|
"grad_norm": 1.134156671769776, |
|
"learning_rate": 5.166023730300036e-07, |
|
"loss": 0.4541, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 2.908745247148289, |
|
"grad_norm": 1.1404768838000066, |
|
"learning_rate": 5.113753611127401e-07, |
|
"loss": 0.4607, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 2.9277566539923954, |
|
"grad_norm": 1.1667213786402724, |
|
"learning_rate": 5.071315944581075e-07, |
|
"loss": 0.4638, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 2.9467680608365017, |
|
"grad_norm": 1.151294180691113, |
|
"learning_rate": 5.038729370067536e-07, |
|
"loss": 0.455, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 2.9657794676806084, |
|
"grad_norm": 1.2244552822051, |
|
"learning_rate": 5.016008200211805e-07, |
|
"loss": 0.4556, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 2.9847908745247147, |
|
"grad_norm": 1.1748003065525463, |
|
"learning_rate": 5.003162414571072e-07, |
|
"loss": 0.455, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.6586939692497253, |
|
"eval_runtime": 55.4914, |
|
"eval_samples_per_second": 255.229, |
|
"eval_steps_per_second": 1.009, |
|
"step": 1578 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 1578, |
|
"total_flos": 2643208773304320.0, |
|
"train_loss": 0.5529848988367426, |
|
"train_runtime": 9902.4641, |
|
"train_samples_per_second": 81.519, |
|
"train_steps_per_second": 0.159 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1578, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2643208773304320.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|