|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 9.977827050997783, |
|
"eval_steps": 500, |
|
"global_step": 2250, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.004434589800443459, |
|
"grad_norm": 10.587126731872559, |
|
"learning_rate": 8.88888888888889e-07, |
|
"loss": 3.0058, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.022172949002217297, |
|
"grad_norm": 10.391400337219238, |
|
"learning_rate": 4.444444444444445e-06, |
|
"loss": 3.0384, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.04434589800443459, |
|
"grad_norm": 10.358064651489258, |
|
"learning_rate": 8.88888888888889e-06, |
|
"loss": 2.9329, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06651884700665188, |
|
"grad_norm": 9.945845603942871, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 2.8718, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.08869179600886919, |
|
"grad_norm": 7.753511905670166, |
|
"learning_rate": 1.777777777777778e-05, |
|
"loss": 2.7228, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.11086474501108648, |
|
"grad_norm": 3.560798406600952, |
|
"learning_rate": 2.2222222222222223e-05, |
|
"loss": 2.5677, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.13303769401330376, |
|
"grad_norm": 1.3562530279159546, |
|
"learning_rate": 2.6666666666666667e-05, |
|
"loss": 2.4311, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.15521064301552107, |
|
"grad_norm": 0.9185900092124939, |
|
"learning_rate": 3.111111111111111e-05, |
|
"loss": 2.335, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.17738359201773837, |
|
"grad_norm": 0.7675415277481079, |
|
"learning_rate": 3.555555555555556e-05, |
|
"loss": 2.2184, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.19955654101995565, |
|
"grad_norm": 0.6789694428443909, |
|
"learning_rate": 4e-05, |
|
"loss": 2.1433, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.22172949002217296, |
|
"grad_norm": 0.583677351474762, |
|
"learning_rate": 4.4444444444444447e-05, |
|
"loss": 2.0729, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.24390243902439024, |
|
"grad_norm": 0.5190356969833374, |
|
"learning_rate": 4.888888888888889e-05, |
|
"loss": 1.9766, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.2660753880266075, |
|
"grad_norm": 0.4500294327735901, |
|
"learning_rate": 5.333333333333333e-05, |
|
"loss": 1.9329, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.28824833702882485, |
|
"grad_norm": 0.4370773136615753, |
|
"learning_rate": 5.7777777777777776e-05, |
|
"loss": 1.8666, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.31042128603104213, |
|
"grad_norm": 0.4395270049571991, |
|
"learning_rate": 6.222222222222222e-05, |
|
"loss": 1.806, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.3325942350332594, |
|
"grad_norm": 0.3201618492603302, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 1.785, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.35476718403547675, |
|
"grad_norm": 0.36038511991500854, |
|
"learning_rate": 7.111111111111112e-05, |
|
"loss": 1.7624, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.376940133037694, |
|
"grad_norm": 0.32811540365219116, |
|
"learning_rate": 7.555555555555556e-05, |
|
"loss": 1.7255, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.3991130820399113, |
|
"grad_norm": 0.3289714753627777, |
|
"learning_rate": 8e-05, |
|
"loss": 1.7171, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.4212860310421286, |
|
"grad_norm": 0.3439875543117523, |
|
"learning_rate": 8.444444444444444e-05, |
|
"loss": 1.69, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.4434589800443459, |
|
"grad_norm": 0.3265211880207062, |
|
"learning_rate": 8.888888888888889e-05, |
|
"loss": 1.6868, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.4656319290465632, |
|
"grad_norm": 0.3699732720851898, |
|
"learning_rate": 9.333333333333334e-05, |
|
"loss": 1.6765, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.4878048780487805, |
|
"grad_norm": 0.3702254891395569, |
|
"learning_rate": 9.777777777777778e-05, |
|
"loss": 1.6533, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.5099778270509978, |
|
"grad_norm": 0.3689574599266052, |
|
"learning_rate": 0.00010222222222222222, |
|
"loss": 1.6488, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.532150776053215, |
|
"grad_norm": 0.42295676469802856, |
|
"learning_rate": 0.00010666666666666667, |
|
"loss": 1.6136, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5543237250554324, |
|
"grad_norm": 0.4599798917770386, |
|
"learning_rate": 0.00011111111111111112, |
|
"loss": 1.5929, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.5764966740576497, |
|
"grad_norm": 0.40791696310043335, |
|
"learning_rate": 0.00011555555555555555, |
|
"loss": 1.5782, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.5986696230598669, |
|
"grad_norm": 0.5501047372817993, |
|
"learning_rate": 0.00012, |
|
"loss": 1.5679, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.6208425720620843, |
|
"grad_norm": 0.4360973536968231, |
|
"learning_rate": 0.00012444444444444444, |
|
"loss": 1.5719, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6430155210643016, |
|
"grad_norm": 0.6553646922111511, |
|
"learning_rate": 0.00012888888888888892, |
|
"loss": 1.5643, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.6651884700665188, |
|
"grad_norm": 0.6760438680648804, |
|
"learning_rate": 0.00013333333333333334, |
|
"loss": 1.5642, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.6873614190687362, |
|
"grad_norm": 0.6198329329490662, |
|
"learning_rate": 0.0001377777777777778, |
|
"loss": 1.5522, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.7095343680709535, |
|
"grad_norm": 0.5623511672019958, |
|
"learning_rate": 0.00014222222222222224, |
|
"loss": 1.5425, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.7317073170731707, |
|
"grad_norm": 0.4507710337638855, |
|
"learning_rate": 0.00014666666666666666, |
|
"loss": 1.52, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.753880266075388, |
|
"grad_norm": 0.4668518602848053, |
|
"learning_rate": 0.0001511111111111111, |
|
"loss": 1.5275, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.7760532150776053, |
|
"grad_norm": 0.40650302171707153, |
|
"learning_rate": 0.00015555555555555556, |
|
"loss": 1.5356, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.7982261640798226, |
|
"grad_norm": 0.47852110862731934, |
|
"learning_rate": 0.00016, |
|
"loss": 1.5367, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.8203991130820399, |
|
"grad_norm": 0.41115882992744446, |
|
"learning_rate": 0.00016444444444444444, |
|
"loss": 1.5142, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.8425720620842572, |
|
"grad_norm": 0.48749256134033203, |
|
"learning_rate": 0.00016888888888888889, |
|
"loss": 1.5132, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.8647450110864745, |
|
"grad_norm": 0.45032379031181335, |
|
"learning_rate": 0.00017333333333333334, |
|
"loss": 1.4971, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.8869179600886918, |
|
"grad_norm": 0.4163128435611725, |
|
"learning_rate": 0.00017777777777777779, |
|
"loss": 1.511, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.9090909090909091, |
|
"grad_norm": 0.43136799335479736, |
|
"learning_rate": 0.00018222222222222224, |
|
"loss": 1.4945, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.9312638580931264, |
|
"grad_norm": 0.4096197783946991, |
|
"learning_rate": 0.0001866666666666667, |
|
"loss": 1.4786, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.9534368070953437, |
|
"grad_norm": 0.5415340065956116, |
|
"learning_rate": 0.00019111111111111114, |
|
"loss": 1.4733, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.975609756097561, |
|
"grad_norm": 0.5372570753097534, |
|
"learning_rate": 0.00019555555555555556, |
|
"loss": 1.4994, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.9977827050997783, |
|
"grad_norm": 0.4384077787399292, |
|
"learning_rate": 0.0002, |
|
"loss": 1.4961, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.9977827050997783, |
|
"eval_loss": 1.770818829536438, |
|
"eval_runtime": 0.3325, |
|
"eval_samples_per_second": 3.008, |
|
"eval_steps_per_second": 3.008, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.0199556541019956, |
|
"grad_norm": 0.592540979385376, |
|
"learning_rate": 0.00019999699144797678, |
|
"loss": 1.4664, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.042128603104213, |
|
"grad_norm": 0.4885026514530182, |
|
"learning_rate": 0.00019998796597293477, |
|
"loss": 1.4644, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.06430155210643, |
|
"grad_norm": 0.3968167006969452, |
|
"learning_rate": 0.00019997292411794618, |
|
"loss": 1.4751, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.0864745011086474, |
|
"grad_norm": 0.37674403190612793, |
|
"learning_rate": 0.00019995186678809513, |
|
"loss": 1.4596, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.1086474501108647, |
|
"grad_norm": 0.4001663029193878, |
|
"learning_rate": 0.00019992479525042303, |
|
"loss": 1.4788, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.130820399113082, |
|
"grad_norm": 0.3921445608139038, |
|
"learning_rate": 0.0001998917111338525, |
|
"loss": 1.4625, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 1.1529933481152994, |
|
"grad_norm": 0.3944418132305145, |
|
"learning_rate": 0.00019985261642908918, |
|
"loss": 1.4488, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.1751662971175167, |
|
"grad_norm": 0.4203372001647949, |
|
"learning_rate": 0.00019980751348850219, |
|
"loss": 1.4565, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 1.1973392461197339, |
|
"grad_norm": 0.4535711109638214, |
|
"learning_rate": 0.00019975640502598244, |
|
"loss": 1.4549, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.2195121951219512, |
|
"grad_norm": 0.41691839694976807, |
|
"learning_rate": 0.0001996992941167792, |
|
"loss": 1.432, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.2416851441241685, |
|
"grad_norm": 0.38892751932144165, |
|
"learning_rate": 0.00019963618419731538, |
|
"loss": 1.4378, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.2638580931263859, |
|
"grad_norm": 0.3704243302345276, |
|
"learning_rate": 0.00019956707906498044, |
|
"loss": 1.4547, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 1.2860310421286032, |
|
"grad_norm": 0.4061298668384552, |
|
"learning_rate": 0.00019949198287790213, |
|
"loss": 1.4309, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.3082039911308203, |
|
"grad_norm": 0.41798922419548035, |
|
"learning_rate": 0.00019941090015469613, |
|
"loss": 1.4253, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 1.3303769401330376, |
|
"grad_norm": 0.4110168516635895, |
|
"learning_rate": 0.00019932383577419432, |
|
"loss": 1.4368, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.352549889135255, |
|
"grad_norm": 0.43604379892349243, |
|
"learning_rate": 0.00019923079497515098, |
|
"loss": 1.429, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 1.3747228381374723, |
|
"grad_norm": 0.3890831172466278, |
|
"learning_rate": 0.00019913178335592783, |
|
"loss": 1.4165, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.3968957871396896, |
|
"grad_norm": 0.3936193585395813, |
|
"learning_rate": 0.00019902680687415705, |
|
"loss": 1.4299, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.4190687361419068, |
|
"grad_norm": 0.4455225467681885, |
|
"learning_rate": 0.00019891587184638272, |
|
"loss": 1.4098, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.441241685144124, |
|
"grad_norm": 0.3836926817893982, |
|
"learning_rate": 0.00019879898494768093, |
|
"loss": 1.4046, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.4634146341463414, |
|
"grad_norm": 0.41238486766815186, |
|
"learning_rate": 0.00019867615321125795, |
|
"loss": 1.42, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.4855875831485588, |
|
"grad_norm": 0.3580898940563202, |
|
"learning_rate": 0.00019854738402802716, |
|
"loss": 1.4199, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.507760532150776, |
|
"grad_norm": 0.3761023283004761, |
|
"learning_rate": 0.00019841268514616433, |
|
"loss": 1.4222, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.5299334811529932, |
|
"grad_norm": 0.3899005055427551, |
|
"learning_rate": 0.00019827206467064133, |
|
"loss": 1.4159, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.5521064301552108, |
|
"grad_norm": 0.38497841358184814, |
|
"learning_rate": 0.00019812553106273847, |
|
"loss": 1.4089, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.5742793791574279, |
|
"grad_norm": 0.3658756911754608, |
|
"learning_rate": 0.00019797309313953541, |
|
"loss": 1.4181, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.5964523281596452, |
|
"grad_norm": 0.3782854378223419, |
|
"learning_rate": 0.00019781476007338058, |
|
"loss": 1.4143, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.6186252771618626, |
|
"grad_norm": 0.5063923001289368, |
|
"learning_rate": 0.0001976505413913393, |
|
"loss": 1.4154, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.6407982261640797, |
|
"grad_norm": 0.38294726610183716, |
|
"learning_rate": 0.00019748044697462058, |
|
"loss": 1.4268, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.6629711751662972, |
|
"grad_norm": 0.35021841526031494, |
|
"learning_rate": 0.00019730448705798239, |
|
"loss": 1.4065, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.6851441241685143, |
|
"grad_norm": 0.3741037845611572, |
|
"learning_rate": 0.00019712267222911606, |
|
"loss": 1.4079, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.7073170731707317, |
|
"grad_norm": 0.3381967842578888, |
|
"learning_rate": 0.00019693501342800892, |
|
"loss": 1.4137, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.729490022172949, |
|
"grad_norm": 0.38490810990333557, |
|
"learning_rate": 0.00019674152194628638, |
|
"loss": 1.4005, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.7516629711751663, |
|
"grad_norm": 0.36805257201194763, |
|
"learning_rate": 0.0001965422094265322, |
|
"loss": 1.3987, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.7738359201773837, |
|
"grad_norm": 0.3764079213142395, |
|
"learning_rate": 0.00019633708786158806, |
|
"loss": 1.4124, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.7960088691796008, |
|
"grad_norm": 0.3698968291282654, |
|
"learning_rate": 0.0001961261695938319, |
|
"loss": 1.4082, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.8181818181818183, |
|
"grad_norm": 0.38398101925849915, |
|
"learning_rate": 0.0001959094673144354, |
|
"loss": 1.3864, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.8403547671840355, |
|
"grad_norm": 0.3575521409511566, |
|
"learning_rate": 0.00019568699406260015, |
|
"loss": 1.3982, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.8625277161862528, |
|
"grad_norm": 0.4014054834842682, |
|
"learning_rate": 0.0001954587632247732, |
|
"loss": 1.3862, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.8847006651884701, |
|
"grad_norm": 0.42844465374946594, |
|
"learning_rate": 0.00019522478853384155, |
|
"loss": 1.4102, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.9068736141906872, |
|
"grad_norm": 0.43658122420310974, |
|
"learning_rate": 0.00019498508406830576, |
|
"loss": 1.4125, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.9290465631929048, |
|
"grad_norm": 0.37285885214805603, |
|
"learning_rate": 0.00019473966425143292, |
|
"loss": 1.3878, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.951219512195122, |
|
"grad_norm": 0.39875471591949463, |
|
"learning_rate": 0.0001944885438503888, |
|
"loss": 1.3991, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.9733924611973392, |
|
"grad_norm": 0.384593665599823, |
|
"learning_rate": 0.00019423173797534924, |
|
"loss": 1.3938, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.9955654101995566, |
|
"grad_norm": 0.4125775098800659, |
|
"learning_rate": 0.00019396926207859084, |
|
"loss": 1.3952, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.7769551277160645, |
|
"eval_runtime": 0.3268, |
|
"eval_samples_per_second": 3.06, |
|
"eval_steps_per_second": 3.06, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 2.0177383592017737, |
|
"grad_norm": 0.40251436829566956, |
|
"learning_rate": 0.0001937011319535615, |
|
"loss": 1.3874, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 2.0399113082039912, |
|
"grad_norm": 0.4194943904876709, |
|
"learning_rate": 0.00019342736373392975, |
|
"loss": 1.3625, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.0620842572062084, |
|
"grad_norm": 0.3550240099430084, |
|
"learning_rate": 0.00019314797389261424, |
|
"loss": 1.3776, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 2.084257206208426, |
|
"grad_norm": 0.360365629196167, |
|
"learning_rate": 0.00019286297924079243, |
|
"loss": 1.3611, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.106430155210643, |
|
"grad_norm": 0.39108407497406006, |
|
"learning_rate": 0.00019257239692688907, |
|
"loss": 1.3671, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.12860310421286, |
|
"grad_norm": 0.4020346701145172, |
|
"learning_rate": 0.00019227624443554425, |
|
"loss": 1.3647, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.1507760532150777, |
|
"grad_norm": 0.3887872099876404, |
|
"learning_rate": 0.00019197453958656158, |
|
"loss": 1.346, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 2.172949002217295, |
|
"grad_norm": 0.374032199382782, |
|
"learning_rate": 0.00019166730053383568, |
|
"loss": 1.3442, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.1951219512195124, |
|
"grad_norm": 0.3590153157711029, |
|
"learning_rate": 0.0001913545457642601, |
|
"loss": 1.3695, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 2.2172949002217295, |
|
"grad_norm": 0.3909887969493866, |
|
"learning_rate": 0.0001910362940966147, |
|
"loss": 1.3629, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.2394678492239466, |
|
"grad_norm": 0.35918113589286804, |
|
"learning_rate": 0.00019071256468043338, |
|
"loss": 1.3759, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 2.261640798226164, |
|
"grad_norm": 0.3407536447048187, |
|
"learning_rate": 0.00019038337699485208, |
|
"loss": 1.3631, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.2838137472283813, |
|
"grad_norm": 0.3984699845314026, |
|
"learning_rate": 0.00019004875084743624, |
|
"loss": 1.3561, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 2.305986696230599, |
|
"grad_norm": 0.4278836250305176, |
|
"learning_rate": 0.00018970870637298934, |
|
"loss": 1.3511, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.328159645232816, |
|
"grad_norm": 0.3488403260707855, |
|
"learning_rate": 0.00018936326403234125, |
|
"loss": 1.3579, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 2.3503325942350335, |
|
"grad_norm": 0.3680429756641388, |
|
"learning_rate": 0.00018901244461111695, |
|
"loss": 1.3565, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.3725055432372506, |
|
"grad_norm": 0.34674274921417236, |
|
"learning_rate": 0.00018865626921848615, |
|
"loss": 1.363, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 2.3946784922394677, |
|
"grad_norm": 0.404840886592865, |
|
"learning_rate": 0.00018829475928589271, |
|
"loss": 1.3662, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.4168514412416853, |
|
"grad_norm": 0.3971501588821411, |
|
"learning_rate": 0.00018792793656576543, |
|
"loss": 1.3684, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 2.4390243902439024, |
|
"grad_norm": 0.4021252393722534, |
|
"learning_rate": 0.0001875558231302091, |
|
"loss": 1.3597, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.4611973392461195, |
|
"grad_norm": 0.49484843015670776, |
|
"learning_rate": 0.00018717844136967624, |
|
"loss": 1.3632, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 2.483370288248337, |
|
"grad_norm": 0.425786554813385, |
|
"learning_rate": 0.00018679581399162008, |
|
"loss": 1.3433, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.505543237250554, |
|
"grad_norm": 0.35825788974761963, |
|
"learning_rate": 0.00018640796401912807, |
|
"loss": 1.3605, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 2.5277161862527717, |
|
"grad_norm": 0.3355163335800171, |
|
"learning_rate": 0.00018601491478953657, |
|
"loss": 1.355, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.549889135254989, |
|
"grad_norm": 0.3719523251056671, |
|
"learning_rate": 0.00018561668995302667, |
|
"loss": 1.3596, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 2.5720620842572064, |
|
"grad_norm": 0.37250471115112305, |
|
"learning_rate": 0.00018521331347120114, |
|
"loss": 1.3569, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.5942350332594235, |
|
"grad_norm": 0.3406711518764496, |
|
"learning_rate": 0.0001848048096156426, |
|
"loss": 1.3493, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 2.6164079822616406, |
|
"grad_norm": 0.3547060489654541, |
|
"learning_rate": 0.00018439120296645308, |
|
"loss": 1.3702, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.638580931263858, |
|
"grad_norm": 0.3630983531475067, |
|
"learning_rate": 0.000183972518410775, |
|
"loss": 1.3761, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 2.6607538802660753, |
|
"grad_norm": 0.36963704228401184, |
|
"learning_rate": 0.00018354878114129367, |
|
"loss": 1.3614, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.682926829268293, |
|
"grad_norm": 0.3835570812225342, |
|
"learning_rate": 0.00018312001665472146, |
|
"loss": 1.3464, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 2.70509977827051, |
|
"grad_norm": 0.33551275730133057, |
|
"learning_rate": 0.00018268625075026375, |
|
"loss": 1.338, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.7272727272727275, |
|
"grad_norm": 0.34833574295043945, |
|
"learning_rate": 0.00018224750952806624, |
|
"loss": 1.3413, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 2.7494456762749446, |
|
"grad_norm": 0.3694487512111664, |
|
"learning_rate": 0.00018180381938764478, |
|
"loss": 1.3616, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.7716186252771617, |
|
"grad_norm": 0.36651611328125, |
|
"learning_rate": 0.00018135520702629675, |
|
"loss": 1.3556, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 2.7937915742793793, |
|
"grad_norm": 0.36336463689804077, |
|
"learning_rate": 0.00018090169943749476, |
|
"loss": 1.3505, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.8159645232815964, |
|
"grad_norm": 0.3422740399837494, |
|
"learning_rate": 0.00018044332390926223, |
|
"loss": 1.3473, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 2.8381374722838135, |
|
"grad_norm": 0.32323065400123596, |
|
"learning_rate": 0.0001799801080225316, |
|
"loss": 1.3328, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.860310421286031, |
|
"grad_norm": 0.32829082012176514, |
|
"learning_rate": 0.0001795120796494848, |
|
"loss": 1.3483, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 2.882483370288248, |
|
"grad_norm": 0.3510291576385498, |
|
"learning_rate": 0.00017903926695187595, |
|
"loss": 1.3489, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.9046563192904657, |
|
"grad_norm": 0.36246925592422485, |
|
"learning_rate": 0.000178561698379337, |
|
"loss": 1.3463, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 2.926829268292683, |
|
"grad_norm": 0.3481374680995941, |
|
"learning_rate": 0.00017807940266766593, |
|
"loss": 1.358, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.9490022172949004, |
|
"grad_norm": 0.32869285345077515, |
|
"learning_rate": 0.00017759240883709744, |
|
"loss": 1.3644, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 2.9711751662971175, |
|
"grad_norm": 0.3213963806629181, |
|
"learning_rate": 0.00017710074619055707, |
|
"loss": 1.3428, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.9933481152993346, |
|
"grad_norm": 0.3256833553314209, |
|
"learning_rate": 0.0001766044443118978, |
|
"loss": 1.3491, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 2.9977827050997785, |
|
"eval_loss": 1.7484476566314697, |
|
"eval_runtime": 0.3309, |
|
"eval_samples_per_second": 3.022, |
|
"eval_steps_per_second": 3.022, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 3.015521064301552, |
|
"grad_norm": 0.34128889441490173, |
|
"learning_rate": 0.00017610353306412007, |
|
"loss": 1.3328, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 3.0376940133037693, |
|
"grad_norm": 0.34881681203842163, |
|
"learning_rate": 0.0001755980425875748, |
|
"loss": 1.3193, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 3.059866962305987, |
|
"grad_norm": 0.3570633828639984, |
|
"learning_rate": 0.00017508800329814995, |
|
"loss": 1.3245, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 3.082039911308204, |
|
"grad_norm": 0.33681634068489075, |
|
"learning_rate": 0.00017457344588544017, |
|
"loss": 1.3074, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 3.104212860310421, |
|
"grad_norm": 0.34801220893859863, |
|
"learning_rate": 0.00017405440131090048, |
|
"loss": 1.3033, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 3.1263858093126387, |
|
"grad_norm": 0.34218576550483704, |
|
"learning_rate": 0.0001735309008059829, |
|
"loss": 1.3135, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 3.1485587583148558, |
|
"grad_norm": 0.37351498007774353, |
|
"learning_rate": 0.00017300297587025747, |
|
"loss": 1.3212, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 3.1707317073170733, |
|
"grad_norm": 0.3948318064212799, |
|
"learning_rate": 0.00017247065826951695, |
|
"loss": 1.3142, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 3.1929046563192904, |
|
"grad_norm": 0.35082143545150757, |
|
"learning_rate": 0.0001719339800338651, |
|
"loss": 1.3088, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 3.2150776053215075, |
|
"grad_norm": 0.34117066860198975, |
|
"learning_rate": 0.00017139297345578994, |
|
"loss": 1.3219, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 3.237250554323725, |
|
"grad_norm": 0.35672473907470703, |
|
"learning_rate": 0.00017084767108822, |
|
"loss": 1.3296, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 3.259423503325942, |
|
"grad_norm": 0.37271085381507874, |
|
"learning_rate": 0.0001702981057425662, |
|
"loss": 1.3428, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 3.2815964523281598, |
|
"grad_norm": 0.38227108120918274, |
|
"learning_rate": 0.00016974431048674715, |
|
"loss": 1.327, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 3.303769401330377, |
|
"grad_norm": 0.34224027395248413, |
|
"learning_rate": 0.00016918631864319957, |
|
"loss": 1.3148, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 3.3259423503325944, |
|
"grad_norm": 0.3518960177898407, |
|
"learning_rate": 0.0001686241637868734, |
|
"loss": 1.3175, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 3.3481152993348116, |
|
"grad_norm": 0.3325146734714508, |
|
"learning_rate": 0.00016805787974321105, |
|
"loss": 1.2974, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 3.3702882483370287, |
|
"grad_norm": 0.34566330909729004, |
|
"learning_rate": 0.00016748750058611278, |
|
"loss": 1.3285, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 3.3924611973392462, |
|
"grad_norm": 0.4051843285560608, |
|
"learning_rate": 0.00016691306063588583, |
|
"loss": 1.3045, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 3.4146341463414633, |
|
"grad_norm": 0.36097583174705505, |
|
"learning_rate": 0.00016633459445717974, |
|
"loss": 1.3254, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 3.436807095343681, |
|
"grad_norm": 0.36703112721443176, |
|
"learning_rate": 0.0001657521368569064, |
|
"loss": 1.3131, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 3.458980044345898, |
|
"grad_norm": 0.3478531539440155, |
|
"learning_rate": 0.00016516572288214552, |
|
"loss": 1.3242, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 3.481152993348115, |
|
"grad_norm": 0.34251999855041504, |
|
"learning_rate": 0.00016457538781803623, |
|
"loss": 1.3092, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 3.5033259423503327, |
|
"grad_norm": 0.3575889766216278, |
|
"learning_rate": 0.00016398116718565348, |
|
"loss": 1.3022, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 3.52549889135255, |
|
"grad_norm": 0.36192360520362854, |
|
"learning_rate": 0.00016338309673987101, |
|
"loss": 1.316, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 3.5476718403547673, |
|
"grad_norm": 0.3578505218029022, |
|
"learning_rate": 0.00016278121246720987, |
|
"loss": 1.3237, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 3.5698447893569845, |
|
"grad_norm": 0.34649041295051575, |
|
"learning_rate": 0.0001621755505836729, |
|
"loss": 1.317, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 3.5920177383592016, |
|
"grad_norm": 0.36542415618896484, |
|
"learning_rate": 0.0001615661475325658, |
|
"loss": 1.3146, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 3.614190687361419, |
|
"grad_norm": 0.352225124835968, |
|
"learning_rate": 0.00016095303998230433, |
|
"loss": 1.3146, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 3.6363636363636362, |
|
"grad_norm": 0.336407333612442, |
|
"learning_rate": 0.00016033626482420758, |
|
"loss": 1.3344, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 3.658536585365854, |
|
"grad_norm": 0.3517133295536041, |
|
"learning_rate": 0.00015971585917027862, |
|
"loss": 1.3175, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 3.680709534368071, |
|
"grad_norm": 0.3601805567741394, |
|
"learning_rate": 0.00015909186035097111, |
|
"loss": 1.3143, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 3.7028824833702885, |
|
"grad_norm": 0.33734479546546936, |
|
"learning_rate": 0.0001584643059129433, |
|
"loss": 1.3187, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 3.7250554323725056, |
|
"grad_norm": 0.3447887897491455, |
|
"learning_rate": 0.00015783323361679864, |
|
"loss": 1.3058, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 3.7472283813747227, |
|
"grad_norm": 0.332766056060791, |
|
"learning_rate": 0.00015719868143481384, |
|
"loss": 1.3096, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 3.7694013303769403, |
|
"grad_norm": 0.348834753036499, |
|
"learning_rate": 0.00015656068754865387, |
|
"loss": 1.3168, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 3.7915742793791574, |
|
"grad_norm": 0.35416728258132935, |
|
"learning_rate": 0.0001559192903470747, |
|
"loss": 1.3226, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 3.8137472283813745, |
|
"grad_norm": 0.360474556684494, |
|
"learning_rate": 0.00015527452842361327, |
|
"loss": 1.3109, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 3.835920177383592, |
|
"grad_norm": 0.33958253264427185, |
|
"learning_rate": 0.0001546264405742654, |
|
"loss": 1.3182, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 3.858093126385809, |
|
"grad_norm": 0.35495725274086, |
|
"learning_rate": 0.0001539750657951513, |
|
"loss": 1.3166, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 3.8802660753880267, |
|
"grad_norm": 0.3402771055698395, |
|
"learning_rate": 0.00015332044328016914, |
|
"loss": 1.3271, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 3.902439024390244, |
|
"grad_norm": 0.3348073661327362, |
|
"learning_rate": 0.00015266261241863674, |
|
"loss": 1.3155, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 3.9246119733924614, |
|
"grad_norm": 0.3448599576950073, |
|
"learning_rate": 0.00015200161279292155, |
|
"loss": 1.3353, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 3.9467849223946785, |
|
"grad_norm": 0.34724217653274536, |
|
"learning_rate": 0.00015133748417605876, |
|
"loss": 1.3302, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 3.9689578713968956, |
|
"grad_norm": 0.3367406725883484, |
|
"learning_rate": 0.00015067026652935823, |
|
"loss": 1.3155, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 3.991130820399113, |
|
"grad_norm": 0.33979812264442444, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 1.3025, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 1.7902376651763916, |
|
"eval_runtime": 0.3268, |
|
"eval_samples_per_second": 3.06, |
|
"eval_steps_per_second": 3.06, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 4.013303769401331, |
|
"grad_norm": 0.3638916313648224, |
|
"learning_rate": 0.00014932672491861854, |
|
"loss": 1.2902, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 4.035476718403547, |
|
"grad_norm": 0.33381396532058716, |
|
"learning_rate": 0.000148650481796876, |
|
"loss": 1.2739, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 4.057649667405765, |
|
"grad_norm": 0.3387119472026825, |
|
"learning_rate": 0.00014797131132502465, |
|
"loss": 1.2744, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 4.0798226164079825, |
|
"grad_norm": 0.3494380712509155, |
|
"learning_rate": 0.00014728925436945838, |
|
"loss": 1.2835, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 4.101995565410199, |
|
"grad_norm": 0.3876653015613556, |
|
"learning_rate": 0.0001466043519702539, |
|
"loss": 1.294, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 4.124168514412417, |
|
"grad_norm": 0.3451765477657318, |
|
"learning_rate": 0.00014591664533870118, |
|
"loss": 1.2773, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 4.146341463414634, |
|
"grad_norm": 0.34489771723747253, |
|
"learning_rate": 0.00014522617585482377, |
|
"loss": 1.2798, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 4.168514412416852, |
|
"grad_norm": 0.3645501136779785, |
|
"learning_rate": 0.00014453298506488896, |
|
"loss": 1.2822, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 4.1906873614190685, |
|
"grad_norm": 0.3511221706867218, |
|
"learning_rate": 0.00014383711467890774, |
|
"loss": 1.2959, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 4.212860310421286, |
|
"grad_norm": 0.34734046459198, |
|
"learning_rate": 0.00014313860656812536, |
|
"loss": 1.2804, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 4.235033259423504, |
|
"grad_norm": 0.34553733468055725, |
|
"learning_rate": 0.00014243750276250153, |
|
"loss": 1.2938, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 4.25720620842572, |
|
"grad_norm": 0.3475314676761627, |
|
"learning_rate": 0.0001417338454481818, |
|
"loss": 1.2882, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 4.279379157427938, |
|
"grad_norm": 0.3424859941005707, |
|
"learning_rate": 0.00014102767696495884, |
|
"loss": 1.2957, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 4.301552106430155, |
|
"grad_norm": 0.35711121559143066, |
|
"learning_rate": 0.00014031903980372504, |
|
"loss": 1.3019, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 4.323725055432373, |
|
"grad_norm": 0.3883892595767975, |
|
"learning_rate": 0.0001396079766039157, |
|
"loss": 1.2627, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 4.34589800443459, |
|
"grad_norm": 0.3351362347602844, |
|
"learning_rate": 0.00013889453015094338, |
|
"loss": 1.2894, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 4.368070953436807, |
|
"grad_norm": 0.35356107354164124, |
|
"learning_rate": 0.0001381787433736235, |
|
"loss": 1.2972, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 4.390243902439025, |
|
"grad_norm": 0.34767022728919983, |
|
"learning_rate": 0.00013746065934159123, |
|
"loss": 1.2728, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 4.412416851441241, |
|
"grad_norm": 0.35094380378723145, |
|
"learning_rate": 0.00013674032126270982, |
|
"loss": 1.2875, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 4.434589800443459, |
|
"grad_norm": 0.342939555644989, |
|
"learning_rate": 0.00013601777248047105, |
|
"loss": 1.2851, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 4.4567627494456765, |
|
"grad_norm": 0.37955859303474426, |
|
"learning_rate": 0.00013529305647138687, |
|
"loss": 1.2848, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 4.478935698447893, |
|
"grad_norm": 0.3524990677833557, |
|
"learning_rate": 0.00013456621684237367, |
|
"loss": 1.2889, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 4.501108647450111, |
|
"grad_norm": 0.354220986366272, |
|
"learning_rate": 0.00013383729732812814, |
|
"loss": 1.2989, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 4.523281596452328, |
|
"grad_norm": 0.34707173705101013, |
|
"learning_rate": 0.0001331063417884958, |
|
"loss": 1.2899, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 4.545454545454545, |
|
"grad_norm": 0.358018159866333, |
|
"learning_rate": 0.00013237339420583212, |
|
"loss": 1.2926, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 4.5676274944567625, |
|
"grad_norm": 0.3700306713581085, |
|
"learning_rate": 0.00013163849868235564, |
|
"loss": 1.2942, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 4.58980044345898, |
|
"grad_norm": 0.38958704471588135, |
|
"learning_rate": 0.00013090169943749476, |
|
"loss": 1.2684, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 4.611973392461198, |
|
"grad_norm": 0.34153124690055847, |
|
"learning_rate": 0.00013016304080522656, |
|
"loss": 1.2885, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 4.634146341463414, |
|
"grad_norm": 0.3536767363548279, |
|
"learning_rate": 0.00012942256723140952, |
|
"loss": 1.299, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 4.656319290465632, |
|
"grad_norm": 0.34350860118865967, |
|
"learning_rate": 0.00012868032327110904, |
|
"loss": 1.2869, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 4.678492239467849, |
|
"grad_norm": 0.3380618691444397, |
|
"learning_rate": 0.00012793635358591645, |
|
"loss": 1.2935, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 4.700665188470067, |
|
"grad_norm": 0.3540153205394745, |
|
"learning_rate": 0.00012719070294126182, |
|
"loss": 1.2875, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 4.722838137472284, |
|
"grad_norm": 0.33930808305740356, |
|
"learning_rate": 0.00012644341620372023, |
|
"loss": 1.285, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 4.745011086474501, |
|
"grad_norm": 0.38791534304618835, |
|
"learning_rate": 0.00012569453833831222, |
|
"loss": 1.2695, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 4.767184035476719, |
|
"grad_norm": 0.34770122170448303, |
|
"learning_rate": 0.00012494411440579814, |
|
"loss": 1.3086, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 4.789356984478935, |
|
"grad_norm": 0.3471126854419708, |
|
"learning_rate": 0.00012419218955996676, |
|
"loss": 1.2935, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 4.811529933481153, |
|
"grad_norm": 0.33490991592407227, |
|
"learning_rate": 0.00012343880904491848, |
|
"loss": 1.2831, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 4.8337028824833705, |
|
"grad_norm": 0.3470149040222168, |
|
"learning_rate": 0.0001226840181923427, |
|
"loss": 1.2775, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 4.855875831485587, |
|
"grad_norm": 0.33513540029525757, |
|
"learning_rate": 0.00012192786241879033, |
|
"loss": 1.2868, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 4.878048780487805, |
|
"grad_norm": 0.3534790873527527, |
|
"learning_rate": 0.0001211703872229411, |
|
"loss": 1.3015, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 4.900221729490022, |
|
"grad_norm": 0.42340680956840515, |
|
"learning_rate": 0.00012041163818286559, |
|
"loss": 1.2753, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 4.922394678492239, |
|
"grad_norm": 0.343871146440506, |
|
"learning_rate": 0.00011965166095328301, |
|
"loss": 1.2776, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 4.9445676274944566, |
|
"grad_norm": 0.3540928363800049, |
|
"learning_rate": 0.00011889050126281405, |
|
"loss": 1.3005, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 4.966740576496674, |
|
"grad_norm": 0.3478156626224518, |
|
"learning_rate": 0.00011812820491122918, |
|
"loss": 1.2888, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 4.988913525498892, |
|
"grad_norm": 0.35215553641319275, |
|
"learning_rate": 0.00011736481776669306, |
|
"loss": 1.2904, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 4.9977827050997785, |
|
"eval_loss": 1.79966139793396, |
|
"eval_runtime": 0.3309, |
|
"eval_samples_per_second": 3.022, |
|
"eval_steps_per_second": 3.022, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 5.011086474501108, |
|
"grad_norm": 0.36425259709358215, |
|
"learning_rate": 0.00011660038576300443, |
|
"loss": 1.2833, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 5.033259423503326, |
|
"grad_norm": 0.36249682307243347, |
|
"learning_rate": 0.00011583495489683229, |
|
"loss": 1.2641, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 5.0554323725055434, |
|
"grad_norm": 0.35734277963638306, |
|
"learning_rate": 0.00011506857122494831, |
|
"loss": 1.2565, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 5.07760532150776, |
|
"grad_norm": 0.3540472686290741, |
|
"learning_rate": 0.00011430128086145542, |
|
"loss": 1.2788, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 5.099778270509978, |
|
"grad_norm": 0.3659023940563202, |
|
"learning_rate": 0.00011353312997501313, |
|
"loss": 1.2464, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 5.121951219512195, |
|
"grad_norm": 0.3884827494621277, |
|
"learning_rate": 0.00011276416478605949, |
|
"loss": 1.2621, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 5.144124168514413, |
|
"grad_norm": 0.35218942165374756, |
|
"learning_rate": 0.00011199443156402998, |
|
"loss": 1.2643, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 5.1662971175166295, |
|
"grad_norm": 0.35376182198524475, |
|
"learning_rate": 0.0001112239766245735, |
|
"loss": 1.25, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 5.188470066518847, |
|
"grad_norm": 0.3779812455177307, |
|
"learning_rate": 0.00011045284632676536, |
|
"loss": 1.2623, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 5.210643015521065, |
|
"grad_norm": 0.40983307361602783, |
|
"learning_rate": 0.00010968108707031792, |
|
"loss": 1.2573, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 5.232815964523281, |
|
"grad_norm": 0.36376869678497314, |
|
"learning_rate": 0.00010890874529278865, |
|
"loss": 1.2556, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 5.254988913525499, |
|
"grad_norm": 0.36532121896743774, |
|
"learning_rate": 0.00010813586746678583, |
|
"loss": 1.2672, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 5.277161862527716, |
|
"grad_norm": 0.3567180633544922, |
|
"learning_rate": 0.00010736250009717247, |
|
"loss": 1.2623, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 5.299334811529933, |
|
"grad_norm": 0.3723650276660919, |
|
"learning_rate": 0.00010658868971826785, |
|
"loss": 1.2609, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 5.321507760532151, |
|
"grad_norm": 0.36513036489486694, |
|
"learning_rate": 0.00010581448289104758, |
|
"loss": 1.2605, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 5.343680709534368, |
|
"grad_norm": 0.34975701570510864, |
|
"learning_rate": 0.00010503992620034202, |
|
"loss": 1.258, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 5.365853658536586, |
|
"grad_norm": 0.3734095096588135, |
|
"learning_rate": 0.00010426506625203307, |
|
"loss": 1.2722, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 5.388026607538802, |
|
"grad_norm": 0.3856619596481323, |
|
"learning_rate": 0.00010348994967025012, |
|
"loss": 1.2637, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 5.41019955654102, |
|
"grad_norm": 0.37565699219703674, |
|
"learning_rate": 0.0001027146230945643, |
|
"loss": 1.2559, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 5.4323725055432375, |
|
"grad_norm": 0.37365949153900146, |
|
"learning_rate": 0.00010193913317718244, |
|
"loss": 1.2604, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 5.454545454545454, |
|
"grad_norm": 0.35709646344184875, |
|
"learning_rate": 0.00010116352658013973, |
|
"loss": 1.2677, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 5.476718403547672, |
|
"grad_norm": 0.3517821133136749, |
|
"learning_rate": 0.00010038784997249205, |
|
"loss": 1.2496, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 5.498891352549889, |
|
"grad_norm": 0.37304961681365967, |
|
"learning_rate": 9.961215002750799e-05, |
|
"loss": 1.2818, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 5.521064301552107, |
|
"grad_norm": 0.3745460510253906, |
|
"learning_rate": 9.883647341986032e-05, |
|
"loss": 1.2817, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 5.5432372505543235, |
|
"grad_norm": 0.35062116384506226, |
|
"learning_rate": 9.806086682281758e-05, |
|
"loss": 1.2515, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 5.565410199556541, |
|
"grad_norm": 0.3595265746116638, |
|
"learning_rate": 9.728537690543572e-05, |
|
"loss": 1.2665, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 5.587583148558759, |
|
"grad_norm": 0.36035382747650146, |
|
"learning_rate": 9.651005032974994e-05, |
|
"loss": 1.2578, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 5.609756097560975, |
|
"grad_norm": 0.3572876751422882, |
|
"learning_rate": 9.573493374796693e-05, |
|
"loss": 1.2488, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 5.631929046563193, |
|
"grad_norm": 0.3515053391456604, |
|
"learning_rate": 9.496007379965801e-05, |
|
"loss": 1.2483, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 5.65410199556541, |
|
"grad_norm": 0.36254850029945374, |
|
"learning_rate": 9.418551710895243e-05, |
|
"loss": 1.2709, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 5.676274944567627, |
|
"grad_norm": 0.36927610635757446, |
|
"learning_rate": 9.341131028173214e-05, |
|
"loss": 1.2587, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 5.698447893569845, |
|
"grad_norm": 0.38191670179367065, |
|
"learning_rate": 9.263749990282754e-05, |
|
"loss": 1.2712, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 5.720620842572062, |
|
"grad_norm": 0.376602441072464, |
|
"learning_rate": 9.186413253321418e-05, |
|
"loss": 1.2758, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 5.74279379157428, |
|
"grad_norm": 0.3506106734275818, |
|
"learning_rate": 9.10912547072114e-05, |
|
"loss": 1.2704, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 5.764966740576496, |
|
"grad_norm": 0.38899239897727966, |
|
"learning_rate": 9.03189129296821e-05, |
|
"loss": 1.2454, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 5.787139689578714, |
|
"grad_norm": 0.34721580147743225, |
|
"learning_rate": 8.954715367323468e-05, |
|
"loss": 1.2823, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 5.8093126385809315, |
|
"grad_norm": 0.37420088052749634, |
|
"learning_rate": 8.877602337542655e-05, |
|
"loss": 1.2583, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 5.831485587583149, |
|
"grad_norm": 0.3575841784477234, |
|
"learning_rate": 8.800556843597002e-05, |
|
"loss": 1.2654, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 5.853658536585366, |
|
"grad_norm": 0.36751991510391235, |
|
"learning_rate": 8.723583521394054e-05, |
|
"loss": 1.259, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 5.875831485587583, |
|
"grad_norm": 0.41278401017189026, |
|
"learning_rate": 8.646687002498692e-05, |
|
"loss": 1.2639, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 5.898004434589801, |
|
"grad_norm": 0.36573198437690735, |
|
"learning_rate": 8.569871913854458e-05, |
|
"loss": 1.2682, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 5.9201773835920175, |
|
"grad_norm": 0.3588218092918396, |
|
"learning_rate": 8.49314287750517e-05, |
|
"loss": 1.2461, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 5.942350332594235, |
|
"grad_norm": 0.3589507043361664, |
|
"learning_rate": 8.416504510316773e-05, |
|
"loss": 1.2569, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 5.964523281596453, |
|
"grad_norm": 0.3812873959541321, |
|
"learning_rate": 8.339961423699562e-05, |
|
"loss": 1.2496, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 5.986696230598669, |
|
"grad_norm": 0.35918959975242615, |
|
"learning_rate": 8.263518223330697e-05, |
|
"loss": 1.2729, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 1.817025065422058, |
|
"eval_runtime": 0.3269, |
|
"eval_samples_per_second": 3.059, |
|
"eval_steps_per_second": 3.059, |
|
"step": 1353 |
|
}, |
|
{ |
|
"epoch": 6.008869179600887, |
|
"grad_norm": 0.365013062953949, |
|
"learning_rate": 8.187179508877085e-05, |
|
"loss": 1.2526, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 6.031042128603104, |
|
"grad_norm": 0.3869335949420929, |
|
"learning_rate": 8.1109498737186e-05, |
|
"loss": 1.2482, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 6.053215077605321, |
|
"grad_norm": 0.3877875804901123, |
|
"learning_rate": 8.034833904671698e-05, |
|
"loss": 1.2583, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 6.075388026607539, |
|
"grad_norm": 0.3739064931869507, |
|
"learning_rate": 7.958836181713445e-05, |
|
"loss": 1.2355, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 6.097560975609756, |
|
"grad_norm": 0.37528711557388306, |
|
"learning_rate": 7.882961277705895e-05, |
|
"loss": 1.2267, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 6.119733924611974, |
|
"grad_norm": 0.36762019991874695, |
|
"learning_rate": 7.807213758120966e-05, |
|
"loss": 1.2174, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 6.14190687361419, |
|
"grad_norm": 0.36748361587524414, |
|
"learning_rate": 7.731598180765732e-05, |
|
"loss": 1.2372, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 6.164079822616408, |
|
"grad_norm": 0.3874417245388031, |
|
"learning_rate": 7.656119095508154e-05, |
|
"loss": 1.2373, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 6.1862527716186255, |
|
"grad_norm": 0.37527891993522644, |
|
"learning_rate": 7.580781044003324e-05, |
|
"loss": 1.2413, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 6.208425720620842, |
|
"grad_norm": 0.35388851165771484, |
|
"learning_rate": 7.505588559420189e-05, |
|
"loss": 1.2348, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 6.23059866962306, |
|
"grad_norm": 0.37881141901016235, |
|
"learning_rate": 7.43054616616878e-05, |
|
"loss": 1.2617, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 6.252771618625277, |
|
"grad_norm": 0.36428573727607727, |
|
"learning_rate": 7.35565837962798e-05, |
|
"loss": 1.2363, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 6.274944567627495, |
|
"grad_norm": 0.373054176568985, |
|
"learning_rate": 7.280929705873818e-05, |
|
"loss": 1.2405, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 6.2971175166297115, |
|
"grad_norm": 0.3633226454257965, |
|
"learning_rate": 7.206364641408357e-05, |
|
"loss": 1.2453, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 6.319290465631929, |
|
"grad_norm": 0.3570459485054016, |
|
"learning_rate": 7.131967672889101e-05, |
|
"loss": 1.2399, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 6.341463414634147, |
|
"grad_norm": 0.3895583748817444, |
|
"learning_rate": 7.057743276859048e-05, |
|
"loss": 1.2508, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 6.363636363636363, |
|
"grad_norm": 0.3734217882156372, |
|
"learning_rate": 6.983695919477345e-05, |
|
"loss": 1.2461, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 6.385809312638581, |
|
"grad_norm": 0.35446175932884216, |
|
"learning_rate": 6.909830056250527e-05, |
|
"loss": 1.2571, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 6.407982261640798, |
|
"grad_norm": 0.35295209288597107, |
|
"learning_rate": 6.836150131764434e-05, |
|
"loss": 1.2395, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 6.430155210643015, |
|
"grad_norm": 0.38218680024147034, |
|
"learning_rate": 6.762660579416791e-05, |
|
"loss": 1.234, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 6.452328159645233, |
|
"grad_norm": 0.38666704297065735, |
|
"learning_rate": 6.68936582115042e-05, |
|
"loss": 1.2461, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 6.47450110864745, |
|
"grad_norm": 0.35704880952835083, |
|
"learning_rate": 6.61627026718719e-05, |
|
"loss": 1.236, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 6.496674057649668, |
|
"grad_norm": 0.37637725472450256, |
|
"learning_rate": 6.543378315762634e-05, |
|
"loss": 1.2354, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 6.518847006651884, |
|
"grad_norm": 0.37539881467819214, |
|
"learning_rate": 6.470694352861312e-05, |
|
"loss": 1.2471, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 6.541019955654102, |
|
"grad_norm": 0.3633996844291687, |
|
"learning_rate": 6.398222751952899e-05, |
|
"loss": 1.2298, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 6.5631929046563195, |
|
"grad_norm": 0.35865873098373413, |
|
"learning_rate": 6.325967873729018e-05, |
|
"loss": 1.2382, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 6.585365853658536, |
|
"grad_norm": 0.37201085686683655, |
|
"learning_rate": 6.25393406584088e-05, |
|
"loss": 1.2366, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 6.607538802660754, |
|
"grad_norm": 0.38256916403770447, |
|
"learning_rate": 6.18212566263765e-05, |
|
"loss": 1.2415, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 6.629711751662971, |
|
"grad_norm": 0.3845401704311371, |
|
"learning_rate": 6.110546984905661e-05, |
|
"loss": 1.2381, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 6.651884700665189, |
|
"grad_norm": 0.36959657073020935, |
|
"learning_rate": 6.039202339608432e-05, |
|
"loss": 1.2406, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 6.674057649667406, |
|
"grad_norm": 0.3971165120601654, |
|
"learning_rate": 5.9680960196274994e-05, |
|
"loss": 1.2421, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 6.696230598669623, |
|
"grad_norm": 0.3609052002429962, |
|
"learning_rate": 5.89723230350412e-05, |
|
"loss": 1.2359, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 6.718403547671841, |
|
"grad_norm": 0.3867206871509552, |
|
"learning_rate": 5.8266154551818216e-05, |
|
"loss": 1.2419, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 6.740576496674057, |
|
"grad_norm": 0.35991984605789185, |
|
"learning_rate": 5.756249723749847e-05, |
|
"loss": 1.2379, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 6.762749445676275, |
|
"grad_norm": 0.3713914752006531, |
|
"learning_rate": 5.6861393431874675e-05, |
|
"loss": 1.2346, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 6.7849223946784925, |
|
"grad_norm": 0.37790387868881226, |
|
"learning_rate": 5.616288532109225e-05, |
|
"loss": 1.253, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 6.807095343680709, |
|
"grad_norm": 0.39418914914131165, |
|
"learning_rate": 5.546701493511106e-05, |
|
"loss": 1.2572, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 6.829268292682927, |
|
"grad_norm": 0.3760906159877777, |
|
"learning_rate": 5.477382414517624e-05, |
|
"loss": 1.2436, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 6.851441241685144, |
|
"grad_norm": 0.37196049094200134, |
|
"learning_rate": 5.4083354661298814e-05, |
|
"loss": 1.2525, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 6.873614190687362, |
|
"grad_norm": 0.37352004647254944, |
|
"learning_rate": 5.339564802974615e-05, |
|
"loss": 1.2381, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 6.8957871396895785, |
|
"grad_norm": 0.3716587722301483, |
|
"learning_rate": 5.2710745630541666e-05, |
|
"loss": 1.2383, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 6.917960088691796, |
|
"grad_norm": 0.3760850131511688, |
|
"learning_rate": 5.2028688674975415e-05, |
|
"loss": 1.244, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 6.940133037694014, |
|
"grad_norm": 0.36280307173728943, |
|
"learning_rate": 5.134951820312401e-05, |
|
"loss": 1.2391, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 6.96230598669623, |
|
"grad_norm": 0.36244943737983704, |
|
"learning_rate": 5.0673275081381475e-05, |
|
"loss": 1.2465, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 6.984478935698448, |
|
"grad_norm": 0.3694084584712982, |
|
"learning_rate": 5.000000000000002e-05, |
|
"loss": 1.2451, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 6.9977827050997785, |
|
"eval_loss": 1.8179672956466675, |
|
"eval_runtime": 0.3324, |
|
"eval_samples_per_second": 3.008, |
|
"eval_steps_per_second": 3.008, |
|
"step": 1578 |
|
}, |
|
{ |
|
"epoch": 7.006651884700665, |
|
"grad_norm": 0.3736945688724518, |
|
"learning_rate": 4.932973347064177e-05, |
|
"loss": 1.2391, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 7.028824833702883, |
|
"grad_norm": 0.40935948491096497, |
|
"learning_rate": 4.8662515823941255e-05, |
|
"loss": 1.2378, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 7.0509977827051, |
|
"grad_norm": 0.37913885712623596, |
|
"learning_rate": 4.799838720707846e-05, |
|
"loss": 1.2151, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 7.073170731707317, |
|
"grad_norm": 0.3685004413127899, |
|
"learning_rate": 4.733738758136327e-05, |
|
"loss": 1.2104, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 7.095343680709535, |
|
"grad_norm": 0.3836255669593811, |
|
"learning_rate": 4.66795567198309e-05, |
|
"loss": 1.22, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 7.117516629711751, |
|
"grad_norm": 0.4032094478607178, |
|
"learning_rate": 4.6024934204848745e-05, |
|
"loss": 1.2326, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 7.139689578713969, |
|
"grad_norm": 0.39435017108917236, |
|
"learning_rate": 4.537355942573463e-05, |
|
"loss": 1.2165, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 7.1618625277161865, |
|
"grad_norm": 0.4004512131214142, |
|
"learning_rate": 4.4725471576386735e-05, |
|
"loss": 1.2291, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 7.184035476718403, |
|
"grad_norm": 0.37496936321258545, |
|
"learning_rate": 4.4080709652925336e-05, |
|
"loss": 1.221, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 7.206208425720621, |
|
"grad_norm": 0.3921966552734375, |
|
"learning_rate": 4.343931245134616e-05, |
|
"loss": 1.2243, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 7.228381374722838, |
|
"grad_norm": 0.41742879152297974, |
|
"learning_rate": 4.2801318565186165e-05, |
|
"loss": 1.2299, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 7.250554323725056, |
|
"grad_norm": 0.38385558128356934, |
|
"learning_rate": 4.216676638320135e-05, |
|
"loss": 1.2364, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 7.2727272727272725, |
|
"grad_norm": 0.36501455307006836, |
|
"learning_rate": 4.15356940870567e-05, |
|
"loss": 1.2195, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 7.29490022172949, |
|
"grad_norm": 0.38054943084716797, |
|
"learning_rate": 4.090813964902889e-05, |
|
"loss": 1.2293, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 7.317073170731708, |
|
"grad_norm": 0.36994317173957825, |
|
"learning_rate": 4.028414082972141e-05, |
|
"loss": 1.2198, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 7.339246119733924, |
|
"grad_norm": 0.3893408477306366, |
|
"learning_rate": 3.966373517579244e-05, |
|
"loss": 1.2384, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 7.361419068736142, |
|
"grad_norm": 0.3703598082065582, |
|
"learning_rate": 3.904696001769571e-05, |
|
"loss": 1.2348, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 7.383592017738359, |
|
"grad_norm": 0.3875352740287781, |
|
"learning_rate": 3.843385246743417e-05, |
|
"loss": 1.2402, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 7.405764966740577, |
|
"grad_norm": 0.399080365896225, |
|
"learning_rate": 3.7824449416327126e-05, |
|
"loss": 1.2268, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 7.427937915742794, |
|
"grad_norm": 0.37249988317489624, |
|
"learning_rate": 3.721878753279017e-05, |
|
"loss": 1.227, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 7.450110864745011, |
|
"grad_norm": 0.3863629996776581, |
|
"learning_rate": 3.661690326012897e-05, |
|
"loss": 1.22, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 7.472283813747229, |
|
"grad_norm": 0.3829827904701233, |
|
"learning_rate": 3.601883281434652e-05, |
|
"loss": 1.2137, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 7.494456762749445, |
|
"grad_norm": 0.3759503960609436, |
|
"learning_rate": 3.542461218196379e-05, |
|
"loss": 1.2077, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 7.516629711751663, |
|
"grad_norm": 0.38627204298973083, |
|
"learning_rate": 3.483427711785449e-05, |
|
"loss": 1.2233, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 7.5388026607538805, |
|
"grad_norm": 0.38867396116256714, |
|
"learning_rate": 3.424786314309365e-05, |
|
"loss": 1.2226, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 7.560975609756097, |
|
"grad_norm": 0.38329002261161804, |
|
"learning_rate": 3.366540554282028e-05, |
|
"loss": 1.2247, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 7.583148558758315, |
|
"grad_norm": 0.3766213059425354, |
|
"learning_rate": 3.308693936411421e-05, |
|
"loss": 1.2376, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 7.605321507760532, |
|
"grad_norm": 0.3749763071537018, |
|
"learning_rate": 3.2512499413887255e-05, |
|
"loss": 1.2099, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 7.627494456762749, |
|
"grad_norm": 0.3896552324295044, |
|
"learning_rate": 3.194212025678896e-05, |
|
"loss": 1.2119, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 7.6496674057649665, |
|
"grad_norm": 0.37435224652290344, |
|
"learning_rate": 3.137583621312665e-05, |
|
"loss": 1.229, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 7.671840354767184, |
|
"grad_norm": 0.3764979839324951, |
|
"learning_rate": 3.0813681356800405e-05, |
|
"loss": 1.2309, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 7.694013303769402, |
|
"grad_norm": 0.39203503727912903, |
|
"learning_rate": 3.025568951325287e-05, |
|
"loss": 1.2204, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 7.716186252771618, |
|
"grad_norm": 0.3817265033721924, |
|
"learning_rate": 2.9701894257433826e-05, |
|
"loss": 1.2297, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 7.738359201773836, |
|
"grad_norm": 0.38842707872390747, |
|
"learning_rate": 2.9152328911780026e-05, |
|
"loss": 1.2134, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 7.760532150776053, |
|
"grad_norm": 0.38738590478897095, |
|
"learning_rate": 2.8607026544210114e-05, |
|
"loss": 1.235, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 7.782705099778271, |
|
"grad_norm": 0.3651287257671356, |
|
"learning_rate": 2.8066019966134904e-05, |
|
"loss": 1.2327, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 7.804878048780488, |
|
"grad_norm": 0.3861520290374756, |
|
"learning_rate": 2.7529341730483117e-05, |
|
"loss": 1.2271, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 7.827050997782705, |
|
"grad_norm": 0.39386945962905884, |
|
"learning_rate": 2.6997024129742542e-05, |
|
"loss": 1.2255, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 7.849223946784923, |
|
"grad_norm": 0.3876512944698334, |
|
"learning_rate": 2.6469099194017143e-05, |
|
"loss": 1.2323, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 7.871396895787139, |
|
"grad_norm": 0.38528311252593994, |
|
"learning_rate": 2.594559868909956e-05, |
|
"loss": 1.2244, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 7.893569844789357, |
|
"grad_norm": 0.3802613615989685, |
|
"learning_rate": 2.542655411455982e-05, |
|
"loss": 1.2207, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 7.9157427937915745, |
|
"grad_norm": 0.3857831060886383, |
|
"learning_rate": 2.491199670185008e-05, |
|
"loss": 1.2183, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 7.937915742793791, |
|
"grad_norm": 0.3909415304660797, |
|
"learning_rate": 2.4401957412425214e-05, |
|
"loss": 1.2336, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 7.960088691796009, |
|
"grad_norm": 0.3930368423461914, |
|
"learning_rate": 2.389646693587996e-05, |
|
"loss": 1.2278, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 7.982261640798226, |
|
"grad_norm": 0.37647300958633423, |
|
"learning_rate": 2.339555568810221e-05, |
|
"loss": 1.229, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 1.8372095823287964, |
|
"eval_runtime": 0.3265, |
|
"eval_samples_per_second": 3.063, |
|
"eval_steps_per_second": 3.063, |
|
"step": 1804 |
|
}, |
|
{ |
|
"epoch": 8.004434589800443, |
|
"grad_norm": 0.3868635594844818, |
|
"learning_rate": 2.2899253809442944e-05, |
|
"loss": 1.233, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 8.026607538802661, |
|
"grad_norm": 0.37865373492240906, |
|
"learning_rate": 2.2407591162902573e-05, |
|
"loss": 1.2044, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 8.048780487804878, |
|
"grad_norm": 0.3961890935897827, |
|
"learning_rate": 2.192059733233408e-05, |
|
"loss": 1.2192, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 8.070953436807095, |
|
"grad_norm": 0.3783002197742462, |
|
"learning_rate": 2.1438301620662993e-05, |
|
"loss": 1.2196, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 8.093126385809313, |
|
"grad_norm": 0.3858291506767273, |
|
"learning_rate": 2.0960733048124083e-05, |
|
"loss": 1.2285, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 8.11529933481153, |
|
"grad_norm": 0.37426653504371643, |
|
"learning_rate": 2.0487920350515212e-05, |
|
"loss": 1.2164, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 8.137472283813747, |
|
"grad_norm": 0.3853384554386139, |
|
"learning_rate": 2.0019891977468408e-05, |
|
"loss": 1.2119, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 8.159645232815965, |
|
"grad_norm": 0.3728988766670227, |
|
"learning_rate": 1.95566760907378e-05, |
|
"loss": 1.2032, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 8.181818181818182, |
|
"grad_norm": 0.38846054673194885, |
|
"learning_rate": 1.9098300562505266e-05, |
|
"loss": 1.2232, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 8.203991130820398, |
|
"grad_norm": 0.38317111134529114, |
|
"learning_rate": 1.864479297370325e-05, |
|
"loss": 1.218, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 8.226164079822617, |
|
"grad_norm": 0.38290783762931824, |
|
"learning_rate": 1.819618061235525e-05, |
|
"loss": 1.2037, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 8.248337028824833, |
|
"grad_norm": 0.38224560022354126, |
|
"learning_rate": 1.775249047193377e-05, |
|
"loss": 1.2055, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 8.27050997782705, |
|
"grad_norm": 0.3762098252773285, |
|
"learning_rate": 1.7313749249736267e-05, |
|
"loss": 1.2073, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 8.292682926829269, |
|
"grad_norm": 0.38844189047813416, |
|
"learning_rate": 1.687998334527853e-05, |
|
"loss": 1.2326, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 8.314855875831485, |
|
"grad_norm": 0.3953261971473694, |
|
"learning_rate": 1.6451218858706374e-05, |
|
"loss": 1.2224, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 8.337028824833704, |
|
"grad_norm": 0.37807291746139526, |
|
"learning_rate": 1.6027481589225026e-05, |
|
"loss": 1.2009, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 8.35920177383592, |
|
"grad_norm": 0.3885143995285034, |
|
"learning_rate": 1.560879703354693e-05, |
|
"loss": 1.224, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 8.381374722838137, |
|
"grad_norm": 0.38399428129196167, |
|
"learning_rate": 1.5195190384357404e-05, |
|
"loss": 1.2082, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 8.403547671840355, |
|
"grad_norm": 0.3910374045372009, |
|
"learning_rate": 1.4786686528798876e-05, |
|
"loss": 1.2202, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 8.425720620842572, |
|
"grad_norm": 0.3843975067138672, |
|
"learning_rate": 1.4383310046973365e-05, |
|
"loss": 1.2202, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 8.447893569844789, |
|
"grad_norm": 0.38334450125694275, |
|
"learning_rate": 1.3985085210463477e-05, |
|
"loss": 1.2099, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 8.470066518847007, |
|
"grad_norm": 0.3954315185546875, |
|
"learning_rate": 1.3592035980871953e-05, |
|
"loss": 1.2161, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 8.492239467849224, |
|
"grad_norm": 0.37927913665771484, |
|
"learning_rate": 1.3204186008379927e-05, |
|
"loss": 1.2153, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 8.51441241685144, |
|
"grad_norm": 0.3803432285785675, |
|
"learning_rate": 1.2821558630323772e-05, |
|
"loss": 1.2016, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 8.536585365853659, |
|
"grad_norm": 0.38970625400543213, |
|
"learning_rate": 1.2444176869790925e-05, |
|
"loss": 1.2011, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 8.558758314855876, |
|
"grad_norm": 0.40996474027633667, |
|
"learning_rate": 1.2072063434234559e-05, |
|
"loss": 1.2238, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 8.580931263858092, |
|
"grad_norm": 0.3958864212036133, |
|
"learning_rate": 1.1705240714107302e-05, |
|
"loss": 1.2321, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 8.60310421286031, |
|
"grad_norm": 0.37584978342056274, |
|
"learning_rate": 1.1343730781513895e-05, |
|
"loss": 1.2138, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 8.625277161862527, |
|
"grad_norm": 0.39090994000434875, |
|
"learning_rate": 1.0987555388883041e-05, |
|
"loss": 1.2095, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 8.647450110864746, |
|
"grad_norm": 0.38422369956970215, |
|
"learning_rate": 1.0636735967658784e-05, |
|
"loss": 1.2115, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 8.669623059866963, |
|
"grad_norm": 0.3975456655025482, |
|
"learning_rate": 1.029129362701068e-05, |
|
"loss": 1.2078, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 8.69179600886918, |
|
"grad_norm": 0.3793989419937134, |
|
"learning_rate": 9.95124915256378e-06, |
|
"loss": 1.2183, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 8.713968957871398, |
|
"grad_norm": 0.3810500502586365, |
|
"learning_rate": 9.616623005147951e-06, |
|
"loss": 1.2218, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 8.736141906873614, |
|
"grad_norm": 0.3820514380931854, |
|
"learning_rate": 9.287435319566618e-06, |
|
"loss": 1.2122, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 8.758314855875831, |
|
"grad_norm": 0.3797374963760376, |
|
"learning_rate": 8.963705903385345e-06, |
|
"loss": 1.2278, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 8.78048780487805, |
|
"grad_norm": 0.376920223236084, |
|
"learning_rate": 8.645454235739903e-06, |
|
"loss": 1.2098, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 8.802660753880266, |
|
"grad_norm": 0.37694764137268066, |
|
"learning_rate": 8.332699466164306e-06, |
|
"loss": 1.2194, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 8.824833702882483, |
|
"grad_norm": 0.382021427154541, |
|
"learning_rate": 8.025460413438457e-06, |
|
"loss": 1.2077, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 8.847006651884701, |
|
"grad_norm": 0.5127786993980408, |
|
"learning_rate": 7.72375556445577e-06, |
|
"loss": 1.203, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 8.869179600886918, |
|
"grad_norm": 0.3939337134361267, |
|
"learning_rate": 7.427603073110967e-06, |
|
"loss": 1.2157, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 8.891352549889135, |
|
"grad_norm": 0.39616692066192627, |
|
"learning_rate": 7.13702075920758e-06, |
|
"loss": 1.2097, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 8.913525498891353, |
|
"grad_norm": 0.3867185115814209, |
|
"learning_rate": 6.852026107385756e-06, |
|
"loss": 1.2186, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 8.93569844789357, |
|
"grad_norm": 0.39439383149147034, |
|
"learning_rate": 6.572636266070264e-06, |
|
"loss": 1.2125, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 8.957871396895786, |
|
"grad_norm": 0.40186840295791626, |
|
"learning_rate": 6.298868046438533e-06, |
|
"loss": 1.214, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 8.980044345898005, |
|
"grad_norm": 0.45572689175605774, |
|
"learning_rate": 6.030737921409169e-06, |
|
"loss": 1.2239, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 8.997782705099779, |
|
"eval_loss": 1.848176121711731, |
|
"eval_runtime": 0.5386, |
|
"eval_samples_per_second": 1.857, |
|
"eval_steps_per_second": 1.857, |
|
"step": 2029 |
|
}, |
|
{ |
|
"epoch": 9.002217294900221, |
|
"grad_norm": 0.3748861849308014, |
|
"learning_rate": 5.768262024650773e-06, |
|
"loss": 1.2133, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 9.024390243902438, |
|
"grad_norm": 0.37916404008865356, |
|
"learning_rate": 5.511456149611194e-06, |
|
"loss": 1.222, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 9.046563192904657, |
|
"grad_norm": 0.3922421634197235, |
|
"learning_rate": 5.26033574856708e-06, |
|
"loss": 1.2187, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 9.068736141906873, |
|
"grad_norm": 0.4379100501537323, |
|
"learning_rate": 5.014915931694253e-06, |
|
"loss": 1.2046, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 9.090909090909092, |
|
"grad_norm": 0.3986884653568268, |
|
"learning_rate": 4.775211466158469e-06, |
|
"loss": 1.2112, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 9.113082039911308, |
|
"grad_norm": 0.3886760473251343, |
|
"learning_rate": 4.541236775226809e-06, |
|
"loss": 1.2125, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 9.135254988913525, |
|
"grad_norm": 0.3771812617778778, |
|
"learning_rate": 4.313005937399861e-06, |
|
"loss": 1.1871, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 9.157427937915743, |
|
"grad_norm": 0.39025792479515076, |
|
"learning_rate": 4.0905326855646185e-06, |
|
"loss": 1.2039, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 9.17960088691796, |
|
"grad_norm": 0.3777172267436981, |
|
"learning_rate": 3.873830406168111e-06, |
|
"loss": 1.2145, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 9.201773835920177, |
|
"grad_norm": 0.39793717861175537, |
|
"learning_rate": 3.662912138411967e-06, |
|
"loss": 1.1983, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 9.223946784922395, |
|
"grad_norm": 0.3853585124015808, |
|
"learning_rate": 3.457790573467812e-06, |
|
"loss": 1.2223, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 9.246119733924612, |
|
"grad_norm": 0.3928294777870178, |
|
"learning_rate": 3.2584780537136207e-06, |
|
"loss": 1.2319, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 9.268292682926829, |
|
"grad_norm": 0.39439529180526733, |
|
"learning_rate": 3.0649865719910798e-06, |
|
"loss": 1.198, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 9.290465631929047, |
|
"grad_norm": 0.3766527771949768, |
|
"learning_rate": 2.877327770883964e-06, |
|
"loss": 1.2137, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 9.312638580931264, |
|
"grad_norm": 0.39711084961891174, |
|
"learning_rate": 2.6955129420176196e-06, |
|
"loss": 1.2232, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 9.33481152993348, |
|
"grad_norm": 0.37520113587379456, |
|
"learning_rate": 2.5195530253794396e-06, |
|
"loss": 1.2156, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 9.356984478935699, |
|
"grad_norm": 0.3946742117404938, |
|
"learning_rate": 2.349458608660704e-06, |
|
"loss": 1.2009, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 9.379157427937916, |
|
"grad_norm": 0.3818342685699463, |
|
"learning_rate": 2.1852399266194314e-06, |
|
"loss": 1.2066, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 9.401330376940132, |
|
"grad_norm": 0.3952968716621399, |
|
"learning_rate": 2.026906860464606e-06, |
|
"loss": 1.2346, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 9.42350332594235, |
|
"grad_norm": 0.3948782682418823, |
|
"learning_rate": 1.874468937261531e-06, |
|
"loss": 1.2154, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 9.445676274944567, |
|
"grad_norm": 0.3829614520072937, |
|
"learning_rate": 1.7279353293586765e-06, |
|
"loss": 1.2089, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 9.467849223946786, |
|
"grad_norm": 0.38727471232414246, |
|
"learning_rate": 1.5873148538356753e-06, |
|
"loss": 1.205, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 9.490022172949002, |
|
"grad_norm": 0.39005568623542786, |
|
"learning_rate": 1.4526159719728594e-06, |
|
"loss": 1.2023, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 9.512195121951219, |
|
"grad_norm": 0.3805971145629883, |
|
"learning_rate": 1.323846788742078e-06, |
|
"loss": 1.2022, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 9.534368070953438, |
|
"grad_norm": 0.3740484416484833, |
|
"learning_rate": 1.201015052319099e-06, |
|
"loss": 1.2251, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 9.556541019955654, |
|
"grad_norm": 0.3876747190952301, |
|
"learning_rate": 1.084128153617292e-06, |
|
"loss": 1.2114, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 9.57871396895787, |
|
"grad_norm": 0.37821733951568604, |
|
"learning_rate": 9.731931258429638e-07, |
|
"loss": 1.2099, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 9.60088691796009, |
|
"grad_norm": 0.38440433144569397, |
|
"learning_rate": 8.682166440721728e-07, |
|
"loss": 1.2067, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 9.623059866962306, |
|
"grad_norm": 0.3851392865180969, |
|
"learning_rate": 7.69205024849029e-07, |
|
"loss": 1.2126, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 9.645232815964523, |
|
"grad_norm": 0.39497315883636475, |
|
"learning_rate": 6.761642258056978e-07, |
|
"loss": 1.219, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 9.667405764966741, |
|
"grad_norm": 0.3837352395057678, |
|
"learning_rate": 5.890998453038644e-07, |
|
"loss": 1.2034, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 9.689578713968958, |
|
"grad_norm": 0.38280948996543884, |
|
"learning_rate": 5.080171220978813e-07, |
|
"loss": 1.218, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 9.711751662971174, |
|
"grad_norm": 0.3924092650413513, |
|
"learning_rate": 4.329209350195651e-07, |
|
"loss": 1.1972, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 9.733924611973393, |
|
"grad_norm": 0.38612061738967896, |
|
"learning_rate": 3.638158026846306e-07, |
|
"loss": 1.2058, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 9.75609756097561, |
|
"grad_norm": 0.38143137097358704, |
|
"learning_rate": 3.007058832207976e-07, |
|
"loss": 1.1958, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 9.778270509977826, |
|
"grad_norm": 0.39435523748397827, |
|
"learning_rate": 2.4359497401758024e-07, |
|
"loss": 1.1962, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 9.800443458980045, |
|
"grad_norm": 0.3891894221305847, |
|
"learning_rate": 1.924865114978025e-07, |
|
"loss": 1.2133, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 9.822616407982261, |
|
"grad_norm": 0.38457366824150085, |
|
"learning_rate": 1.4738357091084176e-07, |
|
"loss": 1.2151, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 9.84478935698448, |
|
"grad_norm": 0.3849795162677765, |
|
"learning_rate": 1.0828886614754341e-07, |
|
"loss": 1.2058, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 9.866962305986696, |
|
"grad_norm": 0.38724285364151, |
|
"learning_rate": 7.520474957699586e-08, |
|
"loss": 1.2102, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 9.889135254988913, |
|
"grad_norm": 0.3877987265586853, |
|
"learning_rate": 4.8133211904888285e-08, |
|
"loss": 1.2122, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 9.911308203991132, |
|
"grad_norm": 0.3823870122432709, |
|
"learning_rate": 2.7075882053828605e-08, |
|
"loss": 1.213, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 9.933481152993348, |
|
"grad_norm": 0.3904678523540497, |
|
"learning_rate": 1.2034027065255249e-08, |
|
"loss": 1.2077, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 9.955654101995565, |
|
"grad_norm": 0.37739241123199463, |
|
"learning_rate": 3.0085520232425722e-09, |
|
"loss": 1.1993, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 9.977827050997783, |
|
"grad_norm": 0.3910938799381256, |
|
"learning_rate": 0.0, |
|
"loss": 1.2051, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 9.977827050997783, |
|
"eval_loss": 1.8520119190216064, |
|
"eval_runtime": 0.3296, |
|
"eval_samples_per_second": 3.034, |
|
"eval_steps_per_second": 3.034, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 9.977827050997783, |
|
"step": 2250, |
|
"total_flos": 3.3259687719144e+18, |
|
"train_loss": 1.3362829395929972, |
|
"train_runtime": 6815.0283, |
|
"train_samples_per_second": 10.572, |
|
"train_steps_per_second": 0.33 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 2250, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.3259687719144e+18, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|