|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 9.992193598750976, |
|
"eval_steps": 500, |
|
"global_step": 6400, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00156128024980484, |
|
"grad_norm": 15.554603576660156, |
|
"learning_rate": 3.125e-07, |
|
"loss": 1.2028, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0078064012490242, |
|
"grad_norm": 14.548426628112793, |
|
"learning_rate": 1.5625e-06, |
|
"loss": 1.23, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0156128024980484, |
|
"grad_norm": 10.735854148864746, |
|
"learning_rate": 3.125e-06, |
|
"loss": 1.1711, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0234192037470726, |
|
"grad_norm": 9.330694198608398, |
|
"learning_rate": 4.6875000000000004e-06, |
|
"loss": 1.137, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0312256049960968, |
|
"grad_norm": 5.181936740875244, |
|
"learning_rate": 6.25e-06, |
|
"loss": 1.0235, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.039032006245121, |
|
"grad_norm": 3.862912654876709, |
|
"learning_rate": 7.8125e-06, |
|
"loss": 0.9359, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0468384074941452, |
|
"grad_norm": 3.142946720123291, |
|
"learning_rate": 9.375000000000001e-06, |
|
"loss": 0.8746, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0546448087431694, |
|
"grad_norm": 1.8955553770065308, |
|
"learning_rate": 1.09375e-05, |
|
"loss": 0.8135, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0624512099921936, |
|
"grad_norm": 1.3491796255111694, |
|
"learning_rate": 1.25e-05, |
|
"loss": 0.7632, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0702576112412178, |
|
"grad_norm": 1.4237151145935059, |
|
"learning_rate": 1.4062500000000001e-05, |
|
"loss": 0.7152, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.078064012490242, |
|
"grad_norm": 1.992169737815857, |
|
"learning_rate": 1.5625e-05, |
|
"loss": 0.6802, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0858704137392662, |
|
"grad_norm": 1.0017743110656738, |
|
"learning_rate": 1.71875e-05, |
|
"loss": 0.6617, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.0936768149882904, |
|
"grad_norm": 0.9323798418045044, |
|
"learning_rate": 1.8750000000000002e-05, |
|
"loss": 0.6236, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.1014832162373146, |
|
"grad_norm": 0.7461961507797241, |
|
"learning_rate": 2.0312500000000002e-05, |
|
"loss": 0.6, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.1092896174863388, |
|
"grad_norm": 0.7259904742240906, |
|
"learning_rate": 2.1875e-05, |
|
"loss": 0.5814, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.117096018735363, |
|
"grad_norm": 0.5570379495620728, |
|
"learning_rate": 2.34375e-05, |
|
"loss": 0.5752, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.1249024199843872, |
|
"grad_norm": 0.5848241448402405, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.5455, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.1327088212334114, |
|
"grad_norm": 0.9087640643119812, |
|
"learning_rate": 2.6562500000000002e-05, |
|
"loss": 0.549, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.1405152224824356, |
|
"grad_norm": 0.6453709602355957, |
|
"learning_rate": 2.8125000000000003e-05, |
|
"loss": 0.5382, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.1483216237314598, |
|
"grad_norm": 0.6244491934776306, |
|
"learning_rate": 2.96875e-05, |
|
"loss": 0.5416, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.156128024980484, |
|
"grad_norm": 0.6923064589500427, |
|
"learning_rate": 3.125e-05, |
|
"loss": 0.5209, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.16393442622950818, |
|
"grad_norm": 0.8899492025375366, |
|
"learning_rate": 3.2812500000000005e-05, |
|
"loss": 0.5193, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.1717408274785324, |
|
"grad_norm": 0.6282894611358643, |
|
"learning_rate": 3.4375e-05, |
|
"loss": 0.5283, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.1795472287275566, |
|
"grad_norm": 0.637545645236969, |
|
"learning_rate": 3.59375e-05, |
|
"loss": 0.5161, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.1873536299765808, |
|
"grad_norm": 0.6604511141777039, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"loss": 0.5024, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.195160031225605, |
|
"grad_norm": 0.5799457430839539, |
|
"learning_rate": 3.90625e-05, |
|
"loss": 0.5113, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2029664324746292, |
|
"grad_norm": 0.745388388633728, |
|
"learning_rate": 4.0625000000000005e-05, |
|
"loss": 0.509, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.2107728337236534, |
|
"grad_norm": 0.8749244809150696, |
|
"learning_rate": 4.21875e-05, |
|
"loss": 0.5085, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.2185792349726776, |
|
"grad_norm": 0.832904040813446, |
|
"learning_rate": 4.375e-05, |
|
"loss": 0.4885, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2263856362217018, |
|
"grad_norm": 0.7716369032859802, |
|
"learning_rate": 4.5312500000000004e-05, |
|
"loss": 0.4886, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.234192037470726, |
|
"grad_norm": 0.6549221277236938, |
|
"learning_rate": 4.6875e-05, |
|
"loss": 0.4879, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2419984387197502, |
|
"grad_norm": 0.5921856164932251, |
|
"learning_rate": 4.8437500000000005e-05, |
|
"loss": 0.4902, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.2498048399687744, |
|
"grad_norm": 0.5582632422447205, |
|
"learning_rate": 5e-05, |
|
"loss": 0.4954, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.2576112412177986, |
|
"grad_norm": 0.5605296492576599, |
|
"learning_rate": 5.15625e-05, |
|
"loss": 0.4882, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.2654176424668228, |
|
"grad_norm": 0.7804284691810608, |
|
"learning_rate": 5.3125000000000004e-05, |
|
"loss": 0.4751, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.273224043715847, |
|
"grad_norm": 0.9444372653961182, |
|
"learning_rate": 5.46875e-05, |
|
"loss": 0.4777, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.2810304449648712, |
|
"grad_norm": 0.5533527135848999, |
|
"learning_rate": 5.6250000000000005e-05, |
|
"loss": 0.4889, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.2888368462138954, |
|
"grad_norm": 0.5322621464729309, |
|
"learning_rate": 5.78125e-05, |
|
"loss": 0.4837, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.2966432474629196, |
|
"grad_norm": 0.5449932217597961, |
|
"learning_rate": 5.9375e-05, |
|
"loss": 0.4695, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.3044496487119438, |
|
"grad_norm": 0.6229299306869507, |
|
"learning_rate": 6.0937500000000004e-05, |
|
"loss": 0.4891, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.312256049960968, |
|
"grad_norm": 0.5832580327987671, |
|
"learning_rate": 6.25e-05, |
|
"loss": 0.4697, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3200624512099922, |
|
"grad_norm": 0.6701719164848328, |
|
"learning_rate": 6.40625e-05, |
|
"loss": 0.466, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.32786885245901637, |
|
"grad_norm": 1.088781476020813, |
|
"learning_rate": 6.562500000000001e-05, |
|
"loss": 0.4684, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.3356752537080406, |
|
"grad_norm": 0.5555649399757385, |
|
"learning_rate": 6.71875e-05, |
|
"loss": 0.471, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.3434816549570648, |
|
"grad_norm": 0.7747379541397095, |
|
"learning_rate": 6.875e-05, |
|
"loss": 0.4749, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.351288056206089, |
|
"grad_norm": 0.5692268013954163, |
|
"learning_rate": 7.031250000000001e-05, |
|
"loss": 0.4716, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.3590944574551132, |
|
"grad_norm": 0.6532680988311768, |
|
"learning_rate": 7.1875e-05, |
|
"loss": 0.4658, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.3669008587041374, |
|
"grad_norm": 0.5181140303611755, |
|
"learning_rate": 7.34375e-05, |
|
"loss": 0.4631, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.3747072599531616, |
|
"grad_norm": 0.5234228372573853, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 0.4632, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.3825136612021858, |
|
"grad_norm": 0.7775989174842834, |
|
"learning_rate": 7.65625e-05, |
|
"loss": 0.4767, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.39032006245121, |
|
"grad_norm": 0.736663281917572, |
|
"learning_rate": 7.8125e-05, |
|
"loss": 0.4552, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.3981264637002342, |
|
"grad_norm": 0.49683699011802673, |
|
"learning_rate": 7.96875e-05, |
|
"loss": 0.458, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.4059328649492584, |
|
"grad_norm": 0.5580967664718628, |
|
"learning_rate": 8.125000000000001e-05, |
|
"loss": 0.4553, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.4137392661982826, |
|
"grad_norm": 0.5070011019706726, |
|
"learning_rate": 8.28125e-05, |
|
"loss": 0.4667, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.4215456674473068, |
|
"grad_norm": 0.5351033806800842, |
|
"learning_rate": 8.4375e-05, |
|
"loss": 0.4633, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.42935206869633097, |
|
"grad_norm": 0.5223821997642517, |
|
"learning_rate": 8.593750000000001e-05, |
|
"loss": 0.4614, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.4371584699453552, |
|
"grad_norm": 0.5035797953605652, |
|
"learning_rate": 8.75e-05, |
|
"loss": 0.4577, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.4449648711943794, |
|
"grad_norm": 0.5388506054878235, |
|
"learning_rate": 8.90625e-05, |
|
"loss": 0.4647, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.4527712724434036, |
|
"grad_norm": 0.5311102867126465, |
|
"learning_rate": 9.062500000000001e-05, |
|
"loss": 0.4537, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.4605776736924278, |
|
"grad_norm": 0.5614621043205261, |
|
"learning_rate": 9.21875e-05, |
|
"loss": 0.4635, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.468384074941452, |
|
"grad_norm": 0.7148458957672119, |
|
"learning_rate": 9.375e-05, |
|
"loss": 0.457, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.47619047619047616, |
|
"grad_norm": 0.7659621834754944, |
|
"learning_rate": 9.53125e-05, |
|
"loss": 0.4522, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.4839968774395004, |
|
"grad_norm": 0.6256824731826782, |
|
"learning_rate": 9.687500000000001e-05, |
|
"loss": 0.4547, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.4918032786885246, |
|
"grad_norm": 0.5664246082305908, |
|
"learning_rate": 9.84375e-05, |
|
"loss": 0.456, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.4996096799375488, |
|
"grad_norm": 0.6836294531822205, |
|
"learning_rate": 0.0001, |
|
"loss": 0.4525, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.507416081186573, |
|
"grad_norm": 0.5858592391014099, |
|
"learning_rate": 0.00010156250000000001, |
|
"loss": 0.4439, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.5152224824355972, |
|
"grad_norm": 0.5709215402603149, |
|
"learning_rate": 0.000103125, |
|
"loss": 0.4591, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5230288836846214, |
|
"grad_norm": 0.5698142647743225, |
|
"learning_rate": 0.0001046875, |
|
"loss": 0.4412, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.5308352849336456, |
|
"grad_norm": 0.5309858918190002, |
|
"learning_rate": 0.00010625000000000001, |
|
"loss": 0.4473, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.5386416861826698, |
|
"grad_norm": 0.49240779876708984, |
|
"learning_rate": 0.00010781250000000001, |
|
"loss": 0.4415, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.546448087431694, |
|
"grad_norm": 0.4742411971092224, |
|
"learning_rate": 0.000109375, |
|
"loss": 0.441, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.5542544886807181, |
|
"grad_norm": 0.4116191864013672, |
|
"learning_rate": 0.0001109375, |
|
"loss": 0.4389, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.5620608899297423, |
|
"grad_norm": 0.5430789589881897, |
|
"learning_rate": 0.00011250000000000001, |
|
"loss": 0.4636, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.5698672911787666, |
|
"grad_norm": 0.6789748668670654, |
|
"learning_rate": 0.0001140625, |
|
"loss": 0.4391, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.5776736924277908, |
|
"grad_norm": 0.4417556822299957, |
|
"learning_rate": 0.000115625, |
|
"loss": 0.4383, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.585480093676815, |
|
"grad_norm": 0.4192645847797394, |
|
"learning_rate": 0.00011718750000000001, |
|
"loss": 0.4392, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.5932864949258392, |
|
"grad_norm": 0.42620405554771423, |
|
"learning_rate": 0.00011875, |
|
"loss": 0.4509, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.6010928961748634, |
|
"grad_norm": 0.44085854291915894, |
|
"learning_rate": 0.0001203125, |
|
"loss": 0.4458, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.6088992974238876, |
|
"grad_norm": 0.45056504011154175, |
|
"learning_rate": 0.00012187500000000001, |
|
"loss": 0.4461, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.6167056986729118, |
|
"grad_norm": 0.5633181929588318, |
|
"learning_rate": 0.0001234375, |
|
"loss": 0.4358, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.624512099921936, |
|
"grad_norm": 0.5012897253036499, |
|
"learning_rate": 0.000125, |
|
"loss": 0.4498, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6323185011709602, |
|
"grad_norm": 0.4812583923339844, |
|
"learning_rate": 0.0001265625, |
|
"loss": 0.4239, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.6401249024199844, |
|
"grad_norm": 0.42709508538246155, |
|
"learning_rate": 0.000128125, |
|
"loss": 0.452, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.6479313036690086, |
|
"grad_norm": 0.5168142914772034, |
|
"learning_rate": 0.0001296875, |
|
"loss": 0.4365, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.6557377049180327, |
|
"grad_norm": 0.4785514771938324, |
|
"learning_rate": 0.00013125000000000002, |
|
"loss": 0.4356, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.663544106167057, |
|
"grad_norm": 0.44413048028945923, |
|
"learning_rate": 0.0001328125, |
|
"loss": 0.4398, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.6713505074160812, |
|
"grad_norm": 0.4412420094013214, |
|
"learning_rate": 0.000134375, |
|
"loss": 0.4363, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.6791569086651054, |
|
"grad_norm": 0.4392535090446472, |
|
"learning_rate": 0.00013593750000000002, |
|
"loss": 0.4339, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.6869633099141296, |
|
"grad_norm": 0.42931196093559265, |
|
"learning_rate": 0.0001375, |
|
"loss": 0.4279, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.6947697111631538, |
|
"grad_norm": 0.41424766182899475, |
|
"learning_rate": 0.0001390625, |
|
"loss": 0.4313, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.702576112412178, |
|
"grad_norm": 0.4159262776374817, |
|
"learning_rate": 0.00014062500000000002, |
|
"loss": 0.4273, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.7103825136612022, |
|
"grad_norm": 0.397058367729187, |
|
"learning_rate": 0.0001421875, |
|
"loss": 0.4416, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.7181889149102264, |
|
"grad_norm": 0.42139485478401184, |
|
"learning_rate": 0.00014375, |
|
"loss": 0.4348, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.7259953161592506, |
|
"grad_norm": 0.47808170318603516, |
|
"learning_rate": 0.00014531250000000002, |
|
"loss": 0.4244, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.7338017174082748, |
|
"grad_norm": 0.4168015122413635, |
|
"learning_rate": 0.000146875, |
|
"loss": 0.4351, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.741608118657299, |
|
"grad_norm": 0.3950536251068115, |
|
"learning_rate": 0.0001484375, |
|
"loss": 0.4338, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.7494145199063232, |
|
"grad_norm": 0.5723301768302917, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 0.4348, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.7572209211553473, |
|
"grad_norm": 0.3689775764942169, |
|
"learning_rate": 0.0001515625, |
|
"loss": 0.4355, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.7650273224043715, |
|
"grad_norm": 0.4008382260799408, |
|
"learning_rate": 0.000153125, |
|
"loss": 0.422, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.7728337236533958, |
|
"grad_norm": 0.49685850739479065, |
|
"learning_rate": 0.0001546875, |
|
"loss": 0.4258, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.78064012490242, |
|
"grad_norm": 0.39265674352645874, |
|
"learning_rate": 0.00015625, |
|
"loss": 0.4334, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.7884465261514442, |
|
"grad_norm": 0.6721764206886292, |
|
"learning_rate": 0.00015781250000000002, |
|
"loss": 0.4283, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.7962529274004684, |
|
"grad_norm": 0.43760669231414795, |
|
"learning_rate": 0.000159375, |
|
"loss": 0.4315, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.8040593286494926, |
|
"grad_norm": 0.3985048532485962, |
|
"learning_rate": 0.0001609375, |
|
"loss": 0.4318, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.8118657298985168, |
|
"grad_norm": 0.3855305016040802, |
|
"learning_rate": 0.00016250000000000002, |
|
"loss": 0.4236, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.819672131147541, |
|
"grad_norm": 0.4576222002506256, |
|
"learning_rate": 0.0001640625, |
|
"loss": 0.4246, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.8274785323965652, |
|
"grad_norm": 0.3545086979866028, |
|
"learning_rate": 0.000165625, |
|
"loss": 0.4248, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.8352849336455894, |
|
"grad_norm": 0.360779345035553, |
|
"learning_rate": 0.00016718750000000002, |
|
"loss": 0.4225, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.8430913348946136, |
|
"grad_norm": 0.344586044549942, |
|
"learning_rate": 0.00016875, |
|
"loss": 0.4231, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.8508977361436377, |
|
"grad_norm": 0.4501068890094757, |
|
"learning_rate": 0.0001703125, |
|
"loss": 0.423, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.8587041373926619, |
|
"grad_norm": 0.5521011352539062, |
|
"learning_rate": 0.00017187500000000002, |
|
"loss": 0.424, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.8665105386416861, |
|
"grad_norm": 0.4081909954547882, |
|
"learning_rate": 0.0001734375, |
|
"loss": 0.4126, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.8743169398907104, |
|
"grad_norm": 0.36447080969810486, |
|
"learning_rate": 0.000175, |
|
"loss": 0.4197, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.8821233411397346, |
|
"grad_norm": 0.43303537368774414, |
|
"learning_rate": 0.00017656250000000002, |
|
"loss": 0.4261, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.8899297423887588, |
|
"grad_norm": 0.35191860795021057, |
|
"learning_rate": 0.000178125, |
|
"loss": 0.4233, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.897736143637783, |
|
"grad_norm": 0.33722788095474243, |
|
"learning_rate": 0.0001796875, |
|
"loss": 0.4149, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.9055425448868072, |
|
"grad_norm": 0.36060288548469543, |
|
"learning_rate": 0.00018125000000000001, |
|
"loss": 0.4244, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.9133489461358314, |
|
"grad_norm": 0.353076696395874, |
|
"learning_rate": 0.0001828125, |
|
"loss": 0.4202, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.9211553473848556, |
|
"grad_norm": 0.39187219738960266, |
|
"learning_rate": 0.000184375, |
|
"loss": 0.4133, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.9289617486338798, |
|
"grad_norm": 0.5925287008285522, |
|
"learning_rate": 0.0001859375, |
|
"loss": 0.4231, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.936768149882904, |
|
"grad_norm": 0.5176442861557007, |
|
"learning_rate": 0.0001875, |
|
"loss": 0.4245, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.9445745511319282, |
|
"grad_norm": 0.33046722412109375, |
|
"learning_rate": 0.00018906250000000002, |
|
"loss": 0.4147, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.9523809523809523, |
|
"grad_norm": 0.3665611147880554, |
|
"learning_rate": 0.000190625, |
|
"loss": 0.4275, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.9601873536299765, |
|
"grad_norm": 0.3403601050376892, |
|
"learning_rate": 0.0001921875, |
|
"loss": 0.4158, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.9679937548790007, |
|
"grad_norm": 0.3726797103881836, |
|
"learning_rate": 0.00019375000000000002, |
|
"loss": 0.4197, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.975800156128025, |
|
"grad_norm": 0.38149702548980713, |
|
"learning_rate": 0.0001953125, |
|
"loss": 0.4102, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.9836065573770492, |
|
"grad_norm": 0.3473910987377167, |
|
"learning_rate": 0.000196875, |
|
"loss": 0.4294, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.9914129586260734, |
|
"grad_norm": 0.32832297682762146, |
|
"learning_rate": 0.00019843750000000002, |
|
"loss": 0.4162, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.9992193598750976, |
|
"grad_norm": 0.33780425786972046, |
|
"learning_rate": 0.0002, |
|
"loss": 0.419, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.9992193598750976, |
|
"eval_loss": 1.0502797365188599, |
|
"eval_runtime": 0.3387, |
|
"eval_samples_per_second": 14.76, |
|
"eval_steps_per_second": 2.952, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.0070257611241218, |
|
"grad_norm": 0.34341296553611755, |
|
"learning_rate": 0.00019999962815279418, |
|
"loss": 0.4058, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.014832162373146, |
|
"grad_norm": 0.3360665440559387, |
|
"learning_rate": 0.00019999851261394218, |
|
"loss": 0.4051, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.0226385636221702, |
|
"grad_norm": 0.33649712800979614, |
|
"learning_rate": 0.00019999665339174013, |
|
"loss": 0.4055, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.0304449648711944, |
|
"grad_norm": 0.33783483505249023, |
|
"learning_rate": 0.00019999405050001497, |
|
"loss": 0.4124, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.0382513661202186, |
|
"grad_norm": 0.35244283080101013, |
|
"learning_rate": 0.0001999907039581243, |
|
"loss": 0.4148, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.0460577673692428, |
|
"grad_norm": 0.3347136974334717, |
|
"learning_rate": 0.0001999866137909562, |
|
"loss": 0.4032, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.053864168618267, |
|
"grad_norm": 0.34604641795158386, |
|
"learning_rate": 0.0001999817800289289, |
|
"loss": 0.4093, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.0616705698672912, |
|
"grad_norm": 0.36726704239845276, |
|
"learning_rate": 0.00019997620270799092, |
|
"loss": 0.4022, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.0694769711163155, |
|
"grad_norm": 0.410674512386322, |
|
"learning_rate": 0.00019996988186962041, |
|
"loss": 0.4095, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.0772833723653397, |
|
"grad_norm": 0.3476228713989258, |
|
"learning_rate": 0.00019996281756082517, |
|
"loss": 0.4141, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.0850897736143639, |
|
"grad_norm": 0.3787543475627899, |
|
"learning_rate": 0.00019995500983414202, |
|
"loss": 0.3973, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.092896174863388, |
|
"grad_norm": 0.35693004727363586, |
|
"learning_rate": 0.00019994645874763658, |
|
"loss": 0.4155, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.100702576112412, |
|
"grad_norm": 0.3264383375644684, |
|
"learning_rate": 0.0001999371643649028, |
|
"loss": 0.3979, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.1085089773614363, |
|
"grad_norm": 0.32450568675994873, |
|
"learning_rate": 0.00019992712675506253, |
|
"loss": 0.41, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.1163153786104605, |
|
"grad_norm": 0.3869469165802002, |
|
"learning_rate": 0.00019991634599276486, |
|
"loss": 0.4127, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.1241217798594847, |
|
"grad_norm": 0.46077024936676025, |
|
"learning_rate": 0.0001999048221581858, |
|
"loss": 0.4067, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.131928181108509, |
|
"grad_norm": 0.3199751675128937, |
|
"learning_rate": 0.00019989255533702736, |
|
"loss": 0.4092, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.139734582357533, |
|
"grad_norm": 0.341310054063797, |
|
"learning_rate": 0.00019987954562051725, |
|
"loss": 0.4192, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.1475409836065573, |
|
"grad_norm": 0.3507387340068817, |
|
"learning_rate": 0.000199865793105408, |
|
"loss": 0.4063, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 1.1553473848555815, |
|
"grad_norm": 0.33445993065834045, |
|
"learning_rate": 0.00019985129789397633, |
|
"loss": 0.4178, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.1631537861046057, |
|
"grad_norm": 0.3056560158729553, |
|
"learning_rate": 0.00019983606009402224, |
|
"loss": 0.395, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 1.17096018735363, |
|
"grad_norm": 0.31608232855796814, |
|
"learning_rate": 0.00019982007981886847, |
|
"loss": 0.4089, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.1787665886026542, |
|
"grad_norm": 0.44849371910095215, |
|
"learning_rate": 0.00019980335718735944, |
|
"loss": 0.4042, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 1.1865729898516784, |
|
"grad_norm": 0.34841999411582947, |
|
"learning_rate": 0.00019978589232386035, |
|
"loss": 0.4017, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.1943793911007026, |
|
"grad_norm": 0.30834051966667175, |
|
"learning_rate": 0.0001997676853582565, |
|
"loss": 0.4059, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 1.2021857923497268, |
|
"grad_norm": 0.319698303937912, |
|
"learning_rate": 0.000199748736425952, |
|
"loss": 0.4133, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.209992193598751, |
|
"grad_norm": 0.3290104269981384, |
|
"learning_rate": 0.00019972904566786903, |
|
"loss": 0.4025, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.2177985948477752, |
|
"grad_norm": 0.3340936601161957, |
|
"learning_rate": 0.00019970861323044667, |
|
"loss": 0.4056, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.2256049960967994, |
|
"grad_norm": 0.32487761974334717, |
|
"learning_rate": 0.00019968743926563983, |
|
"loss": 0.4043, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 1.2334113973458236, |
|
"grad_norm": 0.3178825378417969, |
|
"learning_rate": 0.00019966552393091804, |
|
"loss": 0.4063, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.2412177985948478, |
|
"grad_norm": 0.3003855347633362, |
|
"learning_rate": 0.00019964286738926448, |
|
"loss": 0.3961, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 1.249024199843872, |
|
"grad_norm": 0.32938238978385925, |
|
"learning_rate": 0.00019961946980917456, |
|
"loss": 0.3991, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.2568306010928962, |
|
"grad_norm": 0.3072170615196228, |
|
"learning_rate": 0.0001995953313646548, |
|
"loss": 0.3943, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 1.2646370023419204, |
|
"grad_norm": 0.36918073892593384, |
|
"learning_rate": 0.0001995704522352214, |
|
"loss": 0.3998, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.2724434035909447, |
|
"grad_norm": 0.3280356228351593, |
|
"learning_rate": 0.00019954483260589911, |
|
"loss": 0.4035, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 1.2802498048399689, |
|
"grad_norm": 0.29702678322792053, |
|
"learning_rate": 0.0001995184726672197, |
|
"loss": 0.3901, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.288056206088993, |
|
"grad_norm": 0.30307725071907043, |
|
"learning_rate": 0.00019949137261522052, |
|
"loss": 0.4053, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 1.2958626073380173, |
|
"grad_norm": 0.31273552775382996, |
|
"learning_rate": 0.00019946353265144312, |
|
"loss": 0.4136, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.3036690085870415, |
|
"grad_norm": 0.30774056911468506, |
|
"learning_rate": 0.00019943495298293181, |
|
"loss": 0.3995, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 1.3114754098360657, |
|
"grad_norm": 0.3068085312843323, |
|
"learning_rate": 0.00019940563382223197, |
|
"loss": 0.4136, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.31928181108509, |
|
"grad_norm": 0.33068013191223145, |
|
"learning_rate": 0.00019937557538738854, |
|
"loss": 0.3957, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 1.327088212334114, |
|
"grad_norm": 0.36017951369285583, |
|
"learning_rate": 0.00019934477790194445, |
|
"loss": 0.4094, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.334894613583138, |
|
"grad_norm": 0.3393058776855469, |
|
"learning_rate": 0.00019931324159493886, |
|
"loss": 0.4072, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 1.3427010148321623, |
|
"grad_norm": 0.29926055669784546, |
|
"learning_rate": 0.0001992809667009055, |
|
"loss": 0.3991, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.3505074160811865, |
|
"grad_norm": 0.3126608431339264, |
|
"learning_rate": 0.000199247953459871, |
|
"loss": 0.3997, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 1.3583138173302107, |
|
"grad_norm": 0.31976112723350525, |
|
"learning_rate": 0.00019921420211735299, |
|
"loss": 0.4043, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.366120218579235, |
|
"grad_norm": 0.37549304962158203, |
|
"learning_rate": 0.00019917971292435826, |
|
"loss": 0.4112, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 1.3739266198282591, |
|
"grad_norm": 0.3190479278564453, |
|
"learning_rate": 0.00019914448613738106, |
|
"loss": 0.402, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.3817330210772834, |
|
"grad_norm": 0.30075913667678833, |
|
"learning_rate": 0.00019910852201840103, |
|
"loss": 0.4159, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 1.3895394223263076, |
|
"grad_norm": 0.3074491024017334, |
|
"learning_rate": 0.00019907182083488129, |
|
"loss": 0.4045, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.3973458235753318, |
|
"grad_norm": 0.3855604827404022, |
|
"learning_rate": 0.0001990343828597665, |
|
"loss": 0.4045, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 1.405152224824356, |
|
"grad_norm": 0.3558739721775055, |
|
"learning_rate": 0.00019899620837148077, |
|
"loss": 0.3969, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.4129586260733802, |
|
"grad_norm": 0.3128485083580017, |
|
"learning_rate": 0.00019895729765392572, |
|
"loss": 0.403, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 1.4207650273224044, |
|
"grad_norm": 0.30848538875579834, |
|
"learning_rate": 0.0001989176509964781, |
|
"loss": 0.398, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.4285714285714286, |
|
"grad_norm": 0.2908243238925934, |
|
"learning_rate": 0.00019887726869398793, |
|
"loss": 0.394, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 1.4363778298204528, |
|
"grad_norm": 0.29241564869880676, |
|
"learning_rate": 0.00019883615104677608, |
|
"loss": 0.4007, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.444184231069477, |
|
"grad_norm": 0.29839983582496643, |
|
"learning_rate": 0.00019879429836063226, |
|
"loss": 0.3961, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 1.4519906323185012, |
|
"grad_norm": 0.298441618680954, |
|
"learning_rate": 0.00019875171094681248, |
|
"loss": 0.4003, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.4597970335675254, |
|
"grad_norm": 0.30495795607566833, |
|
"learning_rate": 0.00019870838912203705, |
|
"loss": 0.4036, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 1.4676034348165496, |
|
"grad_norm": 0.2891780436038971, |
|
"learning_rate": 0.0001986643332084879, |
|
"loss": 0.3901, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.4754098360655736, |
|
"grad_norm": 0.2930201292037964, |
|
"learning_rate": 0.00019861954353380643, |
|
"loss": 0.3943, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 1.4832162373145978, |
|
"grad_norm": 0.289411336183548, |
|
"learning_rate": 0.0001985740204310909, |
|
"loss": 0.4003, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.491022638563622, |
|
"grad_norm": 0.3018743693828583, |
|
"learning_rate": 0.0001985277642388941, |
|
"loss": 0.3949, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 1.4988290398126463, |
|
"grad_norm": 0.3346049189567566, |
|
"learning_rate": 0.00019848077530122083, |
|
"loss": 0.3949, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.5066354410616705, |
|
"grad_norm": 0.35057857632637024, |
|
"learning_rate": 0.00019843305396752509, |
|
"loss": 0.3916, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 1.5144418423106947, |
|
"grad_norm": 0.33775442838668823, |
|
"learning_rate": 0.00019838460059270775, |
|
"loss": 0.4003, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.5222482435597189, |
|
"grad_norm": 0.3016526997089386, |
|
"learning_rate": 0.00019833541553711395, |
|
"loss": 0.4032, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 1.530054644808743, |
|
"grad_norm": 0.2809526026248932, |
|
"learning_rate": 0.0001982854991665301, |
|
"loss": 0.3983, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.5378610460577673, |
|
"grad_norm": 0.3081335723400116, |
|
"learning_rate": 0.00019823485185218158, |
|
"loss": 0.3997, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 1.5456674473067915, |
|
"grad_norm": 0.28975608944892883, |
|
"learning_rate": 0.00019818347397072955, |
|
"loss": 0.3943, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.5534738485558157, |
|
"grad_norm": 0.32257601618766785, |
|
"learning_rate": 0.00019813136590426844, |
|
"loss": 0.3948, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 1.56128024980484, |
|
"grad_norm": 0.3275687098503113, |
|
"learning_rate": 0.00019807852804032305, |
|
"loss": 0.4001, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.5690866510538641, |
|
"grad_norm": 0.3463219404220581, |
|
"learning_rate": 0.0001980249607718456, |
|
"loss": 0.3992, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 1.5768930523028883, |
|
"grad_norm": 0.3080371022224426, |
|
"learning_rate": 0.00019797066449721295, |
|
"loss": 0.4023, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.5846994535519126, |
|
"grad_norm": 0.29140976071357727, |
|
"learning_rate": 0.00019791563962022338, |
|
"loss": 0.4036, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 1.5925058548009368, |
|
"grad_norm": 0.3033461570739746, |
|
"learning_rate": 0.00019785988655009385, |
|
"loss": 0.4051, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.600312256049961, |
|
"grad_norm": 0.27476766705513, |
|
"learning_rate": 0.0001978034057014568, |
|
"loss": 0.3818, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 1.6081186572989852, |
|
"grad_norm": 0.3026203513145447, |
|
"learning_rate": 0.0001977461974943572, |
|
"loss": 0.3959, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.6159250585480094, |
|
"grad_norm": 0.28781238198280334, |
|
"learning_rate": 0.00019768826235424926, |
|
"loss": 0.4067, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 1.6237314597970336, |
|
"grad_norm": 0.28916335105895996, |
|
"learning_rate": 0.00019762960071199333, |
|
"loss": 0.3999, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.6315378610460578, |
|
"grad_norm": 0.2844083607196808, |
|
"learning_rate": 0.00019757021300385286, |
|
"loss": 0.3958, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 1.639344262295082, |
|
"grad_norm": 0.30581745505332947, |
|
"learning_rate": 0.00019751009967149087, |
|
"loss": 0.4016, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.6471506635441062, |
|
"grad_norm": 0.27910909056663513, |
|
"learning_rate": 0.00019744926116196685, |
|
"loss": 0.3903, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 1.6549570647931304, |
|
"grad_norm": 0.2814294695854187, |
|
"learning_rate": 0.00019738769792773336, |
|
"loss": 0.3896, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.6627634660421546, |
|
"grad_norm": 0.29259413480758667, |
|
"learning_rate": 0.00019732541042663282, |
|
"loss": 0.4039, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 1.6705698672911788, |
|
"grad_norm": 0.2868483364582062, |
|
"learning_rate": 0.00019726239912189382, |
|
"loss": 0.3893, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.678376268540203, |
|
"grad_norm": 0.30284589529037476, |
|
"learning_rate": 0.00019719866448212795, |
|
"loss": 0.3973, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 1.6861826697892273, |
|
"grad_norm": 0.29238229990005493, |
|
"learning_rate": 0.00019713420698132614, |
|
"loss": 0.404, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.6939890710382515, |
|
"grad_norm": 0.2957955300807953, |
|
"learning_rate": 0.00019706902709885524, |
|
"loss": 0.3929, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 1.7017954722872757, |
|
"grad_norm": 0.29362279176712036, |
|
"learning_rate": 0.00019700312531945442, |
|
"loss": 0.393, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.7096018735362999, |
|
"grad_norm": 0.3154332637786865, |
|
"learning_rate": 0.00019693650213323144, |
|
"loss": 0.3971, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 1.717408274785324, |
|
"grad_norm": 0.31909048557281494, |
|
"learning_rate": 0.00019686915803565934, |
|
"loss": 0.3991, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.7252146760343483, |
|
"grad_norm": 0.31985750794410706, |
|
"learning_rate": 0.00019680109352757227, |
|
"loss": 0.4076, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 1.7330210772833725, |
|
"grad_norm": 0.2867513597011566, |
|
"learning_rate": 0.00019673230911516226, |
|
"loss": 0.3957, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.7408274785323967, |
|
"grad_norm": 0.30128851532936096, |
|
"learning_rate": 0.00019666280530997518, |
|
"loss": 0.4041, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 1.748633879781421, |
|
"grad_norm": 0.2900910973548889, |
|
"learning_rate": 0.00019659258262890683, |
|
"loss": 0.3898, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.756440281030445, |
|
"grad_norm": 0.3005325496196747, |
|
"learning_rate": 0.00019652164159419946, |
|
"loss": 0.4086, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 1.7642466822794691, |
|
"grad_norm": 0.2881026268005371, |
|
"learning_rate": 0.00019644998273343753, |
|
"loss": 0.4005, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.7720530835284933, |
|
"grad_norm": 0.2854783535003662, |
|
"learning_rate": 0.000196377606579544, |
|
"loss": 0.3942, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 1.7798594847775175, |
|
"grad_norm": 0.2979474365711212, |
|
"learning_rate": 0.00019630451367077628, |
|
"loss": 0.3935, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.7876658860265418, |
|
"grad_norm": 0.28160732984542847, |
|
"learning_rate": 0.00019623070455072224, |
|
"loss": 0.4153, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 1.795472287275566, |
|
"grad_norm": 0.27550724148750305, |
|
"learning_rate": 0.0001961561797682962, |
|
"loss": 0.3921, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.8032786885245902, |
|
"grad_norm": 0.27477726340293884, |
|
"learning_rate": 0.00019608093987773478, |
|
"loss": 0.4017, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 1.8110850897736144, |
|
"grad_norm": 0.2824558913707733, |
|
"learning_rate": 0.0001960049854385929, |
|
"loss": 0.3897, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.8188914910226386, |
|
"grad_norm": 0.2761404514312744, |
|
"learning_rate": 0.00019592831701573937, |
|
"loss": 0.4095, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 1.8266978922716628, |
|
"grad_norm": 0.28211623430252075, |
|
"learning_rate": 0.00019585093517935305, |
|
"loss": 0.3912, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.834504293520687, |
|
"grad_norm": 0.27295398712158203, |
|
"learning_rate": 0.0001957728405049183, |
|
"loss": 0.4125, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 1.8423106947697112, |
|
"grad_norm": 0.2896745800971985, |
|
"learning_rate": 0.0001956940335732209, |
|
"loss": 0.3927, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.8501170960187352, |
|
"grad_norm": 0.3007790446281433, |
|
"learning_rate": 0.00019561451497034354, |
|
"loss": 0.3932, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 1.8579234972677594, |
|
"grad_norm": 0.29394078254699707, |
|
"learning_rate": 0.00019553428528766163, |
|
"loss": 0.3905, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.8657298985167836, |
|
"grad_norm": 0.2767265737056732, |
|
"learning_rate": 0.00019545334512183885, |
|
"loss": 0.395, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 1.8735362997658078, |
|
"grad_norm": 0.26828575134277344, |
|
"learning_rate": 0.0001953716950748227, |
|
"loss": 0.3996, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.881342701014832, |
|
"grad_norm": 0.2718822658061981, |
|
"learning_rate": 0.00019528933575384, |
|
"loss": 0.3803, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 1.8891491022638562, |
|
"grad_norm": 0.2866761386394501, |
|
"learning_rate": 0.00019520626777139245, |
|
"loss": 0.4056, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.8969555035128804, |
|
"grad_norm": 0.3028269410133362, |
|
"learning_rate": 0.00019512249174525193, |
|
"loss": 0.4, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 1.9047619047619047, |
|
"grad_norm": 0.2755603492259979, |
|
"learning_rate": 0.0001950380082984561, |
|
"loss": 0.3829, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.9125683060109289, |
|
"grad_norm": 0.28492555022239685, |
|
"learning_rate": 0.00019495281805930367, |
|
"loss": 0.3839, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 1.920374707259953, |
|
"grad_norm": 0.2751103639602661, |
|
"learning_rate": 0.00019486692166134964, |
|
"loss": 0.3907, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.9281811085089773, |
|
"grad_norm": 0.2826032042503357, |
|
"learning_rate": 0.00019478031974340073, |
|
"loss": 0.3884, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 1.9359875097580015, |
|
"grad_norm": 0.28445425629615784, |
|
"learning_rate": 0.0001946930129495106, |
|
"loss": 0.3956, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.9437939110070257, |
|
"grad_norm": 0.2712079584598541, |
|
"learning_rate": 0.0001946050019289749, |
|
"loss": 0.3893, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 1.95160031225605, |
|
"grad_norm": 0.27717921137809753, |
|
"learning_rate": 0.0001945162873363268, |
|
"loss": 0.3912, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.9594067135050741, |
|
"grad_norm": 0.2907007336616516, |
|
"learning_rate": 0.00019442686983133168, |
|
"loss": 0.4004, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 1.9672131147540983, |
|
"grad_norm": 0.2781454026699066, |
|
"learning_rate": 0.00019433675007898255, |
|
"loss": 0.4037, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.9750195160031225, |
|
"grad_norm": 0.27712446451187134, |
|
"learning_rate": 0.00019424592874949496, |
|
"loss": 0.3902, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 1.9828259172521467, |
|
"grad_norm": 0.2906958758831024, |
|
"learning_rate": 0.00019415440651830208, |
|
"loss": 0.3955, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.990632318501171, |
|
"grad_norm": 0.28120070695877075, |
|
"learning_rate": 0.00019406218406604965, |
|
"loss": 0.3878, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 1.9984387197501952, |
|
"grad_norm": 0.2804274260997772, |
|
"learning_rate": 0.00019396926207859084, |
|
"loss": 0.3829, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.0417580604553223, |
|
"eval_runtime": 0.3233, |
|
"eval_samples_per_second": 15.465, |
|
"eval_steps_per_second": 3.093, |
|
"step": 1281 |
|
}, |
|
{ |
|
"epoch": 2.0062451209992194, |
|
"grad_norm": 0.2731517255306244, |
|
"learning_rate": 0.00019387564124698133, |
|
"loss": 0.3747, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 2.0140515222482436, |
|
"grad_norm": 0.2837215065956116, |
|
"learning_rate": 0.00019378132226747398, |
|
"loss": 0.3651, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 2.021857923497268, |
|
"grad_norm": 0.29142889380455017, |
|
"learning_rate": 0.0001936863058415138, |
|
"loss": 0.3736, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 2.029664324746292, |
|
"grad_norm": 0.29420003294944763, |
|
"learning_rate": 0.0001935905926757326, |
|
"loss": 0.3764, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 2.037470725995316, |
|
"grad_norm": 0.2812630832195282, |
|
"learning_rate": 0.00019349418348194383, |
|
"loss": 0.3674, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 2.0452771272443404, |
|
"grad_norm": 0.3169541656970978, |
|
"learning_rate": 0.00019339707897713738, |
|
"loss": 0.3625, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 2.0530835284933646, |
|
"grad_norm": 0.31742027401924133, |
|
"learning_rate": 0.0001932992798834739, |
|
"loss": 0.3718, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 2.060889929742389, |
|
"grad_norm": 0.2889857888221741, |
|
"learning_rate": 0.00019320078692827987, |
|
"loss": 0.3798, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 2.068696330991413, |
|
"grad_norm": 0.27867743372917175, |
|
"learning_rate": 0.00019310160084404186, |
|
"loss": 0.3647, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 2.0765027322404372, |
|
"grad_norm": 0.29241612553596497, |
|
"learning_rate": 0.00019300172236840122, |
|
"loss": 0.3761, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 2.0843091334894615, |
|
"grad_norm": 0.28482168912887573, |
|
"learning_rate": 0.00019290115224414862, |
|
"loss": 0.3743, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 2.0921155347384857, |
|
"grad_norm": 0.3034825325012207, |
|
"learning_rate": 0.00019279989121921847, |
|
"loss": 0.3757, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 2.09992193598751, |
|
"grad_norm": 0.3016110956668854, |
|
"learning_rate": 0.0001926979400466833, |
|
"loss": 0.3774, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 2.107728337236534, |
|
"grad_norm": 0.3048888146877289, |
|
"learning_rate": 0.00019259529948474833, |
|
"loss": 0.369, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 2.1155347384855583, |
|
"grad_norm": 0.2792314291000366, |
|
"learning_rate": 0.00019249197029674566, |
|
"loss": 0.3727, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 2.1233411397345825, |
|
"grad_norm": 0.30155235528945923, |
|
"learning_rate": 0.0001923879532511287, |
|
"loss": 0.3691, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 2.1311475409836067, |
|
"grad_norm": 0.2807096838951111, |
|
"learning_rate": 0.0001922832491214664, |
|
"loss": 0.3618, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 2.138953942232631, |
|
"grad_norm": 0.2897432744503021, |
|
"learning_rate": 0.00019217785868643747, |
|
"loss": 0.3644, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 2.146760343481655, |
|
"grad_norm": 0.2933688759803772, |
|
"learning_rate": 0.0001920717827298248, |
|
"loss": 0.3753, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 2.1545667447306793, |
|
"grad_norm": 0.2933692932128906, |
|
"learning_rate": 0.00019196502204050922, |
|
"loss": 0.3725, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 2.1623731459797035, |
|
"grad_norm": 0.28731822967529297, |
|
"learning_rate": 0.00019185757741246412, |
|
"loss": 0.3733, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 2.1701795472287277, |
|
"grad_norm": 0.3193456530570984, |
|
"learning_rate": 0.00019174944964474912, |
|
"loss": 0.3741, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 2.177985948477752, |
|
"grad_norm": 0.2853163480758667, |
|
"learning_rate": 0.00019164063954150447, |
|
"loss": 0.3774, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 2.185792349726776, |
|
"grad_norm": 0.3006320893764496, |
|
"learning_rate": 0.00019153114791194473, |
|
"loss": 0.3677, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.1935987509758004, |
|
"grad_norm": 0.2752683758735657, |
|
"learning_rate": 0.00019142097557035308, |
|
"loss": 0.3675, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 2.201405152224824, |
|
"grad_norm": 0.2900081276893616, |
|
"learning_rate": 0.00019131012333607506, |
|
"loss": 0.375, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 2.209211553473849, |
|
"grad_norm": 0.3021766245365143, |
|
"learning_rate": 0.00019119859203351255, |
|
"loss": 0.3778, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 2.2170179547228726, |
|
"grad_norm": 0.2842157185077667, |
|
"learning_rate": 0.00019108638249211758, |
|
"loss": 0.3668, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 2.2248243559718968, |
|
"grad_norm": 0.27962204813957214, |
|
"learning_rate": 0.0001909734955463863, |
|
"loss": 0.3731, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 2.232630757220921, |
|
"grad_norm": 0.29117563366889954, |
|
"learning_rate": 0.0001908599320358526, |
|
"loss": 0.3726, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 2.240437158469945, |
|
"grad_norm": 0.307558536529541, |
|
"learning_rate": 0.0001907456928050819, |
|
"loss": 0.37, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 2.2482435597189694, |
|
"grad_norm": 0.3160630762577057, |
|
"learning_rate": 0.000190630778703665, |
|
"loss": 0.3769, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 2.2560499609679936, |
|
"grad_norm": 0.2901064157485962, |
|
"learning_rate": 0.0001905151905862117, |
|
"loss": 0.3703, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 2.263856362217018, |
|
"grad_norm": 0.2909891605377197, |
|
"learning_rate": 0.00019039892931234435, |
|
"loss": 0.374, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.271662763466042, |
|
"grad_norm": 0.3095717132091522, |
|
"learning_rate": 0.00019028199574669153, |
|
"loss": 0.3848, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 2.279469164715066, |
|
"grad_norm": 0.2892928421497345, |
|
"learning_rate": 0.00019016439075888162, |
|
"loss": 0.364, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 2.2872755659640904, |
|
"grad_norm": 0.3012056350708008, |
|
"learning_rate": 0.00019004611522353636, |
|
"loss": 0.3704, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 2.2950819672131146, |
|
"grad_norm": 0.28810742497444153, |
|
"learning_rate": 0.00018992717002026433, |
|
"loss": 0.3757, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 2.302888368462139, |
|
"grad_norm": 0.3010272979736328, |
|
"learning_rate": 0.0001898075560336543, |
|
"loss": 0.3695, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 2.310694769711163, |
|
"grad_norm": 0.2966187000274658, |
|
"learning_rate": 0.00018968727415326884, |
|
"loss": 0.3753, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 2.3185011709601873, |
|
"grad_norm": 0.303035169839859, |
|
"learning_rate": 0.00018956632527363757, |
|
"loss": 0.3844, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 2.3263075722092115, |
|
"grad_norm": 0.2904973030090332, |
|
"learning_rate": 0.00018944471029425054, |
|
"loss": 0.3733, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 2.3341139734582357, |
|
"grad_norm": 0.2944584786891937, |
|
"learning_rate": 0.00018932243011955154, |
|
"loss": 0.3705, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 2.34192037470726, |
|
"grad_norm": 0.3047090470790863, |
|
"learning_rate": 0.00018919948565893142, |
|
"loss": 0.3776, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.349726775956284, |
|
"grad_norm": 0.28889063000679016, |
|
"learning_rate": 0.00018907587782672128, |
|
"loss": 0.3748, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 2.3575331772053083, |
|
"grad_norm": 0.29488006234169006, |
|
"learning_rate": 0.00018895160754218561, |
|
"loss": 0.3719, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 2.3653395784543325, |
|
"grad_norm": 0.28277385234832764, |
|
"learning_rate": 0.00018882667572951562, |
|
"loss": 0.3737, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 2.3731459797033567, |
|
"grad_norm": 0.2788912057876587, |
|
"learning_rate": 0.00018870108331782217, |
|
"loss": 0.37, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 2.380952380952381, |
|
"grad_norm": 0.2984495759010315, |
|
"learning_rate": 0.00018857483124112907, |
|
"loss": 0.3816, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 2.388758782201405, |
|
"grad_norm": 0.28799429535865784, |
|
"learning_rate": 0.00018844792043836589, |
|
"loss": 0.3849, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 2.3965651834504293, |
|
"grad_norm": 0.2892215847969055, |
|
"learning_rate": 0.00018832035185336124, |
|
"loss": 0.372, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 2.4043715846994536, |
|
"grad_norm": 0.30197983980178833, |
|
"learning_rate": 0.0001881921264348355, |
|
"loss": 0.3764, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 2.4121779859484778, |
|
"grad_norm": 0.3232642412185669, |
|
"learning_rate": 0.000188063245136394, |
|
"loss": 0.3788, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 2.419984387197502, |
|
"grad_norm": 0.29107898473739624, |
|
"learning_rate": 0.00018793370891651972, |
|
"loss": 0.3786, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 2.427790788446526, |
|
"grad_norm": 0.29140210151672363, |
|
"learning_rate": 0.00018780351873856627, |
|
"loss": 0.3763, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 2.4355971896955504, |
|
"grad_norm": 0.30477312207221985, |
|
"learning_rate": 0.0001876726755707508, |
|
"loss": 0.3866, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 2.4434035909445746, |
|
"grad_norm": 0.3051372170448303, |
|
"learning_rate": 0.00018754118038614653, |
|
"loss": 0.378, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 2.451209992193599, |
|
"grad_norm": 0.2856997847557068, |
|
"learning_rate": 0.0001874090341626759, |
|
"loss": 0.3744, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 2.459016393442623, |
|
"grad_norm": 0.26825210452079773, |
|
"learning_rate": 0.00018727623788310292, |
|
"loss": 0.369, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 2.4668227946916472, |
|
"grad_norm": 0.2855939269065857, |
|
"learning_rate": 0.00018714279253502616, |
|
"loss": 0.3829, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 2.4746291959406714, |
|
"grad_norm": 0.28602635860443115, |
|
"learning_rate": 0.00018700869911087115, |
|
"loss": 0.3657, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 2.4824355971896956, |
|
"grad_norm": 0.30138617753982544, |
|
"learning_rate": 0.00018687395860788323, |
|
"loss": 0.3694, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.49024199843872, |
|
"grad_norm": 0.2945271134376526, |
|
"learning_rate": 0.00018673857202812, |
|
"loss": 0.3663, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 2.498048399687744, |
|
"grad_norm": 0.28233060240745544, |
|
"learning_rate": 0.00018660254037844388, |
|
"loss": 0.3839, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.5058548009367683, |
|
"grad_norm": 0.30950790643692017, |
|
"learning_rate": 0.00018646586467051463, |
|
"loss": 0.3864, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 2.5136612021857925, |
|
"grad_norm": 0.30938637256622314, |
|
"learning_rate": 0.00018632854592078184, |
|
"loss": 0.3703, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.5214676034348167, |
|
"grad_norm": 0.2714425325393677, |
|
"learning_rate": 0.00018619058515047745, |
|
"loss": 0.3748, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 2.529274004683841, |
|
"grad_norm": 0.30516138672828674, |
|
"learning_rate": 0.0001860519833856079, |
|
"loss": 0.3868, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.537080405932865, |
|
"grad_norm": 0.30349424481391907, |
|
"learning_rate": 0.00018591274165694687, |
|
"loss": 0.3682, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 2.5448868071818893, |
|
"grad_norm": 0.29412826895713806, |
|
"learning_rate": 0.00018577286100002723, |
|
"loss": 0.3683, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.552693208430913, |
|
"grad_norm": 0.2976747155189514, |
|
"learning_rate": 0.00018563234245513364, |
|
"loss": 0.3697, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 2.5604996096799377, |
|
"grad_norm": 0.28876587748527527, |
|
"learning_rate": 0.00018549118706729468, |
|
"loss": 0.3751, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 2.5683060109289615, |
|
"grad_norm": 0.2986242175102234, |
|
"learning_rate": 0.00018534939588627503, |
|
"loss": 0.3772, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 2.576112412177986, |
|
"grad_norm": 0.29054194688796997, |
|
"learning_rate": 0.00018520696996656788, |
|
"loss": 0.388, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 2.58391881342701, |
|
"grad_norm": 0.2982032001018524, |
|
"learning_rate": 0.00018506391036738674, |
|
"loss": 0.3705, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 2.5917252146760346, |
|
"grad_norm": 0.2838602364063263, |
|
"learning_rate": 0.0001849202181526579, |
|
"loss": 0.3705, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 2.5995316159250583, |
|
"grad_norm": 0.28826209902763367, |
|
"learning_rate": 0.00018477589439101232, |
|
"loss": 0.3698, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 2.607338017174083, |
|
"grad_norm": 0.2916529178619385, |
|
"learning_rate": 0.00018463094015577775, |
|
"loss": 0.3711, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 2.6151444184231067, |
|
"grad_norm": 0.2822958528995514, |
|
"learning_rate": 0.00018448535652497073, |
|
"loss": 0.3773, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 2.6229508196721314, |
|
"grad_norm": 0.2998236119747162, |
|
"learning_rate": 0.0001843391445812886, |
|
"loss": 0.3769, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 2.630757220921155, |
|
"grad_norm": 0.2773909866809845, |
|
"learning_rate": 0.00018419230541210139, |
|
"loss": 0.3656, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 2.63856362217018, |
|
"grad_norm": 0.2902582585811615, |
|
"learning_rate": 0.0001840448401094438, |
|
"loss": 0.3811, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 2.6463700234192036, |
|
"grad_norm": 0.28828758001327515, |
|
"learning_rate": 0.00018389674977000708, |
|
"loss": 0.3718, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 2.654176424668228, |
|
"grad_norm": 0.31182461977005005, |
|
"learning_rate": 0.0001837480354951308, |
|
"loss": 0.3719, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.661982825917252, |
|
"grad_norm": 0.2795233130455017, |
|
"learning_rate": 0.0001835986983907947, |
|
"loss": 0.3635, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 2.669789227166276, |
|
"grad_norm": 0.28305554389953613, |
|
"learning_rate": 0.00018344873956761045, |
|
"loss": 0.3659, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 2.6775956284153004, |
|
"grad_norm": 0.27487748861312866, |
|
"learning_rate": 0.00018329816014081346, |
|
"loss": 0.3738, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 2.6854020296643246, |
|
"grad_norm": 0.2843542695045471, |
|
"learning_rate": 0.00018314696123025454, |
|
"loss": 0.3729, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 2.693208430913349, |
|
"grad_norm": 0.28278475999832153, |
|
"learning_rate": 0.0001829951439603915, |
|
"loss": 0.3675, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 2.701014832162373, |
|
"grad_norm": 0.28049561381340027, |
|
"learning_rate": 0.00018284270946028092, |
|
"loss": 0.3722, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 2.7088212334113972, |
|
"grad_norm": 0.2751220464706421, |
|
"learning_rate": 0.00018268965886356964, |
|
"loss": 0.3708, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 2.7166276346604215, |
|
"grad_norm": 0.2924399971961975, |
|
"learning_rate": 0.00018253599330848638, |
|
"loss": 0.3728, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 2.7244340359094457, |
|
"grad_norm": 0.27712133526802063, |
|
"learning_rate": 0.00018238171393783335, |
|
"loss": 0.3716, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 2.73224043715847, |
|
"grad_norm": 0.28451547026634216, |
|
"learning_rate": 0.00018222682189897752, |
|
"loss": 0.3826, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 2.740046838407494, |
|
"grad_norm": 0.275689035654068, |
|
"learning_rate": 0.0001820713183438424, |
|
"loss": 0.3809, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 2.7478532396565183, |
|
"grad_norm": 0.2970561683177948, |
|
"learning_rate": 0.0001819152044288992, |
|
"loss": 0.3807, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 2.7556596409055425, |
|
"grad_norm": 0.27083873748779297, |
|
"learning_rate": 0.00018175848131515837, |
|
"loss": 0.3777, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 2.7634660421545667, |
|
"grad_norm": 0.2864970266819, |
|
"learning_rate": 0.000181601150168161, |
|
"loss": 0.3697, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 2.771272443403591, |
|
"grad_norm": 0.2779759466648102, |
|
"learning_rate": 0.00018144321215797, |
|
"loss": 0.371, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 2.779078844652615, |
|
"grad_norm": 0.2953694462776184, |
|
"learning_rate": 0.00018128466845916154, |
|
"loss": 0.3782, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 2.7868852459016393, |
|
"grad_norm": 0.27182242274284363, |
|
"learning_rate": 0.00018112552025081625, |
|
"loss": 0.3688, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 2.7946916471506635, |
|
"grad_norm": 0.28873521089553833, |
|
"learning_rate": 0.00018096576871651042, |
|
"loss": 0.3703, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 2.8024980483996877, |
|
"grad_norm": 0.2740427851676941, |
|
"learning_rate": 0.00018080541504430733, |
|
"loss": 0.3751, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 2.810304449648712, |
|
"grad_norm": 0.2902701199054718, |
|
"learning_rate": 0.00018064446042674828, |
|
"loss": 0.3691, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.818110850897736, |
|
"grad_norm": 0.28125742077827454, |
|
"learning_rate": 0.00018048290606084371, |
|
"loss": 0.3767, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 2.8259172521467604, |
|
"grad_norm": 0.28101617097854614, |
|
"learning_rate": 0.00018032075314806448, |
|
"loss": 0.3767, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 2.8337236533957846, |
|
"grad_norm": 0.31095340847969055, |
|
"learning_rate": 0.00018015800289433273, |
|
"loss": 0.3841, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 2.841530054644809, |
|
"grad_norm": 0.26109251379966736, |
|
"learning_rate": 0.00017999465651001296, |
|
"loss": 0.3821, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 2.849336455893833, |
|
"grad_norm": 0.293575257062912, |
|
"learning_rate": 0.00017983071520990315, |
|
"loss": 0.3809, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 0.26493799686431885, |
|
"learning_rate": 0.00017966618021322557, |
|
"loss": 0.3752, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 2.8649492583918814, |
|
"grad_norm": 0.2778892517089844, |
|
"learning_rate": 0.00017950105274361776, |
|
"loss": 0.369, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 2.8727556596409056, |
|
"grad_norm": 0.28492116928100586, |
|
"learning_rate": 0.00017933533402912354, |
|
"loss": 0.3714, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 2.88056206088993, |
|
"grad_norm": 0.281095415353775, |
|
"learning_rate": 0.00017916902530218368, |
|
"loss": 0.3727, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 2.888368462138954, |
|
"grad_norm": 0.277915358543396, |
|
"learning_rate": 0.0001790021277996269, |
|
"loss": 0.3728, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 2.8961748633879782, |
|
"grad_norm": 0.27478280663490295, |
|
"learning_rate": 0.00017883464276266064, |
|
"loss": 0.3799, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 2.9039812646370025, |
|
"grad_norm": 0.28556427359580994, |
|
"learning_rate": 0.00017866657143686168, |
|
"loss": 0.3762, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 2.9117876658860267, |
|
"grad_norm": 0.2784564197063446, |
|
"learning_rate": 0.0001784979150721672, |
|
"loss": 0.3772, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 2.919594067135051, |
|
"grad_norm": 0.3048378825187683, |
|
"learning_rate": 0.00017832867492286505, |
|
"loss": 0.3691, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 2.927400468384075, |
|
"grad_norm": 0.2872634828090668, |
|
"learning_rate": 0.0001781588522475848, |
|
"loss": 0.3686, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 2.9352068696330993, |
|
"grad_norm": 0.28422531485557556, |
|
"learning_rate": 0.00017798844830928817, |
|
"loss": 0.3814, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 2.9430132708821235, |
|
"grad_norm": 0.30014801025390625, |
|
"learning_rate": 0.0001778174643752598, |
|
"loss": 0.383, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 2.9508196721311473, |
|
"grad_norm": 0.3135896921157837, |
|
"learning_rate": 0.0001776459017170976, |
|
"loss": 0.3665, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 2.958626073380172, |
|
"grad_norm": 0.28260594606399536, |
|
"learning_rate": 0.00017747376161070344, |
|
"loss": 0.3856, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 2.9664324746291957, |
|
"grad_norm": 0.26851585507392883, |
|
"learning_rate": 0.0001773010453362737, |
|
"loss": 0.3838, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 2.9742388758782203, |
|
"grad_norm": 0.2853715121746063, |
|
"learning_rate": 0.00017712775417828968, |
|
"loss": 0.3776, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 2.982045277127244, |
|
"grad_norm": 0.28013601899147034, |
|
"learning_rate": 0.00017695388942550805, |
|
"loss": 0.3741, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 2.9898516783762688, |
|
"grad_norm": 0.270375519990921, |
|
"learning_rate": 0.00017677945237095123, |
|
"loss": 0.3707, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 2.9976580796252925, |
|
"grad_norm": 0.29110994935035706, |
|
"learning_rate": 0.0001766044443118978, |
|
"loss": 0.3774, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 2.999219359875098, |
|
"eval_loss": 1.0361436605453491, |
|
"eval_runtime": 0.337, |
|
"eval_samples_per_second": 14.838, |
|
"eval_steps_per_second": 2.968, |
|
"step": 1921 |
|
}, |
|
{ |
|
"epoch": 3.0054644808743167, |
|
"grad_norm": 0.27365708351135254, |
|
"learning_rate": 0.000176428866549873, |
|
"loss": 0.357, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 3.013270882123341, |
|
"grad_norm": 0.29497936367988586, |
|
"learning_rate": 0.00017625272039063883, |
|
"loss": 0.3487, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 3.021077283372365, |
|
"grad_norm": 0.28680676221847534, |
|
"learning_rate": 0.00017607600714418436, |
|
"loss": 0.3427, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 3.0288836846213893, |
|
"grad_norm": 0.29143229126930237, |
|
"learning_rate": 0.0001758987281247162, |
|
"loss": 0.3472, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 3.0366900858704136, |
|
"grad_norm": 0.28377851843833923, |
|
"learning_rate": 0.00017572088465064848, |
|
"loss": 0.3468, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 3.0444964871194378, |
|
"grad_norm": 0.27560412883758545, |
|
"learning_rate": 0.00017554247804459316, |
|
"loss": 0.3444, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 3.052302888368462, |
|
"grad_norm": 0.2871435880661011, |
|
"learning_rate": 0.00017536350963335022, |
|
"loss": 0.355, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 3.060109289617486, |
|
"grad_norm": 0.29234611988067627, |
|
"learning_rate": 0.00017518398074789775, |
|
"loss": 0.3405, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 3.0679156908665104, |
|
"grad_norm": 0.2977675199508667, |
|
"learning_rate": 0.000175003892723382, |
|
"loss": 0.3522, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 3.0757220921155346, |
|
"grad_norm": 0.2990856468677521, |
|
"learning_rate": 0.00017482324689910758, |
|
"loss": 0.352, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 3.083528493364559, |
|
"grad_norm": 0.28264856338500977, |
|
"learning_rate": 0.00017464204461852738, |
|
"loss": 0.3497, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 3.091334894613583, |
|
"grad_norm": 0.3084184229373932, |
|
"learning_rate": 0.00017446028722923265, |
|
"loss": 0.3586, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 3.0991412958626072, |
|
"grad_norm": 0.29917633533477783, |
|
"learning_rate": 0.00017427797608294288, |
|
"loss": 0.3542, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 3.1069476971116314, |
|
"grad_norm": 0.2877940237522125, |
|
"learning_rate": 0.00017409511253549593, |
|
"loss": 0.3527, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 3.1147540983606556, |
|
"grad_norm": 0.3050166070461273, |
|
"learning_rate": 0.00017391169794683776, |
|
"loss": 0.3481, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 3.12256049960968, |
|
"grad_norm": 0.29156407713890076, |
|
"learning_rate": 0.0001737277336810124, |
|
"loss": 0.3614, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 3.130366900858704, |
|
"grad_norm": 0.3042657971382141, |
|
"learning_rate": 0.00017354322110615188, |
|
"loss": 0.3608, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 3.1381733021077283, |
|
"grad_norm": 0.29401668906211853, |
|
"learning_rate": 0.00017335816159446584, |
|
"loss": 0.3618, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 3.1459797033567525, |
|
"grad_norm": 0.3223430812358856, |
|
"learning_rate": 0.00017317255652223157, |
|
"loss": 0.3637, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 3.1537861046057767, |
|
"grad_norm": 0.3130653500556946, |
|
"learning_rate": 0.00017298640726978357, |
|
"loss": 0.3561, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 3.161592505854801, |
|
"grad_norm": 0.3067520260810852, |
|
"learning_rate": 0.00017279971522150348, |
|
"loss": 0.3536, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 3.169398907103825, |
|
"grad_norm": 0.30372655391693115, |
|
"learning_rate": 0.00017261248176580958, |
|
"loss": 0.3511, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 3.1772053083528493, |
|
"grad_norm": 0.34251293540000916, |
|
"learning_rate": 0.00017242470829514672, |
|
"loss": 0.3591, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 3.1850117096018735, |
|
"grad_norm": 0.34350651502609253, |
|
"learning_rate": 0.00017223639620597556, |
|
"loss": 0.345, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 3.1928181108508977, |
|
"grad_norm": 0.29779571294784546, |
|
"learning_rate": 0.0001720475468987627, |
|
"loss": 0.3496, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 3.200624512099922, |
|
"grad_norm": 0.29767748713493347, |
|
"learning_rate": 0.0001718581617779698, |
|
"loss": 0.35, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 3.208430913348946, |
|
"grad_norm": 0.30058038234710693, |
|
"learning_rate": 0.00017166824225204352, |
|
"loss": 0.3481, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 3.2162373145979704, |
|
"grad_norm": 0.2958409786224365, |
|
"learning_rate": 0.00017147778973340466, |
|
"loss": 0.3554, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 3.2240437158469946, |
|
"grad_norm": 0.29515761137008667, |
|
"learning_rate": 0.000171286805638438, |
|
"loss": 0.3489, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 3.2318501170960188, |
|
"grad_norm": 0.30113205313682556, |
|
"learning_rate": 0.00017109529138748157, |
|
"loss": 0.3437, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 3.239656518345043, |
|
"grad_norm": 0.2853497862815857, |
|
"learning_rate": 0.0001709032484048162, |
|
"loss": 0.3515, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 3.247462919594067, |
|
"grad_norm": 0.3135252892971039, |
|
"learning_rate": 0.00017071067811865476, |
|
"loss": 0.3502, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 3.2552693208430914, |
|
"grad_norm": 0.30913615226745605, |
|
"learning_rate": 0.00017051758196113173, |
|
"loss": 0.3468, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 3.2630757220921156, |
|
"grad_norm": 0.2965530753135681, |
|
"learning_rate": 0.00017032396136829246, |
|
"loss": 0.3565, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 3.27088212334114, |
|
"grad_norm": 0.2974913418292999, |
|
"learning_rate": 0.00017012981778008243, |
|
"loss": 0.3442, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 3.278688524590164, |
|
"grad_norm": 0.2917117476463318, |
|
"learning_rate": 0.00016993515264033672, |
|
"loss": 0.3479, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 3.2864949258391882, |
|
"grad_norm": 0.29723039269447327, |
|
"learning_rate": 0.00016973996739676905, |
|
"loss": 0.3523, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 3.2943013270882124, |
|
"grad_norm": 0.3127197027206421, |
|
"learning_rate": 0.00016954426350096116, |
|
"loss": 0.3574, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 3.3021077283372366, |
|
"grad_norm": 0.28733986616134644, |
|
"learning_rate": 0.00016934804240835205, |
|
"loss": 0.3495, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 3.309914129586261, |
|
"grad_norm": 0.30774375796318054, |
|
"learning_rate": 0.00016915130557822695, |
|
"loss": 0.3515, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 3.317720530835285, |
|
"grad_norm": 0.3072982132434845, |
|
"learning_rate": 0.0001689540544737067, |
|
"loss": 0.3473, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 3.3255269320843093, |
|
"grad_norm": 0.30740848183631897, |
|
"learning_rate": 0.00016875629056173675, |
|
"loss": 0.3631, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 3.3333333333333335, |
|
"grad_norm": 0.2990294396877289, |
|
"learning_rate": 0.00016855801531307624, |
|
"loss": 0.3508, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 3.3411397345823577, |
|
"grad_norm": 0.28540557622909546, |
|
"learning_rate": 0.00016835923020228712, |
|
"loss": 0.353, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 3.348946135831382, |
|
"grad_norm": 0.29948052763938904, |
|
"learning_rate": 0.0001681599367077232, |
|
"loss": 0.3627, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 3.356752537080406, |
|
"grad_norm": 0.30102795362472534, |
|
"learning_rate": 0.00016796013631151897, |
|
"loss": 0.3526, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 3.3645589383294303, |
|
"grad_norm": 0.2947191298007965, |
|
"learning_rate": 0.00016775983049957887, |
|
"loss": 0.3645, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 3.3723653395784545, |
|
"grad_norm": 0.3223245143890381, |
|
"learning_rate": 0.00016755902076156604, |
|
"loss": 0.3623, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 3.3801717408274783, |
|
"grad_norm": 0.28947457671165466, |
|
"learning_rate": 0.00016735770859089123, |
|
"loss": 0.3638, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 3.387978142076503, |
|
"grad_norm": 0.28693681955337524, |
|
"learning_rate": 0.00016715589548470185, |
|
"loss": 0.3614, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 3.3957845433255267, |
|
"grad_norm": 0.29449644684791565, |
|
"learning_rate": 0.00016695358294387065, |
|
"loss": 0.3529, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 3.4035909445745514, |
|
"grad_norm": 0.289899080991745, |
|
"learning_rate": 0.00016675077247298474, |
|
"loss": 0.3556, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 3.411397345823575, |
|
"grad_norm": 0.2969209849834442, |
|
"learning_rate": 0.00016654746558033423, |
|
"loss": 0.3598, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 3.4192037470725998, |
|
"grad_norm": 0.3075496554374695, |
|
"learning_rate": 0.00016634366377790114, |
|
"loss": 0.3509, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 3.4270101483216235, |
|
"grad_norm": 0.28629180788993835, |
|
"learning_rate": 0.00016613936858134802, |
|
"loss": 0.3481, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 3.4348165495706477, |
|
"grad_norm": 0.3028337061405182, |
|
"learning_rate": 0.00016593458151000688, |
|
"loss": 0.358, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 3.442622950819672, |
|
"grad_norm": 0.2874985933303833, |
|
"learning_rate": 0.00016572930408686776, |
|
"loss": 0.3593, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 3.450429352068696, |
|
"grad_norm": 0.29345887899398804, |
|
"learning_rate": 0.00016552353783856732, |
|
"loss": 0.3485, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 3.4582357533177204, |
|
"grad_norm": 0.28079068660736084, |
|
"learning_rate": 0.00016531728429537766, |
|
"loss": 0.3606, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 3.4660421545667446, |
|
"grad_norm": 0.2924216389656067, |
|
"learning_rate": 0.0001651105449911949, |
|
"loss": 0.351, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 3.473848555815769, |
|
"grad_norm": 0.29718244075775146, |
|
"learning_rate": 0.0001649033214635277, |
|
"loss": 0.3521, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 3.481654957064793, |
|
"grad_norm": 0.2965914309024811, |
|
"learning_rate": 0.00016469561525348577, |
|
"loss": 0.3564, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 3.489461358313817, |
|
"grad_norm": 0.2874649167060852, |
|
"learning_rate": 0.0001644874279057686, |
|
"loss": 0.3554, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 3.4972677595628414, |
|
"grad_norm": 0.2855139970779419, |
|
"learning_rate": 0.00016427876096865394, |
|
"loss": 0.3619, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 3.5050741608118656, |
|
"grad_norm": 0.2949666976928711, |
|
"learning_rate": 0.0001640696159939861, |
|
"loss": 0.3558, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 3.51288056206089, |
|
"grad_norm": 0.29184606671333313, |
|
"learning_rate": 0.00016385999453716454, |
|
"loss": 0.3625, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 3.520686963309914, |
|
"grad_norm": 0.29414233565330505, |
|
"learning_rate": 0.00016364989815713233, |
|
"loss": 0.3559, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 3.5284933645589383, |
|
"grad_norm": 0.2970221936702728, |
|
"learning_rate": 0.00016343932841636456, |
|
"loss": 0.3488, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 3.5362997658079625, |
|
"grad_norm": 0.305912047624588, |
|
"learning_rate": 0.00016322828688085657, |
|
"loss": 0.3553, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 3.5441061670569867, |
|
"grad_norm": 0.3016732931137085, |
|
"learning_rate": 0.00016301677512011248, |
|
"loss": 0.359, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 3.551912568306011, |
|
"grad_norm": 0.2802044749259949, |
|
"learning_rate": 0.00016280479470713344, |
|
"loss": 0.347, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 3.559718969555035, |
|
"grad_norm": 0.3013835549354553, |
|
"learning_rate": 0.00016259234721840591, |
|
"loss": 0.3664, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 3.5675253708040593, |
|
"grad_norm": 0.2899465262889862, |
|
"learning_rate": 0.00016237943423389001, |
|
"loss": 0.3638, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 3.5753317720530835, |
|
"grad_norm": 0.2953374981880188, |
|
"learning_rate": 0.00016216605733700775, |
|
"loss": 0.3621, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 3.5831381733021077, |
|
"grad_norm": 0.28580641746520996, |
|
"learning_rate": 0.00016195221811463113, |
|
"loss": 0.3558, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 3.590944574551132, |
|
"grad_norm": 0.2812565565109253, |
|
"learning_rate": 0.00016173791815707051, |
|
"loss": 0.357, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 3.598750975800156, |
|
"grad_norm": 0.3074684739112854, |
|
"learning_rate": 0.00016152315905806268, |
|
"loss": 0.3593, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 3.6065573770491803, |
|
"grad_norm": 0.29887428879737854, |
|
"learning_rate": 0.0001613079424147591, |
|
"loss": 0.3607, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 3.6143637782982045, |
|
"grad_norm": 0.2820027470588684, |
|
"learning_rate": 0.00016109226982771393, |
|
"loss": 0.3535, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 3.6221701795472288, |
|
"grad_norm": 0.28952109813690186, |
|
"learning_rate": 0.00016087614290087208, |
|
"loss": 0.3563, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 3.629976580796253, |
|
"grad_norm": 0.29456931352615356, |
|
"learning_rate": 0.00016065956324155746, |
|
"loss": 0.3601, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 3.637782982045277, |
|
"grad_norm": 0.29095590114593506, |
|
"learning_rate": 0.0001604425324604609, |
|
"loss": 0.3633, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 3.6455893832943014, |
|
"grad_norm": 0.27606719732284546, |
|
"learning_rate": 0.00016022505217162822, |
|
"loss": 0.3621, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 3.6533957845433256, |
|
"grad_norm": 0.2846406400203705, |
|
"learning_rate": 0.0001600071239924481, |
|
"loss": 0.349, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 3.66120218579235, |
|
"grad_norm": 0.28319764137268066, |
|
"learning_rate": 0.00015978874954364033, |
|
"loss": 0.3514, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 3.669008587041374, |
|
"grad_norm": 0.3063547909259796, |
|
"learning_rate": 0.00015956993044924334, |
|
"loss": 0.3504, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 3.676814988290398, |
|
"grad_norm": 0.27655765414237976, |
|
"learning_rate": 0.0001593506683366026, |
|
"loss": 0.3466, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 3.6846213895394224, |
|
"grad_norm": 0.2881523072719574, |
|
"learning_rate": 0.00015913096483635824, |
|
"loss": 0.3466, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 3.6924277907884466, |
|
"grad_norm": 0.2950555682182312, |
|
"learning_rate": 0.00015891082158243282, |
|
"loss": 0.3512, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 3.700234192037471, |
|
"grad_norm": 0.2711721360683441, |
|
"learning_rate": 0.00015869024021201948, |
|
"loss": 0.3545, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 3.708040593286495, |
|
"grad_norm": 0.31127429008483887, |
|
"learning_rate": 0.00015846922236556946, |
|
"loss": 0.3616, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 3.7158469945355193, |
|
"grad_norm": 0.28716492652893066, |
|
"learning_rate": 0.00015824776968678024, |
|
"loss": 0.3548, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 3.7236533957845435, |
|
"grad_norm": 0.28540754318237305, |
|
"learning_rate": 0.00015802588382258292, |
|
"loss": 0.3571, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 3.7314597970335677, |
|
"grad_norm": 0.30483660101890564, |
|
"learning_rate": 0.00015780356642313036, |
|
"loss": 0.3475, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 3.739266198282592, |
|
"grad_norm": 0.28474995493888855, |
|
"learning_rate": 0.00015758081914178456, |
|
"loss": 0.3475, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 3.747072599531616, |
|
"grad_norm": 0.28476497530937195, |
|
"learning_rate": 0.0001573576436351046, |
|
"loss": 0.3525, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 3.7548790007806403, |
|
"grad_norm": 0.29259222745895386, |
|
"learning_rate": 0.00015713404156283435, |
|
"loss": 0.3683, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 3.762685402029664, |
|
"grad_norm": 0.2946162223815918, |
|
"learning_rate": 0.00015691001458788983, |
|
"loss": 0.3587, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 3.7704918032786887, |
|
"grad_norm": 0.2999446988105774, |
|
"learning_rate": 0.0001566855643763472, |
|
"loss": 0.355, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 3.7782982045277125, |
|
"grad_norm": 0.2700735032558441, |
|
"learning_rate": 0.00015646069259743007, |
|
"loss": 0.3531, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 3.786104605776737, |
|
"grad_norm": 0.28989067673683167, |
|
"learning_rate": 0.00015623540092349732, |
|
"loss": 0.3568, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 3.793911007025761, |
|
"grad_norm": 0.2979315221309662, |
|
"learning_rate": 0.00015600969103003056, |
|
"loss": 0.3514, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 3.8017174082747855, |
|
"grad_norm": 0.27748894691467285, |
|
"learning_rate": 0.00015578356459562163, |
|
"loss": 0.3474, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 3.8095238095238093, |
|
"grad_norm": 0.28995972871780396, |
|
"learning_rate": 0.00015555702330196023, |
|
"loss": 0.352, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 3.817330210772834, |
|
"grad_norm": 0.2819332480430603, |
|
"learning_rate": 0.00015533006883382123, |
|
"loss": 0.3482, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 3.8251366120218577, |
|
"grad_norm": 0.2890748083591461, |
|
"learning_rate": 0.0001551027028790524, |
|
"loss": 0.3608, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 3.8329430132708824, |
|
"grad_norm": 0.295684278011322, |
|
"learning_rate": 0.0001548749271285616, |
|
"loss": 0.3541, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 3.840749414519906, |
|
"grad_norm": 0.2782337963581085, |
|
"learning_rate": 0.00015464674327630434, |
|
"loss": 0.3467, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 3.848555815768931, |
|
"grad_norm": 0.2889052629470825, |
|
"learning_rate": 0.00015441815301927122, |
|
"loss": 0.3597, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 3.8563622170179546, |
|
"grad_norm": 0.27629801630973816, |
|
"learning_rate": 0.00015418915805747517, |
|
"loss": 0.3543, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 3.8641686182669788, |
|
"grad_norm": 0.2963428497314453, |
|
"learning_rate": 0.00015395976009393894, |
|
"loss": 0.3637, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 3.871975019516003, |
|
"grad_norm": 0.29081323742866516, |
|
"learning_rate": 0.0001537299608346824, |
|
"loss": 0.3563, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 3.879781420765027, |
|
"grad_norm": 0.29357287287712097, |
|
"learning_rate": 0.00015349976198870973, |
|
"loss": 0.3633, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 3.8875878220140514, |
|
"grad_norm": 0.29132893681526184, |
|
"learning_rate": 0.00015326916526799692, |
|
"loss": 0.3563, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 3.8953942232630756, |
|
"grad_norm": 0.29773959517478943, |
|
"learning_rate": 0.00015303817238747892, |
|
"loss": 0.3511, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 3.9032006245121, |
|
"grad_norm": 0.2842431366443634, |
|
"learning_rate": 0.0001528067850650368, |
|
"loss": 0.3533, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 3.911007025761124, |
|
"grad_norm": 0.277873694896698, |
|
"learning_rate": 0.00015257500502148522, |
|
"loss": 0.3549, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 3.9188134270101482, |
|
"grad_norm": 0.2919045388698578, |
|
"learning_rate": 0.0001523428339805594, |
|
"loss": 0.3521, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 3.9266198282591724, |
|
"grad_norm": 0.289386123418808, |
|
"learning_rate": 0.00015211027366890235, |
|
"loss": 0.3603, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 3.9344262295081966, |
|
"grad_norm": 0.27099329233169556, |
|
"learning_rate": 0.00015187732581605217, |
|
"loss": 0.3553, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 3.942232630757221, |
|
"grad_norm": 0.2989470660686493, |
|
"learning_rate": 0.00015164399215442898, |
|
"loss": 0.3562, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 3.950039032006245, |
|
"grad_norm": 0.2968309223651886, |
|
"learning_rate": 0.00015141027441932216, |
|
"loss": 0.3581, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 3.9578454332552693, |
|
"grad_norm": 0.29724830389022827, |
|
"learning_rate": 0.00015117617434887753, |
|
"loss": 0.3514, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 3.9656518345042935, |
|
"grad_norm": 0.30498626828193665, |
|
"learning_rate": 0.0001509416936840842, |
|
"loss": 0.3691, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 3.9734582357533177, |
|
"grad_norm": 0.28580141067504883, |
|
"learning_rate": 0.00015070683416876172, |
|
"loss": 0.3509, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 3.981264637002342, |
|
"grad_norm": 0.28979378938674927, |
|
"learning_rate": 0.0001504715975495472, |
|
"loss": 0.3598, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 3.989071038251366, |
|
"grad_norm": 0.30698639154434204, |
|
"learning_rate": 0.0001502359855758823, |
|
"loss": 0.357, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 3.9968774395003903, |
|
"grad_norm": 0.2776874303817749, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 0.355, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 1.041857361793518, |
|
"eval_runtime": 0.3229, |
|
"eval_samples_per_second": 15.485, |
|
"eval_steps_per_second": 3.097, |
|
"step": 2562 |
|
}, |
|
{ |
|
"epoch": 4.004683840749414, |
|
"grad_norm": 0.2781620919704437, |
|
"learning_rate": 0.00014976364257691192, |
|
"loss": 0.3413, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 4.012490241998439, |
|
"grad_norm": 0.3088448643684387, |
|
"learning_rate": 0.00014952691506439495, |
|
"loss": 0.3303, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 4.0202966432474625, |
|
"grad_norm": 0.28063511848449707, |
|
"learning_rate": 0.00014928981922297842, |
|
"loss": 0.3268, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 4.028103044496487, |
|
"grad_norm": 0.32527846097946167, |
|
"learning_rate": 0.0001490523568159308, |
|
"loss": 0.3184, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 4.035909445745511, |
|
"grad_norm": 0.29716962575912476, |
|
"learning_rate": 0.00014881452960924677, |
|
"loss": 0.3287, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 4.043715846994536, |
|
"grad_norm": 0.30141621828079224, |
|
"learning_rate": 0.00014857633937163402, |
|
"loss": 0.3331, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 4.051522248243559, |
|
"grad_norm": 0.2995307743549347, |
|
"learning_rate": 0.00014833778787449998, |
|
"loss": 0.3313, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 4.059328649492584, |
|
"grad_norm": 0.30431655049324036, |
|
"learning_rate": 0.00014809887689193877, |
|
"loss": 0.334, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 4.067135050741608, |
|
"grad_norm": 0.3243234157562256, |
|
"learning_rate": 0.0001478596082007181, |
|
"loss": 0.3339, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 4.074941451990632, |
|
"grad_norm": 0.3121318817138672, |
|
"learning_rate": 0.0001476199835802658, |
|
"loss": 0.3306, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 4.082747853239656, |
|
"grad_norm": 0.31301409006118774, |
|
"learning_rate": 0.0001473800048126568, |
|
"loss": 0.3336, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 4.090554254488681, |
|
"grad_norm": 0.29308003187179565, |
|
"learning_rate": 0.0001471396736825998, |
|
"loss": 0.3239, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 4.098360655737705, |
|
"grad_norm": 0.30689167976379395, |
|
"learning_rate": 0.0001468989919774239, |
|
"loss": 0.3342, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 4.106167056986729, |
|
"grad_norm": 0.32097750902175903, |
|
"learning_rate": 0.0001466579614870656, |
|
"loss": 0.3386, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 4.113973458235753, |
|
"grad_norm": 0.3277100622653961, |
|
"learning_rate": 0.00014641658400405515, |
|
"loss": 0.3434, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 4.121779859484778, |
|
"grad_norm": 0.30586951971054077, |
|
"learning_rate": 0.00014617486132350343, |
|
"loss": 0.3328, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 4.129586260733801, |
|
"grad_norm": 0.30447080731391907, |
|
"learning_rate": 0.0001459327952430884, |
|
"loss": 0.3343, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 4.137392661982826, |
|
"grad_norm": 0.31396037340164185, |
|
"learning_rate": 0.00014569038756304207, |
|
"loss": 0.3331, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 4.14519906323185, |
|
"grad_norm": 0.30158740282058716, |
|
"learning_rate": 0.0001454476400861368, |
|
"loss": 0.33, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 4.1530054644808745, |
|
"grad_norm": 0.30618518590927124, |
|
"learning_rate": 0.000145204554617672, |
|
"loss": 0.3297, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 4.160811865729898, |
|
"grad_norm": 0.3204987049102783, |
|
"learning_rate": 0.00014496113296546067, |
|
"loss": 0.3356, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 4.168618266978923, |
|
"grad_norm": 0.31640496850013733, |
|
"learning_rate": 0.0001447173769398161, |
|
"loss": 0.3397, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 4.176424668227947, |
|
"grad_norm": 0.3094062805175781, |
|
"learning_rate": 0.0001444732883535382, |
|
"loss": 0.3359, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 4.184231069476971, |
|
"grad_norm": 0.3020561933517456, |
|
"learning_rate": 0.00014422886902190014, |
|
"loss": 0.3313, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 4.192037470725995, |
|
"grad_norm": 0.3145982027053833, |
|
"learning_rate": 0.00014398412076263486, |
|
"loss": 0.3441, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 4.19984387197502, |
|
"grad_norm": 0.3120006322860718, |
|
"learning_rate": 0.00014373904539592144, |
|
"loss": 0.3334, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 4.2076502732240435, |
|
"grad_norm": 0.319619357585907, |
|
"learning_rate": 0.00014349364474437173, |
|
"loss": 0.3287, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 4.215456674473068, |
|
"grad_norm": 0.3287660479545593, |
|
"learning_rate": 0.00014324792063301662, |
|
"loss": 0.3296, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 4.223263075722092, |
|
"grad_norm": 0.3166675567626953, |
|
"learning_rate": 0.00014300187488929265, |
|
"loss": 0.3433, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 4.231069476971117, |
|
"grad_norm": 0.3454567492008209, |
|
"learning_rate": 0.00014275550934302823, |
|
"loss": 0.3384, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 4.23887587822014, |
|
"grad_norm": 0.30334779620170593, |
|
"learning_rate": 0.00014250882582643012, |
|
"loss": 0.334, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 4.246682279469165, |
|
"grad_norm": 0.31963109970092773, |
|
"learning_rate": 0.00014226182617406996, |
|
"loss": 0.3428, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 4.254488680718189, |
|
"grad_norm": 0.30890125036239624, |
|
"learning_rate": 0.00014201451222287025, |
|
"loss": 0.3326, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 4.262295081967213, |
|
"grad_norm": 0.32028844952583313, |
|
"learning_rate": 0.0001417668858120911, |
|
"loss": 0.333, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 4.270101483216237, |
|
"grad_norm": 0.3037426471710205, |
|
"learning_rate": 0.00014151894878331622, |
|
"loss": 0.3302, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 4.277907884465262, |
|
"grad_norm": 0.305708110332489, |
|
"learning_rate": 0.00014127070298043947, |
|
"loss": 0.3387, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 4.285714285714286, |
|
"grad_norm": 0.3226141333580017, |
|
"learning_rate": 0.00014102215024965105, |
|
"loss": 0.3352, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 4.29352068696331, |
|
"grad_norm": 0.3047237992286682, |
|
"learning_rate": 0.00014077329243942369, |
|
"loss": 0.3377, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 4.301327088212334, |
|
"grad_norm": 0.3010615408420563, |
|
"learning_rate": 0.000140524131400499, |
|
"loss": 0.3325, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 4.309133489461359, |
|
"grad_norm": 0.3127407431602478, |
|
"learning_rate": 0.00014027466898587374, |
|
"loss": 0.3357, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 4.316939890710382, |
|
"grad_norm": 0.31216832995414734, |
|
"learning_rate": 0.00014002490705078592, |
|
"loss": 0.346, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 4.324746291959407, |
|
"grad_norm": 0.31948304176330566, |
|
"learning_rate": 0.00013977484745270112, |
|
"loss": 0.3292, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 4.332552693208431, |
|
"grad_norm": 0.3235209584236145, |
|
"learning_rate": 0.00013952449205129855, |
|
"loss": 0.3415, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 4.3403590944574555, |
|
"grad_norm": 0.3177550137042999, |
|
"learning_rate": 0.0001392738427084574, |
|
"loss": 0.3341, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 4.348165495706479, |
|
"grad_norm": 0.31263765692710876, |
|
"learning_rate": 0.00013902290128824283, |
|
"loss": 0.3348, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 4.355971896955504, |
|
"grad_norm": 0.31187278032302856, |
|
"learning_rate": 0.00013877166965689205, |
|
"loss": 0.3405, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 4.363778298204528, |
|
"grad_norm": 0.3060557246208191, |
|
"learning_rate": 0.00013852014968280082, |
|
"loss": 0.3328, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 4.371584699453552, |
|
"grad_norm": 0.30810457468032837, |
|
"learning_rate": 0.000138268343236509, |
|
"loss": 0.3303, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 4.379391100702576, |
|
"grad_norm": 0.309929758310318, |
|
"learning_rate": 0.0001380162521906871, |
|
"loss": 0.3371, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 4.387197501951601, |
|
"grad_norm": 0.31695547699928284, |
|
"learning_rate": 0.00013776387842012218, |
|
"loss": 0.3393, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 4.3950039032006245, |
|
"grad_norm": 0.31112900376319885, |
|
"learning_rate": 0.0001375112238017038, |
|
"loss": 0.338, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 4.402810304449648, |
|
"grad_norm": 0.29906710982322693, |
|
"learning_rate": 0.0001372582902144103, |
|
"loss": 0.3318, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 4.410616705698673, |
|
"grad_norm": 0.30970847606658936, |
|
"learning_rate": 0.00013700507953929463, |
|
"loss": 0.3387, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 4.418423106947698, |
|
"grad_norm": 0.3163338005542755, |
|
"learning_rate": 0.00013675159365947036, |
|
"loss": 0.3371, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 4.426229508196721, |
|
"grad_norm": 0.31127607822418213, |
|
"learning_rate": 0.0001364978344600979, |
|
"loss": 0.3372, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 4.434035909445745, |
|
"grad_norm": 0.2967492640018463, |
|
"learning_rate": 0.00013624380382837016, |
|
"loss": 0.339, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 4.44184231069477, |
|
"grad_norm": 0.31639572978019714, |
|
"learning_rate": 0.00013598950365349883, |
|
"loss": 0.3411, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 4.4496487119437935, |
|
"grad_norm": 0.3143470883369446, |
|
"learning_rate": 0.00013573493582670003, |
|
"loss": 0.3386, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 4.457455113192818, |
|
"grad_norm": 0.3157510757446289, |
|
"learning_rate": 0.0001354801022411805, |
|
"loss": 0.3339, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 4.465261514441842, |
|
"grad_norm": 0.3069648742675781, |
|
"learning_rate": 0.00013522500479212337, |
|
"loss": 0.3385, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 4.473067915690867, |
|
"grad_norm": 0.31706541776657104, |
|
"learning_rate": 0.0001349696453766741, |
|
"loss": 0.3353, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 4.48087431693989, |
|
"grad_norm": 0.3200201988220215, |
|
"learning_rate": 0.00013471402589392638, |
|
"loss": 0.3419, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 4.488680718188915, |
|
"grad_norm": 0.32029232382774353, |
|
"learning_rate": 0.00013445814824490805, |
|
"loss": 0.3417, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 4.496487119437939, |
|
"grad_norm": 0.3023779094219208, |
|
"learning_rate": 0.00013420201433256689, |
|
"loss": 0.3388, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 4.504293520686963, |
|
"grad_norm": 0.3167456090450287, |
|
"learning_rate": 0.00013394562606175648, |
|
"loss": 0.335, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 4.512099921935987, |
|
"grad_norm": 0.31978872418403625, |
|
"learning_rate": 0.000133688985339222, |
|
"loss": 0.3425, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 4.519906323185012, |
|
"grad_norm": 0.32426413893699646, |
|
"learning_rate": 0.00013343209407358627, |
|
"loss": 0.3325, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 4.527712724434036, |
|
"grad_norm": 0.28932276368141174, |
|
"learning_rate": 0.00013317495417533524, |
|
"loss": 0.3365, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 4.53551912568306, |
|
"grad_norm": 0.31238290667533875, |
|
"learning_rate": 0.00013291756755680388, |
|
"loss": 0.3353, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 4.543325526932084, |
|
"grad_norm": 0.3126748204231262, |
|
"learning_rate": 0.00013265993613216224, |
|
"loss": 0.3363, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 4.551131928181109, |
|
"grad_norm": 0.331447571516037, |
|
"learning_rate": 0.0001324020618174007, |
|
"loss": 0.3434, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 4.558938329430132, |
|
"grad_norm": 0.31981754302978516, |
|
"learning_rate": 0.00013214394653031616, |
|
"loss": 0.3321, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 4.566744730679157, |
|
"grad_norm": 0.30969661474227905, |
|
"learning_rate": 0.0001318855921904976, |
|
"loss": 0.3432, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 4.574551131928181, |
|
"grad_norm": 0.3188318610191345, |
|
"learning_rate": 0.00013162700071931184, |
|
"loss": 0.3397, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 4.5823575331772055, |
|
"grad_norm": 0.3379952907562256, |
|
"learning_rate": 0.00013136817403988917, |
|
"loss": 0.3389, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 4.590163934426229, |
|
"grad_norm": 0.2957393229007721, |
|
"learning_rate": 0.00013110911407710908, |
|
"loss": 0.3294, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 4.597970335675254, |
|
"grad_norm": 0.30884405970573425, |
|
"learning_rate": 0.00013084982275758615, |
|
"loss": 0.3367, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 4.605776736924278, |
|
"grad_norm": 0.31198689341545105, |
|
"learning_rate": 0.00013059030200965536, |
|
"loss": 0.3301, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 4.613583138173302, |
|
"grad_norm": 0.3176429271697998, |
|
"learning_rate": 0.00013033055376335804, |
|
"loss": 0.3356, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 4.621389539422326, |
|
"grad_norm": 0.3074142038822174, |
|
"learning_rate": 0.00013007057995042732, |
|
"loss": 0.3315, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 4.629195940671351, |
|
"grad_norm": 0.30663779377937317, |
|
"learning_rate": 0.00012981038250427403, |
|
"loss": 0.3376, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 4.6370023419203745, |
|
"grad_norm": 0.30579113960266113, |
|
"learning_rate": 0.0001295499633599719, |
|
"loss": 0.3351, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 4.644808743169399, |
|
"grad_norm": 0.3152572214603424, |
|
"learning_rate": 0.00012928932445424365, |
|
"loss": 0.3369, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 4.652615144418423, |
|
"grad_norm": 0.3045886754989624, |
|
"learning_rate": 0.00012902846772544624, |
|
"loss": 0.3317, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 4.660421545667448, |
|
"grad_norm": 0.3164273500442505, |
|
"learning_rate": 0.00012876739511355657, |
|
"loss": 0.3408, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 4.668227946916471, |
|
"grad_norm": 0.3016981780529022, |
|
"learning_rate": 0.0001285061085601571, |
|
"loss": 0.3417, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 4.676034348165496, |
|
"grad_norm": 0.3070276081562042, |
|
"learning_rate": 0.00012824461000842126, |
|
"loss": 0.3283, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 4.68384074941452, |
|
"grad_norm": 0.30681294202804565, |
|
"learning_rate": 0.00012798290140309923, |
|
"loss": 0.3395, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 4.691647150663544, |
|
"grad_norm": 0.3165208697319031, |
|
"learning_rate": 0.0001277209846905032, |
|
"loss": 0.3358, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 4.699453551912568, |
|
"grad_norm": 0.30810585618019104, |
|
"learning_rate": 0.00012745886181849325, |
|
"loss": 0.3409, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 4.707259953161593, |
|
"grad_norm": 0.3146737515926361, |
|
"learning_rate": 0.00012719653473646243, |
|
"loss": 0.3449, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 4.715066354410617, |
|
"grad_norm": 0.29630905389785767, |
|
"learning_rate": 0.00012693400539532263, |
|
"loss": 0.3384, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 4.722872755659641, |
|
"grad_norm": 0.3214413523674011, |
|
"learning_rate": 0.00012667127574748986, |
|
"loss": 0.3347, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 4.730679156908665, |
|
"grad_norm": 0.3026057183742523, |
|
"learning_rate": 0.00012640834774686985, |
|
"loss": 0.3342, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 4.73848555815769, |
|
"grad_norm": 0.31248968839645386, |
|
"learning_rate": 0.00012614522334884344, |
|
"loss": 0.3381, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 4.7462919594067134, |
|
"grad_norm": 0.3201672434806824, |
|
"learning_rate": 0.00012588190451025207, |
|
"loss": 0.3282, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 4.754098360655737, |
|
"grad_norm": 0.32542794942855835, |
|
"learning_rate": 0.00012561839318938326, |
|
"loss": 0.3464, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 4.761904761904762, |
|
"grad_norm": 0.2962396740913391, |
|
"learning_rate": 0.00012535469134595595, |
|
"loss": 0.3451, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 4.7697111631537865, |
|
"grad_norm": 0.3082274794578552, |
|
"learning_rate": 0.00012509080094110604, |
|
"loss": 0.3325, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 4.77751756440281, |
|
"grad_norm": 0.31192833185195923, |
|
"learning_rate": 0.00012482672393737164, |
|
"loss": 0.3348, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 4.785323965651834, |
|
"grad_norm": 0.30983906984329224, |
|
"learning_rate": 0.00012456246229867877, |
|
"loss": 0.3418, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 4.793130366900859, |
|
"grad_norm": 0.34676092863082886, |
|
"learning_rate": 0.0001242980179903264, |
|
"loss": 0.3435, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 4.800936768149883, |
|
"grad_norm": 0.3103564977645874, |
|
"learning_rate": 0.0001240333929789721, |
|
"loss": 0.3392, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 4.808743169398907, |
|
"grad_norm": 0.32498899102211, |
|
"learning_rate": 0.00012376858923261733, |
|
"loss": 0.3387, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 4.816549570647931, |
|
"grad_norm": 0.3146662414073944, |
|
"learning_rate": 0.0001235036087205927, |
|
"loss": 0.3406, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 4.8243559718969555, |
|
"grad_norm": 0.32241055369377136, |
|
"learning_rate": 0.0001232384534135435, |
|
"loss": 0.3421, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 4.83216237314598, |
|
"grad_norm": 0.29581108689308167, |
|
"learning_rate": 0.00012297312528341484, |
|
"loss": 0.3309, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 4.839968774395004, |
|
"grad_norm": 0.3068925440311432, |
|
"learning_rate": 0.00012270762630343734, |
|
"loss": 0.3419, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 4.847775175644028, |
|
"grad_norm": 0.3263617753982544, |
|
"learning_rate": 0.000122441958448112, |
|
"loss": 0.3359, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 4.855581576893052, |
|
"grad_norm": 0.3101602792739868, |
|
"learning_rate": 0.00012217612369319579, |
|
"loss": 0.342, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 4.863387978142076, |
|
"grad_norm": 0.3126956522464752, |
|
"learning_rate": 0.00012191012401568698, |
|
"loss": 0.3386, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 4.871194379391101, |
|
"grad_norm": 0.3099190890789032, |
|
"learning_rate": 0.00012164396139381029, |
|
"loss": 0.3288, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 4.8790007806401245, |
|
"grad_norm": 0.3113436996936798, |
|
"learning_rate": 0.00012137763780700227, |
|
"loss": 0.3412, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 4.886807181889149, |
|
"grad_norm": 0.29880639910697937, |
|
"learning_rate": 0.00012111115523589651, |
|
"loss": 0.3292, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 4.894613583138173, |
|
"grad_norm": 0.31602269411087036, |
|
"learning_rate": 0.00012084451566230906, |
|
"loss": 0.3372, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 4.902419984387198, |
|
"grad_norm": 0.3172509968280792, |
|
"learning_rate": 0.00012057772106922349, |
|
"loss": 0.3395, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 4.910226385636221, |
|
"grad_norm": 0.31268373131752014, |
|
"learning_rate": 0.0001203107734407763, |
|
"loss": 0.339, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 4.918032786885246, |
|
"grad_norm": 0.3162590265274048, |
|
"learning_rate": 0.00012004367476224206, |
|
"loss": 0.3484, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 4.92583918813427, |
|
"grad_norm": 0.34318986535072327, |
|
"learning_rate": 0.00011977642702001875, |
|
"loss": 0.3473, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 4.9336455893832944, |
|
"grad_norm": 0.2913156747817993, |
|
"learning_rate": 0.00011950903220161285, |
|
"loss": 0.3415, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 4.941451990632318, |
|
"grad_norm": 0.29538801312446594, |
|
"learning_rate": 0.0001192414922956247, |
|
"loss": 0.3361, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 4.949258391881343, |
|
"grad_norm": 0.3098941743373871, |
|
"learning_rate": 0.00011897380929173364, |
|
"loss": 0.3384, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 4.957064793130367, |
|
"grad_norm": 0.3089110851287842, |
|
"learning_rate": 0.0001187059851806832, |
|
"loss": 0.335, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 4.964871194379391, |
|
"grad_norm": 0.29929882287979126, |
|
"learning_rate": 0.00011843802195426635, |
|
"loss": 0.3364, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 4.972677595628415, |
|
"grad_norm": 0.31270211935043335, |
|
"learning_rate": 0.00011816992160531062, |
|
"loss": 0.3482, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 4.98048399687744, |
|
"grad_norm": 0.3091076910495758, |
|
"learning_rate": 0.0001179016861276633, |
|
"loss": 0.3387, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 4.9882903981264635, |
|
"grad_norm": 0.3052199184894562, |
|
"learning_rate": 0.00011763331751617663, |
|
"loss": 0.3454, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 4.996096799375488, |
|
"grad_norm": 0.2918078303337097, |
|
"learning_rate": 0.00011736481776669306, |
|
"loss": 0.3413, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 4.999219359875098, |
|
"eval_loss": 1.0547945499420166, |
|
"eval_runtime": 0.336, |
|
"eval_samples_per_second": 14.88, |
|
"eval_steps_per_second": 2.976, |
|
"step": 3202 |
|
}, |
|
{ |
|
"epoch": 5.003903200624512, |
|
"grad_norm": 0.2900161147117615, |
|
"learning_rate": 0.00011709618887603014, |
|
"loss": 0.3236, |
|
"step": 3205 |
|
}, |
|
{ |
|
"epoch": 5.0117096018735365, |
|
"grad_norm": 0.3155598044395447, |
|
"learning_rate": 0.00011682743284196593, |
|
"loss": 0.3129, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 5.01951600312256, |
|
"grad_norm": 0.3263396918773651, |
|
"learning_rate": 0.00011655855166322409, |
|
"loss": 0.3154, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 5.027322404371585, |
|
"grad_norm": 0.3178098797798157, |
|
"learning_rate": 0.0001162895473394589, |
|
"loss": 0.3133, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 5.035128805620609, |
|
"grad_norm": 0.3300285339355469, |
|
"learning_rate": 0.00011602042187124045, |
|
"loss": 0.3103, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 5.042935206869633, |
|
"grad_norm": 0.3271562159061432, |
|
"learning_rate": 0.00011575117726003979, |
|
"loss": 0.3092, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 5.050741608118657, |
|
"grad_norm": 0.3253972828388214, |
|
"learning_rate": 0.00011548181550821411, |
|
"loss": 0.3076, |
|
"step": 3235 |
|
}, |
|
{ |
|
"epoch": 5.058548009367682, |
|
"grad_norm": 0.33159464597702026, |
|
"learning_rate": 0.00011521233861899167, |
|
"loss": 0.3223, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 5.0663544106167056, |
|
"grad_norm": 0.3398463726043701, |
|
"learning_rate": 0.00011494274859645707, |
|
"loss": 0.3163, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 5.07416081186573, |
|
"grad_norm": 0.332904577255249, |
|
"learning_rate": 0.00011467304744553618, |
|
"loss": 0.308, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 5.081967213114754, |
|
"grad_norm": 0.32349440455436707, |
|
"learning_rate": 0.00011440323717198144, |
|
"loss": 0.3045, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 5.089773614363779, |
|
"grad_norm": 0.3356363773345947, |
|
"learning_rate": 0.00011413331978235678, |
|
"loss": 0.3188, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 5.097580015612802, |
|
"grad_norm": 0.3371128737926483, |
|
"learning_rate": 0.00011386329728402268, |
|
"loss": 0.3103, |
|
"step": 3265 |
|
}, |
|
{ |
|
"epoch": 5.105386416861827, |
|
"grad_norm": 0.3300149142742157, |
|
"learning_rate": 0.00011359317168512142, |
|
"loss": 0.3055, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 5.113192818110851, |
|
"grad_norm": 0.3270377814769745, |
|
"learning_rate": 0.000113322944994562, |
|
"loss": 0.3115, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 5.1209992193598755, |
|
"grad_norm": 0.3231879472732544, |
|
"learning_rate": 0.00011305261922200519, |
|
"loss": 0.3075, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 5.128805620608899, |
|
"grad_norm": 0.33695080876350403, |
|
"learning_rate": 0.00011278219637784861, |
|
"loss": 0.316, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 5.136612021857924, |
|
"grad_norm": 0.34212931990623474, |
|
"learning_rate": 0.00011251167847321193, |
|
"loss": 0.3158, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 5.144418423106948, |
|
"grad_norm": 0.3377404808998108, |
|
"learning_rate": 0.00011224106751992163, |
|
"loss": 0.317, |
|
"step": 3295 |
|
}, |
|
{ |
|
"epoch": 5.152224824355972, |
|
"grad_norm": 0.34452709555625916, |
|
"learning_rate": 0.00011197036553049625, |
|
"loss": 0.3228, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 5.160031225604996, |
|
"grad_norm": 0.3356159031391144, |
|
"learning_rate": 0.00011169957451813138, |
|
"loss": 0.3124, |
|
"step": 3305 |
|
}, |
|
{ |
|
"epoch": 5.167837626854021, |
|
"grad_norm": 0.3585182726383209, |
|
"learning_rate": 0.00011142869649668466, |
|
"loss": 0.3168, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 5.1756440281030445, |
|
"grad_norm": 0.3430497348308563, |
|
"learning_rate": 0.00011115773348066074, |
|
"loss": 0.3178, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 5.183450429352069, |
|
"grad_norm": 0.3532644808292389, |
|
"learning_rate": 0.00011088668748519647, |
|
"loss": 0.3192, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 5.191256830601093, |
|
"grad_norm": 0.34457042813301086, |
|
"learning_rate": 0.00011061556052604578, |
|
"loss": 0.3181, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 5.199063231850117, |
|
"grad_norm": 0.33349376916885376, |
|
"learning_rate": 0.00011034435461956466, |
|
"loss": 0.3121, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 5.206869633099141, |
|
"grad_norm": 0.3395507335662842, |
|
"learning_rate": 0.00011007307178269632, |
|
"loss": 0.3183, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 5.214676034348165, |
|
"grad_norm": 0.3544360399246216, |
|
"learning_rate": 0.0001098017140329561, |
|
"loss": 0.3243, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 5.22248243559719, |
|
"grad_norm": 0.3494921326637268, |
|
"learning_rate": 0.0001095302833884163, |
|
"loss": 0.3146, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 5.2302888368462135, |
|
"grad_norm": 0.33652257919311523, |
|
"learning_rate": 0.00010925878186769158, |
|
"loss": 0.3135, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 5.238095238095238, |
|
"grad_norm": 0.3407340943813324, |
|
"learning_rate": 0.00010898721148992351, |
|
"loss": 0.3205, |
|
"step": 3355 |
|
}, |
|
{ |
|
"epoch": 5.245901639344262, |
|
"grad_norm": 0.3510075509548187, |
|
"learning_rate": 0.00010871557427476583, |
|
"loss": 0.3127, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 5.2537080405932866, |
|
"grad_norm": 0.34021666646003723, |
|
"learning_rate": 0.00010844387224236932, |
|
"loss": 0.3192, |
|
"step": 3365 |
|
}, |
|
{ |
|
"epoch": 5.26151444184231, |
|
"grad_norm": 0.32510459423065186, |
|
"learning_rate": 0.00010817210741336684, |
|
"loss": 0.3108, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 5.269320843091335, |
|
"grad_norm": 0.36883097887039185, |
|
"learning_rate": 0.00010790028180885821, |
|
"loss": 0.3212, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 5.277127244340359, |
|
"grad_norm": 0.3306715190410614, |
|
"learning_rate": 0.00010762839745039526, |
|
"loss": 0.325, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 5.284933645589383, |
|
"grad_norm": 0.3390478491783142, |
|
"learning_rate": 0.00010735645635996676, |
|
"loss": 0.3159, |
|
"step": 3385 |
|
}, |
|
{ |
|
"epoch": 5.292740046838407, |
|
"grad_norm": 0.32592055201530457, |
|
"learning_rate": 0.00010708446055998343, |
|
"loss": 0.3127, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 5.300546448087432, |
|
"grad_norm": 0.3241756856441498, |
|
"learning_rate": 0.00010681241207326282, |
|
"loss": 0.3087, |
|
"step": 3395 |
|
}, |
|
{ |
|
"epoch": 5.308352849336456, |
|
"grad_norm": 0.3321205675601959, |
|
"learning_rate": 0.00010654031292301432, |
|
"loss": 0.3191, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 5.31615925058548, |
|
"grad_norm": 0.3427116572856903, |
|
"learning_rate": 0.00010626816513282412, |
|
"loss": 0.3167, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 5.323965651834504, |
|
"grad_norm": 0.33668723702430725, |
|
"learning_rate": 0.00010599597072664012, |
|
"loss": 0.3173, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 5.331772053083529, |
|
"grad_norm": 0.3475385010242462, |
|
"learning_rate": 0.0001057237317287569, |
|
"loss": 0.3211, |
|
"step": 3415 |
|
}, |
|
{ |
|
"epoch": 5.339578454332552, |
|
"grad_norm": 0.3414543867111206, |
|
"learning_rate": 0.00010545145016380065, |
|
"loss": 0.3171, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 5.347384855581577, |
|
"grad_norm": 0.37167057394981384, |
|
"learning_rate": 0.00010517912805671419, |
|
"loss": 0.3199, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 5.355191256830601, |
|
"grad_norm": 0.3445757031440735, |
|
"learning_rate": 0.00010490676743274181, |
|
"loss": 0.3176, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 5.3629976580796255, |
|
"grad_norm": 0.33130553364753723, |
|
"learning_rate": 0.00010463437031741423, |
|
"loss": 0.3171, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 5.370804059328649, |
|
"grad_norm": 0.34514832496643066, |
|
"learning_rate": 0.00010436193873653361, |
|
"loss": 0.3174, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 5.378610460577674, |
|
"grad_norm": 0.3419683575630188, |
|
"learning_rate": 0.00010408947471615835, |
|
"loss": 0.3229, |
|
"step": 3445 |
|
}, |
|
{ |
|
"epoch": 5.386416861826698, |
|
"grad_norm": 0.3344433903694153, |
|
"learning_rate": 0.00010381698028258817, |
|
"loss": 0.3238, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 5.394223263075722, |
|
"grad_norm": 0.3199335038661957, |
|
"learning_rate": 0.00010354445746234894, |
|
"loss": 0.3218, |
|
"step": 3455 |
|
}, |
|
{ |
|
"epoch": 5.402029664324746, |
|
"grad_norm": 0.3573337197303772, |
|
"learning_rate": 0.00010327190828217763, |
|
"loss": 0.3274, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 5.409836065573771, |
|
"grad_norm": 0.3186776638031006, |
|
"learning_rate": 0.00010299933476900727, |
|
"loss": 0.3278, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 5.4176424668227945, |
|
"grad_norm": 0.3291524648666382, |
|
"learning_rate": 0.00010272673894995187, |
|
"loss": 0.3207, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 5.425448868071819, |
|
"grad_norm": 0.3383654057979584, |
|
"learning_rate": 0.00010245412285229124, |
|
"loss": 0.309, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 5.433255269320843, |
|
"grad_norm": 0.33380231261253357, |
|
"learning_rate": 0.00010218148850345613, |
|
"loss": 0.3139, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 5.441061670569868, |
|
"grad_norm": 0.3124459385871887, |
|
"learning_rate": 0.0001019088379310129, |
|
"loss": 0.3167, |
|
"step": 3485 |
|
}, |
|
{ |
|
"epoch": 5.448868071818891, |
|
"grad_norm": 0.34139421582221985, |
|
"learning_rate": 0.00010163617316264869, |
|
"loss": 0.3177, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 5.456674473067916, |
|
"grad_norm": 0.33428099751472473, |
|
"learning_rate": 0.00010136349622615608, |
|
"loss": 0.3185, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 5.46448087431694, |
|
"grad_norm": 0.3460736870765686, |
|
"learning_rate": 0.00010109080914941824, |
|
"loss": 0.3187, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 5.472287275565964, |
|
"grad_norm": 0.35471439361572266, |
|
"learning_rate": 0.00010081811396039373, |
|
"loss": 0.3213, |
|
"step": 3505 |
|
}, |
|
{ |
|
"epoch": 5.480093676814988, |
|
"grad_norm": 0.3312179446220398, |
|
"learning_rate": 0.00010054541268710138, |
|
"loss": 0.3148, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 5.487900078064013, |
|
"grad_norm": 0.34200319647789, |
|
"learning_rate": 0.00010027270735760541, |
|
"loss": 0.3268, |
|
"step": 3515 |
|
}, |
|
{ |
|
"epoch": 5.495706479313037, |
|
"grad_norm": 0.339599072933197, |
|
"learning_rate": 0.0001, |
|
"loss": 0.3148, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 5.503512880562061, |
|
"grad_norm": 0.3320743143558502, |
|
"learning_rate": 9.972729264239461e-05, |
|
"loss": 0.3176, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 5.511319281811085, |
|
"grad_norm": 0.3287266790866852, |
|
"learning_rate": 9.945458731289862e-05, |
|
"loss": 0.3235, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 5.51912568306011, |
|
"grad_norm": 0.3429989814758301, |
|
"learning_rate": 9.918188603960632e-05, |
|
"loss": 0.3257, |
|
"step": 3535 |
|
}, |
|
{ |
|
"epoch": 5.526932084309133, |
|
"grad_norm": 0.3366888165473938, |
|
"learning_rate": 9.890919085058178e-05, |
|
"loss": 0.3194, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 5.534738485558158, |
|
"grad_norm": 0.3251045346260071, |
|
"learning_rate": 9.863650377384395e-05, |
|
"loss": 0.315, |
|
"step": 3545 |
|
}, |
|
{ |
|
"epoch": 5.542544886807182, |
|
"grad_norm": 0.3391668200492859, |
|
"learning_rate": 9.836382683735132e-05, |
|
"loss": 0.3225, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 5.550351288056206, |
|
"grad_norm": 0.3379986584186554, |
|
"learning_rate": 9.809116206898712e-05, |
|
"loss": 0.3251, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 5.55815768930523, |
|
"grad_norm": 0.336603581905365, |
|
"learning_rate": 9.78185114965439e-05, |
|
"loss": 0.3211, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 5.565964090554255, |
|
"grad_norm": 0.3426755368709564, |
|
"learning_rate": 9.754587714770878e-05, |
|
"loss": 0.323, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 5.573770491803279, |
|
"grad_norm": 0.3463711738586426, |
|
"learning_rate": 9.727326105004817e-05, |
|
"loss": 0.3202, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 5.581576893052302, |
|
"grad_norm": 0.36325374245643616, |
|
"learning_rate": 9.700066523099273e-05, |
|
"loss": 0.3143, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 5.589383294301327, |
|
"grad_norm": 0.32558673620224, |
|
"learning_rate": 9.67280917178224e-05, |
|
"loss": 0.3286, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 5.597189695550352, |
|
"grad_norm": 0.33436307311058044, |
|
"learning_rate": 9.64555425376511e-05, |
|
"loss": 0.3168, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 5.6049960967993755, |
|
"grad_norm": 0.33413755893707275, |
|
"learning_rate": 9.618301971741184e-05, |
|
"loss": 0.3135, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 5.612802498048399, |
|
"grad_norm": 0.3465372920036316, |
|
"learning_rate": 9.591052528384167e-05, |
|
"loss": 0.3175, |
|
"step": 3595 |
|
}, |
|
{ |
|
"epoch": 5.620608899297424, |
|
"grad_norm": 0.3320391774177551, |
|
"learning_rate": 9.563806126346642e-05, |
|
"loss": 0.3201, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 5.628415300546449, |
|
"grad_norm": 0.3287176191806793, |
|
"learning_rate": 9.536562968258578e-05, |
|
"loss": 0.3224, |
|
"step": 3605 |
|
}, |
|
{ |
|
"epoch": 5.636221701795472, |
|
"grad_norm": 0.3486092686653137, |
|
"learning_rate": 9.509323256725821e-05, |
|
"loss": 0.3254, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 5.644028103044496, |
|
"grad_norm": 0.3497357666492462, |
|
"learning_rate": 9.482087194328583e-05, |
|
"loss": 0.3178, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 5.651834504293521, |
|
"grad_norm": 0.338351309299469, |
|
"learning_rate": 9.454854983619936e-05, |
|
"loss": 0.3312, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 5.6596409055425445, |
|
"grad_norm": 0.3393017649650574, |
|
"learning_rate": 9.427626827124317e-05, |
|
"loss": 0.3171, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 5.667447306791569, |
|
"grad_norm": 0.34753820300102234, |
|
"learning_rate": 9.400402927335992e-05, |
|
"loss": 0.3231, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 5.675253708040593, |
|
"grad_norm": 0.3426099121570587, |
|
"learning_rate": 9.373183486717592e-05, |
|
"loss": 0.3195, |
|
"step": 3635 |
|
}, |
|
{ |
|
"epoch": 5.683060109289618, |
|
"grad_norm": 0.3154798150062561, |
|
"learning_rate": 9.345968707698569e-05, |
|
"loss": 0.3234, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 5.690866510538641, |
|
"grad_norm": 0.3466772735118866, |
|
"learning_rate": 9.318758792673722e-05, |
|
"loss": 0.3117, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 5.698672911787666, |
|
"grad_norm": 0.3402402997016907, |
|
"learning_rate": 9.29155394400166e-05, |
|
"loss": 0.3197, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 5.70647931303669, |
|
"grad_norm": 0.34711048007011414, |
|
"learning_rate": 9.264354364003327e-05, |
|
"loss": 0.3238, |
|
"step": 3655 |
|
}, |
|
{ |
|
"epoch": 5.714285714285714, |
|
"grad_norm": 0.33199217915534973, |
|
"learning_rate": 9.237160254960476e-05, |
|
"loss": 0.3136, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 5.722092115534738, |
|
"grad_norm": 0.3648219108581543, |
|
"learning_rate": 9.209971819114181e-05, |
|
"loss": 0.3257, |
|
"step": 3665 |
|
}, |
|
{ |
|
"epoch": 5.729898516783763, |
|
"grad_norm": 0.3323144316673279, |
|
"learning_rate": 9.182789258663321e-05, |
|
"loss": 0.3162, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 5.737704918032787, |
|
"grad_norm": 0.32556962966918945, |
|
"learning_rate": 9.155612775763069e-05, |
|
"loss": 0.3191, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 5.745511319281811, |
|
"grad_norm": 0.33691149950027466, |
|
"learning_rate": 9.128442572523417e-05, |
|
"loss": 0.3221, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 5.753317720530835, |
|
"grad_norm": 0.33508819341659546, |
|
"learning_rate": 9.10127885100765e-05, |
|
"loss": 0.3215, |
|
"step": 3685 |
|
}, |
|
{ |
|
"epoch": 5.76112412177986, |
|
"grad_norm": 0.31548982858657837, |
|
"learning_rate": 9.074121813230847e-05, |
|
"loss": 0.3134, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 5.768930523028883, |
|
"grad_norm": 0.33069273829460144, |
|
"learning_rate": 9.046971661158372e-05, |
|
"loss": 0.3138, |
|
"step": 3695 |
|
}, |
|
{ |
|
"epoch": 5.776736924277908, |
|
"grad_norm": 0.3511803150177002, |
|
"learning_rate": 9.019828596704394e-05, |
|
"loss": 0.322, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 5.784543325526932, |
|
"grad_norm": 0.34525299072265625, |
|
"learning_rate": 8.992692821730368e-05, |
|
"loss": 0.3268, |
|
"step": 3705 |
|
}, |
|
{ |
|
"epoch": 5.7923497267759565, |
|
"grad_norm": 0.32578587532043457, |
|
"learning_rate": 8.965564538043535e-05, |
|
"loss": 0.3254, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 5.80015612802498, |
|
"grad_norm": 0.33581212162971497, |
|
"learning_rate": 8.938443947395427e-05, |
|
"loss": 0.3249, |
|
"step": 3715 |
|
}, |
|
{ |
|
"epoch": 5.807962529274005, |
|
"grad_norm": 0.34412479400634766, |
|
"learning_rate": 8.911331251480357e-05, |
|
"loss": 0.3263, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 5.815768930523029, |
|
"grad_norm": 0.3513209819793701, |
|
"learning_rate": 8.884226651933927e-05, |
|
"loss": 0.3199, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 5.823575331772053, |
|
"grad_norm": 0.3510242700576782, |
|
"learning_rate": 8.857130350331535e-05, |
|
"loss": 0.3156, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 5.831381733021077, |
|
"grad_norm": 0.3359852135181427, |
|
"learning_rate": 8.830042548186863e-05, |
|
"loss": 0.3225, |
|
"step": 3735 |
|
}, |
|
{ |
|
"epoch": 5.839188134270102, |
|
"grad_norm": 0.3355961740016937, |
|
"learning_rate": 8.802963446950377e-05, |
|
"loss": 0.3229, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 5.8469945355191255, |
|
"grad_norm": 0.3400302827358246, |
|
"learning_rate": 8.775893248007839e-05, |
|
"loss": 0.331, |
|
"step": 3745 |
|
}, |
|
{ |
|
"epoch": 5.85480093676815, |
|
"grad_norm": 0.3256356120109558, |
|
"learning_rate": 8.74883215267881e-05, |
|
"loss": 0.327, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 5.862607338017174, |
|
"grad_norm": 0.35503053665161133, |
|
"learning_rate": 8.721780362215138e-05, |
|
"loss": 0.3273, |
|
"step": 3755 |
|
}, |
|
{ |
|
"epoch": 5.870413739266199, |
|
"grad_norm": 0.34520792961120605, |
|
"learning_rate": 8.694738077799488e-05, |
|
"loss": 0.3198, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 5.878220140515222, |
|
"grad_norm": 0.32543712854385376, |
|
"learning_rate": 8.667705500543803e-05, |
|
"loss": 0.3171, |
|
"step": 3765 |
|
}, |
|
{ |
|
"epoch": 5.886026541764247, |
|
"grad_norm": 0.33950576186180115, |
|
"learning_rate": 8.64068283148786e-05, |
|
"loss": 0.3207, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 5.893832943013271, |
|
"grad_norm": 0.3403869569301605, |
|
"learning_rate": 8.613670271597733e-05, |
|
"loss": 0.3243, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 5.901639344262295, |
|
"grad_norm": 0.34109991788864136, |
|
"learning_rate": 8.586668021764329e-05, |
|
"loss": 0.3221, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 5.909445745511319, |
|
"grad_norm": 0.3320345878601074, |
|
"learning_rate": 8.559676282801858e-05, |
|
"loss": 0.3171, |
|
"step": 3785 |
|
}, |
|
{ |
|
"epoch": 5.917252146760344, |
|
"grad_norm": 0.34376391768455505, |
|
"learning_rate": 8.532695255446383e-05, |
|
"loss": 0.3191, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 5.925058548009368, |
|
"grad_norm": 0.33876165747642517, |
|
"learning_rate": 8.505725140354294e-05, |
|
"loss": 0.3198, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 5.932864949258392, |
|
"grad_norm": 0.34862759709358215, |
|
"learning_rate": 8.478766138100834e-05, |
|
"loss": 0.3195, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 5.940671350507416, |
|
"grad_norm": 0.34864258766174316, |
|
"learning_rate": 8.451818449178591e-05, |
|
"loss": 0.3271, |
|
"step": 3805 |
|
}, |
|
{ |
|
"epoch": 5.948477751756441, |
|
"grad_norm": 0.35281243920326233, |
|
"learning_rate": 8.424882273996024e-05, |
|
"loss": 0.3246, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 5.956284153005464, |
|
"grad_norm": 0.33554619550704956, |
|
"learning_rate": 8.397957812875959e-05, |
|
"loss": 0.3239, |
|
"step": 3815 |
|
}, |
|
{ |
|
"epoch": 5.964090554254488, |
|
"grad_norm": 0.3569699227809906, |
|
"learning_rate": 8.371045266054114e-05, |
|
"loss": 0.3182, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 5.971896955503513, |
|
"grad_norm": 0.3249468505382538, |
|
"learning_rate": 8.344144833677594e-05, |
|
"loss": 0.3218, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 5.9797033567525375, |
|
"grad_norm": 0.3645569682121277, |
|
"learning_rate": 8.317256715803407e-05, |
|
"loss": 0.3158, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 5.987509758001561, |
|
"grad_norm": 0.32271501421928406, |
|
"learning_rate": 8.290381112396987e-05, |
|
"loss": 0.3194, |
|
"step": 3835 |
|
}, |
|
{ |
|
"epoch": 5.995316159250585, |
|
"grad_norm": 0.3314792215824127, |
|
"learning_rate": 8.263518223330697e-05, |
|
"loss": 0.318, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 1.0760587453842163, |
|
"eval_runtime": 0.3225, |
|
"eval_samples_per_second": 15.506, |
|
"eval_steps_per_second": 3.101, |
|
"step": 3843 |
|
}, |
|
{ |
|
"epoch": 6.00312256049961, |
|
"grad_norm": 0.31583714485168457, |
|
"learning_rate": 8.236668248382337e-05, |
|
"loss": 0.3, |
|
"step": 3845 |
|
}, |
|
{ |
|
"epoch": 6.0109289617486334, |
|
"grad_norm": 0.36563432216644287, |
|
"learning_rate": 8.209831387233676e-05, |
|
"loss": 0.2928, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 6.018735362997658, |
|
"grad_norm": 0.3586893081665039, |
|
"learning_rate": 8.183007839468942e-05, |
|
"loss": 0.2947, |
|
"step": 3855 |
|
}, |
|
{ |
|
"epoch": 6.026541764246682, |
|
"grad_norm": 0.351180762052536, |
|
"learning_rate": 8.156197804573366e-05, |
|
"loss": 0.2974, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 6.0343481654957065, |
|
"grad_norm": 0.3694261610507965, |
|
"learning_rate": 8.12940148193168e-05, |
|
"loss": 0.2943, |
|
"step": 3865 |
|
}, |
|
{ |
|
"epoch": 6.04215456674473, |
|
"grad_norm": 0.3579952120780945, |
|
"learning_rate": 8.102619070826639e-05, |
|
"loss": 0.2993, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 6.049960967993755, |
|
"grad_norm": 0.3623141348361969, |
|
"learning_rate": 8.075850770437534e-05, |
|
"loss": 0.3012, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 6.057767369242779, |
|
"grad_norm": 0.35049453377723694, |
|
"learning_rate": 8.049096779838719e-05, |
|
"loss": 0.2954, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 6.065573770491803, |
|
"grad_norm": 0.36742958426475525, |
|
"learning_rate": 8.022357297998127e-05, |
|
"loss": 0.292, |
|
"step": 3885 |
|
}, |
|
{ |
|
"epoch": 6.073380171740827, |
|
"grad_norm": 0.36501386761665344, |
|
"learning_rate": 7.995632523775795e-05, |
|
"loss": 0.2938, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 6.081186572989852, |
|
"grad_norm": 0.35817086696624756, |
|
"learning_rate": 7.968922655922374e-05, |
|
"loss": 0.302, |
|
"step": 3895 |
|
}, |
|
{ |
|
"epoch": 6.0889929742388755, |
|
"grad_norm": 0.3561493158340454, |
|
"learning_rate": 7.942227893077652e-05, |
|
"loss": 0.3042, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 6.0967993754879, |
|
"grad_norm": 0.3571511507034302, |
|
"learning_rate": 7.915548433769095e-05, |
|
"loss": 0.2971, |
|
"step": 3905 |
|
}, |
|
{ |
|
"epoch": 6.104605776736924, |
|
"grad_norm": 0.34343209862709045, |
|
"learning_rate": 7.888884476410348e-05, |
|
"loss": 0.2956, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 6.112412177985949, |
|
"grad_norm": 0.3509383201599121, |
|
"learning_rate": 7.862236219299777e-05, |
|
"loss": 0.2991, |
|
"step": 3915 |
|
}, |
|
{ |
|
"epoch": 6.120218579234972, |
|
"grad_norm": 0.35262179374694824, |
|
"learning_rate": 7.835603860618972e-05, |
|
"loss": 0.2984, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 6.128024980483997, |
|
"grad_norm": 0.35837092995643616, |
|
"learning_rate": 7.808987598431303e-05, |
|
"loss": 0.2949, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 6.135831381733021, |
|
"grad_norm": 0.3698333501815796, |
|
"learning_rate": 7.782387630680421e-05, |
|
"loss": 0.3056, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 6.143637782982045, |
|
"grad_norm": 0.3528449535369873, |
|
"learning_rate": 7.755804155188802e-05, |
|
"loss": 0.2861, |
|
"step": 3935 |
|
}, |
|
{ |
|
"epoch": 6.151444184231069, |
|
"grad_norm": 0.3811301290988922, |
|
"learning_rate": 7.729237369656269e-05, |
|
"loss": 0.3052, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 6.159250585480094, |
|
"grad_norm": 0.37010228633880615, |
|
"learning_rate": 7.702687471658518e-05, |
|
"loss": 0.2911, |
|
"step": 3945 |
|
}, |
|
{ |
|
"epoch": 6.167056986729118, |
|
"grad_norm": 0.37986573576927185, |
|
"learning_rate": 7.676154658645656e-05, |
|
"loss": 0.2997, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 6.174863387978142, |
|
"grad_norm": 0.34996771812438965, |
|
"learning_rate": 7.649639127940735e-05, |
|
"loss": 0.3029, |
|
"step": 3955 |
|
}, |
|
{ |
|
"epoch": 6.182669789227166, |
|
"grad_norm": 0.3746223449707031, |
|
"learning_rate": 7.623141076738271e-05, |
|
"loss": 0.3064, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 6.190476190476191, |
|
"grad_norm": 0.3640713393688202, |
|
"learning_rate": 7.596660702102791e-05, |
|
"loss": 0.3013, |
|
"step": 3965 |
|
}, |
|
{ |
|
"epoch": 6.1982825917252145, |
|
"grad_norm": 0.3776625096797943, |
|
"learning_rate": 7.570198200967362e-05, |
|
"loss": 0.3063, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 6.206088992974239, |
|
"grad_norm": 0.3669188916683197, |
|
"learning_rate": 7.543753770132127e-05, |
|
"loss": 0.2982, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 6.213895394223263, |
|
"grad_norm": 0.37241482734680176, |
|
"learning_rate": 7.517327606262836e-05, |
|
"loss": 0.3, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 6.2217017954722875, |
|
"grad_norm": 0.36023199558258057, |
|
"learning_rate": 7.490919905889403e-05, |
|
"loss": 0.2997, |
|
"step": 3985 |
|
}, |
|
{ |
|
"epoch": 6.229508196721311, |
|
"grad_norm": 0.36012396216392517, |
|
"learning_rate": 7.464530865404407e-05, |
|
"loss": 0.2926, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 6.237314597970336, |
|
"grad_norm": 0.3803520202636719, |
|
"learning_rate": 7.438160681061676e-05, |
|
"loss": 0.3045, |
|
"step": 3995 |
|
}, |
|
{ |
|
"epoch": 6.24512099921936, |
|
"grad_norm": 0.34980347752571106, |
|
"learning_rate": 7.411809548974792e-05, |
|
"loss": 0.2991, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 6.252927400468384, |
|
"grad_norm": 0.3561069369316101, |
|
"learning_rate": 7.385477665115658e-05, |
|
"loss": 0.3094, |
|
"step": 4005 |
|
}, |
|
{ |
|
"epoch": 6.260733801717408, |
|
"grad_norm": 0.36508041620254517, |
|
"learning_rate": 7.359165225313019e-05, |
|
"loss": 0.2969, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 6.268540202966433, |
|
"grad_norm": 0.36842647194862366, |
|
"learning_rate": 7.332872425251018e-05, |
|
"loss": 0.3009, |
|
"step": 4015 |
|
}, |
|
{ |
|
"epoch": 6.2763466042154565, |
|
"grad_norm": 0.3771759569644928, |
|
"learning_rate": 7.30659946046774e-05, |
|
"loss": 0.3027, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 6.284153005464481, |
|
"grad_norm": 0.36950060725212097, |
|
"learning_rate": 7.280346526353759e-05, |
|
"loss": 0.3052, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 6.291959406713505, |
|
"grad_norm": 0.36776700615882874, |
|
"learning_rate": 7.25411381815068e-05, |
|
"loss": 0.298, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 6.29976580796253, |
|
"grad_norm": 0.3699568510055542, |
|
"learning_rate": 7.22790153094968e-05, |
|
"loss": 0.2961, |
|
"step": 4035 |
|
}, |
|
{ |
|
"epoch": 6.307572209211553, |
|
"grad_norm": 0.36055341362953186, |
|
"learning_rate": 7.20170985969008e-05, |
|
"loss": 0.297, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 6.315378610460578, |
|
"grad_norm": 0.37468183040618896, |
|
"learning_rate": 7.175538999157876e-05, |
|
"loss": 0.2989, |
|
"step": 4045 |
|
}, |
|
{ |
|
"epoch": 6.323185011709602, |
|
"grad_norm": 0.3778867721557617, |
|
"learning_rate": 7.149389143984295e-05, |
|
"loss": 0.2994, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 6.330991412958626, |
|
"grad_norm": 0.37846189737319946, |
|
"learning_rate": 7.123260488644345e-05, |
|
"loss": 0.2949, |
|
"step": 4055 |
|
}, |
|
{ |
|
"epoch": 6.33879781420765, |
|
"grad_norm": 0.3595486581325531, |
|
"learning_rate": 7.097153227455379e-05, |
|
"loss": 0.3063, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 6.346604215456675, |
|
"grad_norm": 0.38495534658432007, |
|
"learning_rate": 7.071067554575637e-05, |
|
"loss": 0.3037, |
|
"step": 4065 |
|
}, |
|
{ |
|
"epoch": 6.354410616705699, |
|
"grad_norm": 0.37678053975105286, |
|
"learning_rate": 7.045003664002809e-05, |
|
"loss": 0.3111, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 6.362217017954723, |
|
"grad_norm": 0.353000283241272, |
|
"learning_rate": 7.018961749572604e-05, |
|
"loss": 0.3003, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 6.370023419203747, |
|
"grad_norm": 0.381798654794693, |
|
"learning_rate": 6.992942004957271e-05, |
|
"loss": 0.3037, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 6.377829820452771, |
|
"grad_norm": 0.34018296003341675, |
|
"learning_rate": 6.9669446236642e-05, |
|
"loss": 0.2956, |
|
"step": 4085 |
|
}, |
|
{ |
|
"epoch": 6.3856362217017955, |
|
"grad_norm": 0.3533703684806824, |
|
"learning_rate": 6.940969799034465e-05, |
|
"loss": 0.298, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 6.39344262295082, |
|
"grad_norm": 0.36671626567840576, |
|
"learning_rate": 6.915017724241389e-05, |
|
"loss": 0.3004, |
|
"step": 4095 |
|
}, |
|
{ |
|
"epoch": 6.401249024199844, |
|
"grad_norm": 0.3814579248428345, |
|
"learning_rate": 6.889088592289093e-05, |
|
"loss": 0.2992, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 6.409055425448868, |
|
"grad_norm": 0.38459497690200806, |
|
"learning_rate": 6.863182596011087e-05, |
|
"loss": 0.3036, |
|
"step": 4105 |
|
}, |
|
{ |
|
"epoch": 6.416861826697892, |
|
"grad_norm": 0.3617273271083832, |
|
"learning_rate": 6.837299928068817e-05, |
|
"loss": 0.3037, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 6.424668227946917, |
|
"grad_norm": 0.3558304011821747, |
|
"learning_rate": 6.811440780950237e-05, |
|
"loss": 0.2934, |
|
"step": 4115 |
|
}, |
|
{ |
|
"epoch": 6.432474629195941, |
|
"grad_norm": 0.36529627442359924, |
|
"learning_rate": 6.785605346968386e-05, |
|
"loss": 0.3034, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 6.4402810304449645, |
|
"grad_norm": 0.3953593373298645, |
|
"learning_rate": 6.759793818259933e-05, |
|
"loss": 0.3005, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 6.448087431693989, |
|
"grad_norm": 0.38395586609840393, |
|
"learning_rate": 6.73400638678378e-05, |
|
"loss": 0.3009, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 6.455893832943013, |
|
"grad_norm": 0.35242578387260437, |
|
"learning_rate": 6.708243244319611e-05, |
|
"loss": 0.3053, |
|
"step": 4135 |
|
}, |
|
{ |
|
"epoch": 6.4637002341920375, |
|
"grad_norm": 0.3590448200702667, |
|
"learning_rate": 6.682504582466482e-05, |
|
"loss": 0.2994, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 6.471506635441061, |
|
"grad_norm": 0.37192437052726746, |
|
"learning_rate": 6.656790592641375e-05, |
|
"loss": 0.3056, |
|
"step": 4145 |
|
}, |
|
{ |
|
"epoch": 6.479313036690086, |
|
"grad_norm": 0.35877934098243713, |
|
"learning_rate": 6.6311014660778e-05, |
|
"loss": 0.2958, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 6.48711943793911, |
|
"grad_norm": 0.3730364143848419, |
|
"learning_rate": 6.605437393824356e-05, |
|
"loss": 0.3047, |
|
"step": 4155 |
|
}, |
|
{ |
|
"epoch": 6.494925839188134, |
|
"grad_norm": 0.3790695071220398, |
|
"learning_rate": 6.579798566743314e-05, |
|
"loss": 0.3027, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 6.502732240437158, |
|
"grad_norm": 0.36754992604255676, |
|
"learning_rate": 6.554185175509197e-05, |
|
"loss": 0.2916, |
|
"step": 4165 |
|
}, |
|
{ |
|
"epoch": 6.510538641686183, |
|
"grad_norm": 0.37978771328926086, |
|
"learning_rate": 6.528597410607364e-05, |
|
"loss": 0.3023, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 6.518345042935207, |
|
"grad_norm": 0.35783836245536804, |
|
"learning_rate": 6.503035462332592e-05, |
|
"loss": 0.3031, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 6.526151444184231, |
|
"grad_norm": 0.3850097060203552, |
|
"learning_rate": 6.477499520787665e-05, |
|
"loss": 0.3025, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 6.533957845433255, |
|
"grad_norm": 0.38699930906295776, |
|
"learning_rate": 6.451989775881955e-05, |
|
"loss": 0.3032, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 6.54176424668228, |
|
"grad_norm": 0.3620811998844147, |
|
"learning_rate": 6.42650641733e-05, |
|
"loss": 0.3029, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 6.549570647931303, |
|
"grad_norm": 0.3833983838558197, |
|
"learning_rate": 6.401049634650118e-05, |
|
"loss": 0.3018, |
|
"step": 4195 |
|
}, |
|
{ |
|
"epoch": 6.557377049180328, |
|
"grad_norm": 0.3642504811286926, |
|
"learning_rate": 6.375619617162985e-05, |
|
"loss": 0.3052, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 6.565183450429352, |
|
"grad_norm": 0.34987977147102356, |
|
"learning_rate": 6.350216553990212e-05, |
|
"loss": 0.3027, |
|
"step": 4205 |
|
}, |
|
{ |
|
"epoch": 6.5729898516783765, |
|
"grad_norm": 0.3789624571800232, |
|
"learning_rate": 6.324840634052967e-05, |
|
"loss": 0.2983, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 6.5807962529274, |
|
"grad_norm": 0.39051589369773865, |
|
"learning_rate": 6.29949204607054e-05, |
|
"loss": 0.3099, |
|
"step": 4215 |
|
}, |
|
{ |
|
"epoch": 6.588602654176425, |
|
"grad_norm": 0.36654597520828247, |
|
"learning_rate": 6.27417097855897e-05, |
|
"loss": 0.3067, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 6.596409055425449, |
|
"grad_norm": 0.3750831186771393, |
|
"learning_rate": 6.248877619829619e-05, |
|
"loss": 0.304, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 6.604215456674473, |
|
"grad_norm": 0.3867753744125366, |
|
"learning_rate": 6.223612157987786e-05, |
|
"loss": 0.3082, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 6.612021857923497, |
|
"grad_norm": 0.3690248429775238, |
|
"learning_rate": 6.198374780931293e-05, |
|
"loss": 0.3044, |
|
"step": 4235 |
|
}, |
|
{ |
|
"epoch": 6.619828259172522, |
|
"grad_norm": 0.37476563453674316, |
|
"learning_rate": 6.173165676349103e-05, |
|
"loss": 0.3027, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 6.6276346604215455, |
|
"grad_norm": 0.36592650413513184, |
|
"learning_rate": 6.14798503171992e-05, |
|
"loss": 0.3015, |
|
"step": 4245 |
|
}, |
|
{ |
|
"epoch": 6.63544106167057, |
|
"grad_norm": 0.3869721293449402, |
|
"learning_rate": 6.122833034310793e-05, |
|
"loss": 0.2986, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 6.643247462919594, |
|
"grad_norm": 0.3800086975097656, |
|
"learning_rate": 6.097709871175723e-05, |
|
"loss": 0.3026, |
|
"step": 4255 |
|
}, |
|
{ |
|
"epoch": 6.6510538641686185, |
|
"grad_norm": 0.3717212677001953, |
|
"learning_rate": 6.0726157291542605e-05, |
|
"loss": 0.303, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 6.658860265417642, |
|
"grad_norm": 0.35861897468566895, |
|
"learning_rate": 6.047550794870145e-05, |
|
"loss": 0.3008, |
|
"step": 4265 |
|
}, |
|
{ |
|
"epoch": 6.666666666666667, |
|
"grad_norm": 0.3755732476711273, |
|
"learning_rate": 6.02251525472989e-05, |
|
"loss": 0.3006, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 6.674473067915691, |
|
"grad_norm": 0.3563028872013092, |
|
"learning_rate": 5.9975092949214116e-05, |
|
"loss": 0.3041, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 6.682279469164715, |
|
"grad_norm": 0.3711301386356354, |
|
"learning_rate": 5.9725331014126294e-05, |
|
"loss": 0.3047, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 6.690085870413739, |
|
"grad_norm": 0.3619491457939148, |
|
"learning_rate": 5.947586859950103e-05, |
|
"loss": 0.3034, |
|
"step": 4285 |
|
}, |
|
{ |
|
"epoch": 6.697892271662764, |
|
"grad_norm": 0.3621855080127716, |
|
"learning_rate": 5.922670756057633e-05, |
|
"loss": 0.2993, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 6.705698672911788, |
|
"grad_norm": 0.3833586573600769, |
|
"learning_rate": 5.8977849750348944e-05, |
|
"loss": 0.2969, |
|
"step": 4295 |
|
}, |
|
{ |
|
"epoch": 6.713505074160812, |
|
"grad_norm": 0.36535149812698364, |
|
"learning_rate": 5.872929701956054e-05, |
|
"loss": 0.3109, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 6.721311475409836, |
|
"grad_norm": 0.34852883219718933, |
|
"learning_rate": 5.848105121668381e-05, |
|
"loss": 0.2996, |
|
"step": 4305 |
|
}, |
|
{ |
|
"epoch": 6.729117876658861, |
|
"grad_norm": 0.35054901242256165, |
|
"learning_rate": 5.8233114187908935e-05, |
|
"loss": 0.3038, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 6.736924277907884, |
|
"grad_norm": 0.38017308712005615, |
|
"learning_rate": 5.7985487777129765e-05, |
|
"loss": 0.3023, |
|
"step": 4315 |
|
}, |
|
{ |
|
"epoch": 6.744730679156909, |
|
"grad_norm": 0.3646107017993927, |
|
"learning_rate": 5.773817382593008e-05, |
|
"loss": 0.3097, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 6.752537080405933, |
|
"grad_norm": 0.3594595491886139, |
|
"learning_rate": 5.749117417356988e-05, |
|
"loss": 0.3088, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 6.760343481654957, |
|
"grad_norm": 0.39560645818710327, |
|
"learning_rate": 5.7244490656971815e-05, |
|
"loss": 0.3058, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 6.768149882903981, |
|
"grad_norm": 0.3735295236110687, |
|
"learning_rate": 5.699812511070734e-05, |
|
"loss": 0.3109, |
|
"step": 4335 |
|
}, |
|
{ |
|
"epoch": 6.775956284153006, |
|
"grad_norm": 0.3793189823627472, |
|
"learning_rate": 5.675207936698337e-05, |
|
"loss": 0.2962, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 6.78376268540203, |
|
"grad_norm": 0.36545222997665405, |
|
"learning_rate": 5.6506355255628284e-05, |
|
"loss": 0.3034, |
|
"step": 4345 |
|
}, |
|
{ |
|
"epoch": 6.791569086651053, |
|
"grad_norm": 0.3819272816181183, |
|
"learning_rate": 5.6260954604078585e-05, |
|
"loss": 0.3055, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 6.799375487900078, |
|
"grad_norm": 0.40963688492774963, |
|
"learning_rate": 5.601587923736518e-05, |
|
"loss": 0.297, |
|
"step": 4355 |
|
}, |
|
{ |
|
"epoch": 6.807181889149103, |
|
"grad_norm": 0.3582998812198639, |
|
"learning_rate": 5.577113097809989e-05, |
|
"loss": 0.3024, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 6.8149882903981265, |
|
"grad_norm": 0.3594261407852173, |
|
"learning_rate": 5.5526711646461835e-05, |
|
"loss": 0.2976, |
|
"step": 4365 |
|
}, |
|
{ |
|
"epoch": 6.82279469164715, |
|
"grad_norm": 0.36158230900764465, |
|
"learning_rate": 5.528262306018395e-05, |
|
"loss": 0.3015, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 6.830601092896175, |
|
"grad_norm": 0.37753212451934814, |
|
"learning_rate": 5.503886703453933e-05, |
|
"loss": 0.2942, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 6.8384074941451995, |
|
"grad_norm": 0.37955158948898315, |
|
"learning_rate": 5.4795445382328037e-05, |
|
"loss": 0.3017, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 6.846213895394223, |
|
"grad_norm": 0.3600054979324341, |
|
"learning_rate": 5.4552359913863214e-05, |
|
"loss": 0.3061, |
|
"step": 4385 |
|
}, |
|
{ |
|
"epoch": 6.854020296643247, |
|
"grad_norm": 0.3673628270626068, |
|
"learning_rate": 5.4309612436957937e-05, |
|
"loss": 0.3049, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 6.861826697892272, |
|
"grad_norm": 0.36428987979888916, |
|
"learning_rate": 5.4067204756911624e-05, |
|
"loss": 0.3096, |
|
"step": 4395 |
|
}, |
|
{ |
|
"epoch": 6.8696330991412955, |
|
"grad_norm": 0.3699190616607666, |
|
"learning_rate": 5.382513867649663e-05, |
|
"loss": 0.3008, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 6.87743950039032, |
|
"grad_norm": 0.36936306953430176, |
|
"learning_rate": 5.358341599594483e-05, |
|
"loss": 0.2979, |
|
"step": 4405 |
|
}, |
|
{ |
|
"epoch": 6.885245901639344, |
|
"grad_norm": 0.3822793960571289, |
|
"learning_rate": 5.3342038512934424e-05, |
|
"loss": 0.3009, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 6.893052302888369, |
|
"grad_norm": 0.37353184819221497, |
|
"learning_rate": 5.3101008022576126e-05, |
|
"loss": 0.3061, |
|
"step": 4415 |
|
}, |
|
{ |
|
"epoch": 6.900858704137392, |
|
"grad_norm": 0.37411707639694214, |
|
"learning_rate": 5.286032631740023e-05, |
|
"loss": 0.3022, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 6.908665105386417, |
|
"grad_norm": 0.3575087785720825, |
|
"learning_rate": 5.261999518734322e-05, |
|
"loss": 0.302, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 6.916471506635441, |
|
"grad_norm": 0.3561764359474182, |
|
"learning_rate": 5.238001641973422e-05, |
|
"loss": 0.2998, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 6.924277907884465, |
|
"grad_norm": 0.3756863474845886, |
|
"learning_rate": 5.214039179928194e-05, |
|
"loss": 0.3009, |
|
"step": 4435 |
|
}, |
|
{ |
|
"epoch": 6.932084309133489, |
|
"grad_norm": 0.36556264758110046, |
|
"learning_rate": 5.190112310806126e-05, |
|
"loss": 0.296, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 6.939890710382514, |
|
"grad_norm": 0.37201789021492004, |
|
"learning_rate": 5.1662212125500075e-05, |
|
"loss": 0.2993, |
|
"step": 4445 |
|
}, |
|
{ |
|
"epoch": 6.947697111631538, |
|
"grad_norm": 0.34447523951530457, |
|
"learning_rate": 5.142366062836599e-05, |
|
"loss": 0.3002, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 6.955503512880562, |
|
"grad_norm": 0.3652222156524658, |
|
"learning_rate": 5.118547039075326e-05, |
|
"loss": 0.3015, |
|
"step": 4455 |
|
}, |
|
{ |
|
"epoch": 6.963309914129586, |
|
"grad_norm": 0.3540942072868347, |
|
"learning_rate": 5.09476431840692e-05, |
|
"loss": 0.3031, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 6.971116315378611, |
|
"grad_norm": 0.36961525678634644, |
|
"learning_rate": 5.071018077702161e-05, |
|
"loss": 0.3134, |
|
"step": 4465 |
|
}, |
|
{ |
|
"epoch": 6.978922716627634, |
|
"grad_norm": 0.3560226559638977, |
|
"learning_rate": 5.047308493560506e-05, |
|
"loss": 0.2953, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 6.986729117876659, |
|
"grad_norm": 0.3567667007446289, |
|
"learning_rate": 5.023635742308807e-05, |
|
"loss": 0.3081, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 6.994535519125683, |
|
"grad_norm": 0.3771991431713104, |
|
"learning_rate": 5.000000000000002e-05, |
|
"loss": 0.3056, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 6.999219359875098, |
|
"eval_loss": 1.1023039817810059, |
|
"eval_runtime": 0.3388, |
|
"eval_samples_per_second": 14.759, |
|
"eval_steps_per_second": 2.952, |
|
"step": 4483 |
|
}, |
|
{ |
|
"epoch": 7.0023419203747075, |
|
"grad_norm": 0.34990745782852173, |
|
"learning_rate": 4.976401442411774e-05, |
|
"loss": 0.2964, |
|
"step": 4485 |
|
}, |
|
{ |
|
"epoch": 7.010148321623731, |
|
"grad_norm": 0.3766123354434967, |
|
"learning_rate": 4.952840245045278e-05, |
|
"loss": 0.2868, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 7.017954722872756, |
|
"grad_norm": 0.3829561769962311, |
|
"learning_rate": 4.92931658312383e-05, |
|
"loss": 0.2865, |
|
"step": 4495 |
|
}, |
|
{ |
|
"epoch": 7.02576112412178, |
|
"grad_norm": 0.3801458179950714, |
|
"learning_rate": 4.9058306315915826e-05, |
|
"loss": 0.277, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 7.033567525370804, |
|
"grad_norm": 0.3700219690799713, |
|
"learning_rate": 4.882382565112248e-05, |
|
"loss": 0.2803, |
|
"step": 4505 |
|
}, |
|
{ |
|
"epoch": 7.041373926619828, |
|
"grad_norm": 0.3695278763771057, |
|
"learning_rate": 4.8589725580677835e-05, |
|
"loss": 0.2817, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 7.049180327868853, |
|
"grad_norm": 0.3825577199459076, |
|
"learning_rate": 4.835600784557106e-05, |
|
"loss": 0.2791, |
|
"step": 4515 |
|
}, |
|
{ |
|
"epoch": 7.0569867291178765, |
|
"grad_norm": 0.36848512291908264, |
|
"learning_rate": 4.8122674183947836e-05, |
|
"loss": 0.2808, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 7.064793130366901, |
|
"grad_norm": 0.3669551610946655, |
|
"learning_rate": 4.7889726331097686e-05, |
|
"loss": 0.2813, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 7.072599531615925, |
|
"grad_norm": 0.3677949905395508, |
|
"learning_rate": 4.7657166019440614e-05, |
|
"loss": 0.2761, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 7.08040593286495, |
|
"grad_norm": 0.39766794443130493, |
|
"learning_rate": 4.742499497851478e-05, |
|
"loss": 0.2888, |
|
"step": 4535 |
|
}, |
|
{ |
|
"epoch": 7.088212334113973, |
|
"grad_norm": 0.39841169118881226, |
|
"learning_rate": 4.7193214934963206e-05, |
|
"loss": 0.2874, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 7.096018735362998, |
|
"grad_norm": 0.37418073415756226, |
|
"learning_rate": 4.696182761252112e-05, |
|
"loss": 0.2794, |
|
"step": 4545 |
|
}, |
|
{ |
|
"epoch": 7.103825136612022, |
|
"grad_norm": 0.3763929307460785, |
|
"learning_rate": 4.6730834732003104e-05, |
|
"loss": 0.2851, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 7.111631537861046, |
|
"grad_norm": 0.39079606533050537, |
|
"learning_rate": 4.6500238011290295e-05, |
|
"loss": 0.2804, |
|
"step": 4555 |
|
}, |
|
{ |
|
"epoch": 7.11943793911007, |
|
"grad_norm": 0.3997804522514343, |
|
"learning_rate": 4.6270039165317605e-05, |
|
"loss": 0.2827, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 7.127244340359095, |
|
"grad_norm": 0.3801449239253998, |
|
"learning_rate": 4.604023990606105e-05, |
|
"loss": 0.2845, |
|
"step": 4565 |
|
}, |
|
{ |
|
"epoch": 7.135050741608119, |
|
"grad_norm": 0.3795543611049652, |
|
"learning_rate": 4.5810841942524864e-05, |
|
"loss": 0.2793, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 7.142857142857143, |
|
"grad_norm": 0.4058801829814911, |
|
"learning_rate": 4.5581846980728794e-05, |
|
"loss": 0.2844, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 7.150663544106167, |
|
"grad_norm": 0.38217005133628845, |
|
"learning_rate": 4.535325672369567e-05, |
|
"loss": 0.2879, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 7.158469945355192, |
|
"grad_norm": 0.4147985279560089, |
|
"learning_rate": 4.512507287143842e-05, |
|
"loss": 0.2829, |
|
"step": 4585 |
|
}, |
|
{ |
|
"epoch": 7.166276346604215, |
|
"grad_norm": 0.3910820484161377, |
|
"learning_rate": 4.4897297120947624e-05, |
|
"loss": 0.2898, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 7.17408274785324, |
|
"grad_norm": 0.3985479176044464, |
|
"learning_rate": 4.466993116617878e-05, |
|
"loss": 0.2851, |
|
"step": 4595 |
|
}, |
|
{ |
|
"epoch": 7.181889149102264, |
|
"grad_norm": 0.3941868245601654, |
|
"learning_rate": 4.444297669803981e-05, |
|
"loss": 0.2828, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 7.1896955503512885, |
|
"grad_norm": 0.3968215882778168, |
|
"learning_rate": 4.4216435404378355e-05, |
|
"loss": 0.2917, |
|
"step": 4605 |
|
}, |
|
{ |
|
"epoch": 7.197501951600312, |
|
"grad_norm": 0.3764592707157135, |
|
"learning_rate": 4.399030896996945e-05, |
|
"loss": 0.2811, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 7.205308352849336, |
|
"grad_norm": 0.3845047056674957, |
|
"learning_rate": 4.3764599076502696e-05, |
|
"loss": 0.2765, |
|
"step": 4615 |
|
}, |
|
{ |
|
"epoch": 7.213114754098361, |
|
"grad_norm": 0.39302965998649597, |
|
"learning_rate": 4.353930740256996e-05, |
|
"loss": 0.2839, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 7.220921155347384, |
|
"grad_norm": 0.38965940475463867, |
|
"learning_rate": 4.331443562365285e-05, |
|
"loss": 0.2834, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 7.228727556596409, |
|
"grad_norm": 0.36734485626220703, |
|
"learning_rate": 4.308998541211015e-05, |
|
"loss": 0.2846, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 7.236533957845433, |
|
"grad_norm": 0.41275784373283386, |
|
"learning_rate": 4.286595843716569e-05, |
|
"loss": 0.2903, |
|
"step": 4635 |
|
}, |
|
{ |
|
"epoch": 7.2443403590944575, |
|
"grad_norm": 0.38159552216529846, |
|
"learning_rate": 4.264235636489542e-05, |
|
"loss": 0.288, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 7.252146760343481, |
|
"grad_norm": 0.38839638233184814, |
|
"learning_rate": 4.241918085821547e-05, |
|
"loss": 0.2831, |
|
"step": 4645 |
|
}, |
|
{ |
|
"epoch": 7.259953161592506, |
|
"grad_norm": 0.42755383253097534, |
|
"learning_rate": 4.219643357686967e-05, |
|
"loss": 0.2908, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 7.26775956284153, |
|
"grad_norm": 0.43094560503959656, |
|
"learning_rate": 4.19741161774171e-05, |
|
"loss": 0.2841, |
|
"step": 4655 |
|
}, |
|
{ |
|
"epoch": 7.275565964090554, |
|
"grad_norm": 0.3952085077762604, |
|
"learning_rate": 4.17522303132198e-05, |
|
"loss": 0.2858, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 7.283372365339578, |
|
"grad_norm": 0.401659220457077, |
|
"learning_rate": 4.153077763443055e-05, |
|
"loss": 0.2801, |
|
"step": 4665 |
|
}, |
|
{ |
|
"epoch": 7.291178766588603, |
|
"grad_norm": 0.3862559199333191, |
|
"learning_rate": 4.1309759787980565e-05, |
|
"loss": 0.2818, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 7.2989851678376265, |
|
"grad_norm": 0.39351990818977356, |
|
"learning_rate": 4.1089178417567164e-05, |
|
"loss": 0.2916, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 7.306791569086651, |
|
"grad_norm": 0.39333629608154297, |
|
"learning_rate": 4.086903516364179e-05, |
|
"loss": 0.2885, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 7.314597970335675, |
|
"grad_norm": 0.40127068758010864, |
|
"learning_rate": 4.064933166339737e-05, |
|
"loss": 0.2892, |
|
"step": 4685 |
|
}, |
|
{ |
|
"epoch": 7.3224043715847, |
|
"grad_norm": 0.4141682982444763, |
|
"learning_rate": 4.0430069550756665e-05, |
|
"loss": 0.2823, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 7.330210772833723, |
|
"grad_norm": 0.39413705468177795, |
|
"learning_rate": 4.021125045635973e-05, |
|
"loss": 0.2775, |
|
"step": 4695 |
|
}, |
|
{ |
|
"epoch": 7.338017174082748, |
|
"grad_norm": 0.3831061124801636, |
|
"learning_rate": 3.999287600755192e-05, |
|
"loss": 0.2867, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 7.345823575331772, |
|
"grad_norm": 0.3949141204357147, |
|
"learning_rate": 3.977494782837182e-05, |
|
"loss": 0.2895, |
|
"step": 4705 |
|
}, |
|
{ |
|
"epoch": 7.353629976580796, |
|
"grad_norm": 0.41716331243515015, |
|
"learning_rate": 3.9557467539539115e-05, |
|
"loss": 0.2857, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 7.36143637782982, |
|
"grad_norm": 0.385775625705719, |
|
"learning_rate": 3.9340436758442536e-05, |
|
"loss": 0.2819, |
|
"step": 4715 |
|
}, |
|
{ |
|
"epoch": 7.369242779078845, |
|
"grad_norm": 0.40228816866874695, |
|
"learning_rate": 3.9123857099127936e-05, |
|
"loss": 0.2891, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 7.377049180327869, |
|
"grad_norm": 0.38272202014923096, |
|
"learning_rate": 3.8907730172286124e-05, |
|
"loss": 0.284, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 7.384855581576893, |
|
"grad_norm": 0.39481300115585327, |
|
"learning_rate": 3.8692057585240905e-05, |
|
"loss": 0.2822, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 7.392661982825917, |
|
"grad_norm": 0.3933485150337219, |
|
"learning_rate": 3.847684094193733e-05, |
|
"loss": 0.294, |
|
"step": 4735 |
|
}, |
|
{ |
|
"epoch": 7.400468384074942, |
|
"grad_norm": 0.542770504951477, |
|
"learning_rate": 3.826208184292952e-05, |
|
"loss": 0.2887, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 7.408274785323965, |
|
"grad_norm": 0.4088205397129059, |
|
"learning_rate": 3.804778188536887e-05, |
|
"loss": 0.2891, |
|
"step": 4745 |
|
}, |
|
{ |
|
"epoch": 7.41608118657299, |
|
"grad_norm": 0.3968678116798401, |
|
"learning_rate": 3.783394266299228e-05, |
|
"loss": 0.2853, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 7.423887587822014, |
|
"grad_norm": 0.41700446605682373, |
|
"learning_rate": 3.7620565766109975e-05, |
|
"loss": 0.2903, |
|
"step": 4755 |
|
}, |
|
{ |
|
"epoch": 7.4316939890710385, |
|
"grad_norm": 0.399271160364151, |
|
"learning_rate": 3.7407652781594095e-05, |
|
"loss": 0.289, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 7.439500390320062, |
|
"grad_norm": 0.3900463879108429, |
|
"learning_rate": 3.719520529286659e-05, |
|
"loss": 0.2846, |
|
"step": 4765 |
|
}, |
|
{ |
|
"epoch": 7.447306791569087, |
|
"grad_norm": 0.43415582180023193, |
|
"learning_rate": 3.698322487988755e-05, |
|
"loss": 0.2898, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 7.455113192818111, |
|
"grad_norm": 0.38591185212135315, |
|
"learning_rate": 3.677171311914346e-05, |
|
"loss": 0.2832, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 7.462919594067135, |
|
"grad_norm": 0.3995843827724457, |
|
"learning_rate": 3.6560671583635467e-05, |
|
"loss": 0.2904, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 7.470725995316159, |
|
"grad_norm": 0.4198484718799591, |
|
"learning_rate": 3.6350101842867645e-05, |
|
"loss": 0.2769, |
|
"step": 4785 |
|
}, |
|
{ |
|
"epoch": 7.478532396565184, |
|
"grad_norm": 0.39973580837249756, |
|
"learning_rate": 3.614000546283547e-05, |
|
"loss": 0.2867, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 7.4863387978142075, |
|
"grad_norm": 0.37141865491867065, |
|
"learning_rate": 3.593038400601395e-05, |
|
"loss": 0.2834, |
|
"step": 4795 |
|
}, |
|
{ |
|
"epoch": 7.494145199063232, |
|
"grad_norm": 0.4066362679004669, |
|
"learning_rate": 3.5721239031346066e-05, |
|
"loss": 0.2847, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 7.501951600312256, |
|
"grad_norm": 0.40516340732574463, |
|
"learning_rate": 3.55125720942314e-05, |
|
"loss": 0.2854, |
|
"step": 4805 |
|
}, |
|
{ |
|
"epoch": 7.509758001561281, |
|
"grad_norm": 0.428204208612442, |
|
"learning_rate": 3.530438474651428e-05, |
|
"loss": 0.2816, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 7.517564402810304, |
|
"grad_norm": 0.4114013612270355, |
|
"learning_rate": 3.509667853647235e-05, |
|
"loss": 0.2783, |
|
"step": 4815 |
|
}, |
|
{ |
|
"epoch": 7.525370804059329, |
|
"grad_norm": 0.4049510955810547, |
|
"learning_rate": 3.4889455008805106e-05, |
|
"loss": 0.2904, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 7.533177205308353, |
|
"grad_norm": 0.4051942825317383, |
|
"learning_rate": 3.468271570462235e-05, |
|
"loss": 0.2893, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 7.540983606557377, |
|
"grad_norm": 0.4027831554412842, |
|
"learning_rate": 3.447646216143268e-05, |
|
"loss": 0.2903, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 7.548790007806401, |
|
"grad_norm": 0.3996686339378357, |
|
"learning_rate": 3.427069591313226e-05, |
|
"loss": 0.2823, |
|
"step": 4835 |
|
}, |
|
{ |
|
"epoch": 7.556596409055425, |
|
"grad_norm": 0.40199506282806396, |
|
"learning_rate": 3.406541848999312e-05, |
|
"loss": 0.2817, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 7.56440281030445, |
|
"grad_norm": 0.4444729685783386, |
|
"learning_rate": 3.3860631418652e-05, |
|
"loss": 0.2859, |
|
"step": 4845 |
|
}, |
|
{ |
|
"epoch": 7.572209211553474, |
|
"grad_norm": 0.40995243191719055, |
|
"learning_rate": 3.365633622209891e-05, |
|
"loss": 0.2869, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 7.580015612802498, |
|
"grad_norm": 0.4094167649745941, |
|
"learning_rate": 3.345253441966579e-05, |
|
"loss": 0.2862, |
|
"step": 4855 |
|
}, |
|
{ |
|
"epoch": 7.587822014051522, |
|
"grad_norm": 0.38202327489852905, |
|
"learning_rate": 3.324922752701528e-05, |
|
"loss": 0.2842, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 7.595628415300546, |
|
"grad_norm": 0.4048561155796051, |
|
"learning_rate": 3.3046417056129366e-05, |
|
"loss": 0.2829, |
|
"step": 4865 |
|
}, |
|
{ |
|
"epoch": 7.603434816549571, |
|
"grad_norm": 0.3873648941516876, |
|
"learning_rate": 3.2844104515298155e-05, |
|
"loss": 0.2837, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 7.611241217798595, |
|
"grad_norm": 0.39025747776031494, |
|
"learning_rate": 3.2642291409108775e-05, |
|
"loss": 0.2777, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 7.619047619047619, |
|
"grad_norm": 0.39962899684906006, |
|
"learning_rate": 3.244097923843398e-05, |
|
"loss": 0.2897, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 7.626854020296643, |
|
"grad_norm": 0.39887768030166626, |
|
"learning_rate": 3.2240169500421135e-05, |
|
"loss": 0.2777, |
|
"step": 4885 |
|
}, |
|
{ |
|
"epoch": 7.634660421545668, |
|
"grad_norm": 0.42155277729034424, |
|
"learning_rate": 3.2039863688481055e-05, |
|
"loss": 0.2882, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 7.642466822794692, |
|
"grad_norm": 0.40622079372406006, |
|
"learning_rate": 3.184006329227684e-05, |
|
"loss": 0.292, |
|
"step": 4895 |
|
}, |
|
{ |
|
"epoch": 7.6502732240437155, |
|
"grad_norm": 0.41190096735954285, |
|
"learning_rate": 3.164076979771287e-05, |
|
"loss": 0.283, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 7.65807962529274, |
|
"grad_norm": 0.41508612036705017, |
|
"learning_rate": 3.144198468692379e-05, |
|
"loss": 0.279, |
|
"step": 4905 |
|
}, |
|
{ |
|
"epoch": 7.665886026541764, |
|
"grad_norm": 0.39837446808815, |
|
"learning_rate": 3.1243709438263255e-05, |
|
"loss": 0.2843, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 7.6736924277907885, |
|
"grad_norm": 0.4004950523376465, |
|
"learning_rate": 3.104594552629331e-05, |
|
"loss": 0.2866, |
|
"step": 4915 |
|
}, |
|
{ |
|
"epoch": 7.681498829039812, |
|
"grad_norm": 0.3863111436367035, |
|
"learning_rate": 3.0848694421773075e-05, |
|
"loss": 0.2803, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 7.689305230288837, |
|
"grad_norm": 0.39881014823913574, |
|
"learning_rate": 3.065195759164797e-05, |
|
"loss": 0.2832, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 7.697111631537861, |
|
"grad_norm": 0.3900108337402344, |
|
"learning_rate": 3.0455736499038845e-05, |
|
"loss": 0.2955, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 7.704918032786885, |
|
"grad_norm": 0.4136905372142792, |
|
"learning_rate": 3.0260032603230982e-05, |
|
"loss": 0.2938, |
|
"step": 4935 |
|
}, |
|
{ |
|
"epoch": 7.712724434035909, |
|
"grad_norm": 0.4050344228744507, |
|
"learning_rate": 3.0064847359663284e-05, |
|
"loss": 0.2903, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 7.720530835284934, |
|
"grad_norm": 0.4190680980682373, |
|
"learning_rate": 2.9870182219917564e-05, |
|
"loss": 0.2944, |
|
"step": 4945 |
|
}, |
|
{ |
|
"epoch": 7.7283372365339575, |
|
"grad_norm": 0.39599907398223877, |
|
"learning_rate": 2.9676038631707593e-05, |
|
"loss": 0.2911, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 7.736143637782982, |
|
"grad_norm": 0.39728233218193054, |
|
"learning_rate": 2.9482418038868275e-05, |
|
"loss": 0.2861, |
|
"step": 4955 |
|
}, |
|
{ |
|
"epoch": 7.743950039032006, |
|
"grad_norm": 0.39015698432922363, |
|
"learning_rate": 2.9289321881345254e-05, |
|
"loss": 0.2846, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 7.751756440281031, |
|
"grad_norm": 0.394378662109375, |
|
"learning_rate": 2.9096751595183823e-05, |
|
"loss": 0.2879, |
|
"step": 4965 |
|
}, |
|
{ |
|
"epoch": 7.759562841530054, |
|
"grad_norm": 0.39528360962867737, |
|
"learning_rate": 2.8904708612518404e-05, |
|
"loss": 0.2899, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 7.767369242779079, |
|
"grad_norm": 0.41304799914360046, |
|
"learning_rate": 2.8713194361562036e-05, |
|
"loss": 0.2898, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 7.775175644028103, |
|
"grad_norm": 0.4061037003993988, |
|
"learning_rate": 2.8522210266595384e-05, |
|
"loss": 0.2901, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 7.7829820452771274, |
|
"grad_norm": 0.42109227180480957, |
|
"learning_rate": 2.833175774795651e-05, |
|
"loss": 0.2868, |
|
"step": 4985 |
|
}, |
|
{ |
|
"epoch": 7.790788446526151, |
|
"grad_norm": 0.38924074172973633, |
|
"learning_rate": 2.814183822203019e-05, |
|
"loss": 0.2887, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 7.798594847775176, |
|
"grad_norm": 0.4080544114112854, |
|
"learning_rate": 2.795245310123732e-05, |
|
"loss": 0.2884, |
|
"step": 4995 |
|
}, |
|
{ |
|
"epoch": 7.8064012490242, |
|
"grad_norm": 0.39689868688583374, |
|
"learning_rate": 2.776360379402445e-05, |
|
"loss": 0.2859, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 7.814207650273224, |
|
"grad_norm": 0.40437766909599304, |
|
"learning_rate": 2.7575291704853323e-05, |
|
"loss": 0.287, |
|
"step": 5005 |
|
}, |
|
{ |
|
"epoch": 7.822014051522248, |
|
"grad_norm": 0.4085007905960083, |
|
"learning_rate": 2.738751823419041e-05, |
|
"loss": 0.2875, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 7.829820452771273, |
|
"grad_norm": 0.3995596468448639, |
|
"learning_rate": 2.720028477849652e-05, |
|
"loss": 0.2818, |
|
"step": 5015 |
|
}, |
|
{ |
|
"epoch": 7.8376268540202965, |
|
"grad_norm": 0.3908860683441162, |
|
"learning_rate": 2.7013592730216465e-05, |
|
"loss": 0.283, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 7.845433255269321, |
|
"grad_norm": 0.39162319898605347, |
|
"learning_rate": 2.6827443477768454e-05, |
|
"loss": 0.2833, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 7.853239656518345, |
|
"grad_norm": 0.3898274600505829, |
|
"learning_rate": 2.664183840553417e-05, |
|
"loss": 0.2903, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 7.8610460577673695, |
|
"grad_norm": 0.4035649597644806, |
|
"learning_rate": 2.6456778893848144e-05, |
|
"loss": 0.2826, |
|
"step": 5035 |
|
}, |
|
{ |
|
"epoch": 7.868852459016393, |
|
"grad_norm": 0.4092506170272827, |
|
"learning_rate": 2.6272266318987603e-05, |
|
"loss": 0.284, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 7.876658860265418, |
|
"grad_norm": 0.40052658319473267, |
|
"learning_rate": 2.6088302053162272e-05, |
|
"loss": 0.2894, |
|
"step": 5045 |
|
}, |
|
{ |
|
"epoch": 7.884465261514442, |
|
"grad_norm": 0.3952447175979614, |
|
"learning_rate": 2.5904887464504114e-05, |
|
"loss": 0.281, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 7.892271662763466, |
|
"grad_norm": 0.3892611563205719, |
|
"learning_rate": 2.5722023917057125e-05, |
|
"loss": 0.2824, |
|
"step": 5055 |
|
}, |
|
{ |
|
"epoch": 7.90007806401249, |
|
"grad_norm": 0.3946845531463623, |
|
"learning_rate": 2.5539712770767376e-05, |
|
"loss": 0.2865, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 7.907884465261515, |
|
"grad_norm": 0.39248526096343994, |
|
"learning_rate": 2.535795538147262e-05, |
|
"loss": 0.2863, |
|
"step": 5065 |
|
}, |
|
{ |
|
"epoch": 7.9156908665105385, |
|
"grad_norm": 0.4364306330680847, |
|
"learning_rate": 2.5176753100892426e-05, |
|
"loss": 0.2916, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 7.923497267759563, |
|
"grad_norm": 0.3956867754459381, |
|
"learning_rate": 2.4996107276618008e-05, |
|
"loss": 0.2819, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 7.931303669008587, |
|
"grad_norm": 0.3908245265483856, |
|
"learning_rate": 2.4816019252102273e-05, |
|
"loss": 0.2864, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 7.939110070257612, |
|
"grad_norm": 0.3847108483314514, |
|
"learning_rate": 2.46364903666498e-05, |
|
"loss": 0.2937, |
|
"step": 5085 |
|
}, |
|
{ |
|
"epoch": 7.946916471506635, |
|
"grad_norm": 0.412410169839859, |
|
"learning_rate": 2.445752195540687e-05, |
|
"loss": 0.2795, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 7.95472287275566, |
|
"grad_norm": 0.39699193835258484, |
|
"learning_rate": 2.4279115349351543e-05, |
|
"loss": 0.2909, |
|
"step": 5095 |
|
}, |
|
{ |
|
"epoch": 7.962529274004684, |
|
"grad_norm": 0.4221264719963074, |
|
"learning_rate": 2.4101271875283817e-05, |
|
"loss": 0.2862, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 7.970335675253708, |
|
"grad_norm": 0.38855311274528503, |
|
"learning_rate": 2.3923992855815647e-05, |
|
"loss": 0.2847, |
|
"step": 5105 |
|
}, |
|
{ |
|
"epoch": 7.978142076502732, |
|
"grad_norm": 0.4506378173828125, |
|
"learning_rate": 2.3747279609361196e-05, |
|
"loss": 0.2904, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 7.985948477751757, |
|
"grad_norm": 0.4097312390804291, |
|
"learning_rate": 2.3571133450127002e-05, |
|
"loss": 0.2903, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 7.993754879000781, |
|
"grad_norm": 0.4036017060279846, |
|
"learning_rate": 2.339555568810221e-05, |
|
"loss": 0.2857, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 1.1256442070007324, |
|
"eval_runtime": 0.3225, |
|
"eval_samples_per_second": 15.502, |
|
"eval_steps_per_second": 3.1, |
|
"step": 5124 |
|
}, |
|
{ |
|
"epoch": 8.001561280249804, |
|
"grad_norm": 0.3814958333969116, |
|
"learning_rate": 2.3220547629048796e-05, |
|
"loss": 0.2894, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 8.009367681498828, |
|
"grad_norm": 0.3972679376602173, |
|
"learning_rate": 2.3046110574491985e-05, |
|
"loss": 0.2791, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 8.017174082747854, |
|
"grad_norm": 0.3934348225593567, |
|
"learning_rate": 2.2872245821710346e-05, |
|
"loss": 0.2719, |
|
"step": 5135 |
|
}, |
|
{ |
|
"epoch": 8.024980483996877, |
|
"grad_norm": 0.4403572380542755, |
|
"learning_rate": 2.26989546637263e-05, |
|
"loss": 0.2688, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 8.032786885245901, |
|
"grad_norm": 0.43504083156585693, |
|
"learning_rate": 2.2526238389296585e-05, |
|
"loss": 0.2727, |
|
"step": 5145 |
|
}, |
|
{ |
|
"epoch": 8.040593286494925, |
|
"grad_norm": 0.40251481533050537, |
|
"learning_rate": 2.2354098282902446e-05, |
|
"loss": 0.2706, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 8.04839968774395, |
|
"grad_norm": 0.3875437080860138, |
|
"learning_rate": 2.218253562474023e-05, |
|
"loss": 0.2684, |
|
"step": 5155 |
|
}, |
|
{ |
|
"epoch": 8.056206088992974, |
|
"grad_norm": 0.39775198698043823, |
|
"learning_rate": 2.201155169071184e-05, |
|
"loss": 0.2685, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 8.064012490241998, |
|
"grad_norm": 0.4047357439994812, |
|
"learning_rate": 2.1841147752415235e-05, |
|
"loss": 0.2686, |
|
"step": 5165 |
|
}, |
|
{ |
|
"epoch": 8.071818891491022, |
|
"grad_norm": 0.4089979827404022, |
|
"learning_rate": 2.1671325077134963e-05, |
|
"loss": 0.2695, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 8.079625292740047, |
|
"grad_norm": 0.40754222869873047, |
|
"learning_rate": 2.1502084927832845e-05, |
|
"loss": 0.2801, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 8.087431693989071, |
|
"grad_norm": 0.4052791893482208, |
|
"learning_rate": 2.1333428563138303e-05, |
|
"loss": 0.2708, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 8.095238095238095, |
|
"grad_norm": 0.4029444754123688, |
|
"learning_rate": 2.116535723733938e-05, |
|
"loss": 0.2688, |
|
"step": 5185 |
|
}, |
|
{ |
|
"epoch": 8.103044496487119, |
|
"grad_norm": 0.43849509954452515, |
|
"learning_rate": 2.0997872200373116e-05, |
|
"loss": 0.2749, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 8.110850897736144, |
|
"grad_norm": 0.4012914001941681, |
|
"learning_rate": 2.083097469781632e-05, |
|
"loss": 0.2726, |
|
"step": 5195 |
|
}, |
|
{ |
|
"epoch": 8.118657298985168, |
|
"grad_norm": 0.42504939436912537, |
|
"learning_rate": 2.0664665970876496e-05, |
|
"loss": 0.2791, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 8.126463700234192, |
|
"grad_norm": 0.41797903180122375, |
|
"learning_rate": 2.0498947256382273e-05, |
|
"loss": 0.2728, |
|
"step": 5205 |
|
}, |
|
{ |
|
"epoch": 8.134270101483215, |
|
"grad_norm": 0.3988747000694275, |
|
"learning_rate": 2.0333819786774444e-05, |
|
"loss": 0.2757, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 8.142076502732241, |
|
"grad_norm": 0.4232763350009918, |
|
"learning_rate": 2.0169284790096853e-05, |
|
"loss": 0.2749, |
|
"step": 5215 |
|
}, |
|
{ |
|
"epoch": 8.149882903981265, |
|
"grad_norm": 0.41461604833602905, |
|
"learning_rate": 2.000534348998704e-05, |
|
"loss": 0.2747, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 8.157689305230289, |
|
"grad_norm": 0.4166080355644226, |
|
"learning_rate": 1.9841997105667275e-05, |
|
"loss": 0.2721, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 8.165495706479312, |
|
"grad_norm": 0.41639989614486694, |
|
"learning_rate": 1.967924685193552e-05, |
|
"loss": 0.2645, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 8.173302107728338, |
|
"grad_norm": 0.4174750745296478, |
|
"learning_rate": 1.9517093939156295e-05, |
|
"loss": 0.2732, |
|
"step": 5235 |
|
}, |
|
{ |
|
"epoch": 8.181108508977362, |
|
"grad_norm": 0.4450904130935669, |
|
"learning_rate": 1.9355539573251734e-05, |
|
"loss": 0.2777, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 8.188914910226385, |
|
"grad_norm": 0.4246358573436737, |
|
"learning_rate": 1.9194584955692706e-05, |
|
"loss": 0.2791, |
|
"step": 5245 |
|
}, |
|
{ |
|
"epoch": 8.19672131147541, |
|
"grad_norm": 0.4177229106426239, |
|
"learning_rate": 1.903423128348959e-05, |
|
"loss": 0.2743, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 8.204527712724435, |
|
"grad_norm": 0.42174917459487915, |
|
"learning_rate": 1.8874479749183782e-05, |
|
"loss": 0.2754, |
|
"step": 5255 |
|
}, |
|
{ |
|
"epoch": 8.212334113973458, |
|
"grad_norm": 0.4163845181465149, |
|
"learning_rate": 1.8715331540838487e-05, |
|
"loss": 0.2718, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 8.220140515222482, |
|
"grad_norm": 0.4283103346824646, |
|
"learning_rate": 1.855678784203002e-05, |
|
"loss": 0.2704, |
|
"step": 5265 |
|
}, |
|
{ |
|
"epoch": 8.227946916471506, |
|
"grad_norm": 0.4160711467266083, |
|
"learning_rate": 1.8398849831839014e-05, |
|
"loss": 0.2679, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 8.235753317720532, |
|
"grad_norm": 0.4049237370491028, |
|
"learning_rate": 1.824151868484164e-05, |
|
"loss": 0.269, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 8.243559718969555, |
|
"grad_norm": 0.4092932641506195, |
|
"learning_rate": 1.808479557110081e-05, |
|
"loss": 0.2725, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 8.251366120218579, |
|
"grad_norm": 0.42792901396751404, |
|
"learning_rate": 1.792868165615762e-05, |
|
"loss": 0.2778, |
|
"step": 5285 |
|
}, |
|
{ |
|
"epoch": 8.259172521467603, |
|
"grad_norm": 0.4112985134124756, |
|
"learning_rate": 1.7773178101022514e-05, |
|
"loss": 0.2708, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 8.266978922716628, |
|
"grad_norm": 0.4147091209888458, |
|
"learning_rate": 1.7618286062166677e-05, |
|
"loss": 0.2772, |
|
"step": 5295 |
|
}, |
|
{ |
|
"epoch": 8.274785323965652, |
|
"grad_norm": 0.4274720847606659, |
|
"learning_rate": 1.7464006691513623e-05, |
|
"loss": 0.273, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 8.282591725214676, |
|
"grad_norm": 0.40364208817481995, |
|
"learning_rate": 1.7310341136430385e-05, |
|
"loss": 0.2778, |
|
"step": 5305 |
|
}, |
|
{ |
|
"epoch": 8.2903981264637, |
|
"grad_norm": 0.42018428444862366, |
|
"learning_rate": 1.7157290539719106e-05, |
|
"loss": 0.2743, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 8.298204527712725, |
|
"grad_norm": 0.4169312119483948, |
|
"learning_rate": 1.700485603960853e-05, |
|
"loss": 0.275, |
|
"step": 5315 |
|
}, |
|
{ |
|
"epoch": 8.306010928961749, |
|
"grad_norm": 0.40692824125289917, |
|
"learning_rate": 1.6853038769745467e-05, |
|
"loss": 0.2683, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 8.313817330210773, |
|
"grad_norm": 0.41851934790611267, |
|
"learning_rate": 1.6701839859186542e-05, |
|
"loss": 0.2677, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 8.321623731459797, |
|
"grad_norm": 0.4233751595020294, |
|
"learning_rate": 1.655126043238957e-05, |
|
"loss": 0.27, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 8.329430132708822, |
|
"grad_norm": 0.4326801002025604, |
|
"learning_rate": 1.6401301609205335e-05, |
|
"loss": 0.2774, |
|
"step": 5335 |
|
}, |
|
{ |
|
"epoch": 8.337236533957846, |
|
"grad_norm": 0.42542752623558044, |
|
"learning_rate": 1.6251964504869222e-05, |
|
"loss": 0.2692, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 8.34504293520687, |
|
"grad_norm": 0.4217033386230469, |
|
"learning_rate": 1.6103250229992927e-05, |
|
"loss": 0.2804, |
|
"step": 5345 |
|
}, |
|
{ |
|
"epoch": 8.352849336455893, |
|
"grad_norm": 0.4149416387081146, |
|
"learning_rate": 1.595515989055618e-05, |
|
"loss": 0.2732, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 8.360655737704919, |
|
"grad_norm": 0.42739465832710266, |
|
"learning_rate": 1.5807694587898648e-05, |
|
"loss": 0.2782, |
|
"step": 5355 |
|
}, |
|
{ |
|
"epoch": 8.368462138953943, |
|
"grad_norm": 0.42222627997398376, |
|
"learning_rate": 1.566085541871145e-05, |
|
"loss": 0.2699, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 8.376268540202966, |
|
"grad_norm": 0.410786509513855, |
|
"learning_rate": 1.551464347502929e-05, |
|
"loss": 0.2743, |
|
"step": 5365 |
|
}, |
|
{ |
|
"epoch": 8.38407494145199, |
|
"grad_norm": 0.42929258942604065, |
|
"learning_rate": 1.5369059844222278e-05, |
|
"loss": 0.2785, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 8.391881342701016, |
|
"grad_norm": 0.4153624475002289, |
|
"learning_rate": 1.5224105608987704e-05, |
|
"loss": 0.2687, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 8.39968774395004, |
|
"grad_norm": 0.3873469829559326, |
|
"learning_rate": 1.5079781847342123e-05, |
|
"loss": 0.2678, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 8.407494145199063, |
|
"grad_norm": 0.43077465891838074, |
|
"learning_rate": 1.4936089632613281e-05, |
|
"loss": 0.2765, |
|
"step": 5385 |
|
}, |
|
{ |
|
"epoch": 8.415300546448087, |
|
"grad_norm": 0.4113737940788269, |
|
"learning_rate": 1.4793030033432142e-05, |
|
"loss": 0.2688, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 8.42310694769711, |
|
"grad_norm": 0.40002962946891785, |
|
"learning_rate": 1.4650604113724953e-05, |
|
"loss": 0.2693, |
|
"step": 5395 |
|
}, |
|
{ |
|
"epoch": 8.430913348946136, |
|
"grad_norm": 0.4060341417789459, |
|
"learning_rate": 1.4508812932705363e-05, |
|
"loss": 0.2754, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 8.43871975019516, |
|
"grad_norm": 0.43747276067733765, |
|
"learning_rate": 1.4367657544866375e-05, |
|
"loss": 0.273, |
|
"step": 5405 |
|
}, |
|
{ |
|
"epoch": 8.446526151444184, |
|
"grad_norm": 0.41913843154907227, |
|
"learning_rate": 1.42271389999728e-05, |
|
"loss": 0.2756, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 8.454332552693208, |
|
"grad_norm": 0.39348214864730835, |
|
"learning_rate": 1.4087258343053167e-05, |
|
"loss": 0.2701, |
|
"step": 5415 |
|
}, |
|
{ |
|
"epoch": 8.462138953942233, |
|
"grad_norm": 0.4029614329338074, |
|
"learning_rate": 1.3948016614392112e-05, |
|
"loss": 0.269, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 8.469945355191257, |
|
"grad_norm": 0.45508867502212524, |
|
"learning_rate": 1.3809414849522584e-05, |
|
"loss": 0.2738, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 8.47775175644028, |
|
"grad_norm": 0.4153214395046234, |
|
"learning_rate": 1.367145407921817e-05, |
|
"loss": 0.2713, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 8.485558157689304, |
|
"grad_norm": 0.4336087703704834, |
|
"learning_rate": 1.3534135329485376e-05, |
|
"loss": 0.2786, |
|
"step": 5435 |
|
}, |
|
{ |
|
"epoch": 8.49336455893833, |
|
"grad_norm": 0.4236217141151428, |
|
"learning_rate": 1.339745962155613e-05, |
|
"loss": 0.277, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 8.501170960187354, |
|
"grad_norm": 0.43844056129455566, |
|
"learning_rate": 1.3261427971880025e-05, |
|
"loss": 0.2709, |
|
"step": 5445 |
|
}, |
|
{ |
|
"epoch": 8.508977361436378, |
|
"grad_norm": 0.43588918447494507, |
|
"learning_rate": 1.3126041392116772e-05, |
|
"loss": 0.2759, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 8.516783762685401, |
|
"grad_norm": 0.4094368517398834, |
|
"learning_rate": 1.2991300889128866e-05, |
|
"loss": 0.2776, |
|
"step": 5455 |
|
}, |
|
{ |
|
"epoch": 8.524590163934427, |
|
"grad_norm": 0.41443923115730286, |
|
"learning_rate": 1.2857207464973875e-05, |
|
"loss": 0.2746, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 8.53239656518345, |
|
"grad_norm": 0.44371187686920166, |
|
"learning_rate": 1.2723762116897075e-05, |
|
"loss": 0.2821, |
|
"step": 5465 |
|
}, |
|
{ |
|
"epoch": 8.540202966432474, |
|
"grad_norm": 0.3933749198913574, |
|
"learning_rate": 1.2590965837324131e-05, |
|
"loss": 0.2713, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 8.548009367681498, |
|
"grad_norm": 0.4023985266685486, |
|
"learning_rate": 1.2458819613853468e-05, |
|
"loss": 0.2714, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 8.555815768930524, |
|
"grad_norm": 0.4138771891593933, |
|
"learning_rate": 1.2327324429249232e-05, |
|
"loss": 0.2749, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 8.563622170179547, |
|
"grad_norm": 0.4046929180622101, |
|
"learning_rate": 1.2196481261433735e-05, |
|
"loss": 0.2732, |
|
"step": 5485 |
|
}, |
|
{ |
|
"epoch": 8.571428571428571, |
|
"grad_norm": 0.4175087511539459, |
|
"learning_rate": 1.2066291083480296e-05, |
|
"loss": 0.2691, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 8.579234972677595, |
|
"grad_norm": 0.4213341176509857, |
|
"learning_rate": 1.1936754863606014e-05, |
|
"loss": 0.2769, |
|
"step": 5495 |
|
}, |
|
{ |
|
"epoch": 8.58704137392662, |
|
"grad_norm": 0.43556031584739685, |
|
"learning_rate": 1.1807873565164506e-05, |
|
"loss": 0.2712, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 8.594847775175644, |
|
"grad_norm": 0.39848992228507996, |
|
"learning_rate": 1.1679648146638766e-05, |
|
"loss": 0.2718, |
|
"step": 5505 |
|
}, |
|
{ |
|
"epoch": 8.602654176424668, |
|
"grad_norm": 0.4068874418735504, |
|
"learning_rate": 1.155207956163411e-05, |
|
"loss": 0.2714, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 8.610460577673692, |
|
"grad_norm": 0.41433241963386536, |
|
"learning_rate": 1.1425168758870964e-05, |
|
"loss": 0.275, |
|
"step": 5515 |
|
}, |
|
{ |
|
"epoch": 8.618266978922717, |
|
"grad_norm": 0.42830371856689453, |
|
"learning_rate": 1.129891668217783e-05, |
|
"loss": 0.2743, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 8.626073380171741, |
|
"grad_norm": 0.4217815697193146, |
|
"learning_rate": 1.1173324270484397e-05, |
|
"loss": 0.2756, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 8.633879781420765, |
|
"grad_norm": 0.3956458270549774, |
|
"learning_rate": 1.1048392457814405e-05, |
|
"loss": 0.2743, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 8.641686182669789, |
|
"grad_norm": 0.42598819732666016, |
|
"learning_rate": 1.0924122173278751e-05, |
|
"loss": 0.2741, |
|
"step": 5535 |
|
}, |
|
{ |
|
"epoch": 8.649492583918814, |
|
"grad_norm": 0.44088292121887207, |
|
"learning_rate": 1.080051434106859e-05, |
|
"loss": 0.2784, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 8.657298985167838, |
|
"grad_norm": 0.40059834718704224, |
|
"learning_rate": 1.067756988044848e-05, |
|
"loss": 0.2736, |
|
"step": 5545 |
|
}, |
|
{ |
|
"epoch": 8.665105386416862, |
|
"grad_norm": 0.41283807158470154, |
|
"learning_rate": 1.0555289705749483e-05, |
|
"loss": 0.2781, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 8.672911787665885, |
|
"grad_norm": 0.40279173851013184, |
|
"learning_rate": 1.0433674726362442e-05, |
|
"loss": 0.2689, |
|
"step": 5555 |
|
}, |
|
{ |
|
"epoch": 8.680718188914911, |
|
"grad_norm": 0.4150051176548004, |
|
"learning_rate": 1.0312725846731175e-05, |
|
"loss": 0.2733, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 8.688524590163935, |
|
"grad_norm": 0.4160909950733185, |
|
"learning_rate": 1.019244396634571e-05, |
|
"loss": 0.272, |
|
"step": 5565 |
|
}, |
|
{ |
|
"epoch": 8.696330991412959, |
|
"grad_norm": 0.4090689718723297, |
|
"learning_rate": 1.0072829979735699e-05, |
|
"loss": 0.2779, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 8.704137392661982, |
|
"grad_norm": 0.43035534024238586, |
|
"learning_rate": 9.953884776463652e-06, |
|
"loss": 0.2799, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 8.711943793911008, |
|
"grad_norm": 0.4087963402271271, |
|
"learning_rate": 9.835609241118404e-06, |
|
"loss": 0.2724, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 8.719750195160032, |
|
"grad_norm": 0.4282284677028656, |
|
"learning_rate": 9.71800425330851e-06, |
|
"loss": 0.2766, |
|
"step": 5585 |
|
}, |
|
{ |
|
"epoch": 8.727556596409055, |
|
"grad_norm": 0.39750391244888306, |
|
"learning_rate": 9.601070687655667e-06, |
|
"loss": 0.2751, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 8.735362997658079, |
|
"grad_norm": 0.43247750401496887, |
|
"learning_rate": 9.48480941378831e-06, |
|
"loss": 0.2701, |
|
"step": 5595 |
|
}, |
|
{ |
|
"epoch": 8.743169398907105, |
|
"grad_norm": 0.42617279291152954, |
|
"learning_rate": 9.369221296335006e-06, |
|
"loss": 0.2774, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 8.750975800156128, |
|
"grad_norm": 0.4098498225212097, |
|
"learning_rate": 9.254307194918144e-06, |
|
"loss": 0.2779, |
|
"step": 5605 |
|
}, |
|
{ |
|
"epoch": 8.758782201405152, |
|
"grad_norm": 0.4429432153701782, |
|
"learning_rate": 9.140067964147447e-06, |
|
"loss": 0.2809, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 8.766588602654176, |
|
"grad_norm": 0.42042461037635803, |
|
"learning_rate": 9.026504453613726e-06, |
|
"loss": 0.2755, |
|
"step": 5615 |
|
}, |
|
{ |
|
"epoch": 8.774395003903201, |
|
"grad_norm": 0.41905051469802856, |
|
"learning_rate": 8.91361750788241e-06, |
|
"loss": 0.2679, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 8.782201405152225, |
|
"grad_norm": 0.4195885956287384, |
|
"learning_rate": 8.801407966487486e-06, |
|
"loss": 0.2734, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 8.790007806401249, |
|
"grad_norm": 0.4253963232040405, |
|
"learning_rate": 8.689876663924957e-06, |
|
"loss": 0.2674, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 8.797814207650273, |
|
"grad_norm": 0.43018728494644165, |
|
"learning_rate": 8.579024429646932e-06, |
|
"loss": 0.2762, |
|
"step": 5635 |
|
}, |
|
{ |
|
"epoch": 8.805620608899297, |
|
"grad_norm": 0.41742637753486633, |
|
"learning_rate": 8.46885208805529e-06, |
|
"loss": 0.2768, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 8.813427010148322, |
|
"grad_norm": 0.41150280833244324, |
|
"learning_rate": 8.359360458495557e-06, |
|
"loss": 0.2777, |
|
"step": 5645 |
|
}, |
|
{ |
|
"epoch": 8.821233411397346, |
|
"grad_norm": 0.4265673756599426, |
|
"learning_rate": 8.250550355250875e-06, |
|
"loss": 0.2743, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 8.82903981264637, |
|
"grad_norm": 0.4104432463645935, |
|
"learning_rate": 8.142422587535903e-06, |
|
"loss": 0.279, |
|
"step": 5655 |
|
}, |
|
{ |
|
"epoch": 8.836846213895395, |
|
"grad_norm": 0.4088299572467804, |
|
"learning_rate": 8.034977959490775e-06, |
|
"loss": 0.2723, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 8.844652615144419, |
|
"grad_norm": 0.4326065182685852, |
|
"learning_rate": 7.92821727017523e-06, |
|
"loss": 0.2748, |
|
"step": 5665 |
|
}, |
|
{ |
|
"epoch": 8.852459016393443, |
|
"grad_norm": 0.4133962392807007, |
|
"learning_rate": 7.822141313562547e-06, |
|
"loss": 0.2699, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 8.860265417642466, |
|
"grad_norm": 0.42703020572662354, |
|
"learning_rate": 7.71675087853364e-06, |
|
"loss": 0.2735, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 8.86807181889149, |
|
"grad_norm": 0.43023622035980225, |
|
"learning_rate": 7.612046748871327e-06, |
|
"loss": 0.2764, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 8.875878220140516, |
|
"grad_norm": 0.4223518371582031, |
|
"learning_rate": 7.50802970325436e-06, |
|
"loss": 0.276, |
|
"step": 5685 |
|
}, |
|
{ |
|
"epoch": 8.88368462138954, |
|
"grad_norm": 0.43024882674217224, |
|
"learning_rate": 7.404700515251672e-06, |
|
"loss": 0.2745, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 8.891491022638563, |
|
"grad_norm": 0.4196966886520386, |
|
"learning_rate": 7.30205995331672e-06, |
|
"loss": 0.2791, |
|
"step": 5695 |
|
}, |
|
{ |
|
"epoch": 8.899297423887587, |
|
"grad_norm": 0.39749372005462646, |
|
"learning_rate": 7.200108780781556e-06, |
|
"loss": 0.2753, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 8.907103825136613, |
|
"grad_norm": 0.4266413748264313, |
|
"learning_rate": 7.0988477558513785e-06, |
|
"loss": 0.2769, |
|
"step": 5705 |
|
}, |
|
{ |
|
"epoch": 8.914910226385636, |
|
"grad_norm": 0.4070897102355957, |
|
"learning_rate": 6.998277631598793e-06, |
|
"loss": 0.2757, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 8.92271662763466, |
|
"grad_norm": 0.39892005920410156, |
|
"learning_rate": 6.898399155958168e-06, |
|
"loss": 0.2741, |
|
"step": 5715 |
|
}, |
|
{ |
|
"epoch": 8.930523028883684, |
|
"grad_norm": 0.42477500438690186, |
|
"learning_rate": 6.7992130717201564e-06, |
|
"loss": 0.2791, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 8.93832943013271, |
|
"grad_norm": 0.44194042682647705, |
|
"learning_rate": 6.700720116526116e-06, |
|
"loss": 0.2786, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 8.946135831381733, |
|
"grad_norm": 0.4111786484718323, |
|
"learning_rate": 6.6029210228626626e-06, |
|
"loss": 0.2728, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 8.953942232630757, |
|
"grad_norm": 0.43993067741394043, |
|
"learning_rate": 6.505816518056162e-06, |
|
"loss": 0.2799, |
|
"step": 5735 |
|
}, |
|
{ |
|
"epoch": 8.96174863387978, |
|
"grad_norm": 0.413835346698761, |
|
"learning_rate": 6.409407324267447e-06, |
|
"loss": 0.2761, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 8.969555035128806, |
|
"grad_norm": 0.4155209958553314, |
|
"learning_rate": 6.313694158486228e-06, |
|
"loss": 0.2726, |
|
"step": 5745 |
|
}, |
|
{ |
|
"epoch": 8.97736143637783, |
|
"grad_norm": 0.43269336223602295, |
|
"learning_rate": 6.218677732526035e-06, |
|
"loss": 0.2697, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 8.985167837626854, |
|
"grad_norm": 0.4364638924598694, |
|
"learning_rate": 6.124358753018689e-06, |
|
"loss": 0.2788, |
|
"step": 5755 |
|
}, |
|
{ |
|
"epoch": 8.992974238875878, |
|
"grad_norm": 0.40509918332099915, |
|
"learning_rate": 6.030737921409169e-06, |
|
"loss": 0.2732, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 8.999219359875097, |
|
"eval_loss": 1.1471812725067139, |
|
"eval_runtime": 0.3388, |
|
"eval_samples_per_second": 14.758, |
|
"eval_steps_per_second": 2.952, |
|
"step": 5764 |
|
}, |
|
{ |
|
"epoch": 9.000780640124903, |
|
"grad_norm": 0.40420717000961304, |
|
"learning_rate": 5.937815933950375e-06, |
|
"loss": 0.2633, |
|
"step": 5765 |
|
}, |
|
{ |
|
"epoch": 9.008587041373927, |
|
"grad_norm": 0.41395288705825806, |
|
"learning_rate": 5.8455934816979305e-06, |
|
"loss": 0.2729, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 9.01639344262295, |
|
"grad_norm": 0.4095096290111542, |
|
"learning_rate": 5.7540712505050444e-06, |
|
"loss": 0.265, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 9.024199843871974, |
|
"grad_norm": 0.41725656390190125, |
|
"learning_rate": 5.663249921017477e-06, |
|
"loss": 0.2632, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 9.032006245121, |
|
"grad_norm": 0.41932108998298645, |
|
"learning_rate": 5.57313016866835e-06, |
|
"loss": 0.2709, |
|
"step": 5785 |
|
}, |
|
{ |
|
"epoch": 9.039812646370024, |
|
"grad_norm": 0.41799718141555786, |
|
"learning_rate": 5.483712663673224e-06, |
|
"loss": 0.2696, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 9.047619047619047, |
|
"grad_norm": 0.41482260823249817, |
|
"learning_rate": 5.394998071025104e-06, |
|
"loss": 0.2623, |
|
"step": 5795 |
|
}, |
|
{ |
|
"epoch": 9.055425448868071, |
|
"grad_norm": 0.43754109740257263, |
|
"learning_rate": 5.306987050489442e-06, |
|
"loss": 0.2673, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 9.063231850117097, |
|
"grad_norm": 0.41212213039398193, |
|
"learning_rate": 5.21968025659928e-06, |
|
"loss": 0.2643, |
|
"step": 5805 |
|
}, |
|
{ |
|
"epoch": 9.07103825136612, |
|
"grad_norm": 0.43795979022979736, |
|
"learning_rate": 5.1330783386503765e-06, |
|
"loss": 0.2672, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 9.078844652615144, |
|
"grad_norm": 0.416830837726593, |
|
"learning_rate": 5.047181940696333e-06, |
|
"loss": 0.2637, |
|
"step": 5815 |
|
}, |
|
{ |
|
"epoch": 9.086651053864168, |
|
"grad_norm": 0.4341103136539459, |
|
"learning_rate": 4.961991701543889e-06, |
|
"loss": 0.2712, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 9.094457455113194, |
|
"grad_norm": 0.4144740402698517, |
|
"learning_rate": 4.877508254748076e-06, |
|
"loss": 0.2665, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 9.102263856362217, |
|
"grad_norm": 0.42530107498168945, |
|
"learning_rate": 4.7937322286075725e-06, |
|
"loss": 0.2678, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 9.110070257611241, |
|
"grad_norm": 0.4266839623451233, |
|
"learning_rate": 4.710664246160013e-06, |
|
"loss": 0.2709, |
|
"step": 5835 |
|
}, |
|
{ |
|
"epoch": 9.117876658860265, |
|
"grad_norm": 0.42478978633880615, |
|
"learning_rate": 4.628304925177318e-06, |
|
"loss": 0.2692, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 9.12568306010929, |
|
"grad_norm": 0.42860227823257446, |
|
"learning_rate": 4.54665487816115e-06, |
|
"loss": 0.2694, |
|
"step": 5845 |
|
}, |
|
{ |
|
"epoch": 9.133489461358314, |
|
"grad_norm": 0.43037670850753784, |
|
"learning_rate": 4.465714712338398e-06, |
|
"loss": 0.2686, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 9.141295862607338, |
|
"grad_norm": 0.4481579661369324, |
|
"learning_rate": 4.385485029656489e-06, |
|
"loss": 0.2672, |
|
"step": 5855 |
|
}, |
|
{ |
|
"epoch": 9.149102263856362, |
|
"grad_norm": 0.406190425157547, |
|
"learning_rate": 4.305966426779118e-06, |
|
"loss": 0.2697, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 9.156908665105387, |
|
"grad_norm": 0.41195255517959595, |
|
"learning_rate": 4.22715949508169e-06, |
|
"loss": 0.268, |
|
"step": 5865 |
|
}, |
|
{ |
|
"epoch": 9.164715066354411, |
|
"grad_norm": 0.41223201155662537, |
|
"learning_rate": 4.149064820646953e-06, |
|
"loss": 0.2625, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 9.172521467603435, |
|
"grad_norm": 0.40198594331741333, |
|
"learning_rate": 4.071682984260638e-06, |
|
"loss": 0.2654, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 9.180327868852459, |
|
"grad_norm": 0.4083414673805237, |
|
"learning_rate": 3.99501456140714e-06, |
|
"loss": 0.2666, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 9.188134270101482, |
|
"grad_norm": 0.43248122930526733, |
|
"learning_rate": 3.919060122265228e-06, |
|
"loss": 0.2712, |
|
"step": 5885 |
|
}, |
|
{ |
|
"epoch": 9.195940671350508, |
|
"grad_norm": 0.4074991047382355, |
|
"learning_rate": 3.8438202317037986e-06, |
|
"loss": 0.2636, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 9.203747072599532, |
|
"grad_norm": 0.410256028175354, |
|
"learning_rate": 3.7692954492777764e-06, |
|
"loss": 0.2663, |
|
"step": 5895 |
|
}, |
|
{ |
|
"epoch": 9.211553473848555, |
|
"grad_norm": 0.41726231575012207, |
|
"learning_rate": 3.6954863292237297e-06, |
|
"loss": 0.2649, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 9.21935987509758, |
|
"grad_norm": 0.40415236353874207, |
|
"learning_rate": 3.622393420456016e-06, |
|
"loss": 0.2691, |
|
"step": 5905 |
|
}, |
|
{ |
|
"epoch": 9.227166276346605, |
|
"grad_norm": 0.40107595920562744, |
|
"learning_rate": 3.550017266562489e-06, |
|
"loss": 0.2636, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 9.234972677595628, |
|
"grad_norm": 0.4478661119937897, |
|
"learning_rate": 3.4783584058005527e-06, |
|
"loss": 0.2711, |
|
"step": 5915 |
|
}, |
|
{ |
|
"epoch": 9.242779078844652, |
|
"grad_norm": 0.40624120831489563, |
|
"learning_rate": 3.40741737109318e-06, |
|
"loss": 0.2652, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 9.250585480093676, |
|
"grad_norm": 0.42705732583999634, |
|
"learning_rate": 3.3371946900248473e-06, |
|
"loss": 0.2675, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 9.258391881342702, |
|
"grad_norm": 0.4034976661205292, |
|
"learning_rate": 3.267690884837726e-06, |
|
"loss": 0.2631, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 9.266198282591725, |
|
"grad_norm": 0.4358353912830353, |
|
"learning_rate": 3.198906472427732e-06, |
|
"loss": 0.2702, |
|
"step": 5935 |
|
}, |
|
{ |
|
"epoch": 9.274004683840749, |
|
"grad_norm": 0.42044761776924133, |
|
"learning_rate": 3.130841964340692e-06, |
|
"loss": 0.269, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 9.281811085089773, |
|
"grad_norm": 0.4694254994392395, |
|
"learning_rate": 3.06349786676855e-06, |
|
"loss": 0.27, |
|
"step": 5945 |
|
}, |
|
{ |
|
"epoch": 9.289617486338798, |
|
"grad_norm": 0.4230342507362366, |
|
"learning_rate": 2.996874680545603e-06, |
|
"loss": 0.2645, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 9.297423887587822, |
|
"grad_norm": 0.436117947101593, |
|
"learning_rate": 2.9309729011447573e-06, |
|
"loss": 0.2713, |
|
"step": 5955 |
|
}, |
|
{ |
|
"epoch": 9.305230288836846, |
|
"grad_norm": 0.4295603632926941, |
|
"learning_rate": 2.8657930186738567e-06, |
|
"loss": 0.2666, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 9.31303669008587, |
|
"grad_norm": 0.42392897605895996, |
|
"learning_rate": 2.8013355178720612e-06, |
|
"loss": 0.2682, |
|
"step": 5965 |
|
}, |
|
{ |
|
"epoch": 9.320843091334895, |
|
"grad_norm": 0.40090325474739075, |
|
"learning_rate": 2.7376008781061835e-06, |
|
"loss": 0.2624, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 9.328649492583919, |
|
"grad_norm": 0.4344378113746643, |
|
"learning_rate": 2.674589573367192e-06, |
|
"loss": 0.274, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 9.336455893832943, |
|
"grad_norm": 0.4555279016494751, |
|
"learning_rate": 2.612302072266637e-06, |
|
"loss": 0.2696, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 9.344262295081966, |
|
"grad_norm": 0.4352642297744751, |
|
"learning_rate": 2.5507388380331843e-06, |
|
"loss": 0.2677, |
|
"step": 5985 |
|
}, |
|
{ |
|
"epoch": 9.352068696330992, |
|
"grad_norm": 0.4107036590576172, |
|
"learning_rate": 2.4899003285091536e-06, |
|
"loss": 0.2692, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 9.359875097580016, |
|
"grad_norm": 0.4284498691558838, |
|
"learning_rate": 2.429786996147154e-06, |
|
"loss": 0.2623, |
|
"step": 5995 |
|
}, |
|
{ |
|
"epoch": 9.36768149882904, |
|
"grad_norm": 0.42979347705841064, |
|
"learning_rate": 2.3703992880066638e-06, |
|
"loss": 0.2705, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 9.375487900078063, |
|
"grad_norm": 0.4255581796169281, |
|
"learning_rate": 2.3117376457507667e-06, |
|
"loss": 0.2678, |
|
"step": 6005 |
|
}, |
|
{ |
|
"epoch": 9.383294301327089, |
|
"grad_norm": 0.4213410019874573, |
|
"learning_rate": 2.2538025056428214e-06, |
|
"loss": 0.2679, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 9.391100702576113, |
|
"grad_norm": 0.42378342151641846, |
|
"learning_rate": 2.1965942985431977e-06, |
|
"loss": 0.2603, |
|
"step": 6015 |
|
}, |
|
{ |
|
"epoch": 9.398907103825136, |
|
"grad_norm": 0.43251511454582214, |
|
"learning_rate": 2.140113449906167e-06, |
|
"loss": 0.2678, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 9.40671350507416, |
|
"grad_norm": 0.428505539894104, |
|
"learning_rate": 2.0843603797766287e-06, |
|
"loss": 0.2716, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 9.414519906323186, |
|
"grad_norm": 0.40141308307647705, |
|
"learning_rate": 2.0293355027870554e-06, |
|
"loss": 0.264, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 9.42232630757221, |
|
"grad_norm": 0.42421501874923706, |
|
"learning_rate": 1.975039228154385e-06, |
|
"loss": 0.2675, |
|
"step": 6035 |
|
}, |
|
{ |
|
"epoch": 9.430132708821233, |
|
"grad_norm": 0.41700848937034607, |
|
"learning_rate": 1.921471959676957e-06, |
|
"loss": 0.2663, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 9.437939110070257, |
|
"grad_norm": 0.4247893989086151, |
|
"learning_rate": 1.8686340957315695e-06, |
|
"loss": 0.2671, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 9.445745511319283, |
|
"grad_norm": 0.4123760163784027, |
|
"learning_rate": 1.8165260292704711e-06, |
|
"loss": 0.264, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 9.453551912568306, |
|
"grad_norm": 0.4423754811286926, |
|
"learning_rate": 1.7651481478184296e-06, |
|
"loss": 0.2617, |
|
"step": 6055 |
|
}, |
|
{ |
|
"epoch": 9.46135831381733, |
|
"grad_norm": 0.4133644998073578, |
|
"learning_rate": 1.7145008334698898e-06, |
|
"loss": 0.2624, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 9.469164715066354, |
|
"grad_norm": 0.4270353317260742, |
|
"learning_rate": 1.6645844628860764e-06, |
|
"loss": 0.2674, |
|
"step": 6065 |
|
}, |
|
{ |
|
"epoch": 9.47697111631538, |
|
"grad_norm": 0.41615259647369385, |
|
"learning_rate": 1.6153994072922506e-06, |
|
"loss": 0.2708, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 9.484777517564403, |
|
"grad_norm": 0.4244523346424103, |
|
"learning_rate": 1.5669460324749586e-06, |
|
"loss": 0.2698, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 9.492583918813427, |
|
"grad_norm": 0.4362230598926544, |
|
"learning_rate": 1.5192246987791981e-06, |
|
"loss": 0.2617, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 9.50039032006245, |
|
"grad_norm": 0.4365394413471222, |
|
"learning_rate": 1.472235761105878e-06, |
|
"loss": 0.2669, |
|
"step": 6085 |
|
}, |
|
{ |
|
"epoch": 9.508196721311476, |
|
"grad_norm": 0.4158194363117218, |
|
"learning_rate": 1.4259795689090972e-06, |
|
"loss": 0.2694, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 9.5160031225605, |
|
"grad_norm": 0.41145652532577515, |
|
"learning_rate": 1.3804564661935915e-06, |
|
"loss": 0.2629, |
|
"step": 6095 |
|
}, |
|
{ |
|
"epoch": 9.523809523809524, |
|
"grad_norm": 0.39693784713745117, |
|
"learning_rate": 1.3356667915121025e-06, |
|
"loss": 0.2683, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 9.531615925058547, |
|
"grad_norm": 0.4337891638278961, |
|
"learning_rate": 1.2916108779629566e-06, |
|
"loss": 0.2691, |
|
"step": 6105 |
|
}, |
|
{ |
|
"epoch": 9.539422326307573, |
|
"grad_norm": 0.431761234998703, |
|
"learning_rate": 1.2482890531875125e-06, |
|
"loss": 0.2704, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 9.547228727556597, |
|
"grad_norm": 0.42788565158843994, |
|
"learning_rate": 1.2057016393677623e-06, |
|
"loss": 0.2721, |
|
"step": 6115 |
|
}, |
|
{ |
|
"epoch": 9.55503512880562, |
|
"grad_norm": 0.4381605088710785, |
|
"learning_rate": 1.163848953223934e-06, |
|
"loss": 0.2731, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 9.562841530054644, |
|
"grad_norm": 0.42890533804893494, |
|
"learning_rate": 1.1227313060120926e-06, |
|
"loss": 0.2683, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 9.570647931303668, |
|
"grad_norm": 0.41614386439323425, |
|
"learning_rate": 1.0823490035218987e-06, |
|
"loss": 0.2634, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 9.578454332552694, |
|
"grad_norm": 0.41682758927345276, |
|
"learning_rate": 1.042702346074287e-06, |
|
"loss": 0.2621, |
|
"step": 6135 |
|
}, |
|
{ |
|
"epoch": 9.586260733801717, |
|
"grad_norm": 0.4195772707462311, |
|
"learning_rate": 1.003791628519213e-06, |
|
"loss": 0.2677, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 9.594067135050741, |
|
"grad_norm": 0.40635907649993896, |
|
"learning_rate": 9.656171402335213e-07, |
|
"loss": 0.2668, |
|
"step": 6145 |
|
}, |
|
{ |
|
"epoch": 9.601873536299767, |
|
"grad_norm": 0.40993356704711914, |
|
"learning_rate": 9.281791651187366e-07, |
|
"loss": 0.2661, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 9.60967993754879, |
|
"grad_norm": 0.4103589355945587, |
|
"learning_rate": 8.914779815989982e-07, |
|
"loss": 0.2691, |
|
"step": 6155 |
|
}, |
|
{ |
|
"epoch": 9.617486338797814, |
|
"grad_norm": 0.44015854597091675, |
|
"learning_rate": 8.555138626189618e-07, |
|
"loss": 0.2649, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 9.625292740046838, |
|
"grad_norm": 0.414385586977005, |
|
"learning_rate": 8.202870756417569e-07, |
|
"loss": 0.2616, |
|
"step": 6165 |
|
}, |
|
{ |
|
"epoch": 9.633099141295862, |
|
"grad_norm": 0.4170667827129364, |
|
"learning_rate": 7.857978826470325e-07, |
|
"loss": 0.2658, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 9.640905542544887, |
|
"grad_norm": 0.42618632316589355, |
|
"learning_rate": 7.520465401290033e-07, |
|
"loss": 0.2691, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 9.648711943793911, |
|
"grad_norm": 0.41920483112335205, |
|
"learning_rate": 7.19033299094496e-07, |
|
"loss": 0.2639, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 9.656518345042935, |
|
"grad_norm": 0.4324471056461334, |
|
"learning_rate": 6.867584050611498e-07, |
|
"loss": 0.2715, |
|
"step": 6185 |
|
}, |
|
{ |
|
"epoch": 9.664324746291959, |
|
"grad_norm": 0.4518769681453705, |
|
"learning_rate": 6.552220980555635e-07, |
|
"loss": 0.2749, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 9.672131147540984, |
|
"grad_norm": 0.4381062984466553, |
|
"learning_rate": 6.244246126114627e-07, |
|
"loss": 0.2734, |
|
"step": 6195 |
|
}, |
|
{ |
|
"epoch": 9.679937548790008, |
|
"grad_norm": 0.4453407824039459, |
|
"learning_rate": 5.943661777680354e-07, |
|
"loss": 0.2692, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 9.687743950039032, |
|
"grad_norm": 0.41984906792640686, |
|
"learning_rate": 5.650470170681876e-07, |
|
"loss": 0.2644, |
|
"step": 6205 |
|
}, |
|
{ |
|
"epoch": 9.695550351288055, |
|
"grad_norm": 0.4198761284351349, |
|
"learning_rate": 5.364673485568794e-07, |
|
"loss": 0.2676, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 9.703356752537081, |
|
"grad_norm": 0.42462319135665894, |
|
"learning_rate": 5.086273847795031e-07, |
|
"loss": 0.2671, |
|
"step": 6215 |
|
}, |
|
{ |
|
"epoch": 9.711163153786105, |
|
"grad_norm": 0.4393913447856903, |
|
"learning_rate": 4.815273327803182e-07, |
|
"loss": 0.2689, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 9.718969555035128, |
|
"grad_norm": 0.4282847046852112, |
|
"learning_rate": 4.5516739410087494e-07, |
|
"loss": 0.2696, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 9.726775956284152, |
|
"grad_norm": 0.41957032680511475, |
|
"learning_rate": 4.2954776477860393e-07, |
|
"loss": 0.2748, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 9.734582357533178, |
|
"grad_norm": 0.4238908886909485, |
|
"learning_rate": 4.0466863534522893e-07, |
|
"loss": 0.2741, |
|
"step": 6235 |
|
}, |
|
{ |
|
"epoch": 9.742388758782202, |
|
"grad_norm": 0.4146591126918793, |
|
"learning_rate": 3.805301908254455e-07, |
|
"loss": 0.2653, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 9.750195160031225, |
|
"grad_norm": 0.4316459000110626, |
|
"learning_rate": 3.571326107355333e-07, |
|
"loss": 0.2687, |
|
"step": 6245 |
|
}, |
|
{ |
|
"epoch": 9.758001561280249, |
|
"grad_norm": 0.4224886894226074, |
|
"learning_rate": 3.3447606908196817e-07, |
|
"loss": 0.2646, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 9.765807962529275, |
|
"grad_norm": 0.439411461353302, |
|
"learning_rate": 3.125607343602011e-07, |
|
"loss": 0.2657, |
|
"step": 6255 |
|
}, |
|
{ |
|
"epoch": 9.773614363778298, |
|
"grad_norm": 0.4261423647403717, |
|
"learning_rate": 2.9138676955333676e-07, |
|
"loss": 0.2606, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 9.781420765027322, |
|
"grad_norm": 0.41440731287002563, |
|
"learning_rate": 2.709543321309793e-07, |
|
"loss": 0.2673, |
|
"step": 6265 |
|
}, |
|
{ |
|
"epoch": 9.789227166276346, |
|
"grad_norm": 0.42373111844062805, |
|
"learning_rate": 2.5126357404802183e-07, |
|
"loss": 0.2751, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 9.797033567525371, |
|
"grad_norm": 0.4395394027233124, |
|
"learning_rate": 2.3231464174352512e-07, |
|
"loss": 0.2747, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 9.804839968774395, |
|
"grad_norm": 0.4383135735988617, |
|
"learning_rate": 2.141076761396521e-07, |
|
"loss": 0.2708, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 9.812646370023419, |
|
"grad_norm": 0.4313544034957886, |
|
"learning_rate": 1.966428126405795e-07, |
|
"loss": 0.2686, |
|
"step": 6285 |
|
}, |
|
{ |
|
"epoch": 9.820452771272443, |
|
"grad_norm": 0.39951956272125244, |
|
"learning_rate": 1.7992018113152098e-07, |
|
"loss": 0.2591, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 9.828259172521468, |
|
"grad_norm": 0.4347188174724579, |
|
"learning_rate": 1.6393990597775022e-07, |
|
"loss": 0.2709, |
|
"step": 6295 |
|
}, |
|
{ |
|
"epoch": 9.836065573770492, |
|
"grad_norm": 0.4294663965702057, |
|
"learning_rate": 1.487021060236904e-07, |
|
"loss": 0.2644, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 9.843871975019516, |
|
"grad_norm": 0.437920480966568, |
|
"learning_rate": 1.3420689459200386e-07, |
|
"loss": 0.2693, |
|
"step": 6305 |
|
}, |
|
{ |
|
"epoch": 9.85167837626854, |
|
"grad_norm": 0.4170249104499817, |
|
"learning_rate": 1.204543794827595e-07, |
|
"loss": 0.2748, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 9.859484777517565, |
|
"grad_norm": 0.40519675612449646, |
|
"learning_rate": 1.0744466297265554e-07, |
|
"loss": 0.267, |
|
"step": 6315 |
|
}, |
|
{ |
|
"epoch": 9.867291178766589, |
|
"grad_norm": 0.40046894550323486, |
|
"learning_rate": 9.517784181422019e-08, |
|
"loss": 0.2614, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 9.875097580015613, |
|
"grad_norm": 0.4310961067676544, |
|
"learning_rate": 8.365400723512328e-08, |
|
"loss": 0.2617, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 9.882903981264636, |
|
"grad_norm": 0.43310293555259705, |
|
"learning_rate": 7.287324493747693e-08, |
|
"loss": 0.2685, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 9.890710382513662, |
|
"grad_norm": 0.43305081129074097, |
|
"learning_rate": 6.283563509719148e-08, |
|
"loss": 0.2607, |
|
"step": 6335 |
|
}, |
|
{ |
|
"epoch": 9.898516783762686, |
|
"grad_norm": 0.42313021421432495, |
|
"learning_rate": 5.354125236343155e-08, |
|
"loss": 0.2714, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 9.90632318501171, |
|
"grad_norm": 0.43003058433532715, |
|
"learning_rate": 4.499016585799431e-08, |
|
"loss": 0.2685, |
|
"step": 6345 |
|
}, |
|
{ |
|
"epoch": 9.914129586260733, |
|
"grad_norm": 0.4060039818286896, |
|
"learning_rate": 3.7182439174832106e-08, |
|
"loss": 0.2728, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 9.921935987509759, |
|
"grad_norm": 0.4126955270767212, |
|
"learning_rate": 3.0118130379575005e-08, |
|
"loss": 0.2675, |
|
"step": 6355 |
|
}, |
|
{ |
|
"epoch": 9.929742388758783, |
|
"grad_norm": 0.43146437406539917, |
|
"learning_rate": 2.379729200908676e-08, |
|
"loss": 0.2695, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 9.937548790007806, |
|
"grad_norm": 0.4640655219554901, |
|
"learning_rate": 1.8219971071098408e-08, |
|
"loss": 0.2713, |
|
"step": 6365 |
|
}, |
|
{ |
|
"epoch": 9.94535519125683, |
|
"grad_norm": 0.41949042677879333, |
|
"learning_rate": 1.3386209043819708e-08, |
|
"loss": 0.2707, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 9.953161592505854, |
|
"grad_norm": 0.42296043038368225, |
|
"learning_rate": 9.296041875683781e-09, |
|
"loss": 0.2636, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 9.96096799375488, |
|
"grad_norm": 0.45912352204322815, |
|
"learning_rate": 5.949499985025142e-09, |
|
"loss": 0.2702, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 9.968774395003903, |
|
"grad_norm": 0.4141922891139984, |
|
"learning_rate": 3.346608259890971e-09, |
|
"loss": 0.2663, |
|
"step": 6385 |
|
}, |
|
{ |
|
"epoch": 9.976580796252927, |
|
"grad_norm": 0.42737066745758057, |
|
"learning_rate": 1.4873860578412668e-09, |
|
"loss": 0.2692, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 9.984387197501952, |
|
"grad_norm": 0.441078782081604, |
|
"learning_rate": 3.7184720581562306e-10, |
|
"loss": 0.2707, |
|
"step": 6395 |
|
}, |
|
{ |
|
"epoch": 9.992193598750976, |
|
"grad_norm": 0.45411616563796997, |
|
"learning_rate": 0.0, |
|
"loss": 0.2714, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 9.992193598750976, |
|
"eval_loss": 1.1572602987289429, |
|
"eval_runtime": 0.3457, |
|
"eval_samples_per_second": 14.464, |
|
"eval_steps_per_second": 2.893, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 9.992193598750976, |
|
"step": 6400, |
|
"total_flos": 8.968401612833817e+18, |
|
"train_loss": 0.34100163986906407, |
|
"train_runtime": 19497.7947, |
|
"train_samples_per_second": 10.509, |
|
"train_steps_per_second": 0.328 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 6400, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 8.968401612833817e+18, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|