XeTute's picture
Upload 14 files
ac38b26
raw
history blame
134 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.04202894743754761,
"eval_steps": 200,
"global_step": 600,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 7.004824572924602e-05,
"grad_norm": 6.222772121429443,
"learning_rate": 9.99930017513135e-05,
"loss": 1.1076,
"num_input_tokens_seen": 16384,
"step": 1
},
{
"epoch": 0.00014009649145849205,
"grad_norm": 6.042057037353516,
"learning_rate": 9.998600350262697e-05,
"loss": 1.1086,
"num_input_tokens_seen": 32768,
"step": 2
},
{
"epoch": 0.00021014473718773804,
"grad_norm": 7.119229316711426,
"learning_rate": 9.997900525394046e-05,
"loss": 1.4047,
"num_input_tokens_seen": 49152,
"step": 3
},
{
"epoch": 0.0002801929829169841,
"grad_norm": 7.133191108703613,
"learning_rate": 9.997200700525395e-05,
"loss": 1.3921,
"num_input_tokens_seen": 65536,
"step": 4
},
{
"epoch": 0.0003502412286462301,
"grad_norm": 6.1078338623046875,
"learning_rate": 9.996500875656743e-05,
"loss": 1.3171,
"num_input_tokens_seen": 81920,
"step": 5
},
{
"epoch": 0.0004202894743754761,
"grad_norm": 6.466420650482178,
"learning_rate": 9.995801050788092e-05,
"loss": 1.0732,
"num_input_tokens_seen": 97344,
"step": 6
},
{
"epoch": 0.0004903377201047221,
"grad_norm": 5.578189849853516,
"learning_rate": 9.99510122591944e-05,
"loss": 0.9929,
"num_input_tokens_seen": 113728,
"step": 7
},
{
"epoch": 0.0005603859658339682,
"grad_norm": 7.197720527648926,
"learning_rate": 9.994401401050789e-05,
"loss": 1.2512,
"num_input_tokens_seen": 129528,
"step": 8
},
{
"epoch": 0.0006304342115632141,
"grad_norm": 6.618913650512695,
"learning_rate": 9.993701576182136e-05,
"loss": 1.3495,
"num_input_tokens_seen": 145704,
"step": 9
},
{
"epoch": 0.0007004824572924602,
"grad_norm": 6.955508232116699,
"learning_rate": 9.993001751313485e-05,
"loss": 1.1823,
"num_input_tokens_seen": 161664,
"step": 10
},
{
"epoch": 0.0007705307030217062,
"grad_norm": 6.6807074546813965,
"learning_rate": 9.992301926444835e-05,
"loss": 1.1693,
"num_input_tokens_seen": 177960,
"step": 11
},
{
"epoch": 0.0008405789487509522,
"grad_norm": 6.784447193145752,
"learning_rate": 9.991602101576183e-05,
"loss": 1.3744,
"num_input_tokens_seen": 194344,
"step": 12
},
{
"epoch": 0.0009106271944801982,
"grad_norm": 6.7418437004089355,
"learning_rate": 9.990902276707532e-05,
"loss": 1.22,
"num_input_tokens_seen": 210728,
"step": 13
},
{
"epoch": 0.0009806754402094443,
"grad_norm": 6.43395471572876,
"learning_rate": 9.990202451838879e-05,
"loss": 1.1772,
"num_input_tokens_seen": 227112,
"step": 14
},
{
"epoch": 0.0010507236859386903,
"grad_norm": 6.09422492980957,
"learning_rate": 9.989502626970228e-05,
"loss": 1.195,
"num_input_tokens_seen": 243496,
"step": 15
},
{
"epoch": 0.0011207719316679364,
"grad_norm": 6.238271236419678,
"learning_rate": 9.988802802101577e-05,
"loss": 1.2623,
"num_input_tokens_seen": 259744,
"step": 16
},
{
"epoch": 0.0011908201773971822,
"grad_norm": 6.56187629699707,
"learning_rate": 9.988102977232926e-05,
"loss": 1.2721,
"num_input_tokens_seen": 276128,
"step": 17
},
{
"epoch": 0.0012608684231264283,
"grad_norm": 6.818358898162842,
"learning_rate": 9.987403152364275e-05,
"loss": 1.2649,
"num_input_tokens_seen": 292512,
"step": 18
},
{
"epoch": 0.0013309166688556743,
"grad_norm": 5.950352191925049,
"learning_rate": 9.986703327495622e-05,
"loss": 1.0024,
"num_input_tokens_seen": 308632,
"step": 19
},
{
"epoch": 0.0014009649145849204,
"grad_norm": 6.387479305267334,
"learning_rate": 9.986003502626971e-05,
"loss": 1.2783,
"num_input_tokens_seen": 325016,
"step": 20
},
{
"epoch": 0.0014710131603141664,
"grad_norm": 6.187346458435059,
"learning_rate": 9.985303677758318e-05,
"loss": 1.1701,
"num_input_tokens_seen": 341384,
"step": 21
},
{
"epoch": 0.0015410614060434125,
"grad_norm": 5.371951103210449,
"learning_rate": 9.984603852889667e-05,
"loss": 1.0483,
"num_input_tokens_seen": 357768,
"step": 22
},
{
"epoch": 0.0016111096517726585,
"grad_norm": 6.2206807136535645,
"learning_rate": 9.983904028021016e-05,
"loss": 1.2516,
"num_input_tokens_seen": 374152,
"step": 23
},
{
"epoch": 0.0016811578975019044,
"grad_norm": 6.121264457702637,
"learning_rate": 9.983204203152365e-05,
"loss": 1.1506,
"num_input_tokens_seen": 390536,
"step": 24
},
{
"epoch": 0.0017512061432311504,
"grad_norm": 6.353756904602051,
"learning_rate": 9.982504378283714e-05,
"loss": 1.3118,
"num_input_tokens_seen": 406920,
"step": 25
},
{
"epoch": 0.0018212543889603965,
"grad_norm": 6.270686149597168,
"learning_rate": 9.981804553415061e-05,
"loss": 1.0883,
"num_input_tokens_seen": 422728,
"step": 26
},
{
"epoch": 0.0018913026346896425,
"grad_norm": 6.117632865905762,
"learning_rate": 9.98110472854641e-05,
"loss": 1.3346,
"num_input_tokens_seen": 439112,
"step": 27
},
{
"epoch": 0.0019613508804188886,
"grad_norm": 6.429015159606934,
"learning_rate": 9.980404903677759e-05,
"loss": 1.2494,
"num_input_tokens_seen": 455144,
"step": 28
},
{
"epoch": 0.0020313991261481346,
"grad_norm": 6.4467620849609375,
"learning_rate": 9.979705078809107e-05,
"loss": 1.3335,
"num_input_tokens_seen": 470360,
"step": 29
},
{
"epoch": 0.0021014473718773807,
"grad_norm": 6.57926082611084,
"learning_rate": 9.979005253940455e-05,
"loss": 1.2126,
"num_input_tokens_seen": 486120,
"step": 30
},
{
"epoch": 0.0021714956176066267,
"grad_norm": 5.650569915771484,
"learning_rate": 9.978305429071804e-05,
"loss": 1.1363,
"num_input_tokens_seen": 501896,
"step": 31
},
{
"epoch": 0.0022415438633358728,
"grad_norm": 6.380292892456055,
"learning_rate": 9.977605604203153e-05,
"loss": 1.2251,
"num_input_tokens_seen": 517752,
"step": 32
},
{
"epoch": 0.002311592109065119,
"grad_norm": 5.704173564910889,
"learning_rate": 9.976905779334502e-05,
"loss": 1.1685,
"num_input_tokens_seen": 534136,
"step": 33
},
{
"epoch": 0.0023816403547943644,
"grad_norm": 5.342978000640869,
"learning_rate": 9.97620595446585e-05,
"loss": 1.2012,
"num_input_tokens_seen": 550216,
"step": 34
},
{
"epoch": 0.0024516886005236105,
"grad_norm": 5.7014241218566895,
"learning_rate": 9.975506129597198e-05,
"loss": 1.2342,
"num_input_tokens_seen": 566600,
"step": 35
},
{
"epoch": 0.0025217368462528565,
"grad_norm": 6.26229190826416,
"learning_rate": 9.974806304728546e-05,
"loss": 1.2041,
"num_input_tokens_seen": 582984,
"step": 36
},
{
"epoch": 0.0025917850919821026,
"grad_norm": 6.583463191986084,
"learning_rate": 9.974106479859896e-05,
"loss": 1.3021,
"num_input_tokens_seen": 598968,
"step": 37
},
{
"epoch": 0.0026618333377113486,
"grad_norm": 5.58498477935791,
"learning_rate": 9.973406654991245e-05,
"loss": 1.1622,
"num_input_tokens_seen": 614840,
"step": 38
},
{
"epoch": 0.0027318815834405947,
"grad_norm": 5.906906604766846,
"learning_rate": 9.972706830122592e-05,
"loss": 1.1971,
"num_input_tokens_seen": 631224,
"step": 39
},
{
"epoch": 0.0028019298291698407,
"grad_norm": 5.962359428405762,
"learning_rate": 9.972007005253941e-05,
"loss": 1.1326,
"num_input_tokens_seen": 647000,
"step": 40
},
{
"epoch": 0.002871978074899087,
"grad_norm": 6.447500705718994,
"learning_rate": 9.971307180385289e-05,
"loss": 1.0905,
"num_input_tokens_seen": 662480,
"step": 41
},
{
"epoch": 0.002942026320628333,
"grad_norm": 5.7290520668029785,
"learning_rate": 9.970607355516638e-05,
"loss": 1.3585,
"num_input_tokens_seen": 678480,
"step": 42
},
{
"epoch": 0.003012074566357579,
"grad_norm": 6.063445568084717,
"learning_rate": 9.969907530647987e-05,
"loss": 1.2841,
"num_input_tokens_seen": 694256,
"step": 43
},
{
"epoch": 0.003082122812086825,
"grad_norm": 5.302809238433838,
"learning_rate": 9.969207705779335e-05,
"loss": 1.1168,
"num_input_tokens_seen": 710152,
"step": 44
},
{
"epoch": 0.003152171057816071,
"grad_norm": 5.634128093719482,
"learning_rate": 9.968507880910684e-05,
"loss": 1.0609,
"num_input_tokens_seen": 726184,
"step": 45
},
{
"epoch": 0.003222219303545317,
"grad_norm": 5.652642726898193,
"learning_rate": 9.967808056042032e-05,
"loss": 1.2228,
"num_input_tokens_seen": 742520,
"step": 46
},
{
"epoch": 0.0032922675492745627,
"grad_norm": 5.340751647949219,
"learning_rate": 9.96710823117338e-05,
"loss": 1.0595,
"num_input_tokens_seen": 758904,
"step": 47
},
{
"epoch": 0.0033623157950038087,
"grad_norm": 5.422239780426025,
"learning_rate": 9.966408406304728e-05,
"loss": 1.1161,
"num_input_tokens_seen": 775040,
"step": 48
},
{
"epoch": 0.0034323640407330548,
"grad_norm": 5.29241418838501,
"learning_rate": 9.965708581436077e-05,
"loss": 1.0255,
"num_input_tokens_seen": 790856,
"step": 49
},
{
"epoch": 0.003502412286462301,
"grad_norm": 5.146270275115967,
"learning_rate": 9.965008756567426e-05,
"loss": 0.9762,
"num_input_tokens_seen": 807064,
"step": 50
},
{
"epoch": 0.003572460532191547,
"grad_norm": 5.825758457183838,
"learning_rate": 9.964308931698775e-05,
"loss": 1.2108,
"num_input_tokens_seen": 823448,
"step": 51
},
{
"epoch": 0.003642508777920793,
"grad_norm": 6.179538726806641,
"learning_rate": 9.963609106830124e-05,
"loss": 1.322,
"num_input_tokens_seen": 838888,
"step": 52
},
{
"epoch": 0.003712557023650039,
"grad_norm": 6.464454174041748,
"learning_rate": 9.962909281961471e-05,
"loss": 1.5077,
"num_input_tokens_seen": 855272,
"step": 53
},
{
"epoch": 0.003782605269379285,
"grad_norm": 5.4227294921875,
"learning_rate": 9.96220945709282e-05,
"loss": 1.2679,
"num_input_tokens_seen": 871656,
"step": 54
},
{
"epoch": 0.003852653515108531,
"grad_norm": 5.949041366577148,
"learning_rate": 9.961509632224169e-05,
"loss": 1.3618,
"num_input_tokens_seen": 888040,
"step": 55
},
{
"epoch": 0.003922701760837777,
"grad_norm": 6.050904750823975,
"learning_rate": 9.960809807355516e-05,
"loss": 1.3155,
"num_input_tokens_seen": 904400,
"step": 56
},
{
"epoch": 0.003992750006567023,
"grad_norm": 6.048308849334717,
"learning_rate": 9.960109982486866e-05,
"loss": 1.3131,
"num_input_tokens_seen": 919952,
"step": 57
},
{
"epoch": 0.004062798252296269,
"grad_norm": 5.683863162994385,
"learning_rate": 9.959410157618214e-05,
"loss": 1.1692,
"num_input_tokens_seen": 936336,
"step": 58
},
{
"epoch": 0.004132846498025515,
"grad_norm": 5.449287414550781,
"learning_rate": 9.958710332749563e-05,
"loss": 1.0613,
"num_input_tokens_seen": 952152,
"step": 59
},
{
"epoch": 0.004202894743754761,
"grad_norm": 5.31496524810791,
"learning_rate": 9.958010507880912e-05,
"loss": 0.9605,
"num_input_tokens_seen": 967824,
"step": 60
},
{
"epoch": 0.004272942989484007,
"grad_norm": 5.57105016708374,
"learning_rate": 9.957310683012259e-05,
"loss": 1.1701,
"num_input_tokens_seen": 983864,
"step": 61
},
{
"epoch": 0.004342991235213253,
"grad_norm": 5.3456830978393555,
"learning_rate": 9.956610858143608e-05,
"loss": 1.0995,
"num_input_tokens_seen": 1000248,
"step": 62
},
{
"epoch": 0.004413039480942499,
"grad_norm": 5.453295707702637,
"learning_rate": 9.955911033274957e-05,
"loss": 1.2413,
"num_input_tokens_seen": 1016632,
"step": 63
},
{
"epoch": 0.0044830877266717455,
"grad_norm": 4.975449562072754,
"learning_rate": 9.955211208406306e-05,
"loss": 1.0961,
"num_input_tokens_seen": 1033016,
"step": 64
},
{
"epoch": 0.004553135972400991,
"grad_norm": 5.542137145996094,
"learning_rate": 9.954511383537655e-05,
"loss": 1.1171,
"num_input_tokens_seen": 1049400,
"step": 65
},
{
"epoch": 0.004623184218130238,
"grad_norm": 5.213950157165527,
"learning_rate": 9.953811558669002e-05,
"loss": 1.2228,
"num_input_tokens_seen": 1065784,
"step": 66
},
{
"epoch": 0.004693232463859483,
"grad_norm": 5.496099948883057,
"learning_rate": 9.953111733800351e-05,
"loss": 1.1529,
"num_input_tokens_seen": 1082168,
"step": 67
},
{
"epoch": 0.004763280709588729,
"grad_norm": 5.64145565032959,
"learning_rate": 9.952411908931698e-05,
"loss": 1.2301,
"num_input_tokens_seen": 1098024,
"step": 68
},
{
"epoch": 0.004833328955317975,
"grad_norm": 5.566709995269775,
"learning_rate": 9.951712084063047e-05,
"loss": 1.2679,
"num_input_tokens_seen": 1114408,
"step": 69
},
{
"epoch": 0.004903377201047221,
"grad_norm": 6.443673133850098,
"learning_rate": 9.951012259194396e-05,
"loss": 1.2313,
"num_input_tokens_seen": 1130792,
"step": 70
},
{
"epoch": 0.0049734254467764675,
"grad_norm": 5.882962226867676,
"learning_rate": 9.950312434325745e-05,
"loss": 1.4304,
"num_input_tokens_seen": 1147176,
"step": 71
},
{
"epoch": 0.005043473692505713,
"grad_norm": 6.0052666664123535,
"learning_rate": 9.949612609457094e-05,
"loss": 1.3027,
"num_input_tokens_seen": 1160968,
"step": 72
},
{
"epoch": 0.0051135219382349596,
"grad_norm": 5.260256767272949,
"learning_rate": 9.948912784588441e-05,
"loss": 1.1526,
"num_input_tokens_seen": 1177352,
"step": 73
},
{
"epoch": 0.005183570183964205,
"grad_norm": 5.641814708709717,
"learning_rate": 9.94821295971979e-05,
"loss": 1.0666,
"num_input_tokens_seen": 1193032,
"step": 74
},
{
"epoch": 0.005253618429693452,
"grad_norm": 5.121115207672119,
"learning_rate": 9.947513134851138e-05,
"loss": 1.2404,
"num_input_tokens_seen": 1208952,
"step": 75
},
{
"epoch": 0.005323666675422697,
"grad_norm": 5.63930082321167,
"learning_rate": 9.946813309982487e-05,
"loss": 1.5127,
"num_input_tokens_seen": 1225000,
"step": 76
},
{
"epoch": 0.005393714921151944,
"grad_norm": 4.880716800689697,
"learning_rate": 9.946113485113837e-05,
"loss": 1.1484,
"num_input_tokens_seen": 1241384,
"step": 77
},
{
"epoch": 0.005463763166881189,
"grad_norm": 5.59611177444458,
"learning_rate": 9.945413660245184e-05,
"loss": 1.1678,
"num_input_tokens_seen": 1257680,
"step": 78
},
{
"epoch": 0.005533811412610436,
"grad_norm": 5.052026271820068,
"learning_rate": 9.944713835376533e-05,
"loss": 1.2207,
"num_input_tokens_seen": 1274064,
"step": 79
},
{
"epoch": 0.0056038596583396815,
"grad_norm": 5.285096168518066,
"learning_rate": 9.944014010507881e-05,
"loss": 1.1457,
"num_input_tokens_seen": 1290448,
"step": 80
},
{
"epoch": 0.005673907904068927,
"grad_norm": 5.4286580085754395,
"learning_rate": 9.94331418563923e-05,
"loss": 1.3047,
"num_input_tokens_seen": 1306832,
"step": 81
},
{
"epoch": 0.005743956149798174,
"grad_norm": 5.937953472137451,
"learning_rate": 9.942614360770578e-05,
"loss": 1.4353,
"num_input_tokens_seen": 1323216,
"step": 82
},
{
"epoch": 0.005814004395527419,
"grad_norm": 5.129006385803223,
"learning_rate": 9.941914535901927e-05,
"loss": 1.1434,
"num_input_tokens_seen": 1339408,
"step": 83
},
{
"epoch": 0.005884052641256666,
"grad_norm": 5.179675102233887,
"learning_rate": 9.941214711033276e-05,
"loss": 1.2452,
"num_input_tokens_seen": 1355792,
"step": 84
},
{
"epoch": 0.005954100886985911,
"grad_norm": 4.912832736968994,
"learning_rate": 9.940514886164624e-05,
"loss": 1.1255,
"num_input_tokens_seen": 1372176,
"step": 85
},
{
"epoch": 0.006024149132715158,
"grad_norm": 5.190899848937988,
"learning_rate": 9.939815061295973e-05,
"loss": 1.2543,
"num_input_tokens_seen": 1388560,
"step": 86
},
{
"epoch": 0.006094197378444403,
"grad_norm": 5.1751275062561035,
"learning_rate": 9.939115236427321e-05,
"loss": 1.3145,
"num_input_tokens_seen": 1404944,
"step": 87
},
{
"epoch": 0.00616424562417365,
"grad_norm": 5.450705528259277,
"learning_rate": 9.938415411558669e-05,
"loss": 1.2844,
"num_input_tokens_seen": 1421328,
"step": 88
},
{
"epoch": 0.0062342938699028955,
"grad_norm": 5.593935012817383,
"learning_rate": 9.937715586690018e-05,
"loss": 1.3284,
"num_input_tokens_seen": 1437464,
"step": 89
},
{
"epoch": 0.006304342115632142,
"grad_norm": 5.156428813934326,
"learning_rate": 9.937015761821367e-05,
"loss": 1.1682,
"num_input_tokens_seen": 1452952,
"step": 90
},
{
"epoch": 0.006374390361361388,
"grad_norm": 4.673638820648193,
"learning_rate": 9.936315936952715e-05,
"loss": 1.004,
"num_input_tokens_seen": 1469336,
"step": 91
},
{
"epoch": 0.006444438607090634,
"grad_norm": 4.996700763702393,
"learning_rate": 9.935616112084064e-05,
"loss": 1.087,
"num_input_tokens_seen": 1485448,
"step": 92
},
{
"epoch": 0.00651448685281988,
"grad_norm": 4.817474365234375,
"learning_rate": 9.934916287215412e-05,
"loss": 1.151,
"num_input_tokens_seen": 1501472,
"step": 93
},
{
"epoch": 0.006584535098549125,
"grad_norm": 5.400479316711426,
"learning_rate": 9.934216462346761e-05,
"loss": 1.3144,
"num_input_tokens_seen": 1516424,
"step": 94
},
{
"epoch": 0.006654583344278372,
"grad_norm": 5.232216835021973,
"learning_rate": 9.933516637478108e-05,
"loss": 1.0019,
"num_input_tokens_seen": 1532792,
"step": 95
},
{
"epoch": 0.006724631590007617,
"grad_norm": 5.392521381378174,
"learning_rate": 9.932816812609457e-05,
"loss": 1.3195,
"num_input_tokens_seen": 1548600,
"step": 96
},
{
"epoch": 0.006794679835736864,
"grad_norm": 5.5280866622924805,
"learning_rate": 9.932116987740806e-05,
"loss": 1.283,
"num_input_tokens_seen": 1564088,
"step": 97
},
{
"epoch": 0.0068647280814661095,
"grad_norm": 4.963179588317871,
"learning_rate": 9.931417162872155e-05,
"loss": 1.2716,
"num_input_tokens_seen": 1580040,
"step": 98
},
{
"epoch": 0.006934776327195356,
"grad_norm": 4.920302391052246,
"learning_rate": 9.930717338003504e-05,
"loss": 1.088,
"num_input_tokens_seen": 1595880,
"step": 99
},
{
"epoch": 0.007004824572924602,
"grad_norm": 4.935486793518066,
"learning_rate": 9.930017513134851e-05,
"loss": 1.0122,
"num_input_tokens_seen": 1611864,
"step": 100
},
{
"epoch": 0.007074872818653848,
"grad_norm": 5.099087238311768,
"learning_rate": 9.9293176882662e-05,
"loss": 1.1605,
"num_input_tokens_seen": 1627472,
"step": 101
},
{
"epoch": 0.007144921064383094,
"grad_norm": 5.3764328956604,
"learning_rate": 9.928617863397548e-05,
"loss": 1.2225,
"num_input_tokens_seen": 1643856,
"step": 102
},
{
"epoch": 0.00721496931011234,
"grad_norm": 5.281564712524414,
"learning_rate": 9.927918038528898e-05,
"loss": 1.1483,
"num_input_tokens_seen": 1660240,
"step": 103
},
{
"epoch": 0.007285017555841586,
"grad_norm": 5.395167827606201,
"learning_rate": 9.927218213660247e-05,
"loss": 1.6014,
"num_input_tokens_seen": 1676624,
"step": 104
},
{
"epoch": 0.007355065801570832,
"grad_norm": 5.322319507598877,
"learning_rate": 9.926518388791594e-05,
"loss": 1.0933,
"num_input_tokens_seen": 1693008,
"step": 105
},
{
"epoch": 0.007425114047300078,
"grad_norm": 5.301229953765869,
"learning_rate": 9.925818563922943e-05,
"loss": 1.1998,
"num_input_tokens_seen": 1708424,
"step": 106
},
{
"epoch": 0.0074951622930293236,
"grad_norm": 4.958597183227539,
"learning_rate": 9.92511873905429e-05,
"loss": 1.3285,
"num_input_tokens_seen": 1724808,
"step": 107
},
{
"epoch": 0.00756521053875857,
"grad_norm": 4.3913960456848145,
"learning_rate": 9.924418914185639e-05,
"loss": 0.9017,
"num_input_tokens_seen": 1740752,
"step": 108
},
{
"epoch": 0.007635258784487816,
"grad_norm": 5.401021480560303,
"learning_rate": 9.923719089316988e-05,
"loss": 1.3646,
"num_input_tokens_seen": 1755176,
"step": 109
},
{
"epoch": 0.007705307030217062,
"grad_norm": 4.894444942474365,
"learning_rate": 9.923019264448337e-05,
"loss": 0.9955,
"num_input_tokens_seen": 1771560,
"step": 110
},
{
"epoch": 0.007775355275946308,
"grad_norm": 4.878688335418701,
"learning_rate": 9.922319439579686e-05,
"loss": 1.1766,
"num_input_tokens_seen": 1787944,
"step": 111
},
{
"epoch": 0.007845403521675554,
"grad_norm": 4.9379777908325195,
"learning_rate": 9.921619614711033e-05,
"loss": 1.1631,
"num_input_tokens_seen": 1803568,
"step": 112
},
{
"epoch": 0.0079154517674048,
"grad_norm": 5.101811408996582,
"learning_rate": 9.920919789842382e-05,
"loss": 1.2165,
"num_input_tokens_seen": 1819952,
"step": 113
},
{
"epoch": 0.007985500013134045,
"grad_norm": 5.32574987411499,
"learning_rate": 9.920219964973731e-05,
"loss": 1.3012,
"num_input_tokens_seen": 1835296,
"step": 114
},
{
"epoch": 0.008055548258863293,
"grad_norm": 5.2391180992126465,
"learning_rate": 9.919520140105079e-05,
"loss": 1.2451,
"num_input_tokens_seen": 1851224,
"step": 115
},
{
"epoch": 0.008125596504592538,
"grad_norm": 4.865017890930176,
"learning_rate": 9.918820315236427e-05,
"loss": 1.1683,
"num_input_tokens_seen": 1867608,
"step": 116
},
{
"epoch": 0.008195644750321784,
"grad_norm": 4.943136215209961,
"learning_rate": 9.918120490367776e-05,
"loss": 1.31,
"num_input_tokens_seen": 1883696,
"step": 117
},
{
"epoch": 0.00826569299605103,
"grad_norm": 4.769871711730957,
"learning_rate": 9.917420665499125e-05,
"loss": 1.1212,
"num_input_tokens_seen": 1900080,
"step": 118
},
{
"epoch": 0.008335741241780275,
"grad_norm": 4.785780429840088,
"learning_rate": 9.916720840630474e-05,
"loss": 1.2415,
"num_input_tokens_seen": 1916464,
"step": 119
},
{
"epoch": 0.008405789487509523,
"grad_norm": 4.802333831787109,
"learning_rate": 9.916021015761822e-05,
"loss": 1.0513,
"num_input_tokens_seen": 1932848,
"step": 120
},
{
"epoch": 0.008475837733238768,
"grad_norm": 5.22212553024292,
"learning_rate": 9.91532119089317e-05,
"loss": 1.2574,
"num_input_tokens_seen": 1949232,
"step": 121
},
{
"epoch": 0.008545885978968014,
"grad_norm": 5.104204177856445,
"learning_rate": 9.914621366024518e-05,
"loss": 1.0436,
"num_input_tokens_seen": 1964184,
"step": 122
},
{
"epoch": 0.00861593422469726,
"grad_norm": 5.11055326461792,
"learning_rate": 9.913921541155868e-05,
"loss": 1.1939,
"num_input_tokens_seen": 1980568,
"step": 123
},
{
"epoch": 0.008685982470426507,
"grad_norm": 4.784866809844971,
"learning_rate": 9.913221716287216e-05,
"loss": 1.2056,
"num_input_tokens_seen": 1996952,
"step": 124
},
{
"epoch": 0.008756030716155752,
"grad_norm": 4.763037204742432,
"learning_rate": 9.912521891418564e-05,
"loss": 1.1403,
"num_input_tokens_seen": 2013336,
"step": 125
},
{
"epoch": 0.008826078961884998,
"grad_norm": 4.813408851623535,
"learning_rate": 9.911822066549913e-05,
"loss": 1.1897,
"num_input_tokens_seen": 2029720,
"step": 126
},
{
"epoch": 0.008896127207614244,
"grad_norm": 4.79008674621582,
"learning_rate": 9.911122241681261e-05,
"loss": 1.2315,
"num_input_tokens_seen": 2046104,
"step": 127
},
{
"epoch": 0.008966175453343491,
"grad_norm": 4.843508720397949,
"learning_rate": 9.91042241681261e-05,
"loss": 1.0883,
"num_input_tokens_seen": 2061592,
"step": 128
},
{
"epoch": 0.009036223699072737,
"grad_norm": 4.917592525482178,
"learning_rate": 9.909722591943959e-05,
"loss": 1.2512,
"num_input_tokens_seen": 2077792,
"step": 129
},
{
"epoch": 0.009106271944801982,
"grad_norm": 4.9154133796691895,
"learning_rate": 9.909022767075307e-05,
"loss": 1.3284,
"num_input_tokens_seen": 2094176,
"step": 130
},
{
"epoch": 0.009176320190531228,
"grad_norm": 5.2125420570373535,
"learning_rate": 9.908322942206656e-05,
"loss": 1.3469,
"num_input_tokens_seen": 2110480,
"step": 131
},
{
"epoch": 0.009246368436260475,
"grad_norm": 4.715712547302246,
"learning_rate": 9.907623117338004e-05,
"loss": 1.0844,
"num_input_tokens_seen": 2126864,
"step": 132
},
{
"epoch": 0.009316416681989721,
"grad_norm": 4.805694580078125,
"learning_rate": 9.906923292469353e-05,
"loss": 1.069,
"num_input_tokens_seen": 2142848,
"step": 133
},
{
"epoch": 0.009386464927718966,
"grad_norm": 4.961355209350586,
"learning_rate": 9.9062234676007e-05,
"loss": 1.3387,
"num_input_tokens_seen": 2159232,
"step": 134
},
{
"epoch": 0.009456513173448212,
"grad_norm": 4.582219123840332,
"learning_rate": 9.905523642732049e-05,
"loss": 1.2013,
"num_input_tokens_seen": 2175616,
"step": 135
},
{
"epoch": 0.009526561419177458,
"grad_norm": 5.195998191833496,
"learning_rate": 9.904823817863398e-05,
"loss": 1.2552,
"num_input_tokens_seen": 2191872,
"step": 136
},
{
"epoch": 0.009596609664906705,
"grad_norm": 4.934189319610596,
"learning_rate": 9.904123992994747e-05,
"loss": 1.2961,
"num_input_tokens_seen": 2208208,
"step": 137
},
{
"epoch": 0.00966665791063595,
"grad_norm": 4.981037616729736,
"learning_rate": 9.903424168126096e-05,
"loss": 1.1546,
"num_input_tokens_seen": 2224592,
"step": 138
},
{
"epoch": 0.009736706156365196,
"grad_norm": 5.469496250152588,
"learning_rate": 9.902724343257443e-05,
"loss": 1.3833,
"num_input_tokens_seen": 2240976,
"step": 139
},
{
"epoch": 0.009806754402094442,
"grad_norm": 4.889583587646484,
"learning_rate": 9.902024518388792e-05,
"loss": 1.2095,
"num_input_tokens_seen": 2257360,
"step": 140
},
{
"epoch": 0.00987680264782369,
"grad_norm": 4.532052516937256,
"learning_rate": 9.901324693520141e-05,
"loss": 1.143,
"num_input_tokens_seen": 2272848,
"step": 141
},
{
"epoch": 0.009946850893552935,
"grad_norm": 5.278079032897949,
"learning_rate": 9.900624868651488e-05,
"loss": 1.2849,
"num_input_tokens_seen": 2289232,
"step": 142
},
{
"epoch": 0.01001689913928218,
"grad_norm": 4.549891948699951,
"learning_rate": 9.899925043782839e-05,
"loss": 1.0482,
"num_input_tokens_seen": 2305424,
"step": 143
},
{
"epoch": 0.010086947385011426,
"grad_norm": 4.7777180671691895,
"learning_rate": 9.899225218914186e-05,
"loss": 1.1926,
"num_input_tokens_seen": 2320968,
"step": 144
},
{
"epoch": 0.010156995630740673,
"grad_norm": 4.320313453674316,
"learning_rate": 9.898525394045535e-05,
"loss": 1.0468,
"num_input_tokens_seen": 2337352,
"step": 145
},
{
"epoch": 0.010227043876469919,
"grad_norm": 4.915202617645264,
"learning_rate": 9.897825569176882e-05,
"loss": 1.1326,
"num_input_tokens_seen": 2353064,
"step": 146
},
{
"epoch": 0.010297092122199165,
"grad_norm": 4.569783687591553,
"learning_rate": 9.897125744308231e-05,
"loss": 0.8586,
"num_input_tokens_seen": 2369128,
"step": 147
},
{
"epoch": 0.01036714036792841,
"grad_norm": 4.591664791107178,
"learning_rate": 9.89642591943958e-05,
"loss": 1.1369,
"num_input_tokens_seen": 2385512,
"step": 148
},
{
"epoch": 0.010437188613657656,
"grad_norm": 4.913016319274902,
"learning_rate": 9.895726094570929e-05,
"loss": 1.1564,
"num_input_tokens_seen": 2401208,
"step": 149
},
{
"epoch": 0.010507236859386903,
"grad_norm": 4.908018112182617,
"learning_rate": 9.895026269702278e-05,
"loss": 1.1247,
"num_input_tokens_seen": 2417592,
"step": 150
},
{
"epoch": 0.010577285105116149,
"grad_norm": 4.536910057067871,
"learning_rate": 9.894326444833625e-05,
"loss": 1.014,
"num_input_tokens_seen": 2433976,
"step": 151
},
{
"epoch": 0.010647333350845395,
"grad_norm": 4.899227142333984,
"learning_rate": 9.893626619964974e-05,
"loss": 1.0418,
"num_input_tokens_seen": 2448072,
"step": 152
},
{
"epoch": 0.01071738159657464,
"grad_norm": 4.600861072540283,
"learning_rate": 9.892926795096323e-05,
"loss": 1.0459,
"num_input_tokens_seen": 2464240,
"step": 153
},
{
"epoch": 0.010787429842303888,
"grad_norm": 4.707681179046631,
"learning_rate": 9.89222697022767e-05,
"loss": 1.0859,
"num_input_tokens_seen": 2480624,
"step": 154
},
{
"epoch": 0.010857478088033133,
"grad_norm": 4.748518466949463,
"learning_rate": 9.89152714535902e-05,
"loss": 1.0608,
"num_input_tokens_seen": 2497008,
"step": 155
},
{
"epoch": 0.010927526333762379,
"grad_norm": 4.794179439544678,
"learning_rate": 9.890827320490368e-05,
"loss": 1.2243,
"num_input_tokens_seen": 2513392,
"step": 156
},
{
"epoch": 0.010997574579491624,
"grad_norm": 4.593925476074219,
"learning_rate": 9.890127495621717e-05,
"loss": 1.1002,
"num_input_tokens_seen": 2529776,
"step": 157
},
{
"epoch": 0.011067622825220872,
"grad_norm": 4.318257808685303,
"learning_rate": 9.889427670753066e-05,
"loss": 0.9561,
"num_input_tokens_seen": 2546160,
"step": 158
},
{
"epoch": 0.011137671070950117,
"grad_norm": 4.631777286529541,
"learning_rate": 9.888727845884414e-05,
"loss": 1.1553,
"num_input_tokens_seen": 2562544,
"step": 159
},
{
"epoch": 0.011207719316679363,
"grad_norm": 4.896609783172607,
"learning_rate": 9.888028021015762e-05,
"loss": 1.1779,
"num_input_tokens_seen": 2578088,
"step": 160
},
{
"epoch": 0.011277767562408609,
"grad_norm": 4.3978681564331055,
"learning_rate": 9.88732819614711e-05,
"loss": 1.1778,
"num_input_tokens_seen": 2594416,
"step": 161
},
{
"epoch": 0.011347815808137854,
"grad_norm": 4.82927942276001,
"learning_rate": 9.886628371278459e-05,
"loss": 1.0339,
"num_input_tokens_seen": 2609776,
"step": 162
},
{
"epoch": 0.011417864053867102,
"grad_norm": 4.413319110870361,
"learning_rate": 9.885928546409809e-05,
"loss": 1.0992,
"num_input_tokens_seen": 2626160,
"step": 163
},
{
"epoch": 0.011487912299596347,
"grad_norm": 4.626354694366455,
"learning_rate": 9.885228721541156e-05,
"loss": 1.1948,
"num_input_tokens_seen": 2642464,
"step": 164
},
{
"epoch": 0.011557960545325593,
"grad_norm": 4.328434467315674,
"learning_rate": 9.884528896672505e-05,
"loss": 1.1493,
"num_input_tokens_seen": 2658528,
"step": 165
},
{
"epoch": 0.011628008791054838,
"grad_norm": 4.57839822769165,
"learning_rate": 9.883829071803853e-05,
"loss": 1.0775,
"num_input_tokens_seen": 2674912,
"step": 166
},
{
"epoch": 0.011698057036784086,
"grad_norm": 5.103973865509033,
"learning_rate": 9.883129246935202e-05,
"loss": 1.2458,
"num_input_tokens_seen": 2690792,
"step": 167
},
{
"epoch": 0.011768105282513331,
"grad_norm": 4.558016300201416,
"learning_rate": 9.88242942206655e-05,
"loss": 1.0122,
"num_input_tokens_seen": 2705616,
"step": 168
},
{
"epoch": 0.011838153528242577,
"grad_norm": 4.811260223388672,
"learning_rate": 9.8817295971979e-05,
"loss": 1.2989,
"num_input_tokens_seen": 2721704,
"step": 169
},
{
"epoch": 0.011908201773971823,
"grad_norm": 4.726966857910156,
"learning_rate": 9.881029772329248e-05,
"loss": 1.176,
"num_input_tokens_seen": 2738088,
"step": 170
},
{
"epoch": 0.01197825001970107,
"grad_norm": 4.874902725219727,
"learning_rate": 9.880329947460596e-05,
"loss": 1.2586,
"num_input_tokens_seen": 2754040,
"step": 171
},
{
"epoch": 0.012048298265430316,
"grad_norm": 4.379549980163574,
"learning_rate": 9.879630122591945e-05,
"loss": 1.1771,
"num_input_tokens_seen": 2770424,
"step": 172
},
{
"epoch": 0.012118346511159561,
"grad_norm": 4.455331802368164,
"learning_rate": 9.878930297723292e-05,
"loss": 1.0714,
"num_input_tokens_seen": 2786808,
"step": 173
},
{
"epoch": 0.012188394756888807,
"grad_norm": 4.42273473739624,
"learning_rate": 9.878230472854641e-05,
"loss": 1.1798,
"num_input_tokens_seen": 2803176,
"step": 174
},
{
"epoch": 0.012258443002618052,
"grad_norm": 4.4078874588012695,
"learning_rate": 9.87753064798599e-05,
"loss": 1.1672,
"num_input_tokens_seen": 2819448,
"step": 175
},
{
"epoch": 0.0123284912483473,
"grad_norm": 4.79048490524292,
"learning_rate": 9.876830823117339e-05,
"loss": 1.3331,
"num_input_tokens_seen": 2835832,
"step": 176
},
{
"epoch": 0.012398539494076545,
"grad_norm": 4.212133884429932,
"learning_rate": 9.876130998248688e-05,
"loss": 1.0007,
"num_input_tokens_seen": 2851776,
"step": 177
},
{
"epoch": 0.012468587739805791,
"grad_norm": 5.7587738037109375,
"learning_rate": 9.875431173380035e-05,
"loss": 1.4729,
"num_input_tokens_seen": 2867896,
"step": 178
},
{
"epoch": 0.012538635985535037,
"grad_norm": 4.3469462394714355,
"learning_rate": 9.874731348511384e-05,
"loss": 0.957,
"num_input_tokens_seen": 2884280,
"step": 179
},
{
"epoch": 0.012608684231264284,
"grad_norm": 4.584625244140625,
"learning_rate": 9.874031523642733e-05,
"loss": 1.0753,
"num_input_tokens_seen": 2899208,
"step": 180
},
{
"epoch": 0.01267873247699353,
"grad_norm": 4.544627666473389,
"learning_rate": 9.87333169877408e-05,
"loss": 1.1706,
"num_input_tokens_seen": 2915416,
"step": 181
},
{
"epoch": 0.012748780722722775,
"grad_norm": 4.8749237060546875,
"learning_rate": 9.872631873905429e-05,
"loss": 1.3382,
"num_input_tokens_seen": 2931360,
"step": 182
},
{
"epoch": 0.01281882896845202,
"grad_norm": 4.593903541564941,
"learning_rate": 9.871932049036778e-05,
"loss": 1.1588,
"num_input_tokens_seen": 2947744,
"step": 183
},
{
"epoch": 0.012888877214181268,
"grad_norm": 4.478219509124756,
"learning_rate": 9.871232224168127e-05,
"loss": 1.1013,
"num_input_tokens_seen": 2963664,
"step": 184
},
{
"epoch": 0.012958925459910514,
"grad_norm": 5.028106212615967,
"learning_rate": 9.870532399299476e-05,
"loss": 1.3223,
"num_input_tokens_seen": 2980048,
"step": 185
},
{
"epoch": 0.01302897370563976,
"grad_norm": 4.866946697235107,
"learning_rate": 9.869832574430823e-05,
"loss": 1.2376,
"num_input_tokens_seen": 2995992,
"step": 186
},
{
"epoch": 0.013099021951369005,
"grad_norm": 4.421341419219971,
"learning_rate": 9.869132749562172e-05,
"loss": 1.2252,
"num_input_tokens_seen": 3012000,
"step": 187
},
{
"epoch": 0.01316907019709825,
"grad_norm": 4.88083028793335,
"learning_rate": 9.86843292469352e-05,
"loss": 1.2951,
"num_input_tokens_seen": 3028384,
"step": 188
},
{
"epoch": 0.013239118442827498,
"grad_norm": 4.654318809509277,
"learning_rate": 9.86773309982487e-05,
"loss": 1.2839,
"num_input_tokens_seen": 3044768,
"step": 189
},
{
"epoch": 0.013309166688556744,
"grad_norm": 4.626763820648193,
"learning_rate": 9.867033274956219e-05,
"loss": 1.2389,
"num_input_tokens_seen": 3061152,
"step": 190
},
{
"epoch": 0.01337921493428599,
"grad_norm": 4.178484916687012,
"learning_rate": 9.866333450087566e-05,
"loss": 1.1186,
"num_input_tokens_seen": 3077056,
"step": 191
},
{
"epoch": 0.013449263180015235,
"grad_norm": 4.755034923553467,
"learning_rate": 9.865633625218915e-05,
"loss": 1.0594,
"num_input_tokens_seen": 3093400,
"step": 192
},
{
"epoch": 0.013519311425744482,
"grad_norm": 4.437506198883057,
"learning_rate": 9.864933800350263e-05,
"loss": 1.2078,
"num_input_tokens_seen": 3109784,
"step": 193
},
{
"epoch": 0.013589359671473728,
"grad_norm": 5.140488624572754,
"learning_rate": 9.864233975481611e-05,
"loss": 1.4312,
"num_input_tokens_seen": 3124976,
"step": 194
},
{
"epoch": 0.013659407917202973,
"grad_norm": 4.72155237197876,
"learning_rate": 9.86353415061296e-05,
"loss": 1.1752,
"num_input_tokens_seen": 3140632,
"step": 195
},
{
"epoch": 0.013729456162932219,
"grad_norm": 4.914645671844482,
"learning_rate": 9.862834325744309e-05,
"loss": 1.2464,
"num_input_tokens_seen": 3156616,
"step": 196
},
{
"epoch": 0.013799504408661466,
"grad_norm": 4.23387336730957,
"learning_rate": 9.862134500875658e-05,
"loss": 0.9722,
"num_input_tokens_seen": 3172840,
"step": 197
},
{
"epoch": 0.013869552654390712,
"grad_norm": 4.659370422363281,
"learning_rate": 9.861434676007005e-05,
"loss": 1.1981,
"num_input_tokens_seen": 3188584,
"step": 198
},
{
"epoch": 0.013939600900119958,
"grad_norm": 4.580902576446533,
"learning_rate": 9.860734851138354e-05,
"loss": 1.1913,
"num_input_tokens_seen": 3204432,
"step": 199
},
{
"epoch": 0.014009649145849203,
"grad_norm": 4.208237648010254,
"learning_rate": 9.860035026269702e-05,
"loss": 1.2056,
"num_input_tokens_seen": 3220816,
"step": 200
},
{
"epoch": 0.014009649145849203,
"eval_loss": 1.2226407527923584,
"eval_runtime": 0.3992,
"eval_samples_per_second": 2.505,
"eval_steps_per_second": 2.505,
"num_input_tokens_seen": 3220816,
"step": 200
},
{
"epoch": 0.014079697391578449,
"grad_norm": 4.526260852813721,
"learning_rate": 9.85933520140105e-05,
"loss": 1.0488,
"num_input_tokens_seen": 3237200,
"step": 201
},
{
"epoch": 0.014149745637307696,
"grad_norm": 4.46895170211792,
"learning_rate": 9.8586353765324e-05,
"loss": 1.1101,
"num_input_tokens_seen": 3253336,
"step": 202
},
{
"epoch": 0.014219793883036942,
"grad_norm": 4.367347717285156,
"learning_rate": 9.857935551663748e-05,
"loss": 1.0425,
"num_input_tokens_seen": 3269632,
"step": 203
},
{
"epoch": 0.014289842128766187,
"grad_norm": 4.860860347747803,
"learning_rate": 9.857235726795097e-05,
"loss": 1.4068,
"num_input_tokens_seen": 3285432,
"step": 204
},
{
"epoch": 0.014359890374495433,
"grad_norm": 4.336480617523193,
"learning_rate": 9.856535901926445e-05,
"loss": 1.2579,
"num_input_tokens_seen": 3301632,
"step": 205
},
{
"epoch": 0.01442993862022468,
"grad_norm": 4.587873458862305,
"learning_rate": 9.855836077057794e-05,
"loss": 1.1508,
"num_input_tokens_seen": 3318016,
"step": 206
},
{
"epoch": 0.014499986865953926,
"grad_norm": 4.719262599945068,
"learning_rate": 9.855136252189142e-05,
"loss": 1.0208,
"num_input_tokens_seen": 3333168,
"step": 207
},
{
"epoch": 0.014570035111683172,
"grad_norm": 4.419138431549072,
"learning_rate": 9.85443642732049e-05,
"loss": 1.2576,
"num_input_tokens_seen": 3349384,
"step": 208
},
{
"epoch": 0.014640083357412417,
"grad_norm": 4.3150835037231445,
"learning_rate": 9.85373660245184e-05,
"loss": 1.1786,
"num_input_tokens_seen": 3365768,
"step": 209
},
{
"epoch": 0.014710131603141665,
"grad_norm": 4.5917649269104,
"learning_rate": 9.853036777583188e-05,
"loss": 1.2821,
"num_input_tokens_seen": 3382152,
"step": 210
},
{
"epoch": 0.01478017984887091,
"grad_norm": 4.9094343185424805,
"learning_rate": 9.852336952714537e-05,
"loss": 1.2415,
"num_input_tokens_seen": 3397896,
"step": 211
},
{
"epoch": 0.014850228094600156,
"grad_norm": 4.394861698150635,
"learning_rate": 9.851637127845885e-05,
"loss": 1.1776,
"num_input_tokens_seen": 3414280,
"step": 212
},
{
"epoch": 0.014920276340329401,
"grad_norm": 4.196374416351318,
"learning_rate": 9.850937302977233e-05,
"loss": 1.065,
"num_input_tokens_seen": 3430584,
"step": 213
},
{
"epoch": 0.014990324586058647,
"grad_norm": 4.728682518005371,
"learning_rate": 9.850237478108582e-05,
"loss": 1.2686,
"num_input_tokens_seen": 3446968,
"step": 214
},
{
"epoch": 0.015060372831787894,
"grad_norm": 4.291411876678467,
"learning_rate": 9.84953765323993e-05,
"loss": 1.1877,
"num_input_tokens_seen": 3462568,
"step": 215
},
{
"epoch": 0.01513042107751714,
"grad_norm": 4.405060768127441,
"learning_rate": 9.84883782837128e-05,
"loss": 1.2873,
"num_input_tokens_seen": 3478952,
"step": 216
},
{
"epoch": 0.015200469323246386,
"grad_norm": 4.254365921020508,
"learning_rate": 9.848138003502628e-05,
"loss": 1.1062,
"num_input_tokens_seen": 3495304,
"step": 217
},
{
"epoch": 0.015270517568975631,
"grad_norm": 4.741672039031982,
"learning_rate": 9.847438178633976e-05,
"loss": 1.1983,
"num_input_tokens_seen": 3511688,
"step": 218
},
{
"epoch": 0.015340565814704879,
"grad_norm": 4.352742671966553,
"learning_rate": 9.846738353765325e-05,
"loss": 1.2028,
"num_input_tokens_seen": 3528072,
"step": 219
},
{
"epoch": 0.015410614060434124,
"grad_norm": 4.996603488922119,
"learning_rate": 9.846038528896672e-05,
"loss": 1.1561,
"num_input_tokens_seen": 3542904,
"step": 220
},
{
"epoch": 0.01548066230616337,
"grad_norm": 4.911815166473389,
"learning_rate": 9.845338704028021e-05,
"loss": 1.3375,
"num_input_tokens_seen": 3558352,
"step": 221
},
{
"epoch": 0.015550710551892616,
"grad_norm": 4.638419151306152,
"learning_rate": 9.84463887915937e-05,
"loss": 1.1963,
"num_input_tokens_seen": 3574736,
"step": 222
},
{
"epoch": 0.015620758797621863,
"grad_norm": 4.323521614074707,
"learning_rate": 9.843939054290719e-05,
"loss": 1.1224,
"num_input_tokens_seen": 3591120,
"step": 223
},
{
"epoch": 0.01569080704335111,
"grad_norm": 4.466544151306152,
"learning_rate": 9.843239229422068e-05,
"loss": 1.3988,
"num_input_tokens_seen": 3607392,
"step": 224
},
{
"epoch": 0.015760855289080354,
"grad_norm": 4.476973533630371,
"learning_rate": 9.842539404553415e-05,
"loss": 1.184,
"num_input_tokens_seen": 3623776,
"step": 225
},
{
"epoch": 0.0158309035348096,
"grad_norm": 4.648625373840332,
"learning_rate": 9.841839579684764e-05,
"loss": 1.1768,
"num_input_tokens_seen": 3640008,
"step": 226
},
{
"epoch": 0.015900951780538845,
"grad_norm": 4.364476203918457,
"learning_rate": 9.841139754816112e-05,
"loss": 1.0208,
"num_input_tokens_seen": 3656392,
"step": 227
},
{
"epoch": 0.01597100002626809,
"grad_norm": 4.3054633140563965,
"learning_rate": 9.84043992994746e-05,
"loss": 1.1215,
"num_input_tokens_seen": 3672392,
"step": 228
},
{
"epoch": 0.016041048271997337,
"grad_norm": 4.83436918258667,
"learning_rate": 9.83974010507881e-05,
"loss": 1.2284,
"num_input_tokens_seen": 3688776,
"step": 229
},
{
"epoch": 0.016111096517726586,
"grad_norm": 4.447519779205322,
"learning_rate": 9.839040280210158e-05,
"loss": 1.1765,
"num_input_tokens_seen": 3705080,
"step": 230
},
{
"epoch": 0.01618114476345583,
"grad_norm": 4.269217491149902,
"learning_rate": 9.838340455341507e-05,
"loss": 1.0466,
"num_input_tokens_seen": 3721464,
"step": 231
},
{
"epoch": 0.016251193009185077,
"grad_norm": 4.41223669052124,
"learning_rate": 9.837640630472854e-05,
"loss": 1.2098,
"num_input_tokens_seen": 3737184,
"step": 232
},
{
"epoch": 0.016321241254914323,
"grad_norm": 4.632737159729004,
"learning_rate": 9.836940805604203e-05,
"loss": 1.1562,
"num_input_tokens_seen": 3753192,
"step": 233
},
{
"epoch": 0.016391289500643568,
"grad_norm": 4.379425525665283,
"learning_rate": 9.836240980735552e-05,
"loss": 1.1219,
"num_input_tokens_seen": 3767976,
"step": 234
},
{
"epoch": 0.016461337746372814,
"grad_norm": 4.28551721572876,
"learning_rate": 9.835541155866901e-05,
"loss": 1.0259,
"num_input_tokens_seen": 3784008,
"step": 235
},
{
"epoch": 0.01653138599210206,
"grad_norm": 4.642453670501709,
"learning_rate": 9.83484133099825e-05,
"loss": 1.1684,
"num_input_tokens_seen": 3800000,
"step": 236
},
{
"epoch": 0.016601434237831305,
"grad_norm": 4.367178440093994,
"learning_rate": 9.834141506129597e-05,
"loss": 1.2877,
"num_input_tokens_seen": 3816384,
"step": 237
},
{
"epoch": 0.01667148248356055,
"grad_norm": 4.5724005699157715,
"learning_rate": 9.833441681260946e-05,
"loss": 1.1814,
"num_input_tokens_seen": 3830328,
"step": 238
},
{
"epoch": 0.0167415307292898,
"grad_norm": 4.318159580230713,
"learning_rate": 9.832741856392295e-05,
"loss": 1.1143,
"num_input_tokens_seen": 3846712,
"step": 239
},
{
"epoch": 0.016811578975019045,
"grad_norm": 4.408501625061035,
"learning_rate": 9.832042031523643e-05,
"loss": 1.1508,
"num_input_tokens_seen": 3861776,
"step": 240
},
{
"epoch": 0.01688162722074829,
"grad_norm": 4.20060920715332,
"learning_rate": 9.831342206654991e-05,
"loss": 1.209,
"num_input_tokens_seen": 3877736,
"step": 241
},
{
"epoch": 0.016951675466477537,
"grad_norm": 4.431649208068848,
"learning_rate": 9.83064238178634e-05,
"loss": 1.2458,
"num_input_tokens_seen": 3893320,
"step": 242
},
{
"epoch": 0.017021723712206782,
"grad_norm": 4.000490188598633,
"learning_rate": 9.829942556917689e-05,
"loss": 1.0274,
"num_input_tokens_seen": 3909704,
"step": 243
},
{
"epoch": 0.017091771957936028,
"grad_norm": 4.703495025634766,
"learning_rate": 9.829242732049038e-05,
"loss": 1.1711,
"num_input_tokens_seen": 3925808,
"step": 244
},
{
"epoch": 0.017161820203665273,
"grad_norm": 4.639338970184326,
"learning_rate": 9.828542907180386e-05,
"loss": 1.3046,
"num_input_tokens_seen": 3942192,
"step": 245
},
{
"epoch": 0.01723186844939452,
"grad_norm": 4.414276599884033,
"learning_rate": 9.827843082311734e-05,
"loss": 1.271,
"num_input_tokens_seen": 3958528,
"step": 246
},
{
"epoch": 0.017301916695123768,
"grad_norm": 4.404853820800781,
"learning_rate": 9.827143257443082e-05,
"loss": 1.0693,
"num_input_tokens_seen": 3974912,
"step": 247
},
{
"epoch": 0.017371964940853014,
"grad_norm": 4.519491195678711,
"learning_rate": 9.826443432574431e-05,
"loss": 1.2894,
"num_input_tokens_seen": 3991296,
"step": 248
},
{
"epoch": 0.01744201318658226,
"grad_norm": 4.261727809906006,
"learning_rate": 9.825743607705781e-05,
"loss": 1.2059,
"num_input_tokens_seen": 4006544,
"step": 249
},
{
"epoch": 0.017512061432311505,
"grad_norm": 4.102485656738281,
"learning_rate": 9.825043782837129e-05,
"loss": 0.9365,
"num_input_tokens_seen": 4022320,
"step": 250
},
{
"epoch": 0.01758210967804075,
"grad_norm": 4.804764270782471,
"learning_rate": 9.824343957968477e-05,
"loss": 1.3344,
"num_input_tokens_seen": 4037048,
"step": 251
},
{
"epoch": 0.017652157923769996,
"grad_norm": 4.130600452423096,
"learning_rate": 9.823644133099825e-05,
"loss": 1.2349,
"num_input_tokens_seen": 4053432,
"step": 252
},
{
"epoch": 0.017722206169499242,
"grad_norm": 4.234742641448975,
"learning_rate": 9.822944308231174e-05,
"loss": 1.1371,
"num_input_tokens_seen": 4069816,
"step": 253
},
{
"epoch": 0.017792254415228487,
"grad_norm": 4.754928112030029,
"learning_rate": 9.822244483362521e-05,
"loss": 1.5168,
"num_input_tokens_seen": 4085864,
"step": 254
},
{
"epoch": 0.017862302660957733,
"grad_norm": 4.542768478393555,
"learning_rate": 9.821544658493871e-05,
"loss": 1.1943,
"num_input_tokens_seen": 4102240,
"step": 255
},
{
"epoch": 0.017932350906686982,
"grad_norm": 4.411310195922852,
"learning_rate": 9.82084483362522e-05,
"loss": 1.2694,
"num_input_tokens_seen": 4118544,
"step": 256
},
{
"epoch": 0.018002399152416228,
"grad_norm": 4.205377101898193,
"learning_rate": 9.820145008756568e-05,
"loss": 1.1581,
"num_input_tokens_seen": 4134928,
"step": 257
},
{
"epoch": 0.018072447398145473,
"grad_norm": 4.451165199279785,
"learning_rate": 9.819445183887917e-05,
"loss": 1.089,
"num_input_tokens_seen": 4150848,
"step": 258
},
{
"epoch": 0.01814249564387472,
"grad_norm": 4.366336822509766,
"learning_rate": 9.818745359019264e-05,
"loss": 1.1767,
"num_input_tokens_seen": 4167184,
"step": 259
},
{
"epoch": 0.018212543889603965,
"grad_norm": 4.394649982452393,
"learning_rate": 9.818045534150613e-05,
"loss": 1.0741,
"num_input_tokens_seen": 4183376,
"step": 260
},
{
"epoch": 0.01828259213533321,
"grad_norm": 4.344518184661865,
"learning_rate": 9.817345709281962e-05,
"loss": 1.2282,
"num_input_tokens_seen": 4199760,
"step": 261
},
{
"epoch": 0.018352640381062456,
"grad_norm": 4.403041362762451,
"learning_rate": 9.816645884413311e-05,
"loss": 1.2317,
"num_input_tokens_seen": 4215816,
"step": 262
},
{
"epoch": 0.0184226886267917,
"grad_norm": 4.715320110321045,
"learning_rate": 9.81594605954466e-05,
"loss": 1.3074,
"num_input_tokens_seen": 4231504,
"step": 263
},
{
"epoch": 0.01849273687252095,
"grad_norm": 4.5754265785217285,
"learning_rate": 9.815246234676007e-05,
"loss": 1.253,
"num_input_tokens_seen": 4247888,
"step": 264
},
{
"epoch": 0.018562785118250196,
"grad_norm": 4.2346930503845215,
"learning_rate": 9.814546409807356e-05,
"loss": 1.1727,
"num_input_tokens_seen": 4264248,
"step": 265
},
{
"epoch": 0.018632833363979442,
"grad_norm": 4.186713218688965,
"learning_rate": 9.813846584938705e-05,
"loss": 1.2693,
"num_input_tokens_seen": 4280632,
"step": 266
},
{
"epoch": 0.018702881609708687,
"grad_norm": 4.6356706619262695,
"learning_rate": 9.813146760070052e-05,
"loss": 1.3755,
"num_input_tokens_seen": 4296648,
"step": 267
},
{
"epoch": 0.018772929855437933,
"grad_norm": 4.466466903686523,
"learning_rate": 9.812446935201401e-05,
"loss": 1.283,
"num_input_tokens_seen": 4311408,
"step": 268
},
{
"epoch": 0.01884297810116718,
"grad_norm": 4.3369140625,
"learning_rate": 9.81174711033275e-05,
"loss": 1.1555,
"num_input_tokens_seen": 4326736,
"step": 269
},
{
"epoch": 0.018913026346896424,
"grad_norm": 4.434782028198242,
"learning_rate": 9.811047285464099e-05,
"loss": 1.2859,
"num_input_tokens_seen": 4343120,
"step": 270
},
{
"epoch": 0.01898307459262567,
"grad_norm": 4.346708297729492,
"learning_rate": 9.810347460595448e-05,
"loss": 1.1421,
"num_input_tokens_seen": 4359504,
"step": 271
},
{
"epoch": 0.019053122838354915,
"grad_norm": 4.529878616333008,
"learning_rate": 9.809647635726795e-05,
"loss": 1.2654,
"num_input_tokens_seen": 4375888,
"step": 272
},
{
"epoch": 0.019123171084084165,
"grad_norm": 4.051745891571045,
"learning_rate": 9.808947810858144e-05,
"loss": 1.1469,
"num_input_tokens_seen": 4392224,
"step": 273
},
{
"epoch": 0.01919321932981341,
"grad_norm": 4.403522491455078,
"learning_rate": 9.808247985989492e-05,
"loss": 1.233,
"num_input_tokens_seen": 4408608,
"step": 274
},
{
"epoch": 0.019263267575542656,
"grad_norm": 4.166261196136475,
"learning_rate": 9.807548161120842e-05,
"loss": 1.1697,
"num_input_tokens_seen": 4424992,
"step": 275
},
{
"epoch": 0.0193333158212719,
"grad_norm": 4.29187536239624,
"learning_rate": 9.806848336252191e-05,
"loss": 1.0503,
"num_input_tokens_seen": 4441376,
"step": 276
},
{
"epoch": 0.019403364067001147,
"grad_norm": 4.4056172370910645,
"learning_rate": 9.806148511383538e-05,
"loss": 1.1965,
"num_input_tokens_seen": 4457760,
"step": 277
},
{
"epoch": 0.019473412312730393,
"grad_norm": 4.355875015258789,
"learning_rate": 9.805448686514887e-05,
"loss": 1.1024,
"num_input_tokens_seen": 4474144,
"step": 278
},
{
"epoch": 0.019543460558459638,
"grad_norm": 4.46420955657959,
"learning_rate": 9.804748861646235e-05,
"loss": 1.203,
"num_input_tokens_seen": 4488912,
"step": 279
},
{
"epoch": 0.019613508804188884,
"grad_norm": 4.48052453994751,
"learning_rate": 9.804049036777583e-05,
"loss": 1.2089,
"num_input_tokens_seen": 4505296,
"step": 280
},
{
"epoch": 0.01968355704991813,
"grad_norm": 4.458749294281006,
"learning_rate": 9.803349211908932e-05,
"loss": 1.1557,
"num_input_tokens_seen": 4520576,
"step": 281
},
{
"epoch": 0.01975360529564738,
"grad_norm": 4.551771640777588,
"learning_rate": 9.802649387040281e-05,
"loss": 1.1671,
"num_input_tokens_seen": 4536960,
"step": 282
},
{
"epoch": 0.019823653541376624,
"grad_norm": 4.038064956665039,
"learning_rate": 9.80194956217163e-05,
"loss": 1.1562,
"num_input_tokens_seen": 4553344,
"step": 283
},
{
"epoch": 0.01989370178710587,
"grad_norm": 4.647075653076172,
"learning_rate": 9.801249737302978e-05,
"loss": 1.3069,
"num_input_tokens_seen": 4568928,
"step": 284
},
{
"epoch": 0.019963750032835115,
"grad_norm": 4.258941650390625,
"learning_rate": 9.800549912434326e-05,
"loss": 1.0349,
"num_input_tokens_seen": 4585312,
"step": 285
},
{
"epoch": 0.02003379827856436,
"grad_norm": 4.348769664764404,
"learning_rate": 9.799850087565674e-05,
"loss": 1.1163,
"num_input_tokens_seen": 4601696,
"step": 286
},
{
"epoch": 0.020103846524293607,
"grad_norm": 4.105901718139648,
"learning_rate": 9.799150262697023e-05,
"loss": 1.0313,
"num_input_tokens_seen": 4617312,
"step": 287
},
{
"epoch": 0.020173894770022852,
"grad_norm": 4.079495429992676,
"learning_rate": 9.798450437828372e-05,
"loss": 1.0828,
"num_input_tokens_seen": 4633696,
"step": 288
},
{
"epoch": 0.020243943015752098,
"grad_norm": 4.03472375869751,
"learning_rate": 9.79775061295972e-05,
"loss": 0.9475,
"num_input_tokens_seen": 4650080,
"step": 289
},
{
"epoch": 0.020313991261481347,
"grad_norm": 4.077049732208252,
"learning_rate": 9.797050788091069e-05,
"loss": 1.1323,
"num_input_tokens_seen": 4666328,
"step": 290
},
{
"epoch": 0.020384039507210593,
"grad_norm": 4.086606025695801,
"learning_rate": 9.796350963222417e-05,
"loss": 1.1218,
"num_input_tokens_seen": 4682256,
"step": 291
},
{
"epoch": 0.020454087752939838,
"grad_norm": 4.296900749206543,
"learning_rate": 9.795651138353766e-05,
"loss": 1.2964,
"num_input_tokens_seen": 4698640,
"step": 292
},
{
"epoch": 0.020524135998669084,
"grad_norm": 4.040759086608887,
"learning_rate": 9.794951313485115e-05,
"loss": 1.1077,
"num_input_tokens_seen": 4714928,
"step": 293
},
{
"epoch": 0.02059418424439833,
"grad_norm": 3.8260273933410645,
"learning_rate": 9.794251488616462e-05,
"loss": 0.9667,
"num_input_tokens_seen": 4731312,
"step": 294
},
{
"epoch": 0.020664232490127575,
"grad_norm": 4.294517993927002,
"learning_rate": 9.793551663747811e-05,
"loss": 1.2704,
"num_input_tokens_seen": 4747544,
"step": 295
},
{
"epoch": 0.02073428073585682,
"grad_norm": 4.206037521362305,
"learning_rate": 9.79285183887916e-05,
"loss": 1.1593,
"num_input_tokens_seen": 4763928,
"step": 296
},
{
"epoch": 0.020804328981586066,
"grad_norm": 4.147867202758789,
"learning_rate": 9.792152014010509e-05,
"loss": 1.1256,
"num_input_tokens_seen": 4780312,
"step": 297
},
{
"epoch": 0.020874377227315312,
"grad_norm": 4.23718786239624,
"learning_rate": 9.791452189141857e-05,
"loss": 1.2353,
"num_input_tokens_seen": 4796384,
"step": 298
},
{
"epoch": 0.02094442547304456,
"grad_norm": 4.172685146331787,
"learning_rate": 9.790752364273205e-05,
"loss": 1.1868,
"num_input_tokens_seen": 4812768,
"step": 299
},
{
"epoch": 0.021014473718773807,
"grad_norm": 4.167289733886719,
"learning_rate": 9.790052539404554e-05,
"loss": 1.0606,
"num_input_tokens_seen": 4829152,
"step": 300
},
{
"epoch": 0.021084521964503052,
"grad_norm": 4.096963882446289,
"learning_rate": 9.789352714535903e-05,
"loss": 1.0557,
"num_input_tokens_seen": 4845384,
"step": 301
},
{
"epoch": 0.021154570210232298,
"grad_norm": 4.223779678344727,
"learning_rate": 9.788652889667252e-05,
"loss": 1.1485,
"num_input_tokens_seen": 4861768,
"step": 302
},
{
"epoch": 0.021224618455961543,
"grad_norm": 3.8243472576141357,
"learning_rate": 9.7879530647986e-05,
"loss": 1.004,
"num_input_tokens_seen": 4878152,
"step": 303
},
{
"epoch": 0.02129466670169079,
"grad_norm": 4.092590808868408,
"learning_rate": 9.787253239929948e-05,
"loss": 1.0211,
"num_input_tokens_seen": 4894536,
"step": 304
},
{
"epoch": 0.021364714947420035,
"grad_norm": 4.42412805557251,
"learning_rate": 9.786553415061297e-05,
"loss": 0.9915,
"num_input_tokens_seen": 4910320,
"step": 305
},
{
"epoch": 0.02143476319314928,
"grad_norm": 4.488316535949707,
"learning_rate": 9.785853590192644e-05,
"loss": 1.1782,
"num_input_tokens_seen": 4926704,
"step": 306
},
{
"epoch": 0.021504811438878526,
"grad_norm": 4.110256195068359,
"learning_rate": 9.785153765323993e-05,
"loss": 1.102,
"num_input_tokens_seen": 4943088,
"step": 307
},
{
"epoch": 0.021574859684607775,
"grad_norm": 4.246950149536133,
"learning_rate": 9.784453940455342e-05,
"loss": 1.067,
"num_input_tokens_seen": 4958736,
"step": 308
},
{
"epoch": 0.02164490793033702,
"grad_norm": 4.175214767456055,
"learning_rate": 9.783754115586691e-05,
"loss": 1.0638,
"num_input_tokens_seen": 4975120,
"step": 309
},
{
"epoch": 0.021714956176066266,
"grad_norm": 4.427795886993408,
"learning_rate": 9.78305429071804e-05,
"loss": 1.1347,
"num_input_tokens_seen": 4991504,
"step": 310
},
{
"epoch": 0.021785004421795512,
"grad_norm": 4.158191204071045,
"learning_rate": 9.782354465849387e-05,
"loss": 1.1662,
"num_input_tokens_seen": 5007152,
"step": 311
},
{
"epoch": 0.021855052667524758,
"grad_norm": 4.184347629547119,
"learning_rate": 9.781654640980736e-05,
"loss": 1.0791,
"num_input_tokens_seen": 5023536,
"step": 312
},
{
"epoch": 0.021925100913254003,
"grad_norm": 3.8506295680999756,
"learning_rate": 9.780954816112084e-05,
"loss": 1.0615,
"num_input_tokens_seen": 5039728,
"step": 313
},
{
"epoch": 0.02199514915898325,
"grad_norm": 4.310062408447266,
"learning_rate": 9.780254991243432e-05,
"loss": 1.1363,
"num_input_tokens_seen": 5056008,
"step": 314
},
{
"epoch": 0.022065197404712494,
"grad_norm": 4.215006351470947,
"learning_rate": 9.779555166374781e-05,
"loss": 1.1715,
"num_input_tokens_seen": 5072096,
"step": 315
},
{
"epoch": 0.022135245650441743,
"grad_norm": 4.219073295593262,
"learning_rate": 9.77885534150613e-05,
"loss": 1.219,
"num_input_tokens_seen": 5088432,
"step": 316
},
{
"epoch": 0.02220529389617099,
"grad_norm": 4.319522857666016,
"learning_rate": 9.778155516637479e-05,
"loss": 1.3085,
"num_input_tokens_seen": 5104240,
"step": 317
},
{
"epoch": 0.022275342141900235,
"grad_norm": 4.118961334228516,
"learning_rate": 9.777455691768827e-05,
"loss": 1.0926,
"num_input_tokens_seen": 5120624,
"step": 318
},
{
"epoch": 0.02234539038762948,
"grad_norm": 4.195051193237305,
"learning_rate": 9.776755866900175e-05,
"loss": 1.0894,
"num_input_tokens_seen": 5137008,
"step": 319
},
{
"epoch": 0.022415438633358726,
"grad_norm": 4.114197254180908,
"learning_rate": 9.776056042031524e-05,
"loss": 1.1897,
"num_input_tokens_seen": 5153272,
"step": 320
},
{
"epoch": 0.02248548687908797,
"grad_norm": 4.014908313751221,
"learning_rate": 9.775356217162872e-05,
"loss": 1.0932,
"num_input_tokens_seen": 5169472,
"step": 321
},
{
"epoch": 0.022555535124817217,
"grad_norm": 4.190642356872559,
"learning_rate": 9.774656392294222e-05,
"loss": 1.1413,
"num_input_tokens_seen": 5185856,
"step": 322
},
{
"epoch": 0.022625583370546463,
"grad_norm": 4.562993049621582,
"learning_rate": 9.77395656742557e-05,
"loss": 1.2865,
"num_input_tokens_seen": 5202240,
"step": 323
},
{
"epoch": 0.02269563161627571,
"grad_norm": 4.607022762298584,
"learning_rate": 9.773256742556918e-05,
"loss": 1.1465,
"num_input_tokens_seen": 5218168,
"step": 324
},
{
"epoch": 0.022765679862004957,
"grad_norm": 3.956439256668091,
"learning_rate": 9.772556917688267e-05,
"loss": 1.028,
"num_input_tokens_seen": 5234368,
"step": 325
},
{
"epoch": 0.022835728107734203,
"grad_norm": 4.20713472366333,
"learning_rate": 9.771857092819615e-05,
"loss": 1.2332,
"num_input_tokens_seen": 5249808,
"step": 326
},
{
"epoch": 0.02290577635346345,
"grad_norm": 4.4092864990234375,
"learning_rate": 9.771157267950964e-05,
"loss": 1.104,
"num_input_tokens_seen": 5266120,
"step": 327
},
{
"epoch": 0.022975824599192694,
"grad_norm": 4.529845237731934,
"learning_rate": 9.770457443082312e-05,
"loss": 1.3475,
"num_input_tokens_seen": 5282504,
"step": 328
},
{
"epoch": 0.02304587284492194,
"grad_norm": 4.221986293792725,
"learning_rate": 9.769757618213661e-05,
"loss": 1.4115,
"num_input_tokens_seen": 5298344,
"step": 329
},
{
"epoch": 0.023115921090651186,
"grad_norm": 4.29000186920166,
"learning_rate": 9.76905779334501e-05,
"loss": 1.2855,
"num_input_tokens_seen": 5314728,
"step": 330
},
{
"epoch": 0.02318596933638043,
"grad_norm": 4.426812648773193,
"learning_rate": 9.768357968476358e-05,
"loss": 1.514,
"num_input_tokens_seen": 5330816,
"step": 331
},
{
"epoch": 0.023256017582109677,
"grad_norm": 4.210752964019775,
"learning_rate": 9.767658143607706e-05,
"loss": 1.0854,
"num_input_tokens_seen": 5346552,
"step": 332
},
{
"epoch": 0.023326065827838922,
"grad_norm": 4.216427326202393,
"learning_rate": 9.766958318739054e-05,
"loss": 1.1573,
"num_input_tokens_seen": 5362936,
"step": 333
},
{
"epoch": 0.02339611407356817,
"grad_norm": 4.132325649261475,
"learning_rate": 9.766258493870403e-05,
"loss": 1.0942,
"num_input_tokens_seen": 5379320,
"step": 334
},
{
"epoch": 0.023466162319297417,
"grad_norm": 4.277027130126953,
"learning_rate": 9.765558669001752e-05,
"loss": 1.1227,
"num_input_tokens_seen": 5395704,
"step": 335
},
{
"epoch": 0.023536210565026663,
"grad_norm": 4.228096961975098,
"learning_rate": 9.7648588441331e-05,
"loss": 1.1094,
"num_input_tokens_seen": 5412088,
"step": 336
},
{
"epoch": 0.02360625881075591,
"grad_norm": 4.194522380828857,
"learning_rate": 9.76415901926445e-05,
"loss": 1.2066,
"num_input_tokens_seen": 5428472,
"step": 337
},
{
"epoch": 0.023676307056485154,
"grad_norm": 4.336326599121094,
"learning_rate": 9.763459194395797e-05,
"loss": 1.2251,
"num_input_tokens_seen": 5444856,
"step": 338
},
{
"epoch": 0.0237463553022144,
"grad_norm": 4.2723307609558105,
"learning_rate": 9.762759369527146e-05,
"loss": 1.0927,
"num_input_tokens_seen": 5460304,
"step": 339
},
{
"epoch": 0.023816403547943645,
"grad_norm": 4.190036773681641,
"learning_rate": 9.762059544658493e-05,
"loss": 1.2036,
"num_input_tokens_seen": 5476688,
"step": 340
},
{
"epoch": 0.02388645179367289,
"grad_norm": 4.477560043334961,
"learning_rate": 9.761359719789842e-05,
"loss": 1.362,
"num_input_tokens_seen": 5493072,
"step": 341
},
{
"epoch": 0.02395650003940214,
"grad_norm": 4.160232067108154,
"learning_rate": 9.760659894921192e-05,
"loss": 1.1602,
"num_input_tokens_seen": 5509456,
"step": 342
},
{
"epoch": 0.024026548285131386,
"grad_norm": 3.857335090637207,
"learning_rate": 9.75996007005254e-05,
"loss": 1.0963,
"num_input_tokens_seen": 5525840,
"step": 343
},
{
"epoch": 0.02409659653086063,
"grad_norm": 4.141246318817139,
"learning_rate": 9.759260245183889e-05,
"loss": 1.2009,
"num_input_tokens_seen": 5541888,
"step": 344
},
{
"epoch": 0.024166644776589877,
"grad_norm": 4.50364875793457,
"learning_rate": 9.758560420315236e-05,
"loss": 1.1483,
"num_input_tokens_seen": 5557848,
"step": 345
},
{
"epoch": 0.024236693022319122,
"grad_norm": 4.3343353271484375,
"learning_rate": 9.757860595446585e-05,
"loss": 1.3594,
"num_input_tokens_seen": 5573504,
"step": 346
},
{
"epoch": 0.024306741268048368,
"grad_norm": 4.050408363342285,
"learning_rate": 9.757160770577934e-05,
"loss": 1.0563,
"num_input_tokens_seen": 5589544,
"step": 347
},
{
"epoch": 0.024376789513777614,
"grad_norm": 4.051811695098877,
"learning_rate": 9.756460945709283e-05,
"loss": 1.0288,
"num_input_tokens_seen": 5605368,
"step": 348
},
{
"epoch": 0.02444683775950686,
"grad_norm": 4.365113258361816,
"learning_rate": 9.755761120840632e-05,
"loss": 1.3054,
"num_input_tokens_seen": 5621752,
"step": 349
},
{
"epoch": 0.024516886005236105,
"grad_norm": 4.0057501792907715,
"learning_rate": 9.755061295971979e-05,
"loss": 1.1302,
"num_input_tokens_seen": 5638136,
"step": 350
},
{
"epoch": 0.024586934250965354,
"grad_norm": 4.254896640777588,
"learning_rate": 9.754361471103328e-05,
"loss": 1.0495,
"num_input_tokens_seen": 5653168,
"step": 351
},
{
"epoch": 0.0246569824966946,
"grad_norm": 3.8119771480560303,
"learning_rate": 9.753661646234677e-05,
"loss": 1.0349,
"num_input_tokens_seen": 5669504,
"step": 352
},
{
"epoch": 0.024727030742423845,
"grad_norm": 4.5082621574401855,
"learning_rate": 9.752961821366024e-05,
"loss": 1.2537,
"num_input_tokens_seen": 5685168,
"step": 353
},
{
"epoch": 0.02479707898815309,
"grad_norm": 4.392731189727783,
"learning_rate": 9.752261996497373e-05,
"loss": 1.2534,
"num_input_tokens_seen": 5701240,
"step": 354
},
{
"epoch": 0.024867127233882336,
"grad_norm": 4.293395519256592,
"learning_rate": 9.751562171628722e-05,
"loss": 1.2774,
"num_input_tokens_seen": 5717624,
"step": 355
},
{
"epoch": 0.024937175479611582,
"grad_norm": 4.64813756942749,
"learning_rate": 9.750862346760071e-05,
"loss": 1.2795,
"num_input_tokens_seen": 5733104,
"step": 356
},
{
"epoch": 0.025007223725340828,
"grad_norm": 4.5166778564453125,
"learning_rate": 9.75016252189142e-05,
"loss": 1.1301,
"num_input_tokens_seen": 5749488,
"step": 357
},
{
"epoch": 0.025077271971070073,
"grad_norm": 3.894291400909424,
"learning_rate": 9.749462697022767e-05,
"loss": 0.901,
"num_input_tokens_seen": 5765872,
"step": 358
},
{
"epoch": 0.02514732021679932,
"grad_norm": 4.10056209564209,
"learning_rate": 9.748762872154116e-05,
"loss": 1.0529,
"num_input_tokens_seen": 5780856,
"step": 359
},
{
"epoch": 0.025217368462528568,
"grad_norm": 4.6277666091918945,
"learning_rate": 9.748063047285464e-05,
"loss": 1.3649,
"num_input_tokens_seen": 5796856,
"step": 360
},
{
"epoch": 0.025287416708257814,
"grad_norm": 4.029720306396484,
"learning_rate": 9.747363222416813e-05,
"loss": 0.8863,
"num_input_tokens_seen": 5812176,
"step": 361
},
{
"epoch": 0.02535746495398706,
"grad_norm": 3.7772202491760254,
"learning_rate": 9.746663397548161e-05,
"loss": 1.0448,
"num_input_tokens_seen": 5828064,
"step": 362
},
{
"epoch": 0.025427513199716305,
"grad_norm": 4.379861354827881,
"learning_rate": 9.74596357267951e-05,
"loss": 1.3274,
"num_input_tokens_seen": 5843680,
"step": 363
},
{
"epoch": 0.02549756144544555,
"grad_norm": 4.254587173461914,
"learning_rate": 9.745263747810859e-05,
"loss": 1.1502,
"num_input_tokens_seen": 5859024,
"step": 364
},
{
"epoch": 0.025567609691174796,
"grad_norm": 4.271276473999023,
"learning_rate": 9.744563922942207e-05,
"loss": 1.2785,
"num_input_tokens_seen": 5874320,
"step": 365
},
{
"epoch": 0.02563765793690404,
"grad_norm": 4.224324703216553,
"learning_rate": 9.743864098073555e-05,
"loss": 1.0926,
"num_input_tokens_seen": 5890704,
"step": 366
},
{
"epoch": 0.025707706182633287,
"grad_norm": 4.289444446563721,
"learning_rate": 9.743164273204903e-05,
"loss": 1.1913,
"num_input_tokens_seen": 5906016,
"step": 367
},
{
"epoch": 0.025777754428362536,
"grad_norm": 4.280707359313965,
"learning_rate": 9.742464448336253e-05,
"loss": 1.2238,
"num_input_tokens_seen": 5921784,
"step": 368
},
{
"epoch": 0.025847802674091782,
"grad_norm": 4.554803848266602,
"learning_rate": 9.741764623467602e-05,
"loss": 1.2491,
"num_input_tokens_seen": 5938072,
"step": 369
},
{
"epoch": 0.025917850919821028,
"grad_norm": 4.677784442901611,
"learning_rate": 9.74106479859895e-05,
"loss": 1.2387,
"num_input_tokens_seen": 5954456,
"step": 370
},
{
"epoch": 0.025987899165550273,
"grad_norm": 4.268225193023682,
"learning_rate": 9.740364973730298e-05,
"loss": 1.2983,
"num_input_tokens_seen": 5970664,
"step": 371
},
{
"epoch": 0.02605794741127952,
"grad_norm": 4.361818790435791,
"learning_rate": 9.739665148861646e-05,
"loss": 1.199,
"num_input_tokens_seen": 5987048,
"step": 372
},
{
"epoch": 0.026127995657008764,
"grad_norm": 3.9990735054016113,
"learning_rate": 9.738965323992995e-05,
"loss": 1.0777,
"num_input_tokens_seen": 6003432,
"step": 373
},
{
"epoch": 0.02619804390273801,
"grad_norm": 3.992142915725708,
"learning_rate": 9.738265499124344e-05,
"loss": 1.0443,
"num_input_tokens_seen": 6019816,
"step": 374
},
{
"epoch": 0.026268092148467256,
"grad_norm": 4.270167827606201,
"learning_rate": 9.737565674255693e-05,
"loss": 1.1764,
"num_input_tokens_seen": 6036200,
"step": 375
},
{
"epoch": 0.0263381403941965,
"grad_norm": 4.362086296081543,
"learning_rate": 9.736865849387041e-05,
"loss": 1.2735,
"num_input_tokens_seen": 6052120,
"step": 376
},
{
"epoch": 0.02640818863992575,
"grad_norm": 3.6900475025177,
"learning_rate": 9.736166024518389e-05,
"loss": 0.8729,
"num_input_tokens_seen": 6068264,
"step": 377
},
{
"epoch": 0.026478236885654996,
"grad_norm": 3.8281285762786865,
"learning_rate": 9.735466199649738e-05,
"loss": 1.1096,
"num_input_tokens_seen": 6084504,
"step": 378
},
{
"epoch": 0.02654828513138424,
"grad_norm": 3.9335553646087646,
"learning_rate": 9.734766374781087e-05,
"loss": 1.0763,
"num_input_tokens_seen": 6100592,
"step": 379
},
{
"epoch": 0.026618333377113487,
"grad_norm": 4.332645416259766,
"learning_rate": 9.734066549912434e-05,
"loss": 1.1751,
"num_input_tokens_seen": 6116976,
"step": 380
},
{
"epoch": 0.026688381622842733,
"grad_norm": 4.160863399505615,
"learning_rate": 9.733366725043783e-05,
"loss": 1.0778,
"num_input_tokens_seen": 6133360,
"step": 381
},
{
"epoch": 0.02675842986857198,
"grad_norm": 4.388178825378418,
"learning_rate": 9.732666900175132e-05,
"loss": 1.2214,
"num_input_tokens_seen": 6149744,
"step": 382
},
{
"epoch": 0.026828478114301224,
"grad_norm": 4.354910373687744,
"learning_rate": 9.73196707530648e-05,
"loss": 1.4115,
"num_input_tokens_seen": 6166048,
"step": 383
},
{
"epoch": 0.02689852636003047,
"grad_norm": 4.058071613311768,
"learning_rate": 9.73126725043783e-05,
"loss": 1.0934,
"num_input_tokens_seen": 6181840,
"step": 384
},
{
"epoch": 0.026968574605759715,
"grad_norm": 4.060855865478516,
"learning_rate": 9.730567425569177e-05,
"loss": 1.1395,
"num_input_tokens_seen": 6198224,
"step": 385
},
{
"epoch": 0.027038622851488964,
"grad_norm": 4.316681385040283,
"learning_rate": 9.729867600700526e-05,
"loss": 1.1052,
"num_input_tokens_seen": 6214608,
"step": 386
},
{
"epoch": 0.02710867109721821,
"grad_norm": 4.322516918182373,
"learning_rate": 9.729167775831873e-05,
"loss": 1.2512,
"num_input_tokens_seen": 6230992,
"step": 387
},
{
"epoch": 0.027178719342947456,
"grad_norm": 4.090857028961182,
"learning_rate": 9.728467950963224e-05,
"loss": 1.0772,
"num_input_tokens_seen": 6246760,
"step": 388
},
{
"epoch": 0.0272487675886767,
"grad_norm": 4.0143961906433105,
"learning_rate": 9.727768126094571e-05,
"loss": 1.0578,
"num_input_tokens_seen": 6261968,
"step": 389
},
{
"epoch": 0.027318815834405947,
"grad_norm": 4.911194324493408,
"learning_rate": 9.72706830122592e-05,
"loss": 1.3016,
"num_input_tokens_seen": 6276664,
"step": 390
},
{
"epoch": 0.027388864080135192,
"grad_norm": 4.057498931884766,
"learning_rate": 9.726368476357269e-05,
"loss": 1.026,
"num_input_tokens_seen": 6293048,
"step": 391
},
{
"epoch": 0.027458912325864438,
"grad_norm": 3.9827401638031006,
"learning_rate": 9.725668651488616e-05,
"loss": 1.136,
"num_input_tokens_seen": 6309432,
"step": 392
},
{
"epoch": 0.027528960571593684,
"grad_norm": 4.640822887420654,
"learning_rate": 9.724968826619965e-05,
"loss": 1.2823,
"num_input_tokens_seen": 6325568,
"step": 393
},
{
"epoch": 0.027599008817322933,
"grad_norm": 4.372538089752197,
"learning_rate": 9.724269001751314e-05,
"loss": 1.0354,
"num_input_tokens_seen": 6341952,
"step": 394
},
{
"epoch": 0.02766905706305218,
"grad_norm": 4.018289566040039,
"learning_rate": 9.723569176882663e-05,
"loss": 1.029,
"num_input_tokens_seen": 6358336,
"step": 395
},
{
"epoch": 0.027739105308781424,
"grad_norm": 4.440858364105225,
"learning_rate": 9.722869352014012e-05,
"loss": 1.2272,
"num_input_tokens_seen": 6374680,
"step": 396
},
{
"epoch": 0.02780915355451067,
"grad_norm": 4.246788024902344,
"learning_rate": 9.722169527145359e-05,
"loss": 1.0161,
"num_input_tokens_seen": 6390672,
"step": 397
},
{
"epoch": 0.027879201800239915,
"grad_norm": 4.27274751663208,
"learning_rate": 9.721469702276708e-05,
"loss": 1.293,
"num_input_tokens_seen": 6407056,
"step": 398
},
{
"epoch": 0.02794925004596916,
"grad_norm": 4.171760559082031,
"learning_rate": 9.720769877408056e-05,
"loss": 1.2766,
"num_input_tokens_seen": 6423440,
"step": 399
},
{
"epoch": 0.028019298291698407,
"grad_norm": 4.174622535705566,
"learning_rate": 9.720070052539405e-05,
"loss": 1.049,
"num_input_tokens_seen": 6439824,
"step": 400
},
{
"epoch": 0.028019298291698407,
"eval_loss": 1.1994441747665405,
"eval_runtime": 0.2131,
"eval_samples_per_second": 4.693,
"eval_steps_per_second": 4.693,
"num_input_tokens_seen": 6439824,
"step": 400
},
{
"epoch": 0.028089346537427652,
"grad_norm": 4.199150562286377,
"learning_rate": 9.719370227670753e-05,
"loss": 1.3432,
"num_input_tokens_seen": 6456208,
"step": 401
},
{
"epoch": 0.028159394783156898,
"grad_norm": 3.9011733531951904,
"learning_rate": 9.718670402802102e-05,
"loss": 1.0895,
"num_input_tokens_seen": 6472592,
"step": 402
},
{
"epoch": 0.028229443028886147,
"grad_norm": 4.142306327819824,
"learning_rate": 9.717970577933451e-05,
"loss": 0.9031,
"num_input_tokens_seen": 6488976,
"step": 403
},
{
"epoch": 0.028299491274615392,
"grad_norm": 3.9745633602142334,
"learning_rate": 9.717270753064799e-05,
"loss": 0.9951,
"num_input_tokens_seen": 6505360,
"step": 404
},
{
"epoch": 0.028369539520344638,
"grad_norm": 3.838865280151367,
"learning_rate": 9.716570928196147e-05,
"loss": 0.809,
"num_input_tokens_seen": 6521744,
"step": 405
},
{
"epoch": 0.028439587766073884,
"grad_norm": 4.48146390914917,
"learning_rate": 9.715871103327496e-05,
"loss": 1.4985,
"num_input_tokens_seen": 6538128,
"step": 406
},
{
"epoch": 0.02850963601180313,
"grad_norm": 4.393556594848633,
"learning_rate": 9.715171278458844e-05,
"loss": 1.2355,
"num_input_tokens_seen": 6554512,
"step": 407
},
{
"epoch": 0.028579684257532375,
"grad_norm": 3.970860004425049,
"learning_rate": 9.714471453590194e-05,
"loss": 1.1513,
"num_input_tokens_seen": 6570896,
"step": 408
},
{
"epoch": 0.02864973250326162,
"grad_norm": 4.166610240936279,
"learning_rate": 9.713771628721542e-05,
"loss": 1.108,
"num_input_tokens_seen": 6587216,
"step": 409
},
{
"epoch": 0.028719780748990866,
"grad_norm": 3.9887096881866455,
"learning_rate": 9.71307180385289e-05,
"loss": 1.1639,
"num_input_tokens_seen": 6603600,
"step": 410
},
{
"epoch": 0.028789828994720112,
"grad_norm": 4.195802211761475,
"learning_rate": 9.712371978984239e-05,
"loss": 1.1478,
"num_input_tokens_seen": 6619984,
"step": 411
},
{
"epoch": 0.02885987724044936,
"grad_norm": 4.011331081390381,
"learning_rate": 9.711672154115587e-05,
"loss": 0.9554,
"num_input_tokens_seen": 6635904,
"step": 412
},
{
"epoch": 0.028929925486178606,
"grad_norm": 4.4170026779174805,
"learning_rate": 9.710972329246936e-05,
"loss": 1.1452,
"num_input_tokens_seen": 6651944,
"step": 413
},
{
"epoch": 0.028999973731907852,
"grad_norm": 4.073450088500977,
"learning_rate": 9.710272504378284e-05,
"loss": 1.1187,
"num_input_tokens_seen": 6668096,
"step": 414
},
{
"epoch": 0.029070021977637098,
"grad_norm": 4.161722183227539,
"learning_rate": 9.709572679509633e-05,
"loss": 1.1603,
"num_input_tokens_seen": 6684480,
"step": 415
},
{
"epoch": 0.029140070223366343,
"grad_norm": 4.540097713470459,
"learning_rate": 9.708872854640981e-05,
"loss": 1.2143,
"num_input_tokens_seen": 6700536,
"step": 416
},
{
"epoch": 0.02921011846909559,
"grad_norm": 4.030871868133545,
"learning_rate": 9.70817302977233e-05,
"loss": 0.9791,
"num_input_tokens_seen": 6716920,
"step": 417
},
{
"epoch": 0.029280166714824835,
"grad_norm": 4.1743268966674805,
"learning_rate": 9.707473204903679e-05,
"loss": 0.9818,
"num_input_tokens_seen": 6733304,
"step": 418
},
{
"epoch": 0.02935021496055408,
"grad_norm": 4.227272987365723,
"learning_rate": 9.706773380035026e-05,
"loss": 1.0945,
"num_input_tokens_seen": 6749688,
"step": 419
},
{
"epoch": 0.02942026320628333,
"grad_norm": 4.406428813934326,
"learning_rate": 9.706073555166375e-05,
"loss": 1.0302,
"num_input_tokens_seen": 6766072,
"step": 420
},
{
"epoch": 0.029490311452012575,
"grad_norm": 4.17899227142334,
"learning_rate": 9.705373730297724e-05,
"loss": 1.1048,
"num_input_tokens_seen": 6782456,
"step": 421
},
{
"epoch": 0.02956035969774182,
"grad_norm": 4.034752368927002,
"learning_rate": 9.704673905429073e-05,
"loss": 1.2639,
"num_input_tokens_seen": 6798840,
"step": 422
},
{
"epoch": 0.029630407943471066,
"grad_norm": 4.795727729797363,
"learning_rate": 9.703974080560421e-05,
"loss": 1.2448,
"num_input_tokens_seen": 6814912,
"step": 423
},
{
"epoch": 0.029700456189200312,
"grad_norm": 4.509056568145752,
"learning_rate": 9.703274255691769e-05,
"loss": 1.2157,
"num_input_tokens_seen": 6830720,
"step": 424
},
{
"epoch": 0.029770504434929557,
"grad_norm": 4.064620494842529,
"learning_rate": 9.702574430823118e-05,
"loss": 1.2042,
"num_input_tokens_seen": 6847104,
"step": 425
},
{
"epoch": 0.029840552680658803,
"grad_norm": 3.9060182571411133,
"learning_rate": 9.701874605954465e-05,
"loss": 0.9116,
"num_input_tokens_seen": 6862952,
"step": 426
},
{
"epoch": 0.02991060092638805,
"grad_norm": 3.9900951385498047,
"learning_rate": 9.701174781085814e-05,
"loss": 1.1621,
"num_input_tokens_seen": 6879336,
"step": 427
},
{
"epoch": 0.029980649172117294,
"grad_norm": 4.371436595916748,
"learning_rate": 9.700474956217164e-05,
"loss": 1.2731,
"num_input_tokens_seen": 6895720,
"step": 428
},
{
"epoch": 0.030050697417846543,
"grad_norm": 3.9422085285186768,
"learning_rate": 9.699775131348512e-05,
"loss": 0.9636,
"num_input_tokens_seen": 6912104,
"step": 429
},
{
"epoch": 0.03012074566357579,
"grad_norm": 4.080913543701172,
"learning_rate": 9.699075306479861e-05,
"loss": 1.1507,
"num_input_tokens_seen": 6928488,
"step": 430
},
{
"epoch": 0.030190793909305035,
"grad_norm": 4.493942737579346,
"learning_rate": 9.698375481611208e-05,
"loss": 1.2274,
"num_input_tokens_seen": 6944664,
"step": 431
},
{
"epoch": 0.03026084215503428,
"grad_norm": 4.073723793029785,
"learning_rate": 9.697675656742557e-05,
"loss": 1.0498,
"num_input_tokens_seen": 6960344,
"step": 432
},
{
"epoch": 0.030330890400763526,
"grad_norm": 3.9672274589538574,
"learning_rate": 9.696975831873906e-05,
"loss": 1.007,
"num_input_tokens_seen": 6976720,
"step": 433
},
{
"epoch": 0.03040093864649277,
"grad_norm": 4.497872829437256,
"learning_rate": 9.696276007005255e-05,
"loss": 1.1339,
"num_input_tokens_seen": 6992552,
"step": 434
},
{
"epoch": 0.030470986892222017,
"grad_norm": 4.422168731689453,
"learning_rate": 9.695576182136604e-05,
"loss": 1.34,
"num_input_tokens_seen": 7008936,
"step": 435
},
{
"epoch": 0.030541035137951263,
"grad_norm": 4.3009138107299805,
"learning_rate": 9.694876357267951e-05,
"loss": 1.2479,
"num_input_tokens_seen": 7024512,
"step": 436
},
{
"epoch": 0.030611083383680508,
"grad_norm": 4.04030704498291,
"learning_rate": 9.6941765323993e-05,
"loss": 1.097,
"num_input_tokens_seen": 7040896,
"step": 437
},
{
"epoch": 0.030681131629409757,
"grad_norm": 3.877417802810669,
"learning_rate": 9.693476707530649e-05,
"loss": 1.1363,
"num_input_tokens_seen": 7057280,
"step": 438
},
{
"epoch": 0.030751179875139003,
"grad_norm": 3.8185505867004395,
"learning_rate": 9.692776882661996e-05,
"loss": 0.9067,
"num_input_tokens_seen": 7072544,
"step": 439
},
{
"epoch": 0.03082122812086825,
"grad_norm": 4.028950214385986,
"learning_rate": 9.692077057793345e-05,
"loss": 1.1195,
"num_input_tokens_seen": 7088928,
"step": 440
},
{
"epoch": 0.030891276366597494,
"grad_norm": 4.2786431312561035,
"learning_rate": 9.691377232924694e-05,
"loss": 1.1199,
"num_input_tokens_seen": 7105248,
"step": 441
},
{
"epoch": 0.03096132461232674,
"grad_norm": 4.193462371826172,
"learning_rate": 9.690677408056043e-05,
"loss": 1.1812,
"num_input_tokens_seen": 7121008,
"step": 442
},
{
"epoch": 0.031031372858055985,
"grad_norm": 3.93597412109375,
"learning_rate": 9.68997758318739e-05,
"loss": 1.0677,
"num_input_tokens_seen": 7136944,
"step": 443
},
{
"epoch": 0.03110142110378523,
"grad_norm": 4.3208537101745605,
"learning_rate": 9.68927775831874e-05,
"loss": 1.1358,
"num_input_tokens_seen": 7152928,
"step": 444
},
{
"epoch": 0.031171469349514477,
"grad_norm": 3.9743378162384033,
"learning_rate": 9.688577933450088e-05,
"loss": 1.094,
"num_input_tokens_seen": 7169312,
"step": 445
},
{
"epoch": 0.031241517595243726,
"grad_norm": 4.226114273071289,
"learning_rate": 9.687878108581436e-05,
"loss": 1.1752,
"num_input_tokens_seen": 7185696,
"step": 446
},
{
"epoch": 0.03131156584097297,
"grad_norm": 4.210222244262695,
"learning_rate": 9.687178283712785e-05,
"loss": 1.1262,
"num_input_tokens_seen": 7201784,
"step": 447
},
{
"epoch": 0.03138161408670222,
"grad_norm": 4.311635971069336,
"learning_rate": 9.686478458844133e-05,
"loss": 1.2491,
"num_input_tokens_seen": 7218168,
"step": 448
},
{
"epoch": 0.03145166233243146,
"grad_norm": 4.56603479385376,
"learning_rate": 9.685778633975482e-05,
"loss": 1.3512,
"num_input_tokens_seen": 7233360,
"step": 449
},
{
"epoch": 0.03152171057816071,
"grad_norm": 4.232856750488281,
"learning_rate": 9.685078809106831e-05,
"loss": 0.9387,
"num_input_tokens_seen": 7248280,
"step": 450
},
{
"epoch": 0.031591758823889954,
"grad_norm": 4.512947082519531,
"learning_rate": 9.684378984238179e-05,
"loss": 1.1988,
"num_input_tokens_seen": 7264664,
"step": 451
},
{
"epoch": 0.0316618070696192,
"grad_norm": 4.273897171020508,
"learning_rate": 9.683679159369528e-05,
"loss": 1.2523,
"num_input_tokens_seen": 7281048,
"step": 452
},
{
"epoch": 0.031731855315348445,
"grad_norm": 4.288438320159912,
"learning_rate": 9.682979334500875e-05,
"loss": 1.1692,
"num_input_tokens_seen": 7297424,
"step": 453
},
{
"epoch": 0.03180190356107769,
"grad_norm": 4.27367639541626,
"learning_rate": 9.682279509632225e-05,
"loss": 1.1868,
"num_input_tokens_seen": 7312792,
"step": 454
},
{
"epoch": 0.031871951806806936,
"grad_norm": 3.978926181793213,
"learning_rate": 9.681579684763574e-05,
"loss": 1.0382,
"num_input_tokens_seen": 7329176,
"step": 455
},
{
"epoch": 0.03194200005253618,
"grad_norm": 4.4399919509887695,
"learning_rate": 9.680879859894922e-05,
"loss": 1.2072,
"num_input_tokens_seen": 7345560,
"step": 456
},
{
"epoch": 0.03201204829826543,
"grad_norm": 3.9786529541015625,
"learning_rate": 9.68018003502627e-05,
"loss": 1.1704,
"num_input_tokens_seen": 7361944,
"step": 457
},
{
"epoch": 0.03208209654399467,
"grad_norm": 4.171195030212402,
"learning_rate": 9.679480210157618e-05,
"loss": 1.1307,
"num_input_tokens_seen": 7378328,
"step": 458
},
{
"epoch": 0.032152144789723926,
"grad_norm": 3.9415268898010254,
"learning_rate": 9.678780385288967e-05,
"loss": 0.9971,
"num_input_tokens_seen": 7394208,
"step": 459
},
{
"epoch": 0.03222219303545317,
"grad_norm": 4.066036224365234,
"learning_rate": 9.678080560420316e-05,
"loss": 1.1227,
"num_input_tokens_seen": 7410328,
"step": 460
},
{
"epoch": 0.03229224128118242,
"grad_norm": 4.22513484954834,
"learning_rate": 9.677380735551665e-05,
"loss": 1.0883,
"num_input_tokens_seen": 7426712,
"step": 461
},
{
"epoch": 0.03236228952691166,
"grad_norm": 4.310954570770264,
"learning_rate": 9.676680910683013e-05,
"loss": 1.1695,
"num_input_tokens_seen": 7442736,
"step": 462
},
{
"epoch": 0.03243233777264091,
"grad_norm": 4.2868828773498535,
"learning_rate": 9.675981085814361e-05,
"loss": 1.0594,
"num_input_tokens_seen": 7458560,
"step": 463
},
{
"epoch": 0.032502386018370154,
"grad_norm": 4.318186283111572,
"learning_rate": 9.67528126094571e-05,
"loss": 1.1791,
"num_input_tokens_seen": 7474944,
"step": 464
},
{
"epoch": 0.0325724342640994,
"grad_norm": 4.040421009063721,
"learning_rate": 9.674581436077059e-05,
"loss": 1.0649,
"num_input_tokens_seen": 7490344,
"step": 465
},
{
"epoch": 0.032642482509828645,
"grad_norm": 3.914815902709961,
"learning_rate": 9.673881611208406e-05,
"loss": 1.1381,
"num_input_tokens_seen": 7506728,
"step": 466
},
{
"epoch": 0.03271253075555789,
"grad_norm": 4.054527282714844,
"learning_rate": 9.673181786339755e-05,
"loss": 1.2264,
"num_input_tokens_seen": 7522912,
"step": 467
},
{
"epoch": 0.032782579001287136,
"grad_norm": 4.295147895812988,
"learning_rate": 9.672481961471104e-05,
"loss": 1.1369,
"num_input_tokens_seen": 7539040,
"step": 468
},
{
"epoch": 0.03285262724701638,
"grad_norm": 4.109183311462402,
"learning_rate": 9.671782136602453e-05,
"loss": 1.1676,
"num_input_tokens_seen": 7555424,
"step": 469
},
{
"epoch": 0.03292267549274563,
"grad_norm": 4.131369590759277,
"learning_rate": 9.6710823117338e-05,
"loss": 1.1188,
"num_input_tokens_seen": 7571808,
"step": 470
},
{
"epoch": 0.03299272373847487,
"grad_norm": 3.998414993286133,
"learning_rate": 9.670382486865149e-05,
"loss": 1.0201,
"num_input_tokens_seen": 7587528,
"step": 471
},
{
"epoch": 0.03306277198420412,
"grad_norm": 4.1235551834106445,
"learning_rate": 9.669682661996498e-05,
"loss": 1.1265,
"num_input_tokens_seen": 7603912,
"step": 472
},
{
"epoch": 0.033132820229933364,
"grad_norm": 4.800798416137695,
"learning_rate": 9.668982837127845e-05,
"loss": 1.3634,
"num_input_tokens_seen": 7617512,
"step": 473
},
{
"epoch": 0.03320286847566261,
"grad_norm": 4.068000316619873,
"learning_rate": 9.668283012259196e-05,
"loss": 1.1427,
"num_input_tokens_seen": 7633040,
"step": 474
},
{
"epoch": 0.033272916721391856,
"grad_norm": 4.0715484619140625,
"learning_rate": 9.667583187390543e-05,
"loss": 1.0633,
"num_input_tokens_seen": 7648416,
"step": 475
},
{
"epoch": 0.0333429649671211,
"grad_norm": 3.937807321548462,
"learning_rate": 9.666883362521892e-05,
"loss": 1.1393,
"num_input_tokens_seen": 7664624,
"step": 476
},
{
"epoch": 0.033413013212850354,
"grad_norm": 4.195656776428223,
"learning_rate": 9.666183537653241e-05,
"loss": 1.1801,
"num_input_tokens_seen": 7680480,
"step": 477
},
{
"epoch": 0.0334830614585796,
"grad_norm": 4.227575778961182,
"learning_rate": 9.665483712784588e-05,
"loss": 1.0453,
"num_input_tokens_seen": 7696632,
"step": 478
},
{
"epoch": 0.033553109704308845,
"grad_norm": 4.328822135925293,
"learning_rate": 9.664783887915937e-05,
"loss": 1.221,
"num_input_tokens_seen": 7713016,
"step": 479
},
{
"epoch": 0.03362315795003809,
"grad_norm": 4.086736679077148,
"learning_rate": 9.664084063047286e-05,
"loss": 1.2817,
"num_input_tokens_seen": 7729400,
"step": 480
},
{
"epoch": 0.033693206195767336,
"grad_norm": 4.555233955383301,
"learning_rate": 9.663384238178635e-05,
"loss": 1.483,
"num_input_tokens_seen": 7745784,
"step": 481
},
{
"epoch": 0.03376325444149658,
"grad_norm": 4.118983745574951,
"learning_rate": 9.662684413309984e-05,
"loss": 0.9139,
"num_input_tokens_seen": 7762168,
"step": 482
},
{
"epoch": 0.03383330268722583,
"grad_norm": 4.232059001922607,
"learning_rate": 9.661984588441331e-05,
"loss": 1.1269,
"num_input_tokens_seen": 7777920,
"step": 483
},
{
"epoch": 0.03390335093295507,
"grad_norm": 6.288865089416504,
"learning_rate": 9.66128476357268e-05,
"loss": 1.0642,
"num_input_tokens_seen": 7794304,
"step": 484
},
{
"epoch": 0.03397339917868432,
"grad_norm": 4.133046627044678,
"learning_rate": 9.660584938704028e-05,
"loss": 1.2067,
"num_input_tokens_seen": 7810200,
"step": 485
},
{
"epoch": 0.034043447424413564,
"grad_norm": 4.147965431213379,
"learning_rate": 9.659885113835377e-05,
"loss": 1.0367,
"num_input_tokens_seen": 7826384,
"step": 486
},
{
"epoch": 0.03411349567014281,
"grad_norm": 4.1191020011901855,
"learning_rate": 9.659185288966725e-05,
"loss": 1.0972,
"num_input_tokens_seen": 7841704,
"step": 487
},
{
"epoch": 0.034183543915872056,
"grad_norm": 4.518441677093506,
"learning_rate": 9.658485464098074e-05,
"loss": 1.263,
"num_input_tokens_seen": 7858088,
"step": 488
},
{
"epoch": 0.0342535921616013,
"grad_norm": 4.321181297302246,
"learning_rate": 9.657785639229423e-05,
"loss": 1.1378,
"num_input_tokens_seen": 7874472,
"step": 489
},
{
"epoch": 0.03432364040733055,
"grad_norm": 4.366185665130615,
"learning_rate": 9.65708581436077e-05,
"loss": 1.1636,
"num_input_tokens_seen": 7890856,
"step": 490
},
{
"epoch": 0.03439368865305979,
"grad_norm": 4.042731761932373,
"learning_rate": 9.65638598949212e-05,
"loss": 1.0601,
"num_input_tokens_seen": 7906776,
"step": 491
},
{
"epoch": 0.03446373689878904,
"grad_norm": 3.743668556213379,
"learning_rate": 9.655686164623468e-05,
"loss": 1.0441,
"num_input_tokens_seen": 7923160,
"step": 492
},
{
"epoch": 0.034533785144518284,
"grad_norm": 3.8547139167785645,
"learning_rate": 9.654986339754816e-05,
"loss": 1.0842,
"num_input_tokens_seen": 7939296,
"step": 493
},
{
"epoch": 0.034603833390247536,
"grad_norm": 4.238414287567139,
"learning_rate": 9.654286514886166e-05,
"loss": 1.2498,
"num_input_tokens_seen": 7955504,
"step": 494
},
{
"epoch": 0.03467388163597678,
"grad_norm": 4.134857177734375,
"learning_rate": 9.653586690017514e-05,
"loss": 1.1241,
"num_input_tokens_seen": 7971888,
"step": 495
},
{
"epoch": 0.03474392988170603,
"grad_norm": 4.2501983642578125,
"learning_rate": 9.652886865148862e-05,
"loss": 1.1829,
"num_input_tokens_seen": 7988272,
"step": 496
},
{
"epoch": 0.03481397812743527,
"grad_norm": 7.4397053718566895,
"learning_rate": 9.65218704028021e-05,
"loss": 0.9952,
"num_input_tokens_seen": 8003744,
"step": 497
},
{
"epoch": 0.03488402637316452,
"grad_norm": 4.2750959396362305,
"learning_rate": 9.651487215411559e-05,
"loss": 1.2387,
"num_input_tokens_seen": 8019184,
"step": 498
},
{
"epoch": 0.034954074618893764,
"grad_norm": 4.156162261962891,
"learning_rate": 9.650787390542908e-05,
"loss": 1.1201,
"num_input_tokens_seen": 8035176,
"step": 499
},
{
"epoch": 0.03502412286462301,
"grad_norm": 4.178225040435791,
"learning_rate": 9.650087565674257e-05,
"loss": 1.2026,
"num_input_tokens_seen": 8051560,
"step": 500
},
{
"epoch": 0.035094171110352256,
"grad_norm": 4.147096157073975,
"learning_rate": 9.649387740805605e-05,
"loss": 1.2465,
"num_input_tokens_seen": 8067944,
"step": 501
},
{
"epoch": 0.0351642193560815,
"grad_norm": 4.329249858856201,
"learning_rate": 9.648687915936953e-05,
"loss": 1.2742,
"num_input_tokens_seen": 8083824,
"step": 502
},
{
"epoch": 0.03523426760181075,
"grad_norm": 4.404232978820801,
"learning_rate": 9.647988091068302e-05,
"loss": 1.1511,
"num_input_tokens_seen": 8100208,
"step": 503
},
{
"epoch": 0.03530431584753999,
"grad_norm": 4.190586090087891,
"learning_rate": 9.64728826619965e-05,
"loss": 0.9884,
"num_input_tokens_seen": 8116048,
"step": 504
},
{
"epoch": 0.03537436409326924,
"grad_norm": 4.262845516204834,
"learning_rate": 9.646588441330998e-05,
"loss": 1.1321,
"num_input_tokens_seen": 8132432,
"step": 505
},
{
"epoch": 0.035444412338998484,
"grad_norm": 4.452746391296387,
"learning_rate": 9.645888616462347e-05,
"loss": 1.1667,
"num_input_tokens_seen": 8148816,
"step": 506
},
{
"epoch": 0.03551446058472773,
"grad_norm": 4.111443042755127,
"learning_rate": 9.645188791593696e-05,
"loss": 1.0049,
"num_input_tokens_seen": 8164856,
"step": 507
},
{
"epoch": 0.035584508830456975,
"grad_norm": 4.292227268218994,
"learning_rate": 9.644488966725045e-05,
"loss": 1.1535,
"num_input_tokens_seen": 8181240,
"step": 508
},
{
"epoch": 0.03565455707618622,
"grad_norm": 4.295238971710205,
"learning_rate": 9.643789141856394e-05,
"loss": 1.236,
"num_input_tokens_seen": 8197624,
"step": 509
},
{
"epoch": 0.035724605321915466,
"grad_norm": 3.930659294128418,
"learning_rate": 9.643089316987741e-05,
"loss": 0.9195,
"num_input_tokens_seen": 8213816,
"step": 510
},
{
"epoch": 0.03579465356764472,
"grad_norm": 4.092316150665283,
"learning_rate": 9.64238949211909e-05,
"loss": 1.0799,
"num_input_tokens_seen": 8229632,
"step": 511
},
{
"epoch": 0.035864701813373964,
"grad_norm": 4.2939252853393555,
"learning_rate": 9.641689667250437e-05,
"loss": 1.111,
"num_input_tokens_seen": 8245232,
"step": 512
},
{
"epoch": 0.03593475005910321,
"grad_norm": 4.191503524780273,
"learning_rate": 9.640989842381786e-05,
"loss": 0.9399,
"num_input_tokens_seen": 8260912,
"step": 513
},
{
"epoch": 0.036004798304832455,
"grad_norm": 4.141485214233398,
"learning_rate": 9.640290017513136e-05,
"loss": 1.1334,
"num_input_tokens_seen": 8276864,
"step": 514
},
{
"epoch": 0.0360748465505617,
"grad_norm": 3.890547752380371,
"learning_rate": 9.639590192644484e-05,
"loss": 1.0055,
"num_input_tokens_seen": 8292720,
"step": 515
},
{
"epoch": 0.03614489479629095,
"grad_norm": 4.405922889709473,
"learning_rate": 9.638890367775833e-05,
"loss": 1.2238,
"num_input_tokens_seen": 8309104,
"step": 516
},
{
"epoch": 0.03621494304202019,
"grad_norm": 4.207942485809326,
"learning_rate": 9.63819054290718e-05,
"loss": 1.0688,
"num_input_tokens_seen": 8325304,
"step": 517
},
{
"epoch": 0.03628499128774944,
"grad_norm": 4.174366474151611,
"learning_rate": 9.637490718038529e-05,
"loss": 1.2303,
"num_input_tokens_seen": 8341688,
"step": 518
},
{
"epoch": 0.036355039533478684,
"grad_norm": 3.9641714096069336,
"learning_rate": 9.636790893169878e-05,
"loss": 1.2244,
"num_input_tokens_seen": 8357760,
"step": 519
},
{
"epoch": 0.03642508777920793,
"grad_norm": 5.832678318023682,
"learning_rate": 9.636091068301227e-05,
"loss": 1.0645,
"num_input_tokens_seen": 8372712,
"step": 520
},
{
"epoch": 0.036495136024937175,
"grad_norm": 3.7905161380767822,
"learning_rate": 9.635391243432576e-05,
"loss": 1.0551,
"num_input_tokens_seen": 8389096,
"step": 521
},
{
"epoch": 0.03656518427066642,
"grad_norm": 3.6744072437286377,
"learning_rate": 9.634691418563923e-05,
"loss": 1.0687,
"num_input_tokens_seen": 8405216,
"step": 522
},
{
"epoch": 0.036635232516395666,
"grad_norm": 4.897486209869385,
"learning_rate": 9.633991593695272e-05,
"loss": 1.1968,
"num_input_tokens_seen": 8421600,
"step": 523
},
{
"epoch": 0.03670528076212491,
"grad_norm": 3.821457862854004,
"learning_rate": 9.63329176882662e-05,
"loss": 1.0473,
"num_input_tokens_seen": 8437984,
"step": 524
},
{
"epoch": 0.03677532900785416,
"grad_norm": 3.873832941055298,
"learning_rate": 9.632591943957969e-05,
"loss": 0.9656,
"num_input_tokens_seen": 8453760,
"step": 525
},
{
"epoch": 0.0368453772535834,
"grad_norm": 4.139901161193848,
"learning_rate": 9.631892119089317e-05,
"loss": 1.0881,
"num_input_tokens_seen": 8470144,
"step": 526
},
{
"epoch": 0.03691542549931265,
"grad_norm": 3.9512782096862793,
"learning_rate": 9.631192294220666e-05,
"loss": 1.1093,
"num_input_tokens_seen": 8486528,
"step": 527
},
{
"epoch": 0.0369854737450419,
"grad_norm": 3.8937103748321533,
"learning_rate": 9.630492469352015e-05,
"loss": 0.9722,
"num_input_tokens_seen": 8502912,
"step": 528
},
{
"epoch": 0.03705552199077115,
"grad_norm": 4.482640743255615,
"learning_rate": 9.629792644483363e-05,
"loss": 1.056,
"num_input_tokens_seen": 8519296,
"step": 529
},
{
"epoch": 0.03712557023650039,
"grad_norm": 4.127941131591797,
"learning_rate": 9.629092819614711e-05,
"loss": 1.0285,
"num_input_tokens_seen": 8535160,
"step": 530
},
{
"epoch": 0.03719561848222964,
"grad_norm": 3.973585844039917,
"learning_rate": 9.62839299474606e-05,
"loss": 1.0356,
"num_input_tokens_seen": 8551256,
"step": 531
},
{
"epoch": 0.037265666727958884,
"grad_norm": 4.22855281829834,
"learning_rate": 9.627693169877408e-05,
"loss": 1.134,
"num_input_tokens_seen": 8567640,
"step": 532
},
{
"epoch": 0.03733571497368813,
"grad_norm": 4.144021511077881,
"learning_rate": 9.626993345008757e-05,
"loss": 1.0963,
"num_input_tokens_seen": 8583504,
"step": 533
},
{
"epoch": 0.037405763219417375,
"grad_norm": 3.8666226863861084,
"learning_rate": 9.626293520140106e-05,
"loss": 0.912,
"num_input_tokens_seen": 8599888,
"step": 534
},
{
"epoch": 0.03747581146514662,
"grad_norm": 4.215412616729736,
"learning_rate": 9.625593695271454e-05,
"loss": 1.1055,
"num_input_tokens_seen": 8616256,
"step": 535
},
{
"epoch": 0.037545859710875866,
"grad_norm": 4.353022575378418,
"learning_rate": 9.624893870402803e-05,
"loss": 1.0379,
"num_input_tokens_seen": 8632640,
"step": 536
},
{
"epoch": 0.03761590795660511,
"grad_norm": 3.778947591781616,
"learning_rate": 9.624194045534151e-05,
"loss": 1.0547,
"num_input_tokens_seen": 8648624,
"step": 537
},
{
"epoch": 0.03768595620233436,
"grad_norm": 4.481568336486816,
"learning_rate": 9.6234942206655e-05,
"loss": 1.3407,
"num_input_tokens_seen": 8664200,
"step": 538
},
{
"epoch": 0.0377560044480636,
"grad_norm": 4.066302299499512,
"learning_rate": 9.622794395796847e-05,
"loss": 0.995,
"num_input_tokens_seen": 8680584,
"step": 539
},
{
"epoch": 0.03782605269379285,
"grad_norm": 4.262768268585205,
"learning_rate": 9.622094570928197e-05,
"loss": 1.3054,
"num_input_tokens_seen": 8696968,
"step": 540
},
{
"epoch": 0.037896100939522094,
"grad_norm": 3.777597665786743,
"learning_rate": 9.621394746059546e-05,
"loss": 0.9831,
"num_input_tokens_seen": 8713352,
"step": 541
},
{
"epoch": 0.03796614918525134,
"grad_norm": 3.9732742309570312,
"learning_rate": 9.620694921190894e-05,
"loss": 1.0699,
"num_input_tokens_seen": 8729048,
"step": 542
},
{
"epoch": 0.038036197430980585,
"grad_norm": 4.543329238891602,
"learning_rate": 9.619995096322243e-05,
"loss": 1.1546,
"num_input_tokens_seen": 8745432,
"step": 543
},
{
"epoch": 0.03810624567670983,
"grad_norm": 4.903865814208984,
"learning_rate": 9.61929527145359e-05,
"loss": 1.1548,
"num_input_tokens_seen": 8760296,
"step": 544
},
{
"epoch": 0.03817629392243908,
"grad_norm": 4.197691917419434,
"learning_rate": 9.618595446584939e-05,
"loss": 1.1616,
"num_input_tokens_seen": 8776680,
"step": 545
},
{
"epoch": 0.03824634216816833,
"grad_norm": 3.912689208984375,
"learning_rate": 9.617895621716288e-05,
"loss": 0.9926,
"num_input_tokens_seen": 8793064,
"step": 546
},
{
"epoch": 0.038316390413897575,
"grad_norm": 4.291840076446533,
"learning_rate": 9.617195796847637e-05,
"loss": 1.1943,
"num_input_tokens_seen": 8809448,
"step": 547
},
{
"epoch": 0.03838643865962682,
"grad_norm": 3.9053072929382324,
"learning_rate": 9.616495971978985e-05,
"loss": 1.2437,
"num_input_tokens_seen": 8825536,
"step": 548
},
{
"epoch": 0.038456486905356066,
"grad_norm": 4.860696315765381,
"learning_rate": 9.615796147110333e-05,
"loss": 1.3045,
"num_input_tokens_seen": 8841920,
"step": 549
},
{
"epoch": 0.03852653515108531,
"grad_norm": 3.9394373893737793,
"learning_rate": 9.615096322241682e-05,
"loss": 1.1367,
"num_input_tokens_seen": 8858304,
"step": 550
},
{
"epoch": 0.03859658339681456,
"grad_norm": 3.8160409927368164,
"learning_rate": 9.61439649737303e-05,
"loss": 1.0864,
"num_input_tokens_seen": 8874688,
"step": 551
},
{
"epoch": 0.0386666316425438,
"grad_norm": 4.3792805671691895,
"learning_rate": 9.613696672504378e-05,
"loss": 1.2516,
"num_input_tokens_seen": 8891072,
"step": 552
},
{
"epoch": 0.03873667988827305,
"grad_norm": 4.103452682495117,
"learning_rate": 9.612996847635727e-05,
"loss": 0.9737,
"num_input_tokens_seen": 8907456,
"step": 553
},
{
"epoch": 0.038806728134002294,
"grad_norm": 4.117603302001953,
"learning_rate": 9.612297022767076e-05,
"loss": 1.096,
"num_input_tokens_seen": 8923816,
"step": 554
},
{
"epoch": 0.03887677637973154,
"grad_norm": 4.272468566894531,
"learning_rate": 9.611597197898425e-05,
"loss": 1.161,
"num_input_tokens_seen": 8939344,
"step": 555
},
{
"epoch": 0.038946824625460785,
"grad_norm": 4.323635578155518,
"learning_rate": 9.610897373029772e-05,
"loss": 1.1922,
"num_input_tokens_seen": 8954920,
"step": 556
},
{
"epoch": 0.03901687287119003,
"grad_norm": 3.783510684967041,
"learning_rate": 9.610197548161121e-05,
"loss": 1.0658,
"num_input_tokens_seen": 8971304,
"step": 557
},
{
"epoch": 0.039086921116919277,
"grad_norm": 4.3757548332214355,
"learning_rate": 9.60949772329247e-05,
"loss": 1.3186,
"num_input_tokens_seen": 8987672,
"step": 558
},
{
"epoch": 0.03915696936264852,
"grad_norm": 4.048824787139893,
"learning_rate": 9.608797898423818e-05,
"loss": 1.1452,
"num_input_tokens_seen": 9003896,
"step": 559
},
{
"epoch": 0.03922701760837777,
"grad_norm": 4.06865930557251,
"learning_rate": 9.608098073555168e-05,
"loss": 0.9861,
"num_input_tokens_seen": 9020280,
"step": 560
},
{
"epoch": 0.03929706585410701,
"grad_norm": 3.966737747192383,
"learning_rate": 9.607398248686515e-05,
"loss": 1.0323,
"num_input_tokens_seen": 9036280,
"step": 561
},
{
"epoch": 0.03936711409983626,
"grad_norm": 4.466656684875488,
"learning_rate": 9.606698423817864e-05,
"loss": 1.2462,
"num_input_tokens_seen": 9052664,
"step": 562
},
{
"epoch": 0.03943716234556551,
"grad_norm": 4.312132358551025,
"learning_rate": 9.605998598949213e-05,
"loss": 1.2133,
"num_input_tokens_seen": 9068832,
"step": 563
},
{
"epoch": 0.03950721059129476,
"grad_norm": 3.9202895164489746,
"learning_rate": 9.60529877408056e-05,
"loss": 1.0723,
"num_input_tokens_seen": 9084680,
"step": 564
},
{
"epoch": 0.039577258837024,
"grad_norm": 5.139899730682373,
"learning_rate": 9.604598949211909e-05,
"loss": 1.1165,
"num_input_tokens_seen": 9099792,
"step": 565
},
{
"epoch": 0.03964730708275325,
"grad_norm": 4.398557186126709,
"learning_rate": 9.603899124343258e-05,
"loss": 1.1737,
"num_input_tokens_seen": 9116136,
"step": 566
},
{
"epoch": 0.039717355328482494,
"grad_norm": 4.350982666015625,
"learning_rate": 9.603199299474607e-05,
"loss": 1.2174,
"num_input_tokens_seen": 9132520,
"step": 567
},
{
"epoch": 0.03978740357421174,
"grad_norm": 3.787644386291504,
"learning_rate": 9.602499474605956e-05,
"loss": 0.9914,
"num_input_tokens_seen": 9148856,
"step": 568
},
{
"epoch": 0.039857451819940985,
"grad_norm": 4.630245685577393,
"learning_rate": 9.601799649737303e-05,
"loss": 1.4135,
"num_input_tokens_seen": 9164888,
"step": 569
},
{
"epoch": 0.03992750006567023,
"grad_norm": 4.063969135284424,
"learning_rate": 9.601099824868652e-05,
"loss": 1.1312,
"num_input_tokens_seen": 9181272,
"step": 570
},
{
"epoch": 0.039997548311399476,
"grad_norm": 4.2443413734436035,
"learning_rate": 9.6004e-05,
"loss": 1.1627,
"num_input_tokens_seen": 9197344,
"step": 571
},
{
"epoch": 0.04006759655712872,
"grad_norm": 4.396352767944336,
"learning_rate": 9.599700175131349e-05,
"loss": 1.1222,
"num_input_tokens_seen": 9212312,
"step": 572
},
{
"epoch": 0.04013764480285797,
"grad_norm": 4.364585876464844,
"learning_rate": 9.599000350262697e-05,
"loss": 1.0522,
"num_input_tokens_seen": 9228696,
"step": 573
},
{
"epoch": 0.04020769304858721,
"grad_norm": 3.9348409175872803,
"learning_rate": 9.598300525394046e-05,
"loss": 1.1375,
"num_input_tokens_seen": 9245080,
"step": 574
},
{
"epoch": 0.04027774129431646,
"grad_norm": 4.051416873931885,
"learning_rate": 9.597600700525395e-05,
"loss": 1.0265,
"num_input_tokens_seen": 9260752,
"step": 575
},
{
"epoch": 0.040347789540045705,
"grad_norm": 4.661770820617676,
"learning_rate": 9.596900875656743e-05,
"loss": 1.192,
"num_input_tokens_seen": 9276792,
"step": 576
},
{
"epoch": 0.04041783778577495,
"grad_norm": 4.378422260284424,
"learning_rate": 9.596201050788092e-05,
"loss": 1.0497,
"num_input_tokens_seen": 9292768,
"step": 577
},
{
"epoch": 0.040487886031504196,
"grad_norm": 4.4690399169921875,
"learning_rate": 9.595501225919439e-05,
"loss": 1.2398,
"num_input_tokens_seen": 9309152,
"step": 578
},
{
"epoch": 0.04055793427723344,
"grad_norm": 4.1711273193359375,
"learning_rate": 9.594801401050788e-05,
"loss": 1.097,
"num_input_tokens_seen": 9325536,
"step": 579
},
{
"epoch": 0.040627982522962694,
"grad_norm": 3.8115949630737305,
"learning_rate": 9.594101576182137e-05,
"loss": 1.0317,
"num_input_tokens_seen": 9341920,
"step": 580
},
{
"epoch": 0.04069803076869194,
"grad_norm": 4.072190284729004,
"learning_rate": 9.593401751313486e-05,
"loss": 1.0649,
"num_input_tokens_seen": 9357904,
"step": 581
},
{
"epoch": 0.040768079014421185,
"grad_norm": 3.895766258239746,
"learning_rate": 9.592701926444835e-05,
"loss": 1.1906,
"num_input_tokens_seen": 9373496,
"step": 582
},
{
"epoch": 0.04083812726015043,
"grad_norm": 4.026490688323975,
"learning_rate": 9.592002101576182e-05,
"loss": 0.9913,
"num_input_tokens_seen": 9389824,
"step": 583
},
{
"epoch": 0.040908175505879676,
"grad_norm": 3.612987518310547,
"learning_rate": 9.591302276707531e-05,
"loss": 0.9376,
"num_input_tokens_seen": 9406208,
"step": 584
},
{
"epoch": 0.04097822375160892,
"grad_norm": 4.4619646072387695,
"learning_rate": 9.59060245183888e-05,
"loss": 1.2198,
"num_input_tokens_seen": 9422592,
"step": 585
},
{
"epoch": 0.04104827199733817,
"grad_norm": 3.990372896194458,
"learning_rate": 9.589902626970229e-05,
"loss": 1.082,
"num_input_tokens_seen": 9438816,
"step": 586
},
{
"epoch": 0.04111832024306741,
"grad_norm": 3.7697947025299072,
"learning_rate": 9.589202802101577e-05,
"loss": 1.0173,
"num_input_tokens_seen": 9455200,
"step": 587
},
{
"epoch": 0.04118836848879666,
"grad_norm": 4.066056728363037,
"learning_rate": 9.588502977232925e-05,
"loss": 1.124,
"num_input_tokens_seen": 9471320,
"step": 588
},
{
"epoch": 0.041258416734525905,
"grad_norm": 3.913506507873535,
"learning_rate": 9.587803152364274e-05,
"loss": 1.0501,
"num_input_tokens_seen": 9487304,
"step": 589
},
{
"epoch": 0.04132846498025515,
"grad_norm": 3.9049429893493652,
"learning_rate": 9.587103327495623e-05,
"loss": 1.0563,
"num_input_tokens_seen": 9503688,
"step": 590
},
{
"epoch": 0.041398513225984396,
"grad_norm": 4.316978454589844,
"learning_rate": 9.58640350262697e-05,
"loss": 1.1333,
"num_input_tokens_seen": 9519488,
"step": 591
},
{
"epoch": 0.04146856147171364,
"grad_norm": 3.7818517684936523,
"learning_rate": 9.585703677758319e-05,
"loss": 1.0537,
"num_input_tokens_seen": 9535872,
"step": 592
},
{
"epoch": 0.04153860971744289,
"grad_norm": 3.8751401901245117,
"learning_rate": 9.585003852889668e-05,
"loss": 1.1745,
"num_input_tokens_seen": 9551928,
"step": 593
},
{
"epoch": 0.04160865796317213,
"grad_norm": 4.357265949249268,
"learning_rate": 9.584304028021017e-05,
"loss": 1.1154,
"num_input_tokens_seen": 9568312,
"step": 594
},
{
"epoch": 0.04167870620890138,
"grad_norm": 4.184159755706787,
"learning_rate": 9.583604203152366e-05,
"loss": 1.125,
"num_input_tokens_seen": 9583968,
"step": 595
},
{
"epoch": 0.041748754454630624,
"grad_norm": 3.9540369510650635,
"learning_rate": 9.582904378283713e-05,
"loss": 1.2032,
"num_input_tokens_seen": 9600152,
"step": 596
},
{
"epoch": 0.04181880270035987,
"grad_norm": 4.401122093200684,
"learning_rate": 9.582204553415062e-05,
"loss": 1.4808,
"num_input_tokens_seen": 9615632,
"step": 597
},
{
"epoch": 0.04188885094608912,
"grad_norm": 4.418131351470947,
"learning_rate": 9.58150472854641e-05,
"loss": 1.0077,
"num_input_tokens_seen": 9631712,
"step": 598
},
{
"epoch": 0.04195889919181837,
"grad_norm": 4.362226963043213,
"learning_rate": 9.580804903677758e-05,
"loss": 1.1614,
"num_input_tokens_seen": 9648096,
"step": 599
},
{
"epoch": 0.04202894743754761,
"grad_norm": 4.051177024841309,
"learning_rate": 9.580105078809107e-05,
"loss": 1.0718,
"num_input_tokens_seen": 9663792,
"step": 600
},
{
"epoch": 0.04202894743754761,
"eval_loss": 1.1809133291244507,
"eval_runtime": 0.2062,
"eval_samples_per_second": 4.849,
"eval_steps_per_second": 4.849,
"num_input_tokens_seen": 9663792,
"step": 600
}
],
"logging_steps": 1,
"max_steps": 14275,
"num_input_tokens_seen": 9663792,
"num_train_epochs": 1,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.0751917469364224e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}