other / trainer_state.json
Yhhhhhhhhh's picture
Upload folder using huggingface_hub
7baf389 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.0,
"eval_steps": 50000,
"global_step": 2364,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01692047377326565,
"grad_norm": 4.428616046905518,
"learning_rate": 7.042253521126762e-07,
"loss": 0.6706,
"step": 10
},
{
"epoch": 0.0338409475465313,
"grad_norm": 6.486809730529785,
"learning_rate": 1.4084507042253523e-06,
"loss": 0.6483,
"step": 20
},
{
"epoch": 0.050761421319796954,
"grad_norm": 3.8806605339050293,
"learning_rate": 2.1126760563380285e-06,
"loss": 0.5231,
"step": 30
},
{
"epoch": 0.0676818950930626,
"grad_norm": 2.814377784729004,
"learning_rate": 2.8169014084507046e-06,
"loss": 0.3641,
"step": 40
},
{
"epoch": 0.08460236886632826,
"grad_norm": 2.3236453533172607,
"learning_rate": 3.5211267605633804e-06,
"loss": 0.2923,
"step": 50
},
{
"epoch": 0.10152284263959391,
"grad_norm": 1.732696294784546,
"learning_rate": 4.225352112676057e-06,
"loss": 0.2632,
"step": 60
},
{
"epoch": 0.11844331641285956,
"grad_norm": 1.71094810962677,
"learning_rate": 4.929577464788733e-06,
"loss": 0.2672,
"step": 70
},
{
"epoch": 0.1353637901861252,
"grad_norm": 2.0413691997528076,
"learning_rate": 4.999809944183634e-06,
"loss": 0.2208,
"step": 80
},
{
"epoch": 0.15228426395939088,
"grad_norm": 2.169725179672241,
"learning_rate": 4.999152998213973e-06,
"loss": 0.252,
"step": 90
},
{
"epoch": 0.1692047377326565,
"grad_norm": 1.3290756940841675,
"learning_rate": 4.998026939008151e-06,
"loss": 0.2468,
"step": 100
},
{
"epoch": 0.18612521150592218,
"grad_norm": 2.0767831802368164,
"learning_rate": 4.996431977937491e-06,
"loss": 0.2468,
"step": 110
},
{
"epoch": 0.20304568527918782,
"grad_norm": 1.3648345470428467,
"learning_rate": 4.994368414390368e-06,
"loss": 0.2068,
"step": 120
},
{
"epoch": 0.21996615905245348,
"grad_norm": 1.9443793296813965,
"learning_rate": 4.991836635716003e-06,
"loss": 0.233,
"step": 130
},
{
"epoch": 0.23688663282571912,
"grad_norm": 1.4086923599243164,
"learning_rate": 4.988837117151762e-06,
"loss": 0.2153,
"step": 140
},
{
"epoch": 0.25380710659898476,
"grad_norm": 1.4785760641098022,
"learning_rate": 4.985370421733948e-06,
"loss": 0.2316,
"step": 150
},
{
"epoch": 0.2707275803722504,
"grad_norm": 1.52463960647583,
"learning_rate": 4.981437200192111e-06,
"loss": 0.215,
"step": 160
},
{
"epoch": 0.2876480541455161,
"grad_norm": 1.5714839696884155,
"learning_rate": 4.9770381908269005e-06,
"loss": 0.2366,
"step": 170
},
{
"epoch": 0.30456852791878175,
"grad_norm": 1.8277614116668701,
"learning_rate": 4.972174219371484e-06,
"loss": 0.2443,
"step": 180
},
{
"epoch": 0.32148900169204736,
"grad_norm": 1.175654411315918,
"learning_rate": 4.966846198836546e-06,
"loss": 0.2226,
"step": 190
},
{
"epoch": 0.338409475465313,
"grad_norm": 1.5629938840866089,
"learning_rate": 4.96105512933891e-06,
"loss": 0.2231,
"step": 200
},
{
"epoch": 0.3553299492385787,
"grad_norm": 1.6523264646530151,
"learning_rate": 4.954802097913804e-06,
"loss": 0.2193,
"step": 210
},
{
"epoch": 0.37225042301184436,
"grad_norm": 1.222960114479065,
"learning_rate": 4.94808827831082e-06,
"loss": 0.2534,
"step": 220
},
{
"epoch": 0.38917089678510997,
"grad_norm": 1.26102614402771,
"learning_rate": 4.9409149307735835e-06,
"loss": 0.2283,
"step": 230
},
{
"epoch": 0.40609137055837563,
"grad_norm": 1.4353593587875366,
"learning_rate": 4.9332834018032025e-06,
"loss": 0.2594,
"step": 240
},
{
"epoch": 0.4230118443316413,
"grad_norm": 1.1220771074295044,
"learning_rate": 4.925195123905513e-06,
"loss": 0.2262,
"step": 250
},
{
"epoch": 0.43993231810490696,
"grad_norm": 1.7100379467010498,
"learning_rate": 4.916651615322182e-06,
"loss": 0.2093,
"step": 260
},
{
"epoch": 0.45685279187817257,
"grad_norm": 1.975794792175293,
"learning_rate": 4.907654479745726e-06,
"loss": 0.2101,
"step": 270
},
{
"epoch": 0.47377326565143824,
"grad_norm": 1.3200404644012451,
"learning_rate": 4.89820540601848e-06,
"loss": 0.2462,
"step": 280
},
{
"epoch": 0.4906937394247039,
"grad_norm": 1.254960298538208,
"learning_rate": 4.8883061678155905e-06,
"loss": 0.2051,
"step": 290
},
{
"epoch": 0.5076142131979695,
"grad_norm": 1.7000998258590698,
"learning_rate": 4.877958623312076e-06,
"loss": 0.2531,
"step": 300
},
{
"epoch": 0.5245346869712352,
"grad_norm": 1.2153050899505615,
"learning_rate": 4.8671647148340304e-06,
"loss": 0.2484,
"step": 310
},
{
"epoch": 0.5414551607445008,
"grad_norm": 1.4881746768951416,
"learning_rate": 4.855926468494039e-06,
"loss": 0.2121,
"step": 320
},
{
"epoch": 0.5583756345177665,
"grad_norm": 1.3982009887695312,
"learning_rate": 4.844245993810852e-06,
"loss": 0.237,
"step": 330
},
{
"epoch": 0.5752961082910322,
"grad_norm": 1.4024406671524048,
"learning_rate": 4.832125483313411e-06,
"loss": 0.2126,
"step": 340
},
{
"epoch": 0.5922165820642978,
"grad_norm": 1.4790270328521729,
"learning_rate": 4.819567212129292e-06,
"loss": 0.2491,
"step": 350
},
{
"epoch": 0.6091370558375635,
"grad_norm": 1.5148818492889404,
"learning_rate": 4.806573537557643e-06,
"loss": 0.2359,
"step": 360
},
{
"epoch": 0.626057529610829,
"grad_norm": 1.7280242443084717,
"learning_rate": 4.7931468986267014e-06,
"loss": 0.2363,
"step": 370
},
{
"epoch": 0.6429780033840947,
"grad_norm": 1.5051395893096924,
"learning_rate": 4.779289815635961e-06,
"loss": 0.2132,
"step": 380
},
{
"epoch": 0.6598984771573604,
"grad_norm": 2.0735015869140625,
"learning_rate": 4.765004889683096e-06,
"loss": 0.2307,
"step": 390
},
{
"epoch": 0.676818950930626,
"grad_norm": 1.4238276481628418,
"learning_rate": 4.750294802175703e-06,
"loss": 0.2237,
"step": 400
},
{
"epoch": 0.6937394247038917,
"grad_norm": 1.283726453781128,
"learning_rate": 4.735162314327987e-06,
"loss": 0.2299,
"step": 410
},
{
"epoch": 0.7106598984771574,
"grad_norm": 1.8388890027999878,
"learning_rate": 4.71961026664245e-06,
"loss": 0.23,
"step": 420
},
{
"epoch": 0.727580372250423,
"grad_norm": 1.456938624382019,
"learning_rate": 4.703641578376706e-06,
"loss": 0.1979,
"step": 430
},
{
"epoch": 0.7445008460236887,
"grad_norm": 1.3884906768798828,
"learning_rate": 4.687259246995512e-06,
"loss": 0.2358,
"step": 440
},
{
"epoch": 0.7614213197969543,
"grad_norm": 1.8143761157989502,
"learning_rate": 4.670466347608109e-06,
"loss": 0.2382,
"step": 450
},
{
"epoch": 0.7783417935702199,
"grad_norm": 1.3720914125442505,
"learning_rate": 4.65326603239101e-06,
"loss": 0.1786,
"step": 460
},
{
"epoch": 0.7952622673434856,
"grad_norm": 1.378342866897583,
"learning_rate": 4.6356615299963e-06,
"loss": 0.2109,
"step": 470
},
{
"epoch": 0.8121827411167513,
"grad_norm": 1.6520991325378418,
"learning_rate": 4.617656144945584e-06,
"loss": 0.2296,
"step": 480
},
{
"epoch": 0.8291032148900169,
"grad_norm": 1.617244839668274,
"learning_rate": 4.599253257009716e-06,
"loss": 0.21,
"step": 490
},
{
"epoch": 0.8460236886632826,
"grad_norm": 1.8561798334121704,
"learning_rate": 4.580456320574367e-06,
"loss": 0.1935,
"step": 500
},
{
"epoch": 0.8629441624365483,
"grad_norm": 2.1917858123779297,
"learning_rate": 4.561268863991618e-06,
"loss": 0.2246,
"step": 510
},
{
"epoch": 0.8798646362098139,
"grad_norm": 1.4701884984970093,
"learning_rate": 4.541694488917654e-06,
"loss": 0.2234,
"step": 520
},
{
"epoch": 0.8967851099830795,
"grad_norm": 1.3482675552368164,
"learning_rate": 4.521736869636699e-06,
"loss": 0.2423,
"step": 530
},
{
"epoch": 0.9137055837563451,
"grad_norm": 1.618618130683899,
"learning_rate": 4.501399752371323e-06,
"loss": 0.2384,
"step": 540
},
{
"epoch": 0.9306260575296108,
"grad_norm": 1.7225276231765747,
"learning_rate": 4.480686954579241e-06,
"loss": 0.2208,
"step": 550
},
{
"epoch": 0.9475465313028765,
"grad_norm": 1.5978412628173828,
"learning_rate": 4.459602364236743e-06,
"loss": 0.2421,
"step": 560
},
{
"epoch": 0.9644670050761421,
"grad_norm": 1.34050452709198,
"learning_rate": 4.438149939108887e-06,
"loss": 0.2117,
"step": 570
},
{
"epoch": 0.9813874788494078,
"grad_norm": 1.7396200895309448,
"learning_rate": 4.4163337060065895e-06,
"loss": 0.2395,
"step": 580
},
{
"epoch": 0.9983079526226735,
"grad_norm": 1.5188558101654053,
"learning_rate": 4.394157760030756e-06,
"loss": 0.2414,
"step": 590
},
{
"epoch": 1.015228426395939,
"grad_norm": 1.3064625263214111,
"learning_rate": 4.371626263803601e-06,
"loss": 0.1639,
"step": 600
},
{
"epoch": 1.0321489001692048,
"grad_norm": 1.2497655153274536,
"learning_rate": 4.348743446687279e-06,
"loss": 0.1606,
"step": 610
},
{
"epoch": 1.0490693739424704,
"grad_norm": 1.3260619640350342,
"learning_rate": 4.325513603989998e-06,
"loss": 0.1373,
"step": 620
},
{
"epoch": 1.0659898477157361,
"grad_norm": 1.7090791463851929,
"learning_rate": 4.301941096159757e-06,
"loss": 0.151,
"step": 630
},
{
"epoch": 1.0829103214890017,
"grad_norm": 1.4625272750854492,
"learning_rate": 4.278030347965842e-06,
"loss": 0.1567,
"step": 640
},
{
"epoch": 1.0998307952622675,
"grad_norm": 1.374337911605835,
"learning_rate": 4.253785847668264e-06,
"loss": 0.1412,
"step": 650
},
{
"epoch": 1.116751269035533,
"grad_norm": 1.326503038406372,
"learning_rate": 4.229212146175273e-06,
"loss": 0.1249,
"step": 660
},
{
"epoch": 1.1336717428087986,
"grad_norm": 1.2745444774627686,
"learning_rate": 4.2043138561891086e-06,
"loss": 0.1697,
"step": 670
},
{
"epoch": 1.1505922165820643,
"grad_norm": 1.3895198106765747,
"learning_rate": 4.179095651340155e-06,
"loss": 0.1429,
"step": 680
},
{
"epoch": 1.16751269035533,
"grad_norm": 1.331308126449585,
"learning_rate": 4.153562265309662e-06,
"loss": 0.1196,
"step": 690
},
{
"epoch": 1.1844331641285957,
"grad_norm": 1.7868833541870117,
"learning_rate": 4.1277184909411885e-06,
"loss": 0.1512,
"step": 700
},
{
"epoch": 1.2013536379018612,
"grad_norm": 1.2682002782821655,
"learning_rate": 4.101569179340946e-06,
"loss": 0.1795,
"step": 710
},
{
"epoch": 1.218274111675127,
"grad_norm": 1.1160831451416016,
"learning_rate": 4.075119238967196e-06,
"loss": 0.1353,
"step": 720
},
{
"epoch": 1.2351945854483926,
"grad_norm": 1.2749055624008179,
"learning_rate": 4.048373634708899e-06,
"loss": 0.1452,
"step": 730
},
{
"epoch": 1.252115059221658,
"grad_norm": 1.6326930522918701,
"learning_rate": 4.0213373869537504e-06,
"loss": 0.1498,
"step": 740
},
{
"epoch": 1.2690355329949239,
"grad_norm": 1.2723865509033203,
"learning_rate": 3.994015570645818e-06,
"loss": 0.1221,
"step": 750
},
{
"epoch": 1.2859560067681894,
"grad_norm": 1.4743432998657227,
"learning_rate": 3.966413314332924e-06,
"loss": 0.1436,
"step": 760
},
{
"epoch": 1.3028764805414552,
"grad_norm": 1.3993955850601196,
"learning_rate": 3.938535799203971e-06,
"loss": 0.1456,
"step": 770
},
{
"epoch": 1.3197969543147208,
"grad_norm": 1.7814866304397583,
"learning_rate": 3.91038825811639e-06,
"loss": 0.1329,
"step": 780
},
{
"epoch": 1.3367174280879865,
"grad_norm": 1.551604151725769,
"learning_rate": 3.881975974613888e-06,
"loss": 0.1636,
"step": 790
},
{
"epoch": 1.353637901861252,
"grad_norm": 1.244645118713379,
"learning_rate": 3.85330428193467e-06,
"loss": 0.1291,
"step": 800
},
{
"epoch": 1.3705583756345177,
"grad_norm": 1.368085503578186,
"learning_rate": 3.824378562010358e-06,
"loss": 0.1292,
"step": 810
},
{
"epoch": 1.3874788494077834,
"grad_norm": 1.6551761627197266,
"learning_rate": 3.7952042444557396e-06,
"loss": 0.158,
"step": 820
},
{
"epoch": 1.404399323181049,
"grad_norm": 1.4006320238113403,
"learning_rate": 3.7657868055495917e-06,
"loss": 0.146,
"step": 830
},
{
"epoch": 1.4213197969543148,
"grad_norm": 1.8334702253341675,
"learning_rate": 3.736131767206727e-06,
"loss": 0.1572,
"step": 840
},
{
"epoch": 1.4382402707275803,
"grad_norm": 1.3605976104736328,
"learning_rate": 3.706244695941489e-06,
"loss": 0.1274,
"step": 850
},
{
"epoch": 1.455160744500846,
"grad_norm": 1.306801199913025,
"learning_rate": 3.6761312018228597e-06,
"loss": 0.1582,
"step": 860
},
{
"epoch": 1.4720812182741116,
"grad_norm": 1.4271026849746704,
"learning_rate": 3.645796937421406e-06,
"loss": 0.1197,
"step": 870
},
{
"epoch": 1.4890016920473772,
"grad_norm": 1.3257379531860352,
"learning_rate": 3.6152475967482393e-06,
"loss": 0.1392,
"step": 880
},
{
"epoch": 1.505922165820643,
"grad_norm": 1.4457297325134277,
"learning_rate": 3.5844889141861977e-06,
"loss": 0.1334,
"step": 890
},
{
"epoch": 1.5228426395939088,
"grad_norm": 1.5230379104614258,
"learning_rate": 3.5535266634134556e-06,
"loss": 0.1325,
"step": 900
},
{
"epoch": 1.5397631133671743,
"grad_norm": 1.1343902349472046,
"learning_rate": 3.522366656319748e-06,
"loss": 0.1315,
"step": 910
},
{
"epoch": 1.5566835871404399,
"grad_norm": 1.58329439163208,
"learning_rate": 3.4910147419154306e-06,
"loss": 0.1464,
"step": 920
},
{
"epoch": 1.5736040609137056,
"grad_norm": 1.3953927755355835,
"learning_rate": 3.459476805233567e-06,
"loss": 0.132,
"step": 930
},
{
"epoch": 1.5905245346869712,
"grad_norm": 1.5068899393081665,
"learning_rate": 3.427758766225257e-06,
"loss": 0.1348,
"step": 940
},
{
"epoch": 1.6074450084602367,
"grad_norm": 1.7132457494735718,
"learning_rate": 3.3958665786484063e-06,
"loss": 0.1343,
"step": 950
},
{
"epoch": 1.6243654822335025,
"grad_norm": 1.1419628858566284,
"learning_rate": 3.3638062289501584e-06,
"loss": 0.1381,
"step": 960
},
{
"epoch": 1.6412859560067683,
"grad_norm": 1.8477777242660522,
"learning_rate": 3.331583735143179e-06,
"loss": 0.1642,
"step": 970
},
{
"epoch": 1.6582064297800339,
"grad_norm": 1.5605595111846924,
"learning_rate": 3.2992051456760242e-06,
"loss": 0.1479,
"step": 980
},
{
"epoch": 1.6751269035532994,
"grad_norm": 1.643684983253479,
"learning_rate": 3.266676538297797e-06,
"loss": 0.1564,
"step": 990
},
{
"epoch": 1.6920473773265652,
"grad_norm": 1.6141108274459839,
"learning_rate": 3.234004018917293e-06,
"loss": 0.1496,
"step": 1000
},
{
"epoch": 1.708967851099831,
"grad_norm": 1.8253343105316162,
"learning_rate": 3.20119372045687e-06,
"loss": 0.1535,
"step": 1010
},
{
"epoch": 1.7258883248730963,
"grad_norm": 1.3858433961868286,
"learning_rate": 3.1682518017012452e-06,
"loss": 0.1273,
"step": 1020
},
{
"epoch": 1.742808798646362,
"grad_norm": 1.9302154779434204,
"learning_rate": 3.1351844461414348e-06,
"loss": 0.15,
"step": 1030
},
{
"epoch": 1.7597292724196278,
"grad_norm": 1.4370014667510986,
"learning_rate": 3.1019978608140584e-06,
"loss": 0.1387,
"step": 1040
},
{
"epoch": 1.7766497461928934,
"grad_norm": 1.0620018243789673,
"learning_rate": 3.0686982751362234e-06,
"loss": 0.1261,
"step": 1050
},
{
"epoch": 1.793570219966159,
"grad_norm": 1.401853084564209,
"learning_rate": 3.0352919397362064e-06,
"loss": 0.1492,
"step": 1060
},
{
"epoch": 1.8104906937394247,
"grad_norm": 1.4132765531539917,
"learning_rate": 3.0017851252801574e-06,
"loss": 0.1401,
"step": 1070
},
{
"epoch": 1.8274111675126905,
"grad_norm": 1.540830373764038,
"learning_rate": 2.968184121295038e-06,
"loss": 0.1327,
"step": 1080
},
{
"epoch": 1.844331641285956,
"grad_norm": 1.434097170829773,
"learning_rate": 2.93449523498802e-06,
"loss": 0.1501,
"step": 1090
},
{
"epoch": 1.8612521150592216,
"grad_norm": 1.495847463607788,
"learning_rate": 2.900724790062571e-06,
"loss": 0.1294,
"step": 1100
},
{
"epoch": 1.8781725888324874,
"grad_norm": 1.7926485538482666,
"learning_rate": 2.866879125531437e-06,
"loss": 0.1286,
"step": 1110
},
{
"epoch": 1.895093062605753,
"grad_norm": 1.778939962387085,
"learning_rate": 2.832964594526748e-06,
"loss": 0.1617,
"step": 1120
},
{
"epoch": 1.9120135363790185,
"grad_norm": 1.9225724935531616,
"learning_rate": 2.798987563107488e-06,
"loss": 0.1467,
"step": 1130
},
{
"epoch": 1.9289340101522843,
"grad_norm": 1.0806710720062256,
"learning_rate": 2.7649544090645226e-06,
"loss": 0.156,
"step": 1140
},
{
"epoch": 1.94585448392555,
"grad_norm": 1.5917881727218628,
"learning_rate": 2.7308715207234326e-06,
"loss": 0.1377,
"step": 1150
},
{
"epoch": 1.9627749576988156,
"grad_norm": 1.3541003465652466,
"learning_rate": 2.69674529574537e-06,
"loss": 0.141,
"step": 1160
},
{
"epoch": 1.9796954314720812,
"grad_norm": 1.592405915260315,
"learning_rate": 2.6625821399261562e-06,
"loss": 0.1578,
"step": 1170
},
{
"epoch": 1.996615905245347,
"grad_norm": 1.7678178548812866,
"learning_rate": 2.628388465993864e-06,
"loss": 0.172,
"step": 1180
},
{
"epoch": 2.0135363790186127,
"grad_norm": 0.7831181287765503,
"learning_rate": 2.594170692405083e-06,
"loss": 0.1007,
"step": 1190
},
{
"epoch": 2.030456852791878,
"grad_norm": 1.5708684921264648,
"learning_rate": 2.559935242140125e-06,
"loss": 0.0863,
"step": 1200
},
{
"epoch": 2.047377326565144,
"grad_norm": 1.2178500890731812,
"learning_rate": 2.5256885414973715e-06,
"loss": 0.0815,
"step": 1210
},
{
"epoch": 2.0642978003384096,
"grad_norm": 1.4404493570327759,
"learning_rate": 2.491437018886998e-06,
"loss": 0.0924,
"step": 1220
},
{
"epoch": 2.081218274111675,
"grad_norm": 1.701210618019104,
"learning_rate": 2.457187103624308e-06,
"loss": 0.0843,
"step": 1230
},
{
"epoch": 2.0981387478849407,
"grad_norm": 1.3683816194534302,
"learning_rate": 2.4229452247228895e-06,
"loss": 0.0712,
"step": 1240
},
{
"epoch": 2.1150592216582065,
"grad_norm": 2.7988007068634033,
"learning_rate": 2.3887178096878363e-06,
"loss": 0.0883,
"step": 1250
},
{
"epoch": 2.1319796954314723,
"grad_norm": 1.0649851560592651,
"learning_rate": 2.354511283309244e-06,
"loss": 0.0706,
"step": 1260
},
{
"epoch": 2.1489001692047376,
"grad_norm": 2.3365893363952637,
"learning_rate": 2.320332066456224e-06,
"loss": 0.077,
"step": 1270
},
{
"epoch": 2.1658206429780034,
"grad_norm": 1.3339556455612183,
"learning_rate": 2.2861865748716448e-06,
"loss": 0.0864,
"step": 1280
},
{
"epoch": 2.182741116751269,
"grad_norm": 1.6249768733978271,
"learning_rate": 2.2520812179678422e-06,
"loss": 0.0783,
"step": 1290
},
{
"epoch": 2.199661590524535,
"grad_norm": 1.3000891208648682,
"learning_rate": 2.218022397623517e-06,
"loss": 0.075,
"step": 1300
},
{
"epoch": 2.2165820642978002,
"grad_norm": 1.5419507026672363,
"learning_rate": 2.1840165069820434e-06,
"loss": 0.0745,
"step": 1310
},
{
"epoch": 2.233502538071066,
"grad_norm": 1.4701499938964844,
"learning_rate": 2.150069929251419e-06,
"loss": 0.0652,
"step": 1320
},
{
"epoch": 2.250423011844332,
"grad_norm": 1.1285984516143799,
"learning_rate": 2.1161890365060838e-06,
"loss": 0.0842,
"step": 1330
},
{
"epoch": 2.267343485617597,
"grad_norm": 1.860626459121704,
"learning_rate": 2.082380188490817e-06,
"loss": 0.0824,
"step": 1340
},
{
"epoch": 2.284263959390863,
"grad_norm": 1.9373877048492432,
"learning_rate": 2.048649731426965e-06,
"loss": 0.0861,
"step": 1350
},
{
"epoch": 2.3011844331641287,
"grad_norm": 1.4356682300567627,
"learning_rate": 2.0150039968211958e-06,
"loss": 0.0793,
"step": 1360
},
{
"epoch": 2.3181049069373945,
"grad_norm": 1.4798492193222046,
"learning_rate": 1.9814493002770186e-06,
"loss": 0.0761,
"step": 1370
},
{
"epoch": 2.33502538071066,
"grad_norm": 2.065126895904541,
"learning_rate": 1.9479919403092863e-06,
"loss": 0.0802,
"step": 1380
},
{
"epoch": 2.3519458544839256,
"grad_norm": 1.5476558208465576,
"learning_rate": 1.914638197161914e-06,
"loss": 0.0804,
"step": 1390
},
{
"epoch": 2.3688663282571913,
"grad_norm": 2.891502618789673,
"learning_rate": 1.8813943316290177e-06,
"loss": 0.08,
"step": 1400
},
{
"epoch": 2.3857868020304567,
"grad_norm": 2.407071113586426,
"learning_rate": 1.8482665838797087e-06,
"loss": 0.0958,
"step": 1410
},
{
"epoch": 2.4027072758037225,
"grad_norm": 2.067172050476074,
"learning_rate": 1.8152611722867575e-06,
"loss": 0.0772,
"step": 1420
},
{
"epoch": 2.4196277495769882,
"grad_norm": 1.3346515893936157,
"learning_rate": 1.78238429225935e-06,
"loss": 0.0794,
"step": 1430
},
{
"epoch": 2.436548223350254,
"grad_norm": 1.4311262369155884,
"learning_rate": 1.7496421150801547e-06,
"loss": 0.1038,
"step": 1440
},
{
"epoch": 2.4534686971235193,
"grad_norm": 1.4051713943481445,
"learning_rate": 1.7170407867469185e-06,
"loss": 0.0854,
"step": 1450
},
{
"epoch": 2.470389170896785,
"grad_norm": 1.4341652393341064,
"learning_rate": 1.6845864268188073e-06,
"loss": 0.0722,
"step": 1460
},
{
"epoch": 2.487309644670051,
"grad_norm": 2.1416544914245605,
"learning_rate": 1.6522851272677126e-06,
"loss": 0.0662,
"step": 1470
},
{
"epoch": 2.504230118443316,
"grad_norm": 1.5771617889404297,
"learning_rate": 1.6201429513347275e-06,
"loss": 0.0949,
"step": 1480
},
{
"epoch": 2.521150592216582,
"grad_norm": 1.555201768875122,
"learning_rate": 1.5881659323920329e-06,
"loss": 0.069,
"step": 1490
},
{
"epoch": 2.5380710659898478,
"grad_norm": 2.297422170639038,
"learning_rate": 1.556360072810371e-06,
"loss": 0.0822,
"step": 1500
},
{
"epoch": 2.5549915397631136,
"grad_norm": 1.4423418045043945,
"learning_rate": 1.5247313428323521e-06,
"loss": 0.0739,
"step": 1510
},
{
"epoch": 2.571912013536379,
"grad_norm": 1.7135895490646362,
"learning_rate": 1.4932856794517828e-06,
"loss": 0.0701,
"step": 1520
},
{
"epoch": 2.5888324873096447,
"grad_norm": 1.2873613834381104,
"learning_rate": 1.4620289852992436e-06,
"loss": 0.0765,
"step": 1530
},
{
"epoch": 2.6057529610829104,
"grad_norm": 1.5120519399642944,
"learning_rate": 1.4309671275341115e-06,
"loss": 0.0733,
"step": 1540
},
{
"epoch": 2.6226734348561758,
"grad_norm": 1.4765515327453613,
"learning_rate": 1.4001059367432387e-06,
"loss": 0.0759,
"step": 1550
},
{
"epoch": 2.6395939086294415,
"grad_norm": 1.6146334409713745,
"learning_rate": 1.3694512058465064e-06,
"loss": 0.0731,
"step": 1560
},
{
"epoch": 2.6565143824027073,
"grad_norm": 2.312690258026123,
"learning_rate": 1.3390086890094346e-06,
"loss": 0.0826,
"step": 1570
},
{
"epoch": 2.673434856175973,
"grad_norm": 1.315255880355835,
"learning_rate": 1.3087841005630774e-06,
"loss": 0.0753,
"step": 1580
},
{
"epoch": 2.6903553299492384,
"grad_norm": 1.6095589399337769,
"learning_rate": 1.2787831139313957e-06,
"loss": 0.0942,
"step": 1590
},
{
"epoch": 2.707275803722504,
"grad_norm": 2.9680864810943604,
"learning_rate": 1.249011360566301e-06,
"loss": 0.083,
"step": 1600
},
{
"epoch": 2.72419627749577,
"grad_norm": 1.3204880952835083,
"learning_rate": 1.2194744288905807e-06,
"loss": 0.087,
"step": 1610
},
{
"epoch": 2.7411167512690353,
"grad_norm": 1.4213435649871826,
"learning_rate": 1.1901778632489037e-06,
"loss": 0.0793,
"step": 1620
},
{
"epoch": 2.758037225042301,
"grad_norm": 2.258690595626831,
"learning_rate": 1.1611271628671e-06,
"loss": 0.0802,
"step": 1630
},
{
"epoch": 2.774957698815567,
"grad_norm": 1.5296716690063477,
"learning_rate": 1.132327780819901e-06,
"loss": 0.0789,
"step": 1640
},
{
"epoch": 2.7918781725888326,
"grad_norm": 1.8829957246780396,
"learning_rate": 1.1037851230073563e-06,
"loss": 0.0779,
"step": 1650
},
{
"epoch": 2.808798646362098,
"grad_norm": 1.4642276763916016,
"learning_rate": 1.0755045471400897e-06,
"loss": 0.0813,
"step": 1660
},
{
"epoch": 2.8257191201353637,
"grad_norm": 1.2218215465545654,
"learning_rate": 1.0474913617336184e-06,
"loss": 0.0763,
"step": 1670
},
{
"epoch": 2.8426395939086295,
"grad_norm": 2.2098135948181152,
"learning_rate": 1.0197508251118892e-06,
"loss": 0.0731,
"step": 1680
},
{
"epoch": 2.859560067681895,
"grad_norm": 1.3853759765625,
"learning_rate": 9.922881444202514e-07,
"loss": 0.0861,
"step": 1690
},
{
"epoch": 2.8764805414551606,
"grad_norm": 1.280692458152771,
"learning_rate": 9.651084746480229e-07,
"loss": 0.078,
"step": 1700
},
{
"epoch": 2.8934010152284264,
"grad_norm": 1.077873945236206,
"learning_rate": 9.382169176608568e-07,
"loss": 0.0807,
"step": 1710
},
{
"epoch": 2.910321489001692,
"grad_norm": 1.4427660703659058,
"learning_rate": 9.116185212430765e-07,
"loss": 0.0866,
"step": 1720
},
{
"epoch": 2.927241962774958,
"grad_norm": 1.9770188331604004,
"learning_rate": 8.853182781501598e-07,
"loss": 0.0728,
"step": 1730
},
{
"epoch": 2.9441624365482233,
"grad_norm": 1.3886785507202148,
"learning_rate": 8.593211251715533e-07,
"loss": 0.0726,
"step": 1740
},
{
"epoch": 2.961082910321489,
"grad_norm": 1.6050931215286255,
"learning_rate": 8.336319422039945e-07,
"loss": 0.0806,
"step": 1750
},
{
"epoch": 2.9780033840947544,
"grad_norm": 2.0447897911071777,
"learning_rate": 8.082555513355148e-07,
"loss": 0.0809,
"step": 1760
},
{
"epoch": 2.99492385786802,
"grad_norm": 1.4558676481246948,
"learning_rate": 7.831967159402837e-07,
"loss": 0.0871,
"step": 1770
},
{
"epoch": 3.011844331641286,
"grad_norm": 1.2891719341278076,
"learning_rate": 7.584601397844882e-07,
"loss": 0.0494,
"step": 1780
},
{
"epoch": 3.0287648054145517,
"grad_norm": 1.3743082284927368,
"learning_rate": 7.340504661433918e-07,
"loss": 0.0507,
"step": 1790
},
{
"epoch": 3.045685279187817,
"grad_norm": 1.0931307077407837,
"learning_rate": 7.099722769297445e-07,
"loss": 0.0465,
"step": 1800
},
{
"epoch": 3.062605752961083,
"grad_norm": 1.310681939125061,
"learning_rate": 6.862300918337275e-07,
"loss": 0.0527,
"step": 1810
},
{
"epoch": 3.0795262267343486,
"grad_norm": 1.3822815418243408,
"learning_rate": 6.628283674745608e-07,
"loss": 0.0412,
"step": 1820
},
{
"epoch": 3.0964467005076144,
"grad_norm": 1.4022243022918701,
"learning_rate": 6.397714965639556e-07,
"loss": 0.0557,
"step": 1830
},
{
"epoch": 3.1133671742808797,
"grad_norm": 0.8449741005897522,
"learning_rate": 6.170638070815668e-07,
"loss": 0.0465,
"step": 1840
},
{
"epoch": 3.1302876480541455,
"grad_norm": 1.6387944221496582,
"learning_rate": 5.947095614625894e-07,
"loss": 0.0408,
"step": 1850
},
{
"epoch": 3.1472081218274113,
"grad_norm": 1.1685737371444702,
"learning_rate": 5.727129557976632e-07,
"loss": 0.0369,
"step": 1860
},
{
"epoch": 3.164128595600677,
"grad_norm": 1.1949836015701294,
"learning_rate": 5.510781190452299e-07,
"loss": 0.0413,
"step": 1870
},
{
"epoch": 3.1810490693739424,
"grad_norm": 1.5283178091049194,
"learning_rate": 5.298091122564911e-07,
"loss": 0.0394,
"step": 1880
},
{
"epoch": 3.197969543147208,
"grad_norm": 1.7606452703475952,
"learning_rate": 5.089099278131079e-07,
"loss": 0.0539,
"step": 1890
},
{
"epoch": 3.214890016920474,
"grad_norm": 1.3933011293411255,
"learning_rate": 4.883844886778028e-07,
"loss": 0.0437,
"step": 1900
},
{
"epoch": 3.2318104906937393,
"grad_norm": 1.65366530418396,
"learning_rate": 4.6823664765798136e-07,
"loss": 0.0472,
"step": 1910
},
{
"epoch": 3.248730964467005,
"grad_norm": 1.2451823949813843,
"learning_rate": 4.484701866825231e-07,
"loss": 0.0408,
"step": 1920
},
{
"epoch": 3.265651438240271,
"grad_norm": 0.8365059494972229,
"learning_rate": 4.290888160918855e-07,
"loss": 0.0417,
"step": 1930
},
{
"epoch": 3.2825719120135366,
"grad_norm": 1.6369590759277344,
"learning_rate": 4.1009617394163566e-07,
"loss": 0.0483,
"step": 1940
},
{
"epoch": 3.299492385786802,
"grad_norm": 1.3412925004959106,
"learning_rate": 3.914958253195561e-07,
"loss": 0.0422,
"step": 1950
},
{
"epoch": 3.3164128595600677,
"grad_norm": 1.2426025867462158,
"learning_rate": 3.732912616764478e-07,
"loss": 0.0374,
"step": 1960
},
{
"epoch": 3.3333333333333335,
"grad_norm": 1.1421136856079102,
"learning_rate": 3.554859001707522e-07,
"loss": 0.0427,
"step": 1970
},
{
"epoch": 3.350253807106599,
"grad_norm": 1.1082905530929565,
"learning_rate": 3.3808308302712007e-07,
"loss": 0.0485,
"step": 1980
},
{
"epoch": 3.3671742808798646,
"grad_norm": 1.8885458707809448,
"learning_rate": 3.2108607690904595e-07,
"loss": 0.0465,
"step": 1990
},
{
"epoch": 3.3840947546531304,
"grad_norm": 1.5490193367004395,
"learning_rate": 3.044980723056884e-07,
"loss": 0.0424,
"step": 2000
},
{
"epoch": 3.401015228426396,
"grad_norm": 1.3184294700622559,
"learning_rate": 2.883221829329857e-07,
"loss": 0.0469,
"step": 2010
},
{
"epoch": 3.4179357021996615,
"grad_norm": 2.039376974105835,
"learning_rate": 2.7256144514918204e-07,
"loss": 0.0564,
"step": 2020
},
{
"epoch": 3.4348561759729273,
"grad_norm": 1.3883781433105469,
"learning_rate": 2.5721881738487776e-07,
"loss": 0.0479,
"step": 2030
},
{
"epoch": 3.451776649746193,
"grad_norm": 1.4026315212249756,
"learning_rate": 2.4229717958770685e-07,
"loss": 0.0469,
"step": 2040
},
{
"epoch": 3.4686971235194584,
"grad_norm": 1.3182621002197266,
"learning_rate": 2.277993326817421e-07,
"loss": 0.0733,
"step": 2050
},
{
"epoch": 3.485617597292724,
"grad_norm": 1.252341866493225,
"learning_rate": 2.1372799804173894e-07,
"loss": 0.0401,
"step": 2060
},
{
"epoch": 3.50253807106599,
"grad_norm": 1.346508502960205,
"learning_rate": 2.0008581698231e-07,
"loss": 0.0486,
"step": 2070
},
{
"epoch": 3.5194585448392557,
"grad_norm": 0.9947476983070374,
"learning_rate": 1.8687535026212206e-07,
"loss": 0.0448,
"step": 2080
},
{
"epoch": 3.536379018612521,
"grad_norm": 1.5827783346176147,
"learning_rate": 1.740990776032256e-07,
"loss": 0.0462,
"step": 2090
},
{
"epoch": 3.553299492385787,
"grad_norm": 1.36147940158844,
"learning_rate": 1.617593972255846e-07,
"loss": 0.0474,
"step": 2100
},
{
"epoch": 3.5702199661590526,
"grad_norm": 1.3989553451538086,
"learning_rate": 1.4985862539691033e-07,
"loss": 0.0506,
"step": 2110
},
{
"epoch": 3.587140439932318,
"grad_norm": 2.2407867908477783,
"learning_rate": 1.3839899599787655e-07,
"loss": 0.0563,
"step": 2120
},
{
"epoch": 3.6040609137055837,
"grad_norm": 1.7094990015029907,
"learning_rate": 1.2738266010280275e-07,
"loss": 0.0455,
"step": 2130
},
{
"epoch": 3.6209813874788495,
"grad_norm": 2.9613707065582275,
"learning_rate": 1.168116855758747e-07,
"loss": 0.0462,
"step": 2140
},
{
"epoch": 3.6379018612521152,
"grad_norm": 1.730281949043274,
"learning_rate": 1.066880566829917e-07,
"loss": 0.0478,
"step": 2150
},
{
"epoch": 3.6548223350253806,
"grad_norm": 1.7777799367904663,
"learning_rate": 9.701367371930059e-08,
"loss": 0.0514,
"step": 2160
},
{
"epoch": 3.6717428087986463,
"grad_norm": 1.5566476583480835,
"learning_rate": 8.779035265249064e-08,
"loss": 0.0509,
"step": 2170
},
{
"epoch": 3.688663282571912,
"grad_norm": 1.7991079092025757,
"learning_rate": 7.901982478192572e-08,
"loss": 0.0432,
"step": 2180
},
{
"epoch": 3.7055837563451774,
"grad_norm": 1.1896286010742188,
"learning_rate": 7.07037364136609e-08,
"loss": 0.0468,
"step": 2190
},
{
"epoch": 3.7225042301184432,
"grad_norm": 1.1462163925170898,
"learning_rate": 6.2843648551415e-08,
"loss": 0.0451,
"step": 2200
},
{
"epoch": 3.739424703891709,
"grad_norm": 1.4667885303497314,
"learning_rate": 5.544103660355987e-08,
"loss": 0.0416,
"step": 2210
},
{
"epoch": 3.7563451776649748,
"grad_norm": 1.2911655902862549,
"learning_rate": 4.849729010616949e-08,
"loss": 0.0494,
"step": 2220
},
{
"epoch": 3.77326565143824,
"grad_norm": 1.1728025674819946,
"learning_rate": 4.201371246219388e-08,
"loss": 0.0411,
"step": 2230
},
{
"epoch": 3.790186125211506,
"grad_norm": 1.3054083585739136,
"learning_rate": 3.599152069679812e-08,
"loss": 0.04,
"step": 2240
},
{
"epoch": 3.8071065989847717,
"grad_norm": 1.867836594581604,
"learning_rate": 3.043184522891679e-08,
"loss": 0.0567,
"step": 2250
},
{
"epoch": 3.824027072758037,
"grad_norm": 1.3285189867019653,
"learning_rate": 2.5335729659062002e-08,
"loss": 0.0401,
"step": 2260
},
{
"epoch": 3.8409475465313028,
"grad_norm": 1.121762752532959,
"learning_rate": 2.07041305734329e-08,
"loss": 0.0408,
"step": 2270
},
{
"epoch": 3.8578680203045685,
"grad_norm": 0.9373975992202759,
"learning_rate": 1.6537917364354838e-08,
"loss": 0.0379,
"step": 2280
},
{
"epoch": 3.8747884940778343,
"grad_norm": 1.4871065616607666,
"learning_rate": 1.2837872067085477e-08,
"loss": 0.0454,
"step": 2290
},
{
"epoch": 3.8917089678511,
"grad_norm": 1.1789824962615967,
"learning_rate": 9.604689213022767e-09,
"loss": 0.0558,
"step": 2300
},
{
"epoch": 3.9086294416243654,
"grad_norm": 1.1937594413757324,
"learning_rate": 6.838975699332007e-09,
"loss": 0.0447,
"step": 2310
},
{
"epoch": 3.925549915397631,
"grad_norm": 1.5232540369033813,
"learning_rate": 4.541250675028363e-09,
"loss": 0.0535,
"step": 2320
},
{
"epoch": 3.9424703891708965,
"grad_norm": 1.3633273839950562,
"learning_rate": 2.711945443526209e-09,
"loss": 0.047,
"step": 2330
},
{
"epoch": 3.9593908629441623,
"grad_norm": 1.0834922790527344,
"learning_rate": 1.351403381680827e-09,
"loss": 0.0456,
"step": 2340
},
{
"epoch": 3.976311336717428,
"grad_norm": 1.0907598733901978,
"learning_rate": 4.598798753330247e-10,
"loss": 0.0413,
"step": 2350
},
{
"epoch": 3.993231810490694,
"grad_norm": 1.6505200862884521,
"learning_rate": 3.754227137081623e-11,
"loss": 0.046,
"step": 2360
},
{
"epoch": 4.0,
"step": 2364,
"total_flos": 1.5782755510340813e+17,
"train_loss": 0.13063472979615384,
"train_runtime": 2788.0875,
"train_samples_per_second": 13.558,
"train_steps_per_second": 0.848
}
],
"logging_steps": 10,
"max_steps": 2364,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.5782755510340813e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}