Orca_save / lora /trainer_state.json
pepoo20's picture
Upload lora/trainer_state.json with huggingface_hub
cefa398 verified
raw
history blame
56.3 kB
{
"best_metric": 0.44520583748817444,
"best_model_checkpoint": "saves/Orca/lora/checkpoint-3000",
"epoch": 0.9998413957176844,
"eval_steps": 500,
"global_step": 3152,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0031720856463124504,
"grad_norm": 0.8383375406265259,
"learning_rate": 5.000000000000001e-07,
"loss": 2.0081,
"step": 10
},
{
"epoch": 0.006344171292624901,
"grad_norm": 0.9924185276031494,
"learning_rate": 1.0000000000000002e-06,
"loss": 2.0141,
"step": 20
},
{
"epoch": 0.00951625693893735,
"grad_norm": 1.7123312950134277,
"learning_rate": 1.5e-06,
"loss": 2.0134,
"step": 30
},
{
"epoch": 0.012688342585249802,
"grad_norm": 1.6002644300460815,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.9886,
"step": 40
},
{
"epoch": 0.01586042823156225,
"grad_norm": 1.4284940958023071,
"learning_rate": 2.5e-06,
"loss": 1.9963,
"step": 50
},
{
"epoch": 0.0190325138778747,
"grad_norm": 1.9293030500411987,
"learning_rate": 3e-06,
"loss": 1.9858,
"step": 60
},
{
"epoch": 0.022204599524187154,
"grad_norm": 1.8281705379486084,
"learning_rate": 3.5000000000000004e-06,
"loss": 1.9608,
"step": 70
},
{
"epoch": 0.025376685170499604,
"grad_norm": 1.8404144048690796,
"learning_rate": 4.000000000000001e-06,
"loss": 1.9317,
"step": 80
},
{
"epoch": 0.028548770816812053,
"grad_norm": 1.8080644607543945,
"learning_rate": 4.5e-06,
"loss": 1.8928,
"step": 90
},
{
"epoch": 0.0317208564631245,
"grad_norm": 1.4789849519729614,
"learning_rate": 5e-06,
"loss": 1.8453,
"step": 100
},
{
"epoch": 0.034892942109436956,
"grad_norm": 2.3675060272216797,
"learning_rate": 5.500000000000001e-06,
"loss": 1.7683,
"step": 110
},
{
"epoch": 0.0380650277557494,
"grad_norm": 6.9262309074401855,
"learning_rate": 6e-06,
"loss": 1.6648,
"step": 120
},
{
"epoch": 0.041237113402061855,
"grad_norm": 1.5342968702316284,
"learning_rate": 6.5000000000000004e-06,
"loss": 1.5507,
"step": 130
},
{
"epoch": 0.04440919904837431,
"grad_norm": 1.7138350009918213,
"learning_rate": 7.000000000000001e-06,
"loss": 1.4388,
"step": 140
},
{
"epoch": 0.047581284694686754,
"grad_norm": 0.9868146181106567,
"learning_rate": 7.5e-06,
"loss": 1.3484,
"step": 150
},
{
"epoch": 0.05075337034099921,
"grad_norm": 0.649817705154419,
"learning_rate": 8.000000000000001e-06,
"loss": 1.2671,
"step": 160
},
{
"epoch": 0.05392545598731166,
"grad_norm": 0.5127522945404053,
"learning_rate": 8.500000000000002e-06,
"loss": 1.19,
"step": 170
},
{
"epoch": 0.057097541633624106,
"grad_norm": 0.43639031052589417,
"learning_rate": 9e-06,
"loss": 1.1479,
"step": 180
},
{
"epoch": 0.06026962727993656,
"grad_norm": 0.31427648663520813,
"learning_rate": 9.5e-06,
"loss": 1.0992,
"step": 190
},
{
"epoch": 0.063441712926249,
"grad_norm": 0.2897365689277649,
"learning_rate": 1e-05,
"loss": 1.0673,
"step": 200
},
{
"epoch": 0.06661379857256146,
"grad_norm": 0.43495845794677734,
"learning_rate": 1.05e-05,
"loss": 1.0399,
"step": 210
},
{
"epoch": 0.06978588421887391,
"grad_norm": 0.2536582946777344,
"learning_rate": 1.1000000000000001e-05,
"loss": 1.0043,
"step": 220
},
{
"epoch": 0.07295796986518636,
"grad_norm": 0.3480685353279114,
"learning_rate": 1.1500000000000002e-05,
"loss": 0.9747,
"step": 230
},
{
"epoch": 0.0761300555114988,
"grad_norm": 0.25048044323921204,
"learning_rate": 1.2e-05,
"loss": 0.9477,
"step": 240
},
{
"epoch": 0.07930214115781126,
"grad_norm": 0.31881338357925415,
"learning_rate": 1.25e-05,
"loss": 0.9099,
"step": 250
},
{
"epoch": 0.08247422680412371,
"grad_norm": 0.25372961163520813,
"learning_rate": 1.3000000000000001e-05,
"loss": 0.8661,
"step": 260
},
{
"epoch": 0.08564631245043616,
"grad_norm": 0.2892467975616455,
"learning_rate": 1.3500000000000001e-05,
"loss": 0.8306,
"step": 270
},
{
"epoch": 0.08881839809674862,
"grad_norm": 0.265289306640625,
"learning_rate": 1.4000000000000001e-05,
"loss": 0.79,
"step": 280
},
{
"epoch": 0.09199048374306107,
"grad_norm": 0.2384192794561386,
"learning_rate": 1.45e-05,
"loss": 0.7414,
"step": 290
},
{
"epoch": 0.09516256938937351,
"grad_norm": 0.2318211942911148,
"learning_rate": 1.5e-05,
"loss": 0.6994,
"step": 300
},
{
"epoch": 0.09833465503568596,
"grad_norm": 0.22584764659404755,
"learning_rate": 1.55e-05,
"loss": 0.6663,
"step": 310
},
{
"epoch": 0.10150674068199841,
"grad_norm": 0.2281956821680069,
"learning_rate": 1.6000000000000003e-05,
"loss": 0.6366,
"step": 320
},
{
"epoch": 0.10467882632831087,
"grad_norm": 0.23853954672813416,
"learning_rate": 1.65e-05,
"loss": 0.6105,
"step": 330
},
{
"epoch": 0.10785091197462332,
"grad_norm": 0.2249988317489624,
"learning_rate": 1.7000000000000003e-05,
"loss": 0.5923,
"step": 340
},
{
"epoch": 0.11102299762093576,
"grad_norm": 0.22325117886066437,
"learning_rate": 1.75e-05,
"loss": 0.5782,
"step": 350
},
{
"epoch": 0.11419508326724821,
"grad_norm": 0.26995840668678284,
"learning_rate": 1.8e-05,
"loss": 0.5588,
"step": 360
},
{
"epoch": 0.11736716891356067,
"grad_norm": 0.3128603994846344,
"learning_rate": 1.85e-05,
"loss": 0.5539,
"step": 370
},
{
"epoch": 0.12053925455987312,
"grad_norm": 0.2401484102010727,
"learning_rate": 1.9e-05,
"loss": 0.5463,
"step": 380
},
{
"epoch": 0.12371134020618557,
"grad_norm": 0.252087265253067,
"learning_rate": 1.9500000000000003e-05,
"loss": 0.5396,
"step": 390
},
{
"epoch": 0.126883425852498,
"grad_norm": 0.3551071286201477,
"learning_rate": 2e-05,
"loss": 0.5293,
"step": 400
},
{
"epoch": 0.13005551149881048,
"grad_norm": 0.2852734625339508,
"learning_rate": 2.05e-05,
"loss": 0.5327,
"step": 410
},
{
"epoch": 0.13322759714512292,
"grad_norm": 0.25599339604377747,
"learning_rate": 2.1e-05,
"loss": 0.5215,
"step": 420
},
{
"epoch": 0.13639968279143536,
"grad_norm": 0.2308124303817749,
"learning_rate": 2.15e-05,
"loss": 0.5179,
"step": 430
},
{
"epoch": 0.13957176843774782,
"grad_norm": 0.28417715430259705,
"learning_rate": 2.2000000000000003e-05,
"loss": 0.5122,
"step": 440
},
{
"epoch": 0.14274385408406026,
"grad_norm": 0.32094892859458923,
"learning_rate": 2.25e-05,
"loss": 0.5078,
"step": 450
},
{
"epoch": 0.14591593973037273,
"grad_norm": 0.28217512369155884,
"learning_rate": 2.3000000000000003e-05,
"loss": 0.5064,
"step": 460
},
{
"epoch": 0.14908802537668517,
"grad_norm": 0.3319534659385681,
"learning_rate": 2.35e-05,
"loss": 0.5056,
"step": 470
},
{
"epoch": 0.1522601110229976,
"grad_norm": 0.35069146752357483,
"learning_rate": 2.4e-05,
"loss": 0.5001,
"step": 480
},
{
"epoch": 0.15543219666931007,
"grad_norm": 0.2778172492980957,
"learning_rate": 2.45e-05,
"loss": 0.4994,
"step": 490
},
{
"epoch": 0.1586042823156225,
"grad_norm": 0.2950364053249359,
"learning_rate": 2.5e-05,
"loss": 0.5009,
"step": 500
},
{
"epoch": 0.1586042823156225,
"eval_loss": 0.49637770652770996,
"eval_runtime": 23627.0877,
"eval_samples_per_second": 0.044,
"eval_steps_per_second": 0.022,
"step": 500
},
{
"epoch": 0.16177636796193498,
"grad_norm": 0.26569539308547974,
"learning_rate": 2.5500000000000003e-05,
"loss": 0.4978,
"step": 510
},
{
"epoch": 0.16494845360824742,
"grad_norm": 0.26948851346969604,
"learning_rate": 2.6000000000000002e-05,
"loss": 0.4926,
"step": 520
},
{
"epoch": 0.1681205392545599,
"grad_norm": 0.29083552956581116,
"learning_rate": 2.6500000000000004e-05,
"loss": 0.4893,
"step": 530
},
{
"epoch": 0.17129262490087233,
"grad_norm": 0.2939367890357971,
"learning_rate": 2.7000000000000002e-05,
"loss": 0.491,
"step": 540
},
{
"epoch": 0.17446471054718476,
"grad_norm": 0.28195658326148987,
"learning_rate": 2.7500000000000004e-05,
"loss": 0.4865,
"step": 550
},
{
"epoch": 0.17763679619349723,
"grad_norm": 0.2827642858028412,
"learning_rate": 2.8000000000000003e-05,
"loss": 0.4851,
"step": 560
},
{
"epoch": 0.18080888183980967,
"grad_norm": 0.2715272605419159,
"learning_rate": 2.8499999999999998e-05,
"loss": 0.4833,
"step": 570
},
{
"epoch": 0.18398096748612214,
"grad_norm": 0.30428314208984375,
"learning_rate": 2.9e-05,
"loss": 0.4834,
"step": 580
},
{
"epoch": 0.18715305313243458,
"grad_norm": 0.2857949435710907,
"learning_rate": 2.95e-05,
"loss": 0.4796,
"step": 590
},
{
"epoch": 0.19032513877874702,
"grad_norm": 0.302212655544281,
"learning_rate": 3e-05,
"loss": 0.4826,
"step": 600
},
{
"epoch": 0.19349722442505948,
"grad_norm": 0.34079357981681824,
"learning_rate": 3.05e-05,
"loss": 0.4826,
"step": 610
},
{
"epoch": 0.19666931007137192,
"grad_norm": 0.3239024877548218,
"learning_rate": 3.1e-05,
"loss": 0.4804,
"step": 620
},
{
"epoch": 0.1998413957176844,
"grad_norm": 0.36531686782836914,
"learning_rate": 3.15e-05,
"loss": 0.4782,
"step": 630
},
{
"epoch": 0.20301348136399683,
"grad_norm": 0.3582271337509155,
"learning_rate": 3.2000000000000005e-05,
"loss": 0.4784,
"step": 640
},
{
"epoch": 0.20618556701030927,
"grad_norm": 0.3250937759876251,
"learning_rate": 3.2500000000000004e-05,
"loss": 0.4716,
"step": 650
},
{
"epoch": 0.20935765265662173,
"grad_norm": 0.28494277596473694,
"learning_rate": 3.3e-05,
"loss": 0.4764,
"step": 660
},
{
"epoch": 0.21252973830293417,
"grad_norm": 0.3156846761703491,
"learning_rate": 3.35e-05,
"loss": 0.4742,
"step": 670
},
{
"epoch": 0.21570182394924664,
"grad_norm": 0.27053794264793396,
"learning_rate": 3.4000000000000007e-05,
"loss": 0.4765,
"step": 680
},
{
"epoch": 0.21887390959555908,
"grad_norm": 0.31487756967544556,
"learning_rate": 3.45e-05,
"loss": 0.4749,
"step": 690
},
{
"epoch": 0.22204599524187152,
"grad_norm": 0.2739400267601013,
"learning_rate": 3.5e-05,
"loss": 0.4693,
"step": 700
},
{
"epoch": 0.22521808088818399,
"grad_norm": 0.2924807369709015,
"learning_rate": 3.55e-05,
"loss": 0.4697,
"step": 710
},
{
"epoch": 0.22839016653449642,
"grad_norm": 0.27241086959838867,
"learning_rate": 3.6e-05,
"loss": 0.4711,
"step": 720
},
{
"epoch": 0.2315622521808089,
"grad_norm": 0.249298095703125,
"learning_rate": 3.65e-05,
"loss": 0.4748,
"step": 730
},
{
"epoch": 0.23473433782712133,
"grad_norm": 0.33761996030807495,
"learning_rate": 3.7e-05,
"loss": 0.4692,
"step": 740
},
{
"epoch": 0.23790642347343377,
"grad_norm": 0.3454744517803192,
"learning_rate": 3.7500000000000003e-05,
"loss": 0.4691,
"step": 750
},
{
"epoch": 0.24107850911974624,
"grad_norm": 0.3940749168395996,
"learning_rate": 3.8e-05,
"loss": 0.4694,
"step": 760
},
{
"epoch": 0.24425059476605868,
"grad_norm": 0.2833056151866913,
"learning_rate": 3.85e-05,
"loss": 0.4651,
"step": 770
},
{
"epoch": 0.24742268041237114,
"grad_norm": 0.3439841866493225,
"learning_rate": 3.9000000000000006e-05,
"loss": 0.4718,
"step": 780
},
{
"epoch": 0.2505947660586836,
"grad_norm": 0.28585824370384216,
"learning_rate": 3.9500000000000005e-05,
"loss": 0.4678,
"step": 790
},
{
"epoch": 0.253766851704996,
"grad_norm": 0.2761208415031433,
"learning_rate": 4e-05,
"loss": 0.4672,
"step": 800
},
{
"epoch": 0.25693893735130846,
"grad_norm": 0.287431538105011,
"learning_rate": 4.05e-05,
"loss": 0.4654,
"step": 810
},
{
"epoch": 0.26011102299762096,
"grad_norm": 0.2529178261756897,
"learning_rate": 4.1e-05,
"loss": 0.4658,
"step": 820
},
{
"epoch": 0.2632831086439334,
"grad_norm": 0.22234879434108734,
"learning_rate": 4.15e-05,
"loss": 0.467,
"step": 830
},
{
"epoch": 0.26645519429024583,
"grad_norm": 0.2847014367580414,
"learning_rate": 4.2e-05,
"loss": 0.4643,
"step": 840
},
{
"epoch": 0.2696272799365583,
"grad_norm": 0.27712419629096985,
"learning_rate": 4.25e-05,
"loss": 0.4604,
"step": 850
},
{
"epoch": 0.2727993655828707,
"grad_norm": 0.24886064231395721,
"learning_rate": 4.3e-05,
"loss": 0.4657,
"step": 860
},
{
"epoch": 0.2759714512291832,
"grad_norm": 0.3196752369403839,
"learning_rate": 4.35e-05,
"loss": 0.4652,
"step": 870
},
{
"epoch": 0.27914353687549565,
"grad_norm": 0.27527421712875366,
"learning_rate": 4.4000000000000006e-05,
"loss": 0.4634,
"step": 880
},
{
"epoch": 0.2823156225218081,
"grad_norm": 0.26819467544555664,
"learning_rate": 4.4500000000000004e-05,
"loss": 0.4633,
"step": 890
},
{
"epoch": 0.2854877081681205,
"grad_norm": 0.26196032762527466,
"learning_rate": 4.5e-05,
"loss": 0.4656,
"step": 900
},
{
"epoch": 0.28865979381443296,
"grad_norm": 0.23337453603744507,
"learning_rate": 4.55e-05,
"loss": 0.4622,
"step": 910
},
{
"epoch": 0.29183187946074546,
"grad_norm": 0.2958989143371582,
"learning_rate": 4.600000000000001e-05,
"loss": 0.4627,
"step": 920
},
{
"epoch": 0.2950039651070579,
"grad_norm": 0.23167885839939117,
"learning_rate": 4.6500000000000005e-05,
"loss": 0.4636,
"step": 930
},
{
"epoch": 0.29817605075337034,
"grad_norm": 0.2608003318309784,
"learning_rate": 4.7e-05,
"loss": 0.4628,
"step": 940
},
{
"epoch": 0.3013481363996828,
"grad_norm": 0.2806340456008911,
"learning_rate": 4.75e-05,
"loss": 0.4624,
"step": 950
},
{
"epoch": 0.3045202220459952,
"grad_norm": 0.2206449806690216,
"learning_rate": 4.8e-05,
"loss": 0.4594,
"step": 960
},
{
"epoch": 0.3076923076923077,
"grad_norm": 0.2883213460445404,
"learning_rate": 4.85e-05,
"loss": 0.4612,
"step": 970
},
{
"epoch": 0.31086439333862015,
"grad_norm": 0.23365403711795807,
"learning_rate": 4.9e-05,
"loss": 0.463,
"step": 980
},
{
"epoch": 0.3140364789849326,
"grad_norm": 0.28844788670539856,
"learning_rate": 4.9500000000000004e-05,
"loss": 0.4612,
"step": 990
},
{
"epoch": 0.317208564631245,
"grad_norm": 0.26974356174468994,
"learning_rate": 5e-05,
"loss": 0.4641,
"step": 1000
},
{
"epoch": 0.317208564631245,
"eval_loss": 0.45910927653312683,
"eval_runtime": 419.1696,
"eval_samples_per_second": 2.457,
"eval_steps_per_second": 1.229,
"step": 1000
},
{
"epoch": 0.3203806502775575,
"grad_norm": 0.23146626353263855,
"learning_rate": 4.9997336102492574e-05,
"loss": 0.4583,
"step": 1010
},
{
"epoch": 0.32355273592386996,
"grad_norm": 0.24651704728603363,
"learning_rate": 4.9989344977678285e-05,
"loss": 0.461,
"step": 1020
},
{
"epoch": 0.3267248215701824,
"grad_norm": 0.2995850741863251,
"learning_rate": 4.997602832856013e-05,
"loss": 0.4583,
"step": 1030
},
{
"epoch": 0.32989690721649484,
"grad_norm": 0.2886098325252533,
"learning_rate": 4.995738899307319e-05,
"loss": 0.4565,
"step": 1040
},
{
"epoch": 0.3330689928628073,
"grad_norm": 0.2367076575756073,
"learning_rate": 4.99334309434798e-05,
"loss": 0.4598,
"step": 1050
},
{
"epoch": 0.3362410785091198,
"grad_norm": 0.29130271077156067,
"learning_rate": 4.990415928552305e-05,
"loss": 0.4564,
"step": 1060
},
{
"epoch": 0.3394131641554322,
"grad_norm": 0.23702391982078552,
"learning_rate": 4.9869580257338685e-05,
"loss": 0.4567,
"step": 1070
},
{
"epoch": 0.34258524980174465,
"grad_norm": 0.24812078475952148,
"learning_rate": 4.9829701228125655e-05,
"loss": 0.4587,
"step": 1080
},
{
"epoch": 0.3457573354480571,
"grad_norm": 0.2195880264043808,
"learning_rate": 4.9784530696575684e-05,
"loss": 0.4578,
"step": 1090
},
{
"epoch": 0.34892942109436953,
"grad_norm": 0.2531008720397949,
"learning_rate": 4.973407828906208e-05,
"loss": 0.4551,
"step": 1100
},
{
"epoch": 0.352101506740682,
"grad_norm": 0.2672252655029297,
"learning_rate": 4.967835475758825e-05,
"loss": 0.4596,
"step": 1110
},
{
"epoch": 0.35527359238699446,
"grad_norm": 0.2659941613674164,
"learning_rate": 4.961737197749633e-05,
"loss": 0.4536,
"step": 1120
},
{
"epoch": 0.3584456780333069,
"grad_norm": 0.2517182230949402,
"learning_rate": 4.955114294493639e-05,
"loss": 0.4579,
"step": 1130
},
{
"epoch": 0.36161776367961934,
"grad_norm": 0.2654310166835785,
"learning_rate": 4.947968177409681e-05,
"loss": 0.4548,
"step": 1140
},
{
"epoch": 0.3647898493259318,
"grad_norm": 0.36614441871643066,
"learning_rate": 4.940300369419637e-05,
"loss": 0.4575,
"step": 1150
},
{
"epoch": 0.3679619349722443,
"grad_norm": 0.2583513855934143,
"learning_rate": 4.9321125046238756e-05,
"loss": 0.4598,
"step": 1160
},
{
"epoch": 0.3711340206185567,
"grad_norm": 0.23981873691082,
"learning_rate": 4.923406327953008e-05,
"loss": 0.4546,
"step": 1170
},
{
"epoch": 0.37430610626486915,
"grad_norm": 0.23798957467079163,
"learning_rate": 4.9141836947960165e-05,
"loss": 0.458,
"step": 1180
},
{
"epoch": 0.3774781919111816,
"grad_norm": 0.2931083142757416,
"learning_rate": 4.904446570604863e-05,
"loss": 0.4543,
"step": 1190
},
{
"epoch": 0.38065027755749403,
"grad_norm": 0.27312615513801575,
"learning_rate": 4.8941970304756144e-05,
"loss": 0.4552,
"step": 1200
},
{
"epoch": 0.3838223632038065,
"grad_norm": 0.2659391164779663,
"learning_rate": 4.883437258706224e-05,
"loss": 0.4583,
"step": 1210
},
{
"epoch": 0.38699444885011897,
"grad_norm": 0.23010990023612976,
"learning_rate": 4.8721695483310275e-05,
"loss": 0.4564,
"step": 1220
},
{
"epoch": 0.3901665344964314,
"grad_norm": 0.2815942168235779,
"learning_rate": 4.860396300632072e-05,
"loss": 0.4548,
"step": 1230
},
{
"epoch": 0.39333862014274384,
"grad_norm": 0.36561158299446106,
"learning_rate": 4.848120024627372e-05,
"loss": 0.4528,
"step": 1240
},
{
"epoch": 0.3965107057890563,
"grad_norm": 0.25015613436698914,
"learning_rate": 4.835343336536212e-05,
"loss": 0.4567,
"step": 1250
},
{
"epoch": 0.3996827914353688,
"grad_norm": 0.2373395711183548,
"learning_rate": 4.822068959221598e-05,
"loss": 0.4544,
"step": 1260
},
{
"epoch": 0.4028548770816812,
"grad_norm": 0.26331648230552673,
"learning_rate": 4.80829972160998e-05,
"loss": 0.4532,
"step": 1270
},
{
"epoch": 0.40602696272799366,
"grad_norm": 0.2622736692428589,
"learning_rate": 4.794038558088378e-05,
"loss": 0.4535,
"step": 1280
},
{
"epoch": 0.4091990483743061,
"grad_norm": 0.1999824345111847,
"learning_rate": 4.779288507879031e-05,
"loss": 0.4568,
"step": 1290
},
{
"epoch": 0.41237113402061853,
"grad_norm": 0.24161754548549652,
"learning_rate": 4.764052714391695e-05,
"loss": 0.4534,
"step": 1300
},
{
"epoch": 0.41554321966693103,
"grad_norm": 0.2368980497121811,
"learning_rate": 4.7483344245537545e-05,
"loss": 0.4558,
"step": 1310
},
{
"epoch": 0.41871530531324347,
"grad_norm": 0.1948440819978714,
"learning_rate": 4.7321369881182584e-05,
"loss": 0.4564,
"step": 1320
},
{
"epoch": 0.4218873909595559,
"grad_norm": 0.2509821355342865,
"learning_rate": 4.715463856950053e-05,
"loss": 0.4558,
"step": 1330
},
{
"epoch": 0.42505947660586835,
"grad_norm": 0.23150520026683807,
"learning_rate": 4.698318584290141e-05,
"loss": 0.4538,
"step": 1340
},
{
"epoch": 0.4282315622521808,
"grad_norm": 0.22577635943889618,
"learning_rate": 4.680704823998452e-05,
"loss": 0.4571,
"step": 1350
},
{
"epoch": 0.4314036478984933,
"grad_norm": 0.22882512211799622,
"learning_rate": 4.6626263297751546e-05,
"loss": 0.453,
"step": 1360
},
{
"epoch": 0.4345757335448057,
"grad_norm": 0.1979285478591919,
"learning_rate": 4.644086954360708e-05,
"loss": 0.4542,
"step": 1370
},
{
"epoch": 0.43774781919111816,
"grad_norm": 0.2554062008857727,
"learning_rate": 4.625090648714786e-05,
"loss": 0.4561,
"step": 1380
},
{
"epoch": 0.4409199048374306,
"grad_norm": 0.2330465316772461,
"learning_rate": 4.60564146117429e-05,
"loss": 0.4525,
"step": 1390
},
{
"epoch": 0.44409199048374304,
"grad_norm": 0.22269243001937866,
"learning_rate": 4.585743536590599e-05,
"loss": 0.4544,
"step": 1400
},
{
"epoch": 0.44726407613005553,
"grad_norm": 0.2291986495256424,
"learning_rate": 4.565401115446246e-05,
"loss": 0.4548,
"step": 1410
},
{
"epoch": 0.45043616177636797,
"grad_norm": 0.21675272285938263,
"learning_rate": 4.5446185329512314e-05,
"loss": 0.4518,
"step": 1420
},
{
"epoch": 0.4536082474226804,
"grad_norm": 0.21207275986671448,
"learning_rate": 4.5234002181191304e-05,
"loss": 0.4511,
"step": 1430
},
{
"epoch": 0.45678033306899285,
"grad_norm": 0.22509776055812836,
"learning_rate": 4.5017506928232247e-05,
"loss": 0.4518,
"step": 1440
},
{
"epoch": 0.4599524187153053,
"grad_norm": 0.23429358005523682,
"learning_rate": 4.4796745708328294e-05,
"loss": 0.4556,
"step": 1450
},
{
"epoch": 0.4631245043616178,
"grad_norm": 0.18599826097488403,
"learning_rate": 4.457176556830054e-05,
"loss": 0.4538,
"step": 1460
},
{
"epoch": 0.4662965900079302,
"grad_norm": 0.2500051259994507,
"learning_rate": 4.434261445407172e-05,
"loss": 0.4538,
"step": 1470
},
{
"epoch": 0.46946867565424266,
"grad_norm": 0.213278666138649,
"learning_rate": 4.410934120044838e-05,
"loss": 0.4545,
"step": 1480
},
{
"epoch": 0.4726407613005551,
"grad_norm": 0.2551578879356384,
"learning_rate": 4.387199552071366e-05,
"loss": 0.4512,
"step": 1490
},
{
"epoch": 0.47581284694686754,
"grad_norm": 0.23356303572654724,
"learning_rate": 4.3630627996032706e-05,
"loss": 0.4514,
"step": 1500
},
{
"epoch": 0.47581284694686754,
"eval_loss": 0.4515945613384247,
"eval_runtime": 293.5383,
"eval_samples_per_second": 3.509,
"eval_steps_per_second": 1.754,
"step": 1500
},
{
"epoch": 0.47898493259318004,
"grad_norm": 0.22167843580245972,
"learning_rate": 4.3385290064673316e-05,
"loss": 0.4529,
"step": 1510
},
{
"epoch": 0.4821570182394925,
"grad_norm": 0.2573811709880829,
"learning_rate": 4.31360340110438e-05,
"loss": 0.4517,
"step": 1520
},
{
"epoch": 0.4853291038858049,
"grad_norm": 0.20118848979473114,
"learning_rate": 4.288291295455055e-05,
"loss": 0.4538,
"step": 1530
},
{
"epoch": 0.48850118953211735,
"grad_norm": 0.24297507107257843,
"learning_rate": 4.262598083827769e-05,
"loss": 0.4519,
"step": 1540
},
{
"epoch": 0.4916732751784298,
"grad_norm": 0.18671482801437378,
"learning_rate": 4.236529241749114e-05,
"loss": 0.4515,
"step": 1550
},
{
"epoch": 0.4948453608247423,
"grad_norm": 0.20851927995681763,
"learning_rate": 4.2100903247969644e-05,
"loss": 0.4526,
"step": 1560
},
{
"epoch": 0.4980174464710547,
"grad_norm": 0.23694059252738953,
"learning_rate": 4.1832869674165204e-05,
"loss": 0.4519,
"step": 1570
},
{
"epoch": 0.5011895321173672,
"grad_norm": 0.3004760146141052,
"learning_rate": 4.156124881719533e-05,
"loss": 0.4493,
"step": 1580
},
{
"epoch": 0.5043616177636796,
"grad_norm": 0.20880846679210663,
"learning_rate": 4.1286098562669925e-05,
"loss": 0.4508,
"step": 1590
},
{
"epoch": 0.507533703409992,
"grad_norm": 0.2033112347126007,
"learning_rate": 4.1007477548355185e-05,
"loss": 0.4522,
"step": 1600
},
{
"epoch": 0.5107057890563045,
"grad_norm": 0.24385987222194672,
"learning_rate": 4.072544515167714e-05,
"loss": 0.45,
"step": 1610
},
{
"epoch": 0.5138778747026169,
"grad_norm": 0.19505774974822998,
"learning_rate": 4.044006147706768e-05,
"loss": 0.4528,
"step": 1620
},
{
"epoch": 0.5170499603489295,
"grad_norm": 0.20980720221996307,
"learning_rate": 4.0151387343155545e-05,
"loss": 0.4512,
"step": 1630
},
{
"epoch": 0.5202220459952419,
"grad_norm": 0.2582622766494751,
"learning_rate": 3.985948426980521e-05,
"loss": 0.452,
"step": 1640
},
{
"epoch": 0.5233941316415543,
"grad_norm": 0.18902088701725006,
"learning_rate": 3.956441446500624e-05,
"loss": 0.4498,
"step": 1650
},
{
"epoch": 0.5265662172878668,
"grad_norm": 0.21405866742134094,
"learning_rate": 3.926624081161604e-05,
"loss": 0.4462,
"step": 1660
},
{
"epoch": 0.5297383029341792,
"grad_norm": 0.21340113878250122,
"learning_rate": 3.896502685395876e-05,
"loss": 0.449,
"step": 1670
},
{
"epoch": 0.5329103885804917,
"grad_norm": 0.2296970933675766,
"learning_rate": 3.8660836784283275e-05,
"loss": 0.4513,
"step": 1680
},
{
"epoch": 0.5360824742268041,
"grad_norm": 0.2143164724111557,
"learning_rate": 3.835373542908308e-05,
"loss": 0.4488,
"step": 1690
},
{
"epoch": 0.5392545598731165,
"grad_norm": 0.1945250779390335,
"learning_rate": 3.804378823528093e-05,
"loss": 0.4529,
"step": 1700
},
{
"epoch": 0.542426645519429,
"grad_norm": 0.20191755890846252,
"learning_rate": 3.7731061256281394e-05,
"loss": 0.4507,
"step": 1710
},
{
"epoch": 0.5455987311657414,
"grad_norm": 0.21345514059066772,
"learning_rate": 3.7415621137894056e-05,
"loss": 0.4487,
"step": 1720
},
{
"epoch": 0.548770816812054,
"grad_norm": 0.2062884271144867,
"learning_rate": 3.709753510413052e-05,
"loss": 0.4504,
"step": 1730
},
{
"epoch": 0.5519429024583664,
"grad_norm": 0.20377105474472046,
"learning_rate": 3.6776870942878196e-05,
"loss": 0.4489,
"step": 1740
},
{
"epoch": 0.5551149881046789,
"grad_norm": 0.19985787570476532,
"learning_rate": 3.645369699145387e-05,
"loss": 0.4497,
"step": 1750
},
{
"epoch": 0.5582870737509913,
"grad_norm": 0.2308078110218048,
"learning_rate": 3.6128082122040224e-05,
"loss": 0.453,
"step": 1760
},
{
"epoch": 0.5614591593973037,
"grad_norm": 0.19868697226047516,
"learning_rate": 3.5800095727008395e-05,
"loss": 0.4512,
"step": 1770
},
{
"epoch": 0.5646312450436162,
"grad_norm": 0.20231026411056519,
"learning_rate": 3.54698077041296e-05,
"loss": 0.4482,
"step": 1780
},
{
"epoch": 0.5678033306899286,
"grad_norm": 0.21103453636169434,
"learning_rate": 3.51372884416791e-05,
"loss": 0.452,
"step": 1790
},
{
"epoch": 0.570975416336241,
"grad_norm": 0.21350406110286713,
"learning_rate": 3.4802608803435646e-05,
"loss": 0.4518,
"step": 1800
},
{
"epoch": 0.5741475019825535,
"grad_norm": 0.2247006744146347,
"learning_rate": 3.446584011357957e-05,
"loss": 0.451,
"step": 1810
},
{
"epoch": 0.5773195876288659,
"grad_norm": 0.20522333681583405,
"learning_rate": 3.412705414149276e-05,
"loss": 0.4491,
"step": 1820
},
{
"epoch": 0.5804916732751785,
"grad_norm": 0.20249220728874207,
"learning_rate": 3.3786323086463736e-05,
"loss": 0.4508,
"step": 1830
},
{
"epoch": 0.5836637589214909,
"grad_norm": 0.17756901681423187,
"learning_rate": 3.3443719562301147e-05,
"loss": 0.4493,
"step": 1840
},
{
"epoch": 0.5868358445678034,
"grad_norm": 0.19246827065944672,
"learning_rate": 3.309931658185892e-05,
"loss": 0.4501,
"step": 1850
},
{
"epoch": 0.5900079302141158,
"grad_norm": 0.20991584658622742,
"learning_rate": 3.275318754147636e-05,
"loss": 0.4478,
"step": 1860
},
{
"epoch": 0.5931800158604282,
"grad_norm": 0.21498160064220428,
"learning_rate": 3.240540620533649e-05,
"loss": 0.4454,
"step": 1870
},
{
"epoch": 0.5963521015067407,
"grad_norm": 0.199667289853096,
"learning_rate": 3.205604668974607e-05,
"loss": 0.4469,
"step": 1880
},
{
"epoch": 0.5995241871530531,
"grad_norm": 0.21041718125343323,
"learning_rate": 3.170518344734051e-05,
"loss": 0.4517,
"step": 1890
},
{
"epoch": 0.6026962727993656,
"grad_norm": 0.2041397988796234,
"learning_rate": 3.135289125121718e-05,
"loss": 0.4518,
"step": 1900
},
{
"epoch": 0.605868358445678,
"grad_norm": 0.18860045075416565,
"learning_rate": 3.0999245179000325e-05,
"loss": 0.4511,
"step": 1910
},
{
"epoch": 0.6090404440919904,
"grad_norm": 0.20879918336868286,
"learning_rate": 3.064432059684117e-05,
"loss": 0.4504,
"step": 1920
},
{
"epoch": 0.612212529738303,
"grad_norm": 0.18158119916915894,
"learning_rate": 3.0288193143356484e-05,
"loss": 0.4501,
"step": 1930
},
{
"epoch": 0.6153846153846154,
"grad_norm": 0.20405800640583038,
"learning_rate": 2.9930938713509125e-05,
"loss": 0.4478,
"step": 1940
},
{
"epoch": 0.6185567010309279,
"grad_norm": 0.19368775188922882,
"learning_rate": 2.9572633442433917e-05,
"loss": 0.449,
"step": 1950
},
{
"epoch": 0.6217287866772403,
"grad_norm": 0.19516830146312714,
"learning_rate": 2.9213353689212337e-05,
"loss": 0.4509,
"step": 1960
},
{
"epoch": 0.6249008723235527,
"grad_norm": 0.2207994908094406,
"learning_rate": 2.8853176020599504e-05,
"loss": 0.4524,
"step": 1970
},
{
"epoch": 0.6280729579698652,
"grad_norm": 0.20896515250205994,
"learning_rate": 2.849217719470691e-05,
"loss": 0.447,
"step": 1980
},
{
"epoch": 0.6312450436161776,
"grad_norm": 0.19302915036678314,
"learning_rate": 2.8130434144644364e-05,
"loss": 0.4462,
"step": 1990
},
{
"epoch": 0.63441712926249,
"grad_norm": 0.21044501662254333,
"learning_rate": 2.776802396212461e-05,
"loss": 0.4522,
"step": 2000
},
{
"epoch": 0.63441712926249,
"eval_loss": 0.44816499948501587,
"eval_runtime": 293.5621,
"eval_samples_per_second": 3.509,
"eval_steps_per_second": 1.754,
"step": 2000
},
{
"epoch": 0.6375892149088025,
"grad_norm": 0.2075473517179489,
"learning_rate": 2.7405023881034204e-05,
"loss": 0.4487,
"step": 2010
},
{
"epoch": 0.640761300555115,
"grad_norm": 0.1936168372631073,
"learning_rate": 2.7041511260974028e-05,
"loss": 0.4485,
"step": 2020
},
{
"epoch": 0.6439333862014275,
"grad_norm": 0.2282039374113083,
"learning_rate": 2.66775635707731e-05,
"loss": 0.4478,
"step": 2030
},
{
"epoch": 0.6471054718477399,
"grad_norm": 0.18652349710464478,
"learning_rate": 2.6313258371978994e-05,
"loss": 0.449,
"step": 2040
},
{
"epoch": 0.6502775574940524,
"grad_norm": 0.20407763123512268,
"learning_rate": 2.5948673302328587e-05,
"loss": 0.4486,
"step": 2050
},
{
"epoch": 0.6534496431403648,
"grad_norm": 0.22227470576763153,
"learning_rate": 2.558388605920255e-05,
"loss": 0.4506,
"step": 2060
},
{
"epoch": 0.6566217287866772,
"grad_norm": 0.20888665318489075,
"learning_rate": 2.5218974383067085e-05,
"loss": 0.4487,
"step": 2070
},
{
"epoch": 0.6597938144329897,
"grad_norm": 0.19797295331954956,
"learning_rate": 2.4854016040906572e-05,
"loss": 0.4521,
"step": 2080
},
{
"epoch": 0.6629659000793021,
"grad_norm": 0.22293215990066528,
"learning_rate": 2.4489088809650425e-05,
"loss": 0.449,
"step": 2090
},
{
"epoch": 0.6661379857256146,
"grad_norm": 0.2310066670179367,
"learning_rate": 2.4124270459598007e-05,
"loss": 0.4479,
"step": 2100
},
{
"epoch": 0.669310071371927,
"grad_norm": 0.19995397329330444,
"learning_rate": 2.375963873784478e-05,
"loss": 0.4504,
"step": 2110
},
{
"epoch": 0.6724821570182395,
"grad_norm": 0.18815317749977112,
"learning_rate": 2.3395271351713515e-05,
"loss": 0.4501,
"step": 2120
},
{
"epoch": 0.675654242664552,
"grad_norm": 0.19861653447151184,
"learning_rate": 2.303124595219395e-05,
"loss": 0.4506,
"step": 2130
},
{
"epoch": 0.6788263283108644,
"grad_norm": 0.17454959452152252,
"learning_rate": 2.2667640117394442e-05,
"loss": 0.4453,
"step": 2140
},
{
"epoch": 0.6819984139571769,
"grad_norm": 0.19067569077014923,
"learning_rate": 2.230453133600915e-05,
"loss": 0.4466,
"step": 2150
},
{
"epoch": 0.6851704996034893,
"grad_norm": 0.1979096531867981,
"learning_rate": 2.1941996990804288e-05,
"loss": 0.4479,
"step": 2160
},
{
"epoch": 0.6883425852498017,
"grad_norm": 0.18306080996990204,
"learning_rate": 2.1580114342126933e-05,
"loss": 0.4482,
"step": 2170
},
{
"epoch": 0.6915146708961142,
"grad_norm": 0.18473172187805176,
"learning_rate": 2.1218960511439952e-05,
"loss": 0.4468,
"step": 2180
},
{
"epoch": 0.6946867565424266,
"grad_norm": 0.17530548572540283,
"learning_rate": 2.0858612464886505e-05,
"loss": 0.449,
"step": 2190
},
{
"epoch": 0.6978588421887391,
"grad_norm": 0.1818351149559021,
"learning_rate": 2.0499146996887618e-05,
"loss": 0.4458,
"step": 2200
},
{
"epoch": 0.7010309278350515,
"grad_norm": 0.19750525057315826,
"learning_rate": 2.0140640713776438e-05,
"loss": 0.4489,
"step": 2210
},
{
"epoch": 0.704203013481364,
"grad_norm": 0.20060710608959198,
"learning_rate": 1.978317001747248e-05,
"loss": 0.4482,
"step": 2220
},
{
"epoch": 0.7073750991276765,
"grad_norm": 0.19399641454219818,
"learning_rate": 1.942681108919949e-05,
"loss": 0.4494,
"step": 2230
},
{
"epoch": 0.7105471847739889,
"grad_norm": 0.2124367654323578,
"learning_rate": 1.9071639873250334e-05,
"loss": 0.4491,
"step": 2240
},
{
"epoch": 0.7137192704203014,
"grad_norm": 0.2064894288778305,
"learning_rate": 1.871773206080236e-05,
"loss": 0.4433,
"step": 2250
},
{
"epoch": 0.7168913560666138,
"grad_norm": 0.20361967384815216,
"learning_rate": 1.836516307378671e-05,
"loss": 0.4482,
"step": 2260
},
{
"epoch": 0.7200634417129262,
"grad_norm": 0.22753477096557617,
"learning_rate": 1.801400804881507e-05,
"loss": 0.4434,
"step": 2270
},
{
"epoch": 0.7232355273592387,
"grad_norm": 0.2242439240217209,
"learning_rate": 1.766434182116708e-05,
"loss": 0.4487,
"step": 2280
},
{
"epoch": 0.7264076130055511,
"grad_norm": 0.1731296181678772,
"learning_rate": 1.7316238908842127e-05,
"loss": 0.4456,
"step": 2290
},
{
"epoch": 0.7295796986518636,
"grad_norm": 0.19372820854187012,
"learning_rate": 1.6969773496678647e-05,
"loss": 0.446,
"step": 2300
},
{
"epoch": 0.732751784298176,
"grad_norm": 0.21722117066383362,
"learning_rate": 1.662501942054447e-05,
"loss": 0.4465,
"step": 2310
},
{
"epoch": 0.7359238699444886,
"grad_norm": 0.24181324243545532,
"learning_rate": 1.6282050151601518e-05,
"loss": 0.4496,
"step": 2320
},
{
"epoch": 0.739095955590801,
"grad_norm": 0.2091435343027115,
"learning_rate": 1.594093878064825e-05,
"loss": 0.4469,
"step": 2330
},
{
"epoch": 0.7422680412371134,
"grad_norm": 0.19474650919437408,
"learning_rate": 1.5601758002543137e-05,
"loss": 0.4465,
"step": 2340
},
{
"epoch": 0.7454401268834259,
"grad_norm": 0.21081089973449707,
"learning_rate": 1.5264580100712507e-05,
"loss": 0.4449,
"step": 2350
},
{
"epoch": 0.7486122125297383,
"grad_norm": 0.21220536530017853,
"learning_rate": 1.4929476931746167e-05,
"loss": 0.4461,
"step": 2360
},
{
"epoch": 0.7517842981760507,
"grad_norm": 0.19944216310977936,
"learning_rate": 1.4596519910083825e-05,
"loss": 0.4476,
"step": 2370
},
{
"epoch": 0.7549563838223632,
"grad_norm": 0.19958476722240448,
"learning_rate": 1.4265779992795893e-05,
"loss": 0.4467,
"step": 2380
},
{
"epoch": 0.7581284694686756,
"grad_norm": 0.22247722744941711,
"learning_rate": 1.3937327664461672e-05,
"loss": 0.4427,
"step": 2390
},
{
"epoch": 0.7613005551149881,
"grad_norm": 0.19884702563285828,
"learning_rate": 1.361123292214826e-05,
"loss": 0.4498,
"step": 2400
},
{
"epoch": 0.7644726407613005,
"grad_norm": 0.17867441475391388,
"learning_rate": 1.3287565260493357e-05,
"loss": 0.4484,
"step": 2410
},
{
"epoch": 0.767644726407613,
"grad_norm": 0.17917264997959137,
"learning_rate": 1.2966393656895134e-05,
"loss": 0.4439,
"step": 2420
},
{
"epoch": 0.7708168120539255,
"grad_norm": 0.24791833758354187,
"learning_rate": 1.2647786556812332e-05,
"loss": 0.4463,
"step": 2430
},
{
"epoch": 0.7739888977002379,
"grad_norm": 0.1971379816532135,
"learning_rate": 1.2331811859177722e-05,
"loss": 0.4475,
"step": 2440
},
{
"epoch": 0.7771609833465504,
"grad_norm": 0.19460198283195496,
"learning_rate": 1.2018536901928079e-05,
"loss": 0.4468,
"step": 2450
},
{
"epoch": 0.7803330689928628,
"grad_norm": 0.21018314361572266,
"learning_rate": 1.1708028447653613e-05,
"loss": 0.4489,
"step": 2460
},
{
"epoch": 0.7835051546391752,
"grad_norm": 0.18419425189495087,
"learning_rate": 1.1400352669370115e-05,
"loss": 0.4469,
"step": 2470
},
{
"epoch": 0.7866772402854877,
"grad_norm": 0.192842036485672,
"learning_rate": 1.1095575136416695e-05,
"loss": 0.4457,
"step": 2480
},
{
"epoch": 0.7898493259318001,
"grad_norm": 0.21617653965950012,
"learning_rate": 1.0793760800482179e-05,
"loss": 0.4511,
"step": 2490
},
{
"epoch": 0.7930214115781126,
"grad_norm": 0.1891343593597412,
"learning_rate": 1.0494973981763146e-05,
"loss": 0.4436,
"step": 2500
},
{
"epoch": 0.7930214115781126,
"eval_loss": 0.44588717818260193,
"eval_runtime": 293.564,
"eval_samples_per_second": 3.509,
"eval_steps_per_second": 1.754,
"step": 2500
},
{
"epoch": 0.796193497224425,
"grad_norm": 0.193357452750206,
"learning_rate": 1.0199278355256522e-05,
"loss": 0.4436,
"step": 2510
},
{
"epoch": 0.7993655828707376,
"grad_norm": 0.19498059153556824,
"learning_rate": 9.906736937189697e-06,
"loss": 0.4457,
"step": 2520
},
{
"epoch": 0.80253766851705,
"grad_norm": 0.2086293250322342,
"learning_rate": 9.61741207159104e-06,
"loss": 0.4488,
"step": 2530
},
{
"epoch": 0.8057097541633624,
"grad_norm": 0.1872304081916809,
"learning_rate": 9.331365417003601e-06,
"loss": 0.4436,
"step": 2540
},
{
"epoch": 0.8088818398096749,
"grad_norm": 0.196652352809906,
"learning_rate": 9.048657933345e-06,
"loss": 0.4448,
"step": 2550
},
{
"epoch": 0.8120539254559873,
"grad_norm": 0.21112217009067535,
"learning_rate": 8.769349868916118e-06,
"loss": 0.4481,
"step": 2560
},
{
"epoch": 0.8152260111022998,
"grad_norm": 0.19768798351287842,
"learning_rate": 8.493500747561488e-06,
"loss": 0.4493,
"step": 2570
},
{
"epoch": 0.8183980967486122,
"grad_norm": 0.19622166454792023,
"learning_rate": 8.221169355984052e-06,
"loss": 0.4448,
"step": 2580
},
{
"epoch": 0.8215701823949246,
"grad_norm": 0.2044159471988678,
"learning_rate": 7.952413731217025e-06,
"loss": 0.4472,
"step": 2590
},
{
"epoch": 0.8247422680412371,
"grad_norm": 0.206452876329422,
"learning_rate": 7.687291148255527e-06,
"loss": 0.4452,
"step": 2600
},
{
"epoch": 0.8279143536875495,
"grad_norm": 0.20858535170555115,
"learning_rate": 7.425858107850578e-06,
"loss": 0.4419,
"step": 2610
},
{
"epoch": 0.8310864393338621,
"grad_norm": 0.18986332416534424,
"learning_rate": 7.168170324468171e-06,
"loss": 0.449,
"step": 2620
},
{
"epoch": 0.8342585249801745,
"grad_norm": 0.21051757037639618,
"learning_rate": 6.9142827144158066e-06,
"loss": 0.4468,
"step": 2630
},
{
"epoch": 0.8374306106264869,
"grad_norm": 0.20453451573848724,
"learning_rate": 6.66424938413921e-06,
"loss": 0.4471,
"step": 2640
},
{
"epoch": 0.8406026962727994,
"grad_norm": 0.18574683368206024,
"learning_rate": 6.418123618691607e-06,
"loss": 0.4473,
"step": 2650
},
{
"epoch": 0.8437747819191118,
"grad_norm": 0.2309403419494629,
"learning_rate": 6.175957870378043e-06,
"loss": 0.4477,
"step": 2660
},
{
"epoch": 0.8469468675654243,
"grad_norm": 0.19485293328762054,
"learning_rate": 5.937803747577186e-06,
"loss": 0.4443,
"step": 2670
},
{
"epoch": 0.8501189532117367,
"grad_norm": 0.17677082121372223,
"learning_rate": 5.7037120037429645e-06,
"loss": 0.4501,
"step": 2680
},
{
"epoch": 0.8532910388580491,
"grad_norm": 0.18835794925689697,
"learning_rate": 5.473732526588407e-06,
"loss": 0.4451,
"step": 2690
},
{
"epoch": 0.8564631245043616,
"grad_norm": 0.2117832899093628,
"learning_rate": 5.2479143274539954e-06,
"loss": 0.4464,
"step": 2700
},
{
"epoch": 0.8596352101506741,
"grad_norm": 0.18720006942749023,
"learning_rate": 5.026305530862749e-06,
"loss": 0.4484,
"step": 2710
},
{
"epoch": 0.8628072957969866,
"grad_norm": 0.17910288274288177,
"learning_rate": 4.80895336426434e-06,
"loss": 0.4457,
"step": 2720
},
{
"epoch": 0.865979381443299,
"grad_norm": 0.1965160220861435,
"learning_rate": 4.595904147970356e-06,
"loss": 0.4455,
"step": 2730
},
{
"epoch": 0.8691514670896114,
"grad_norm": 0.16718564927577972,
"learning_rate": 4.3872032852828955e-06,
"loss": 0.4455,
"step": 2740
},
{
"epoch": 0.8723235527359239,
"grad_norm": 0.19451048970222473,
"learning_rate": 4.182895252818589e-06,
"loss": 0.4454,
"step": 2750
},
{
"epoch": 0.8754956383822363,
"grad_norm": 0.2160135954618454,
"learning_rate": 3.983023591030113e-06,
"loss": 0.4462,
"step": 2760
},
{
"epoch": 0.8786677240285488,
"grad_norm": 0.1994054913520813,
"learning_rate": 3.7876308949271995e-06,
"loss": 0.4455,
"step": 2770
},
{
"epoch": 0.8818398096748612,
"grad_norm": 0.20801877975463867,
"learning_rate": 3.5967588049991313e-06,
"loss": 0.4465,
"step": 2780
},
{
"epoch": 0.8850118953211736,
"grad_norm": 0.20264749228954315,
"learning_rate": 3.410447998340688e-06,
"loss": 0.4451,
"step": 2790
},
{
"epoch": 0.8881839809674861,
"grad_norm": 0.173824280500412,
"learning_rate": 3.2287381799833428e-06,
"loss": 0.4453,
"step": 2800
},
{
"epoch": 0.8913560666137986,
"grad_norm": 0.18437042832374573,
"learning_rate": 3.051668074433667e-06,
"loss": 0.449,
"step": 2810
},
{
"epoch": 0.8945281522601111,
"grad_norm": 0.17420920729637146,
"learning_rate": 2.8792754174206905e-06,
"loss": 0.4453,
"step": 2820
},
{
"epoch": 0.8977002379064235,
"grad_norm": 0.18584947288036346,
"learning_rate": 2.7115969478539562e-06,
"loss": 0.4446,
"step": 2830
},
{
"epoch": 0.9008723235527359,
"grad_norm": 0.18006384372711182,
"learning_rate": 2.5486683999940335e-06,
"loss": 0.4449,
"step": 2840
},
{
"epoch": 0.9040444091990484,
"grad_norm": 0.19664862751960754,
"learning_rate": 2.3905244958371263e-06,
"loss": 0.4463,
"step": 2850
},
{
"epoch": 0.9072164948453608,
"grad_norm": 0.20159141719341278,
"learning_rate": 2.2371989377154013e-06,
"loss": 0.444,
"step": 2860
},
{
"epoch": 0.9103885804916733,
"grad_norm": 0.23178640007972717,
"learning_rate": 2.088724401114625e-06,
"loss": 0.4472,
"step": 2870
},
{
"epoch": 0.9135606661379857,
"grad_norm": 0.18047955632209778,
"learning_rate": 1.9451325277106413e-06,
"loss": 0.4489,
"step": 2880
},
{
"epoch": 0.9167327517842981,
"grad_norm": 0.18607738614082336,
"learning_rate": 1.8064539186261387e-06,
"loss": 0.4489,
"step": 2890
},
{
"epoch": 0.9199048374306106,
"grad_norm": 0.19285809993743896,
"learning_rate": 1.6727181279092036e-06,
"loss": 0.4448,
"step": 2900
},
{
"epoch": 0.9230769230769231,
"grad_norm": 0.1946035474538803,
"learning_rate": 1.5439536562349976e-06,
"loss": 0.4474,
"step": 2910
},
{
"epoch": 0.9262490087232356,
"grad_norm": 0.1947924941778183,
"learning_rate": 1.4201879448319355e-06,
"loss": 0.4465,
"step": 2920
},
{
"epoch": 0.929421094369548,
"grad_norm": 0.17740875482559204,
"learning_rate": 1.301447369633621e-06,
"loss": 0.4475,
"step": 2930
},
{
"epoch": 0.9325931800158604,
"grad_norm": 0.18701794743537903,
"learning_rate": 1.187757235657841e-06,
"loss": 0.4488,
"step": 2940
},
{
"epoch": 0.9357652656621729,
"grad_norm": 0.19531536102294922,
"learning_rate": 1.0791417716137565e-06,
"loss": 0.4458,
"step": 2950
},
{
"epoch": 0.9389373513084853,
"grad_norm": 0.17441792786121368,
"learning_rate": 9.756241247384807e-07,
"loss": 0.4426,
"step": 2960
},
{
"epoch": 0.9421094369547978,
"grad_norm": 0.20463122427463531,
"learning_rate": 8.772263558641386e-07,
"loss": 0.4485,
"step": 2970
},
{
"epoch": 0.9452815226011102,
"grad_norm": 0.18601520359516144,
"learning_rate": 7.839694347164223e-07,
"loss": 0.4421,
"step": 2980
},
{
"epoch": 0.9484536082474226,
"grad_norm": 0.18209311366081238,
"learning_rate": 6.958732354457209e-07,
"loss": 0.4499,
"step": 2990
},
{
"epoch": 0.9516256938937351,
"grad_norm": 0.1675465852022171,
"learning_rate": 6.129565323916813e-07,
"loss": 0.4463,
"step": 3000
},
{
"epoch": 0.9516256938937351,
"eval_loss": 0.44520583748817444,
"eval_runtime": 293.5596,
"eval_samples_per_second": 3.509,
"eval_steps_per_second": 1.754,
"step": 3000
},
{
"epoch": 0.9547977795400476,
"grad_norm": 0.21267291903495789,
"learning_rate": 5.352369960821946e-07,
"loss": 0.4428,
"step": 3010
},
{
"epoch": 0.9579698651863601,
"grad_norm": 0.19138512015342712,
"learning_rate": 4.627311894675856e-07,
"loss": 0.446,
"step": 3020
},
{
"epoch": 0.9611419508326725,
"grad_norm": 0.19043755531311035,
"learning_rate": 3.954545643908514e-07,
"loss": 0.4446,
"step": 3030
},
{
"epoch": 0.964314036478985,
"grad_norm": 0.1850534826517105,
"learning_rate": 3.3342145829469983e-07,
"loss": 0.4435,
"step": 3040
},
{
"epoch": 0.9674861221252974,
"grad_norm": 0.21564887464046478,
"learning_rate": 2.7664509116607506e-07,
"loss": 0.4451,
"step": 3050
},
{
"epoch": 0.9706582077716098,
"grad_norm": 0.1970747411251068,
"learning_rate": 2.251375627187996e-07,
"loss": 0.4465,
"step": 3060
},
{
"epoch": 0.9738302934179223,
"grad_norm": 0.1968993842601776,
"learning_rate": 1.789098498150066e-07,
"loss": 0.4479,
"step": 3070
},
{
"epoch": 0.9770023790642347,
"grad_norm": 0.18383654952049255,
"learning_rate": 1.3797180412583322e-07,
"loss": 0.4472,
"step": 3080
},
{
"epoch": 0.9801744647105471,
"grad_norm": 0.18183210492134094,
"learning_rate": 1.0233215003190577e-07,
"loss": 0.4484,
"step": 3090
},
{
"epoch": 0.9833465503568596,
"grad_norm": 0.18428853154182434,
"learning_rate": 7.199848276408238e-08,
"loss": 0.4445,
"step": 3100
},
{
"epoch": 0.9865186360031721,
"grad_norm": 0.21674080193042755,
"learning_rate": 4.6977266784811736e-08,
"loss": 0.4465,
"step": 3110
},
{
"epoch": 0.9896907216494846,
"grad_norm": 0.2023499608039856,
"learning_rate": 2.7273834410485034e-08,
"loss": 0.4448,
"step": 3120
},
{
"epoch": 0.992862807295797,
"grad_norm": 0.19151511788368225,
"learning_rate": 1.2892384675056158e-08,
"loss": 0.4461,
"step": 3130
},
{
"epoch": 0.9960348929421095,
"grad_norm": 0.16607537865638733,
"learning_rate": 3.8359824351685836e-09,
"loss": 0.4453,
"step": 3140
},
{
"epoch": 0.9992069785884219,
"grad_norm": 0.21372340619564056,
"learning_rate": 1.0655771701395534e-10,
"loss": 0.448,
"step": 3150
},
{
"epoch": 0.9998413957176844,
"step": 3152,
"total_flos": 2.0499162760961065e+18,
"train_loss": 0.5496831483344742,
"train_runtime": 71646.9274,
"train_samples_per_second": 0.704,
"train_steps_per_second": 0.044
}
],
"logging_steps": 10,
"max_steps": 3152,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"total_flos": 2.0499162760961065e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}