ruadapt_qwen2.5_1.5B_test / trainer_state.json
Alant2000's picture
Upload 11 files
b2563c7 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 1000,
"global_step": 1317,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0007593014426727411,
"eval_accuracy": 0.14927598308927809,
"eval_loss": 7.863234519958496,
"eval_runtime": 368.109,
"eval_samples_per_second": 192.592,
"eval_steps_per_second": 0.752,
"step": 1
},
{
"epoch": 0.007593014426727411,
"grad_norm": 9.6875,
"learning_rate": 1e-05,
"loss": 7.8621,
"step": 10
},
{
"epoch": 0.015186028853454821,
"grad_norm": 9.625,
"learning_rate": 2e-05,
"loss": 7.8394,
"step": 20
},
{
"epoch": 0.022779043280182234,
"grad_norm": 9.4375,
"learning_rate": 3e-05,
"loss": 7.7726,
"step": 30
},
{
"epoch": 0.030372057706909643,
"grad_norm": 9.0625,
"learning_rate": 4e-05,
"loss": 7.6201,
"step": 40
},
{
"epoch": 0.037965072133637055,
"grad_norm": 8.375,
"learning_rate": 5e-05,
"loss": 7.3344,
"step": 50
},
{
"epoch": 0.04555808656036447,
"grad_norm": 7.3125,
"learning_rate": 6e-05,
"loss": 7.0089,
"step": 60
},
{
"epoch": 0.05315110098709187,
"grad_norm": 5.0,
"learning_rate": 7e-05,
"loss": 6.5946,
"step": 70
},
{
"epoch": 0.060744115413819286,
"grad_norm": 2.234375,
"learning_rate": 8e-05,
"loss": 6.0679,
"step": 80
},
{
"epoch": 0.0683371298405467,
"grad_norm": 0.9296875,
"learning_rate": 9e-05,
"loss": 5.6963,
"step": 90
},
{
"epoch": 0.07593014426727411,
"grad_norm": 0.55859375,
"learning_rate": 0.0001,
"loss": 5.4276,
"step": 100
},
{
"epoch": 0.08352315869400152,
"grad_norm": 0.43359375,
"learning_rate": 9.998334155418425e-05,
"loss": 5.2216,
"step": 110
},
{
"epoch": 0.09111617312072894,
"grad_norm": 0.33203125,
"learning_rate": 9.993337731688963e-05,
"loss": 5.0464,
"step": 120
},
{
"epoch": 0.09870918754745633,
"grad_norm": 0.298828125,
"learning_rate": 9.985014058117778e-05,
"loss": 4.9161,
"step": 130
},
{
"epoch": 0.10630220197418375,
"grad_norm": 0.275390625,
"learning_rate": 9.973368681083473e-05,
"loss": 4.8156,
"step": 140
},
{
"epoch": 0.11389521640091116,
"grad_norm": 0.255859375,
"learning_rate": 9.958409360341343e-05,
"loss": 4.735,
"step": 150
},
{
"epoch": 0.12148823082763857,
"grad_norm": 0.251953125,
"learning_rate": 9.940146063852748e-05,
"loss": 4.6663,
"step": 160
},
{
"epoch": 0.12908124525436598,
"grad_norm": 0.240234375,
"learning_rate": 9.918590961143088e-05,
"loss": 4.6082,
"step": 170
},
{
"epoch": 0.1366742596810934,
"grad_norm": 0.259765625,
"learning_rate": 9.893758415192783e-05,
"loss": 4.5533,
"step": 180
},
{
"epoch": 0.1442672741078208,
"grad_norm": 0.236328125,
"learning_rate": 9.865664972866682e-05,
"loss": 4.5055,
"step": 190
},
{
"epoch": 0.15186028853454822,
"grad_norm": 0.2314453125,
"learning_rate": 9.834329353888254e-05,
"loss": 4.4668,
"step": 200
},
{
"epoch": 0.15945330296127563,
"grad_norm": 0.25390625,
"learning_rate": 9.799772438365936e-05,
"loss": 4.4333,
"step": 210
},
{
"epoch": 0.16704631738800305,
"grad_norm": 0.259765625,
"learning_rate": 9.762017252879917e-05,
"loss": 4.4072,
"step": 220
},
{
"epoch": 0.17463933181473046,
"grad_norm": 0.2333984375,
"learning_rate": 9.721088955138664e-05,
"loss": 4.3759,
"step": 230
},
{
"epoch": 0.18223234624145787,
"grad_norm": 0.255859375,
"learning_rate": 9.67701481721539e-05,
"loss": 4.3542,
"step": 240
},
{
"epoch": 0.18982536066818528,
"grad_norm": 0.220703125,
"learning_rate": 9.629824207375632e-05,
"loss": 4.3318,
"step": 250
},
{
"epoch": 0.19741837509491267,
"grad_norm": 0.2216796875,
"learning_rate": 9.57954857050807e-05,
"loss": 4.3142,
"step": 260
},
{
"epoch": 0.20501138952164008,
"grad_norm": 0.23828125,
"learning_rate": 9.52622140717161e-05,
"loss": 4.2948,
"step": 270
},
{
"epoch": 0.2126044039483675,
"grad_norm": 0.21875,
"learning_rate": 9.46987825127269e-05,
"loss": 4.282,
"step": 280
},
{
"epoch": 0.2201974183750949,
"grad_norm": 0.2109375,
"learning_rate": 9.410556646387694e-05,
"loss": 4.2621,
"step": 290
},
{
"epoch": 0.22779043280182232,
"grad_norm": 0.2216796875,
"learning_rate": 9.348296120746251e-05,
"loss": 4.2483,
"step": 300
},
{
"epoch": 0.23538344722854973,
"grad_norm": 0.205078125,
"learning_rate": 9.283138160892074e-05,
"loss": 4.2307,
"step": 310
},
{
"epoch": 0.24297646165527714,
"grad_norm": 0.2060546875,
"learning_rate": 9.215126184038911e-05,
"loss": 4.2256,
"step": 320
},
{
"epoch": 0.2505694760820046,
"grad_norm": 0.232421875,
"learning_rate": 9.144305509140012e-05,
"loss": 4.2123,
"step": 330
},
{
"epoch": 0.25816249050873197,
"grad_norm": 0.228515625,
"learning_rate": 9.070723326690396e-05,
"loss": 4.1984,
"step": 340
},
{
"epoch": 0.26575550493545935,
"grad_norm": 0.21484375,
"learning_rate": 8.994428667282032e-05,
"loss": 4.1908,
"step": 350
},
{
"epoch": 0.2733485193621868,
"grad_norm": 0.2041015625,
"learning_rate": 8.915472368932916e-05,
"loss": 4.1878,
"step": 360
},
{
"epoch": 0.2809415337889142,
"grad_norm": 0.205078125,
"learning_rate": 8.83390704321176e-05,
"loss": 4.1766,
"step": 370
},
{
"epoch": 0.2885345482156416,
"grad_norm": 0.2119140625,
"learning_rate": 8.749787040180923e-05,
"loss": 4.1719,
"step": 380
},
{
"epoch": 0.296127562642369,
"grad_norm": 0.203125,
"learning_rate": 8.663168412180907e-05,
"loss": 4.1615,
"step": 390
},
{
"epoch": 0.30372057706909644,
"grad_norm": 0.2041015625,
"learning_rate": 8.574108876480557e-05,
"loss": 4.1505,
"step": 400
},
{
"epoch": 0.3113135914958238,
"grad_norm": 0.22265625,
"learning_rate": 8.482667776817868e-05,
"loss": 4.1422,
"step": 410
},
{
"epoch": 0.31890660592255127,
"grad_norm": 0.2236328125,
"learning_rate": 8.388906043857001e-05,
"loss": 4.1427,
"step": 420
},
{
"epoch": 0.32649962034927865,
"grad_norm": 0.2060546875,
"learning_rate": 8.292886154587884e-05,
"loss": 4.14,
"step": 430
},
{
"epoch": 0.3340926347760061,
"grad_norm": 0.2216796875,
"learning_rate": 8.194672090695419e-05,
"loss": 4.127,
"step": 440
},
{
"epoch": 0.3416856492027335,
"grad_norm": 0.232421875,
"learning_rate": 8.094329295926075e-05,
"loss": 4.1251,
"step": 450
},
{
"epoch": 0.3492786636294609,
"grad_norm": 0.2197265625,
"learning_rate": 7.991924632480238e-05,
"loss": 4.1234,
"step": 460
},
{
"epoch": 0.3568716780561883,
"grad_norm": 0.205078125,
"learning_rate": 7.887526336459401e-05,
"loss": 4.1151,
"step": 470
},
{
"epoch": 0.36446469248291574,
"grad_norm": 0.201171875,
"learning_rate": 7.781203972397866e-05,
"loss": 4.1102,
"step": 480
},
{
"epoch": 0.3720577069096431,
"grad_norm": 0.20703125,
"learning_rate": 7.673028386909259e-05,
"loss": 4.1107,
"step": 490
},
{
"epoch": 0.37965072133637057,
"grad_norm": 0.201171875,
"learning_rate": 7.563071661478758e-05,
"loss": 4.1017,
"step": 500
},
{
"epoch": 0.38724373576309795,
"grad_norm": 0.2177734375,
"learning_rate": 7.451407064432471e-05,
"loss": 4.096,
"step": 510
},
{
"epoch": 0.39483675018982534,
"grad_norm": 0.21484375,
"learning_rate": 7.338109002115976e-05,
"loss": 4.0982,
"step": 520
},
{
"epoch": 0.4024297646165528,
"grad_norm": 0.1962890625,
"learning_rate": 7.223252969314557e-05,
"loss": 4.0925,
"step": 530
},
{
"epoch": 0.41002277904328016,
"grad_norm": 0.201171875,
"learning_rate": 7.106915498948175e-05,
"loss": 4.0917,
"step": 540
},
{
"epoch": 0.4176157934700076,
"grad_norm": 0.208984375,
"learning_rate": 6.989174111074688e-05,
"loss": 4.0865,
"step": 550
},
{
"epoch": 0.425208807896735,
"grad_norm": 0.2060546875,
"learning_rate": 6.870107261235304e-05,
"loss": 4.0831,
"step": 560
},
{
"epoch": 0.4328018223234624,
"grad_norm": 0.1845703125,
"learning_rate": 6.749794288176681e-05,
"loss": 4.0833,
"step": 570
},
{
"epoch": 0.4403948367501898,
"grad_norm": 0.1962890625,
"learning_rate": 6.628315360984526e-05,
"loss": 4.0813,
"step": 580
},
{
"epoch": 0.44798785117691725,
"grad_norm": 0.2021484375,
"learning_rate": 6.505751425663892e-05,
"loss": 4.0757,
"step": 590
},
{
"epoch": 0.45558086560364464,
"grad_norm": 0.19140625,
"learning_rate": 6.382184151201804e-05,
"loss": 4.0782,
"step": 600
},
{
"epoch": 0.4631738800303721,
"grad_norm": 0.189453125,
"learning_rate": 6.257695875148106e-05,
"loss": 4.0763,
"step": 610
},
{
"epoch": 0.47076689445709946,
"grad_norm": 0.1904296875,
"learning_rate": 6.132369548750851e-05,
"loss": 4.0728,
"step": 620
},
{
"epoch": 0.4783599088838269,
"grad_norm": 0.1904296875,
"learning_rate": 6.0062886816827494e-05,
"loss": 4.0696,
"step": 630
},
{
"epoch": 0.4859529233105543,
"grad_norm": 0.19921875,
"learning_rate": 5.8795372863954924e-05,
"loss": 4.0634,
"step": 640
},
{
"epoch": 0.4935459377372817,
"grad_norm": 0.189453125,
"learning_rate": 5.7521998221391004e-05,
"loss": 4.0658,
"step": 650
},
{
"epoch": 0.5011389521640092,
"grad_norm": 0.19921875,
"learning_rate": 5.6243611386835205e-05,
"loss": 4.0631,
"step": 660
},
{
"epoch": 0.5087319665907365,
"grad_norm": 0.185546875,
"learning_rate": 5.49610641978001e-05,
"loss": 4.0639,
"step": 670
},
{
"epoch": 0.5163249810174639,
"grad_norm": 0.1962890625,
"learning_rate": 5.36752112639999e-05,
"loss": 4.0631,
"step": 680
},
{
"epoch": 0.5239179954441914,
"grad_norm": 0.205078125,
"learning_rate": 5.2386909397891595e-05,
"loss": 4.0592,
"step": 690
},
{
"epoch": 0.5315110098709187,
"grad_norm": 0.1943359375,
"learning_rate": 5.109701704374841e-05,
"loss": 4.0625,
"step": 700
},
{
"epoch": 0.5391040242976461,
"grad_norm": 0.1884765625,
"learning_rate": 4.980639370564593e-05,
"loss": 4.0632,
"step": 710
},
{
"epoch": 0.5466970387243736,
"grad_norm": 0.185546875,
"learning_rate": 4.851589937474202e-05,
"loss": 4.0557,
"step": 720
},
{
"epoch": 0.554290053151101,
"grad_norm": 0.19140625,
"learning_rate": 4.722639395623215e-05,
"loss": 4.0654,
"step": 730
},
{
"epoch": 0.5618830675778284,
"grad_norm": 0.1875,
"learning_rate": 4.593873669636204e-05,
"loss": 4.0542,
"step": 740
},
{
"epoch": 0.5694760820045558,
"grad_norm": 0.185546875,
"learning_rate": 4.465378560987944e-05,
"loss": 4.0584,
"step": 750
},
{
"epoch": 0.5770690964312832,
"grad_norm": 0.181640625,
"learning_rate": 4.337239690830632e-05,
"loss": 4.058,
"step": 760
},
{
"epoch": 0.5846621108580107,
"grad_norm": 0.181640625,
"learning_rate": 4.209542442941284e-05,
"loss": 4.0574,
"step": 770
},
{
"epoch": 0.592255125284738,
"grad_norm": 0.20703125,
"learning_rate": 4.082371906827295e-05,
"loss": 4.0569,
"step": 780
},
{
"epoch": 0.5998481397114654,
"grad_norm": 0.18359375,
"learning_rate": 3.9558128210280706e-05,
"loss": 4.0512,
"step": 790
},
{
"epoch": 0.6074411541381929,
"grad_norm": 0.1875,
"learning_rate": 3.829949516650543e-05,
"loss": 4.054,
"step": 800
},
{
"epoch": 0.6150341685649203,
"grad_norm": 0.1904296875,
"learning_rate": 3.704865861176159e-05,
"loss": 4.0514,
"step": 810
},
{
"epoch": 0.6226271829916477,
"grad_norm": 0.1865234375,
"learning_rate": 3.5806452025768056e-05,
"loss": 4.0564,
"step": 820
},
{
"epoch": 0.6302201974183751,
"grad_norm": 0.1767578125,
"learning_rate": 3.457370313776897e-05,
"loss": 4.0529,
"step": 830
},
{
"epoch": 0.6378132118451025,
"grad_norm": 0.1796875,
"learning_rate": 3.335123337498662e-05,
"loss": 4.055,
"step": 840
},
{
"epoch": 0.6454062262718299,
"grad_norm": 0.1943359375,
"learning_rate": 3.213985731527313e-05,
"loss": 4.053,
"step": 850
},
{
"epoch": 0.6529992406985573,
"grad_norm": 0.189453125,
"learning_rate": 3.0940382144326666e-05,
"loss": 4.0494,
"step": 860
},
{
"epoch": 0.6605922551252847,
"grad_norm": 0.197265625,
"learning_rate": 2.9753607117832903e-05,
"loss": 4.0488,
"step": 870
},
{
"epoch": 0.6681852695520122,
"grad_norm": 0.19140625,
"learning_rate": 2.858032302889084e-05,
"loss": 4.0548,
"step": 880
},
{
"epoch": 0.6757782839787395,
"grad_norm": 0.189453125,
"learning_rate": 2.7421311681077344e-05,
"loss": 4.0507,
"step": 890
},
{
"epoch": 0.683371298405467,
"grad_norm": 0.189453125,
"learning_rate": 2.627734536750191e-05,
"loss": 4.0524,
"step": 900
},
{
"epoch": 0.6909643128321944,
"grad_norm": 0.1884765625,
"learning_rate": 2.5149186356198562e-05,
"loss": 4.0494,
"step": 910
},
{
"epoch": 0.6985573272589218,
"grad_norm": 0.189453125,
"learning_rate": 2.4037586382197685e-05,
"loss": 4.0516,
"step": 920
},
{
"epoch": 0.7061503416856492,
"grad_norm": 0.197265625,
"learning_rate": 2.2943286146616788e-05,
"loss": 4.0507,
"step": 930
},
{
"epoch": 0.7137433561123766,
"grad_norm": 0.1943359375,
"learning_rate": 2.1867014823103044e-05,
"loss": 4.0507,
"step": 940
},
{
"epoch": 0.721336370539104,
"grad_norm": 0.189453125,
"learning_rate": 2.080948957195749e-05,
"loss": 4.0506,
"step": 950
},
{
"epoch": 0.7289293849658315,
"grad_norm": 0.1796875,
"learning_rate": 1.9771415062263965e-05,
"loss": 4.0502,
"step": 960
},
{
"epoch": 0.7365223993925588,
"grad_norm": 0.189453125,
"learning_rate": 1.875348300234131e-05,
"loss": 4.0503,
"step": 970
},
{
"epoch": 0.7441154138192863,
"grad_norm": 0.185546875,
"learning_rate": 1.7756371678832145e-05,
"loss": 4.049,
"step": 980
},
{
"epoch": 0.7517084282460137,
"grad_norm": 0.18359375,
"learning_rate": 1.6780745504734617e-05,
"loss": 4.0465,
"step": 990
},
{
"epoch": 0.7593014426727411,
"grad_norm": 0.189453125,
"learning_rate": 1.58272545766791e-05,
"loss": 4.0503,
"step": 1000
},
{
"epoch": 0.7593014426727411,
"eval_accuracy": 0.30700041923136506,
"eval_loss": 4.048835277557373,
"eval_runtime": 369.3769,
"eval_samples_per_second": 191.931,
"eval_steps_per_second": 0.75,
"step": 1000
},
{
"epoch": 0.7668944570994685,
"grad_norm": 0.1865234375,
"learning_rate": 1.4896534241743988e-05,
"loss": 4.0501,
"step": 1010
},
{
"epoch": 0.7744874715261959,
"grad_norm": 0.1982421875,
"learning_rate": 1.398920467410002e-05,
"loss": 4.0469,
"step": 1020
},
{
"epoch": 0.7820804859529233,
"grad_norm": 0.1884765625,
"learning_rate": 1.3105870461764847e-05,
"loss": 4.0512,
"step": 1030
},
{
"epoch": 0.7896735003796507,
"grad_norm": 0.185546875,
"learning_rate": 1.2247120203742945e-05,
"loss": 4.0463,
"step": 1040
},
{
"epoch": 0.7972665148063781,
"grad_norm": 0.19140625,
"learning_rate": 1.141352611782005e-05,
"loss": 4.0457,
"step": 1050
},
{
"epoch": 0.8048595292331056,
"grad_norm": 0.189453125,
"learning_rate": 1.0605643659272668e-05,
"loss": 4.0511,
"step": 1060
},
{
"epoch": 0.812452543659833,
"grad_norm": 0.1875,
"learning_rate": 9.824011150747226e-06,
"loss": 4.0533,
"step": 1070
},
{
"epoch": 0.8200455580865603,
"grad_norm": 0.1884765625,
"learning_rate": 9.069149423555378e-06,
"loss": 4.0541,
"step": 1080
},
{
"epoch": 0.8276385725132878,
"grad_norm": 0.181640625,
"learning_rate": 8.341561470624365e-06,
"loss": 4.0464,
"step": 1090
},
{
"epoch": 0.8352315869400152,
"grad_norm": 0.181640625,
"learning_rate": 7.641732111333793e-06,
"loss": 4.0518,
"step": 1100
},
{
"epoch": 0.8428246013667426,
"grad_norm": 0.1875,
"learning_rate": 6.970127668462129e-06,
"loss": 4.0492,
"step": 1110
},
{
"epoch": 0.85041761579347,
"grad_norm": 0.1806640625,
"learning_rate": 6.3271956574582296e-06,
"loss": 4.0489,
"step": 1120
},
{
"epoch": 0.8580106302201974,
"grad_norm": 0.1884765625,
"learning_rate": 5.7133644882448376e-06,
"loss": 4.0462,
"step": 1130
},
{
"epoch": 0.8656036446469249,
"grad_norm": 0.1806640625,
"learning_rate": 5.12904317975283e-06,
"loss": 4.0466,
"step": 1140
},
{
"epoch": 0.8731966590736523,
"grad_norm": 0.181640625,
"learning_rate": 4.574621087376501e-06,
"loss": 4.0496,
"step": 1150
},
{
"epoch": 0.8807896735003796,
"grad_norm": 0.18359375,
"learning_rate": 4.050467643531214e-06,
"loss": 4.0478,
"step": 1160
},
{
"epoch": 0.8883826879271071,
"grad_norm": 0.18359375,
"learning_rate": 3.556932111486727e-06,
"loss": 4.0524,
"step": 1170
},
{
"epoch": 0.8959757023538345,
"grad_norm": 0.1806640625,
"learning_rate": 3.0943433526397838e-06,
"loss": 4.0526,
"step": 1180
},
{
"epoch": 0.9035687167805618,
"grad_norm": 0.18359375,
"learning_rate": 2.663009607381345e-06,
"loss": 4.0484,
"step": 1190
},
{
"epoch": 0.9111617312072893,
"grad_norm": 0.177734375,
"learning_rate": 2.2632182897043773e-06,
"loss": 4.0489,
"step": 1200
},
{
"epoch": 0.9187547456340167,
"grad_norm": 0.1767578125,
"learning_rate": 1.8952357956889988e-06,
"loss": 4.0477,
"step": 1210
},
{
"epoch": 0.9263477600607442,
"grad_norm": 0.1748046875,
"learning_rate": 1.559307325992726e-06,
"loss": 4.0498,
"step": 1220
},
{
"epoch": 0.9339407744874715,
"grad_norm": 0.181640625,
"learning_rate": 1.2556567224639714e-06,
"loss": 4.0451,
"step": 1230
},
{
"epoch": 0.9415337889141989,
"grad_norm": 0.181640625,
"learning_rate": 9.844863189877596e-07,
"loss": 4.0513,
"step": 1240
},
{
"epoch": 0.9491268033409264,
"grad_norm": 0.177734375,
"learning_rate": 7.459768066630291e-07,
"loss": 4.0487,
"step": 1250
},
{
"epoch": 0.9567198177676538,
"grad_norm": 0.177734375,
"learning_rate": 5.402871134012666e-07,
"loss": 4.0509,
"step": 1260
},
{
"epoch": 0.9643128321943811,
"grad_norm": 0.2001953125,
"learning_rate": 3.6755429802690087e-07,
"loss": 4.048,
"step": 1270
},
{
"epoch": 0.9719058466211086,
"grad_norm": 0.1787109375,
"learning_rate": 2.2789345894972945e-07,
"loss": 4.0524,
"step": 1280
},
{
"epoch": 0.979498861047836,
"grad_norm": 0.1806640625,
"learning_rate": 1.213976574705733e-07,
"loss": 4.0501,
"step": 1290
},
{
"epoch": 0.9870918754745635,
"grad_norm": 0.1875,
"learning_rate": 4.813785577097996e-08,
"loss": 4.0455,
"step": 1300
},
{
"epoch": 0.9946848899012908,
"grad_norm": 0.1845703125,
"learning_rate": 8.162869628425274e-09,
"loss": 4.0489,
"step": 1310
},
{
"epoch": 1.0,
"step": 1317,
"total_flos": 1.0857282238836376e+19,
"train_loss": 4.36365500488513,
"train_runtime": 5657.2998,
"train_samples_per_second": 238.343,
"train_steps_per_second": 0.233
}
],
"logging_steps": 10,
"max_steps": 1317,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.0857282238836376e+19,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}