pythia-2_8b_sft-gpt4 / trainer_state.json
DatPySci's picture
Model save
950d057 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 100,
"global_step": 1231,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.008123476848090982,
"grad_norm": 3.203125,
"learning_rate": 1.6129032258064516e-06,
"loss": 1.5961,
"step": 10
},
{
"epoch": 0.016246953696181964,
"grad_norm": 3.453125,
"learning_rate": 3.225806451612903e-06,
"loss": 1.5961,
"step": 20
},
{
"epoch": 0.024370430544272948,
"grad_norm": 2.796875,
"learning_rate": 4.838709677419355e-06,
"loss": 1.5969,
"step": 30
},
{
"epoch": 0.03249390739236393,
"grad_norm": 2.609375,
"learning_rate": 6.451612903225806e-06,
"loss": 1.5125,
"step": 40
},
{
"epoch": 0.04061738424045491,
"grad_norm": 2.015625,
"learning_rate": 8.064516129032258e-06,
"loss": 1.4758,
"step": 50
},
{
"epoch": 0.048740861088545896,
"grad_norm": 2.078125,
"learning_rate": 9.67741935483871e-06,
"loss": 1.3445,
"step": 60
},
{
"epoch": 0.05686433793663688,
"grad_norm": 2.21875,
"learning_rate": 1.1290322580645164e-05,
"loss": 1.4297,
"step": 70
},
{
"epoch": 0.06498781478472786,
"grad_norm": 1.84375,
"learning_rate": 1.2903225806451613e-05,
"loss": 1.3977,
"step": 80
},
{
"epoch": 0.07311129163281885,
"grad_norm": 1.9765625,
"learning_rate": 1.4516129032258066e-05,
"loss": 1.3516,
"step": 90
},
{
"epoch": 0.08123476848090982,
"grad_norm": 1.8828125,
"learning_rate": 1.6129032258064517e-05,
"loss": 1.3461,
"step": 100
},
{
"epoch": 0.08935824532900082,
"grad_norm": 2.015625,
"learning_rate": 1.774193548387097e-05,
"loss": 1.3195,
"step": 110
},
{
"epoch": 0.09748172217709179,
"grad_norm": 2.078125,
"learning_rate": 1.935483870967742e-05,
"loss": 1.3828,
"step": 120
},
{
"epoch": 0.10560519902518278,
"grad_norm": 1.84375,
"learning_rate": 1.989159891598916e-05,
"loss": 1.3586,
"step": 130
},
{
"epoch": 0.11372867587327376,
"grad_norm": 2.0625,
"learning_rate": 1.971093044263776e-05,
"loss": 1.3945,
"step": 140
},
{
"epoch": 0.12185215272136475,
"grad_norm": 1.984375,
"learning_rate": 1.9530261969286363e-05,
"loss": 1.343,
"step": 150
},
{
"epoch": 0.12997562956945571,
"grad_norm": 2.0625,
"learning_rate": 1.934959349593496e-05,
"loss": 1.3172,
"step": 160
},
{
"epoch": 0.13809910641754672,
"grad_norm": 1.8671875,
"learning_rate": 1.916892502258356e-05,
"loss": 1.3141,
"step": 170
},
{
"epoch": 0.1462225832656377,
"grad_norm": 2.171875,
"learning_rate": 1.898825654923216e-05,
"loss": 1.3242,
"step": 180
},
{
"epoch": 0.15434606011372867,
"grad_norm": 1.921875,
"learning_rate": 1.8807588075880762e-05,
"loss": 1.3633,
"step": 190
},
{
"epoch": 0.16246953696181965,
"grad_norm": 1.890625,
"learning_rate": 1.862691960252936e-05,
"loss": 1.3359,
"step": 200
},
{
"epoch": 0.17059301380991065,
"grad_norm": 1.7421875,
"learning_rate": 1.8446251129177958e-05,
"loss": 1.3266,
"step": 210
},
{
"epoch": 0.17871649065800163,
"grad_norm": 1.984375,
"learning_rate": 1.826558265582656e-05,
"loss": 1.3367,
"step": 220
},
{
"epoch": 0.1868399675060926,
"grad_norm": 2.09375,
"learning_rate": 1.808491418247516e-05,
"loss": 1.3023,
"step": 230
},
{
"epoch": 0.19496344435418358,
"grad_norm": 1.8515625,
"learning_rate": 1.790424570912376e-05,
"loss": 1.2961,
"step": 240
},
{
"epoch": 0.20308692120227456,
"grad_norm": 1.859375,
"learning_rate": 1.772357723577236e-05,
"loss": 1.3711,
"step": 250
},
{
"epoch": 0.21121039805036557,
"grad_norm": 1.9375,
"learning_rate": 1.754290876242096e-05,
"loss": 1.3062,
"step": 260
},
{
"epoch": 0.21933387489845654,
"grad_norm": 2.140625,
"learning_rate": 1.7362240289069557e-05,
"loss": 1.3281,
"step": 270
},
{
"epoch": 0.22745735174654752,
"grad_norm": 1.7734375,
"learning_rate": 1.718157181571816e-05,
"loss": 1.2883,
"step": 280
},
{
"epoch": 0.2355808285946385,
"grad_norm": 1.9921875,
"learning_rate": 1.700090334236676e-05,
"loss": 1.2695,
"step": 290
},
{
"epoch": 0.2437043054427295,
"grad_norm": 1.765625,
"learning_rate": 1.6820234869015358e-05,
"loss": 1.3047,
"step": 300
},
{
"epoch": 0.2518277822908205,
"grad_norm": 1.8828125,
"learning_rate": 1.6639566395663956e-05,
"loss": 1.2727,
"step": 310
},
{
"epoch": 0.25995125913891143,
"grad_norm": 2.140625,
"learning_rate": 1.6458897922312558e-05,
"loss": 1.318,
"step": 320
},
{
"epoch": 0.26807473598700243,
"grad_norm": 1.953125,
"learning_rate": 1.627822944896116e-05,
"loss": 1.3477,
"step": 330
},
{
"epoch": 0.27619821283509344,
"grad_norm": 1.921875,
"learning_rate": 1.6097560975609757e-05,
"loss": 1.3273,
"step": 340
},
{
"epoch": 0.2843216896831844,
"grad_norm": 2.09375,
"learning_rate": 1.5916892502258355e-05,
"loss": 1.3086,
"step": 350
},
{
"epoch": 0.2924451665312754,
"grad_norm": 1.78125,
"learning_rate": 1.5736224028906957e-05,
"loss": 1.2727,
"step": 360
},
{
"epoch": 0.3005686433793664,
"grad_norm": 1.8359375,
"learning_rate": 1.555555555555556e-05,
"loss": 1.3195,
"step": 370
},
{
"epoch": 0.30869212022745735,
"grad_norm": 1.8203125,
"learning_rate": 1.5374887082204156e-05,
"loss": 1.2828,
"step": 380
},
{
"epoch": 0.31681559707554835,
"grad_norm": 1.7265625,
"learning_rate": 1.5194218608852756e-05,
"loss": 1.2719,
"step": 390
},
{
"epoch": 0.3249390739236393,
"grad_norm": 2.03125,
"learning_rate": 1.5013550135501356e-05,
"loss": 1.2914,
"step": 400
},
{
"epoch": 0.3330625507717303,
"grad_norm": 1.9140625,
"learning_rate": 1.4832881662149956e-05,
"loss": 1.25,
"step": 410
},
{
"epoch": 0.3411860276198213,
"grad_norm": 1.6328125,
"learning_rate": 1.4652213188798556e-05,
"loss": 1.282,
"step": 420
},
{
"epoch": 0.34930950446791226,
"grad_norm": 2.015625,
"learning_rate": 1.4471544715447157e-05,
"loss": 1.368,
"step": 430
},
{
"epoch": 0.35743298131600326,
"grad_norm": 1.953125,
"learning_rate": 1.4290876242095755e-05,
"loss": 1.3383,
"step": 440
},
{
"epoch": 0.3655564581640942,
"grad_norm": 1.8203125,
"learning_rate": 1.4110207768744355e-05,
"loss": 1.2992,
"step": 450
},
{
"epoch": 0.3736799350121852,
"grad_norm": 1.9375,
"learning_rate": 1.3929539295392955e-05,
"loss": 1.3234,
"step": 460
},
{
"epoch": 0.3818034118602762,
"grad_norm": 1.875,
"learning_rate": 1.3748870822041556e-05,
"loss": 1.2969,
"step": 470
},
{
"epoch": 0.38992688870836717,
"grad_norm": 1.9140625,
"learning_rate": 1.3568202348690154e-05,
"loss": 1.2648,
"step": 480
},
{
"epoch": 0.3980503655564582,
"grad_norm": 2.046875,
"learning_rate": 1.3387533875338754e-05,
"loss": 1.3109,
"step": 490
},
{
"epoch": 0.4061738424045491,
"grad_norm": 1.953125,
"learning_rate": 1.3206865401987354e-05,
"loss": 1.3086,
"step": 500
},
{
"epoch": 0.41429731925264013,
"grad_norm": 1.8046875,
"learning_rate": 1.3026196928635954e-05,
"loss": 1.2188,
"step": 510
},
{
"epoch": 0.42242079610073113,
"grad_norm": 1.8984375,
"learning_rate": 1.2845528455284555e-05,
"loss": 1.2383,
"step": 520
},
{
"epoch": 0.4305442729488221,
"grad_norm": 1.7421875,
"learning_rate": 1.2664859981933153e-05,
"loss": 1.3164,
"step": 530
},
{
"epoch": 0.4386677497969131,
"grad_norm": 2.0,
"learning_rate": 1.2484191508581753e-05,
"loss": 1.3047,
"step": 540
},
{
"epoch": 0.44679122664500404,
"grad_norm": 1.8359375,
"learning_rate": 1.2303523035230353e-05,
"loss": 1.3023,
"step": 550
},
{
"epoch": 0.45491470349309504,
"grad_norm": 1.734375,
"learning_rate": 1.2122854561878954e-05,
"loss": 1.3242,
"step": 560
},
{
"epoch": 0.46303818034118605,
"grad_norm": 2.171875,
"learning_rate": 1.1942186088527553e-05,
"loss": 1.2961,
"step": 570
},
{
"epoch": 0.471161657189277,
"grad_norm": 2.0625,
"learning_rate": 1.1761517615176152e-05,
"loss": 1.325,
"step": 580
},
{
"epoch": 0.479285134037368,
"grad_norm": 2.0625,
"learning_rate": 1.1580849141824752e-05,
"loss": 1.232,
"step": 590
},
{
"epoch": 0.487408610885459,
"grad_norm": 2.15625,
"learning_rate": 1.1400180668473354e-05,
"loss": 1.2805,
"step": 600
},
{
"epoch": 0.49553208773354995,
"grad_norm": 1.8046875,
"learning_rate": 1.1219512195121953e-05,
"loss": 1.2867,
"step": 610
},
{
"epoch": 0.503655564581641,
"grad_norm": 1.765625,
"learning_rate": 1.1038843721770552e-05,
"loss": 1.3094,
"step": 620
},
{
"epoch": 0.511779041429732,
"grad_norm": 1.8515625,
"learning_rate": 1.0858175248419151e-05,
"loss": 1.3133,
"step": 630
},
{
"epoch": 0.5199025182778229,
"grad_norm": 1.8203125,
"learning_rate": 1.0677506775067751e-05,
"loss": 1.3578,
"step": 640
},
{
"epoch": 0.5280259951259139,
"grad_norm": 1.8828125,
"learning_rate": 1.0496838301716353e-05,
"loss": 1.2969,
"step": 650
},
{
"epoch": 0.5361494719740049,
"grad_norm": 1.8359375,
"learning_rate": 1.031616982836495e-05,
"loss": 1.2867,
"step": 660
},
{
"epoch": 0.5442729488220959,
"grad_norm": 1.8984375,
"learning_rate": 1.013550135501355e-05,
"loss": 1.2977,
"step": 670
},
{
"epoch": 0.5523964256701869,
"grad_norm": 2.0625,
"learning_rate": 9.95483288166215e-06,
"loss": 1.2828,
"step": 680
},
{
"epoch": 0.5605199025182778,
"grad_norm": 1.796875,
"learning_rate": 9.77416440831075e-06,
"loss": 1.2984,
"step": 690
},
{
"epoch": 0.5686433793663688,
"grad_norm": 1.796875,
"learning_rate": 9.59349593495935e-06,
"loss": 1.2656,
"step": 700
},
{
"epoch": 0.5767668562144598,
"grad_norm": 1.703125,
"learning_rate": 9.412827461607951e-06,
"loss": 1.2453,
"step": 710
},
{
"epoch": 0.5848903330625508,
"grad_norm": 1.859375,
"learning_rate": 9.23215898825655e-06,
"loss": 1.3477,
"step": 720
},
{
"epoch": 0.5930138099106418,
"grad_norm": 2.171875,
"learning_rate": 9.051490514905151e-06,
"loss": 1.3141,
"step": 730
},
{
"epoch": 0.6011372867587328,
"grad_norm": 1.84375,
"learning_rate": 8.870822041553749e-06,
"loss": 1.2219,
"step": 740
},
{
"epoch": 0.6092607636068237,
"grad_norm": 1.796875,
"learning_rate": 8.690153568202349e-06,
"loss": 1.2766,
"step": 750
},
{
"epoch": 0.6173842404549147,
"grad_norm": 2.0625,
"learning_rate": 8.509485094850949e-06,
"loss": 1.2789,
"step": 760
},
{
"epoch": 0.6255077173030057,
"grad_norm": 2.109375,
"learning_rate": 8.328816621499549e-06,
"loss": 1.2828,
"step": 770
},
{
"epoch": 0.6336311941510967,
"grad_norm": 1.9140625,
"learning_rate": 8.148148148148148e-06,
"loss": 1.2492,
"step": 780
},
{
"epoch": 0.6417546709991877,
"grad_norm": 1.84375,
"learning_rate": 7.967479674796748e-06,
"loss": 1.2836,
"step": 790
},
{
"epoch": 0.6498781478472786,
"grad_norm": 2.15625,
"learning_rate": 7.78681120144535e-06,
"loss": 1.282,
"step": 800
},
{
"epoch": 0.6580016246953696,
"grad_norm": 1.984375,
"learning_rate": 7.6061427280939486e-06,
"loss": 1.268,
"step": 810
},
{
"epoch": 0.6661251015434606,
"grad_norm": 1.8046875,
"learning_rate": 7.425474254742548e-06,
"loss": 1.3234,
"step": 820
},
{
"epoch": 0.6742485783915516,
"grad_norm": 2.3125,
"learning_rate": 7.244805781391147e-06,
"loss": 1.3195,
"step": 830
},
{
"epoch": 0.6823720552396426,
"grad_norm": 1.9453125,
"learning_rate": 7.064137308039748e-06,
"loss": 1.2617,
"step": 840
},
{
"epoch": 0.6904955320877335,
"grad_norm": 1.84375,
"learning_rate": 6.883468834688347e-06,
"loss": 1.3188,
"step": 850
},
{
"epoch": 0.6986190089358245,
"grad_norm": 1.96875,
"learning_rate": 6.7028003613369475e-06,
"loss": 1.2969,
"step": 860
},
{
"epoch": 0.7067424857839155,
"grad_norm": 2.28125,
"learning_rate": 6.5221318879855465e-06,
"loss": 1.3078,
"step": 870
},
{
"epoch": 0.7148659626320065,
"grad_norm": 2.0,
"learning_rate": 6.341463414634147e-06,
"loss": 1.2789,
"step": 880
},
{
"epoch": 0.7229894394800975,
"grad_norm": 2.046875,
"learning_rate": 6.160794941282746e-06,
"loss": 1.2898,
"step": 890
},
{
"epoch": 0.7311129163281884,
"grad_norm": 1.875,
"learning_rate": 5.980126467931347e-06,
"loss": 1.2625,
"step": 900
},
{
"epoch": 0.7392363931762794,
"grad_norm": 2.0625,
"learning_rate": 5.7994579945799465e-06,
"loss": 1.3141,
"step": 910
},
{
"epoch": 0.7473598700243704,
"grad_norm": 1.640625,
"learning_rate": 5.618789521228546e-06,
"loss": 1.2594,
"step": 920
},
{
"epoch": 0.7554833468724614,
"grad_norm": 1.859375,
"learning_rate": 5.438121047877146e-06,
"loss": 1.3594,
"step": 930
},
{
"epoch": 0.7636068237205524,
"grad_norm": 2.015625,
"learning_rate": 5.257452574525745e-06,
"loss": 1.2742,
"step": 940
},
{
"epoch": 0.7717303005686433,
"grad_norm": 2.125,
"learning_rate": 5.076784101174346e-06,
"loss": 1.2391,
"step": 950
},
{
"epoch": 0.7798537774167343,
"grad_norm": 1.90625,
"learning_rate": 4.8961156278229455e-06,
"loss": 1.3109,
"step": 960
},
{
"epoch": 0.7879772542648253,
"grad_norm": 1.734375,
"learning_rate": 4.715447154471545e-06,
"loss": 1.3289,
"step": 970
},
{
"epoch": 0.7961007311129163,
"grad_norm": 2.125,
"learning_rate": 4.534778681120145e-06,
"loss": 1.2641,
"step": 980
},
{
"epoch": 0.8042242079610074,
"grad_norm": 1.8828125,
"learning_rate": 4.354110207768745e-06,
"loss": 1.2391,
"step": 990
},
{
"epoch": 0.8123476848090982,
"grad_norm": 1.6875,
"learning_rate": 4.173441734417345e-06,
"loss": 1.2977,
"step": 1000
},
{
"epoch": 0.8204711616571893,
"grad_norm": 2.078125,
"learning_rate": 3.9927732610659445e-06,
"loss": 1.2625,
"step": 1010
},
{
"epoch": 0.8285946385052803,
"grad_norm": 1.6484375,
"learning_rate": 3.812104787714544e-06,
"loss": 1.2937,
"step": 1020
},
{
"epoch": 0.8367181153533713,
"grad_norm": 1.96875,
"learning_rate": 3.6314363143631437e-06,
"loss": 1.3109,
"step": 1030
},
{
"epoch": 0.8448415922014623,
"grad_norm": 1.84375,
"learning_rate": 3.450767841011744e-06,
"loss": 1.2125,
"step": 1040
},
{
"epoch": 0.8529650690495532,
"grad_norm": 1.8203125,
"learning_rate": 3.2700993676603437e-06,
"loss": 1.2547,
"step": 1050
},
{
"epoch": 0.8610885458976442,
"grad_norm": 1.8515625,
"learning_rate": 3.0894308943089435e-06,
"loss": 1.2547,
"step": 1060
},
{
"epoch": 0.8692120227457352,
"grad_norm": 1.8984375,
"learning_rate": 2.9087624209575433e-06,
"loss": 1.2523,
"step": 1070
},
{
"epoch": 0.8773354995938262,
"grad_norm": 1.8359375,
"learning_rate": 2.728093947606143e-06,
"loss": 1.3062,
"step": 1080
},
{
"epoch": 0.8854589764419172,
"grad_norm": 1.7890625,
"learning_rate": 2.547425474254743e-06,
"loss": 1.2961,
"step": 1090
},
{
"epoch": 0.8935824532900081,
"grad_norm": 1.7890625,
"learning_rate": 2.3667570009033427e-06,
"loss": 1.2406,
"step": 1100
},
{
"epoch": 0.9017059301380991,
"grad_norm": 1.765625,
"learning_rate": 2.186088527551942e-06,
"loss": 1.3203,
"step": 1110
},
{
"epoch": 0.9098294069861901,
"grad_norm": 2.03125,
"learning_rate": 2.0054200542005423e-06,
"loss": 1.3039,
"step": 1120
},
{
"epoch": 0.9179528838342811,
"grad_norm": 2.09375,
"learning_rate": 1.824751580849142e-06,
"loss": 1.2844,
"step": 1130
},
{
"epoch": 0.9260763606823721,
"grad_norm": 1.8359375,
"learning_rate": 1.6440831074977418e-06,
"loss": 1.3031,
"step": 1140
},
{
"epoch": 0.934199837530463,
"grad_norm": 1.9609375,
"learning_rate": 1.4634146341463414e-06,
"loss": 1.2391,
"step": 1150
},
{
"epoch": 0.942323314378554,
"grad_norm": 1.875,
"learning_rate": 1.2827461607949414e-06,
"loss": 1.2508,
"step": 1160
},
{
"epoch": 0.950446791226645,
"grad_norm": 1.859375,
"learning_rate": 1.102077687443541e-06,
"loss": 1.3148,
"step": 1170
},
{
"epoch": 0.958570268074736,
"grad_norm": 1.875,
"learning_rate": 9.21409214092141e-07,
"loss": 1.2883,
"step": 1180
},
{
"epoch": 0.966693744922827,
"grad_norm": 2.03125,
"learning_rate": 7.407407407407407e-07,
"loss": 1.282,
"step": 1190
},
{
"epoch": 0.974817221770918,
"grad_norm": 1.75,
"learning_rate": 5.600722673893405e-07,
"loss": 1.318,
"step": 1200
},
{
"epoch": 0.9829406986190089,
"grad_norm": 2.0,
"learning_rate": 3.794037940379404e-07,
"loss": 1.2641,
"step": 1210
},
{
"epoch": 0.9910641754670999,
"grad_norm": 2.1875,
"learning_rate": 1.987353206865402e-07,
"loss": 1.3148,
"step": 1220
},
{
"epoch": 0.9991876523151909,
"grad_norm": 1.953125,
"learning_rate": 1.806684733514002e-08,
"loss": 1.2688,
"step": 1230
},
{
"epoch": 1.0,
"step": 1231,
"total_flos": 5.907000409747046e+17,
"train_loss": 1.310145461007311,
"train_runtime": 4855.0115,
"train_samples_per_second": 8.111,
"train_steps_per_second": 0.254
}
],
"logging_steps": 10,
"max_steps": 1231,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 5.907000409747046e+17,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}