dada22231's picture
Training in progress, step 75, checkpoint
28275dd verified
raw
history blame
14.7 kB
{
"best_metric": 3.767355442047119,
"best_model_checkpoint": "miner_id_24/checkpoint-75",
"epoch": 1.9639934533551555,
"eval_steps": 25,
"global_step": 75,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02618657937806874,
"grad_norm": 0.8365817070007324,
"learning_rate": 3.3333333333333335e-05,
"loss": 4.3524,
"step": 1
},
{
"epoch": 0.02618657937806874,
"eval_loss": 4.789937496185303,
"eval_runtime": 0.2765,
"eval_samples_per_second": 180.833,
"eval_steps_per_second": 47.017,
"step": 1
},
{
"epoch": 0.05237315875613748,
"grad_norm": 1.151999831199646,
"learning_rate": 6.666666666666667e-05,
"loss": 4.528,
"step": 2
},
{
"epoch": 0.07855973813420622,
"grad_norm": 1.1144686937332153,
"learning_rate": 0.0001,
"loss": 4.5486,
"step": 3
},
{
"epoch": 0.10474631751227496,
"grad_norm": 0.9449473023414612,
"learning_rate": 9.997376600647783e-05,
"loss": 4.6597,
"step": 4
},
{
"epoch": 0.1309328968903437,
"grad_norm": 0.8100228905677795,
"learning_rate": 9.989509461357426e-05,
"loss": 4.6623,
"step": 5
},
{
"epoch": 0.15711947626841244,
"grad_norm": 0.8471514582633972,
"learning_rate": 9.976407754861426e-05,
"loss": 4.6003,
"step": 6
},
{
"epoch": 0.18330605564648117,
"grad_norm": 0.9861735701560974,
"learning_rate": 9.958086757163489e-05,
"loss": 4.5851,
"step": 7
},
{
"epoch": 0.20949263502454993,
"grad_norm": 1.1125617027282715,
"learning_rate": 9.934567829727386e-05,
"loss": 4.5956,
"step": 8
},
{
"epoch": 0.23567921440261866,
"grad_norm": 1.0278071165084839,
"learning_rate": 9.905878394570453e-05,
"loss": 4.7214,
"step": 9
},
{
"epoch": 0.2618657937806874,
"grad_norm": 2.4233837127685547,
"learning_rate": 9.872051902290737e-05,
"loss": 4.0809,
"step": 10
},
{
"epoch": 0.28805237315875615,
"grad_norm": 0.6828261613845825,
"learning_rate": 9.833127793065098e-05,
"loss": 4.1133,
"step": 11
},
{
"epoch": 0.3142389525368249,
"grad_norm": 1.080686330795288,
"learning_rate": 9.789151450663723e-05,
"loss": 4.1603,
"step": 12
},
{
"epoch": 0.3404255319148936,
"grad_norm": 0.7597512602806091,
"learning_rate": 9.740174149534693e-05,
"loss": 4.2075,
"step": 13
},
{
"epoch": 0.36661211129296234,
"grad_norm": 0.6978819370269775,
"learning_rate": 9.686252995020249e-05,
"loss": 4.155,
"step": 14
},
{
"epoch": 0.39279869067103107,
"grad_norm": 0.5029189586639404,
"learning_rate": 9.627450856774539e-05,
"loss": 4.2456,
"step": 15
},
{
"epoch": 0.41898527004909986,
"grad_norm": 0.49850398302078247,
"learning_rate": 9.563836295460398e-05,
"loss": 4.375,
"step": 16
},
{
"epoch": 0.4451718494271686,
"grad_norm": 0.6559662222862244,
"learning_rate": 9.495483482810688e-05,
"loss": 4.3,
"step": 17
},
{
"epoch": 0.4713584288052373,
"grad_norm": 1.1999343633651733,
"learning_rate": 9.422472115147382e-05,
"loss": 4.4429,
"step": 18
},
{
"epoch": 0.49754500818330605,
"grad_norm": 1.2498600482940674,
"learning_rate": 9.3448873204592e-05,
"loss": 4.5719,
"step": 19
},
{
"epoch": 0.5237315875613748,
"grad_norm": 0.8708962798118591,
"learning_rate": 9.2628195591462e-05,
"loss": 3.7549,
"step": 20
},
{
"epoch": 0.5499181669394435,
"grad_norm": 0.6062394380569458,
"learning_rate": 9.176364518546989e-05,
"loss": 3.9636,
"step": 21
},
{
"epoch": 0.5761047463175123,
"grad_norm": 0.46776407957077026,
"learning_rate": 9.08562300137157e-05,
"loss": 4.0005,
"step": 22
},
{
"epoch": 0.602291325695581,
"grad_norm": 0.41210252046585083,
"learning_rate": 8.990700808169889e-05,
"loss": 3.9414,
"step": 23
},
{
"epoch": 0.6284779050736498,
"grad_norm": 0.4796847701072693,
"learning_rate": 8.891708613973126e-05,
"loss": 4.0629,
"step": 24
},
{
"epoch": 0.6546644844517185,
"grad_norm": 0.42416390776634216,
"learning_rate": 8.788761839251559e-05,
"loss": 4.0834,
"step": 25
},
{
"epoch": 0.6546644844517185,
"eval_loss": 4.142909526824951,
"eval_runtime": 0.2882,
"eval_samples_per_second": 173.488,
"eval_steps_per_second": 45.107,
"step": 25
},
{
"epoch": 0.6808510638297872,
"grad_norm": 0.4640260636806488,
"learning_rate": 8.681980515339464e-05,
"loss": 4.2203,
"step": 26
},
{
"epoch": 0.707037643207856,
"grad_norm": 0.5100923776626587,
"learning_rate": 8.571489144483944e-05,
"loss": 4.181,
"step": 27
},
{
"epoch": 0.7332242225859247,
"grad_norm": 0.6287262439727783,
"learning_rate": 8.457416554680877e-05,
"loss": 4.3732,
"step": 28
},
{
"epoch": 0.7594108019639935,
"grad_norm": 0.5281463265419006,
"learning_rate": 8.339895749467238e-05,
"loss": 3.7263,
"step": 29
},
{
"epoch": 0.7855973813420621,
"grad_norm": 0.41669386625289917,
"learning_rate": 8.219063752844926e-05,
"loss": 3.7891,
"step": 30
},
{
"epoch": 0.8117839607201309,
"grad_norm": 0.39933446049690247,
"learning_rate": 8.095061449516903e-05,
"loss": 3.8458,
"step": 31
},
{
"epoch": 0.8379705400981997,
"grad_norm": 0.42396020889282227,
"learning_rate": 7.968033420621935e-05,
"loss": 3.8991,
"step": 32
},
{
"epoch": 0.8641571194762684,
"grad_norm": 0.45608440041542053,
"learning_rate": 7.838127775159452e-05,
"loss": 3.9113,
"step": 33
},
{
"epoch": 0.8903436988543372,
"grad_norm": 0.5071250200271606,
"learning_rate": 7.705495977301078e-05,
"loss": 3.9696,
"step": 34
},
{
"epoch": 0.9165302782324058,
"grad_norm": 0.408988356590271,
"learning_rate": 7.570292669790186e-05,
"loss": 4.0264,
"step": 35
},
{
"epoch": 0.9427168576104746,
"grad_norm": 0.41267451643943787,
"learning_rate": 7.43267549363537e-05,
"loss": 4.0024,
"step": 36
},
{
"epoch": 0.9689034369885434,
"grad_norm": 0.6320325136184692,
"learning_rate": 7.292804904308087e-05,
"loss": 4.0503,
"step": 37
},
{
"epoch": 0.9950900163666121,
"grad_norm": 0.7328307628631592,
"learning_rate": 7.150843984658754e-05,
"loss": 4.1336,
"step": 38
},
{
"epoch": 1.0212765957446808,
"grad_norm": 1.987341284751892,
"learning_rate": 7.006958254769438e-05,
"loss": 6.7162,
"step": 39
},
{
"epoch": 1.0474631751227497,
"grad_norm": 1.042544960975647,
"learning_rate": 6.861315478964841e-05,
"loss": 3.6702,
"step": 40
},
{
"epoch": 1.0736497545008183,
"grad_norm": 0.6465303301811218,
"learning_rate": 6.714085470206609e-05,
"loss": 3.8179,
"step": 41
},
{
"epoch": 1.099836333878887,
"grad_norm": 0.3951312005519867,
"learning_rate": 6.56543989209901e-05,
"loss": 3.7864,
"step": 42
},
{
"epoch": 1.1260229132569557,
"grad_norm": 0.41512835025787354,
"learning_rate": 6.415552058736854e-05,
"loss": 3.7806,
"step": 43
},
{
"epoch": 1.1522094926350246,
"grad_norm": 0.534665584564209,
"learning_rate": 6.264596732629e-05,
"loss": 3.8374,
"step": 44
},
{
"epoch": 1.1783960720130933,
"grad_norm": 0.5684964060783386,
"learning_rate": 6.112749920933111e-05,
"loss": 3.9551,
"step": 45
},
{
"epoch": 1.204582651391162,
"grad_norm": 0.6400809288024902,
"learning_rate": 5.960188670239154e-05,
"loss": 4.0032,
"step": 46
},
{
"epoch": 1.2307692307692308,
"grad_norm": 0.7185765504837036,
"learning_rate": 5.80709086014102e-05,
"loss": 3.9989,
"step": 47
},
{
"epoch": 1.2569558101472995,
"grad_norm": 0.5369796752929688,
"learning_rate": 5.653634995836856e-05,
"loss": 3.3087,
"step": 48
},
{
"epoch": 1.2831423895253682,
"grad_norm": 0.3849294185638428,
"learning_rate": 5.500000000000001e-05,
"loss": 3.8981,
"step": 49
},
{
"epoch": 1.3093289689034369,
"grad_norm": 0.34771353006362915,
"learning_rate": 5.346365004163145e-05,
"loss": 3.6507,
"step": 50
},
{
"epoch": 1.3093289689034369,
"eval_loss": 3.856473922729492,
"eval_runtime": 0.2803,
"eval_samples_per_second": 178.389,
"eval_steps_per_second": 46.381,
"step": 50
},
{
"epoch": 1.3355155482815058,
"grad_norm": 0.37725579738616943,
"learning_rate": 5.192909139858981e-05,
"loss": 3.7317,
"step": 51
},
{
"epoch": 1.3617021276595744,
"grad_norm": 0.4553833603858948,
"learning_rate": 5.0398113297608465e-05,
"loss": 3.7528,
"step": 52
},
{
"epoch": 1.3878887070376433,
"grad_norm": 0.4247891306877136,
"learning_rate": 4.887250079066892e-05,
"loss": 3.8082,
"step": 53
},
{
"epoch": 1.414075286415712,
"grad_norm": 0.45754310488700867,
"learning_rate": 4.7354032673710005e-05,
"loss": 3.8597,
"step": 54
},
{
"epoch": 1.4402618657937807,
"grad_norm": 0.40189892053604126,
"learning_rate": 4.584447941263149e-05,
"loss": 3.8544,
"step": 55
},
{
"epoch": 1.4664484451718494,
"grad_norm": 0.4472440779209137,
"learning_rate": 4.43456010790099e-05,
"loss": 3.902,
"step": 56
},
{
"epoch": 1.492635024549918,
"grad_norm": 0.7986133694648743,
"learning_rate": 4.285914529793391e-05,
"loss": 4.3171,
"step": 57
},
{
"epoch": 1.518821603927987,
"grad_norm": 0.47378429770469666,
"learning_rate": 4.13868452103516e-05,
"loss": 3.2801,
"step": 58
},
{
"epoch": 1.5450081833060556,
"grad_norm": 0.609893798828125,
"learning_rate": 3.9930417452305626e-05,
"loss": 3.5901,
"step": 59
},
{
"epoch": 1.5711947626841245,
"grad_norm": 0.5858629941940308,
"learning_rate": 3.8491560153412466e-05,
"loss": 3.601,
"step": 60
},
{
"epoch": 1.5973813420621932,
"grad_norm": 0.37681153416633606,
"learning_rate": 3.707195095691913e-05,
"loss": 3.7124,
"step": 61
},
{
"epoch": 1.6235679214402619,
"grad_norm": 0.338227778673172,
"learning_rate": 3.567324506364632e-05,
"loss": 3.7552,
"step": 62
},
{
"epoch": 1.6497545008183305,
"grad_norm": 0.35423874855041504,
"learning_rate": 3.4297073302098156e-05,
"loss": 3.7876,
"step": 63
},
{
"epoch": 1.6759410801963992,
"grad_norm": 0.4390277862548828,
"learning_rate": 3.2945040226989244e-05,
"loss": 3.8012,
"step": 64
},
{
"epoch": 1.702127659574468,
"grad_norm": 0.563541829586029,
"learning_rate": 3.16187222484055e-05,
"loss": 3.8544,
"step": 65
},
{
"epoch": 1.7283142389525368,
"grad_norm": 0.7141799330711365,
"learning_rate": 3.0319665793780648e-05,
"loss": 4.0238,
"step": 66
},
{
"epoch": 1.7545008183306057,
"grad_norm": 0.4013899862766266,
"learning_rate": 2.9049385504830985e-05,
"loss": 3.1346,
"step": 67
},
{
"epoch": 1.7806873977086743,
"grad_norm": 0.3977162539958954,
"learning_rate": 2.7809362471550748e-05,
"loss": 3.8809,
"step": 68
},
{
"epoch": 1.806873977086743,
"grad_norm": 0.35375577211380005,
"learning_rate": 2.660104250532764e-05,
"loss": 3.6103,
"step": 69
},
{
"epoch": 1.8330605564648117,
"grad_norm": 0.3449622094631195,
"learning_rate": 2.5425834453191232e-05,
"loss": 3.6076,
"step": 70
},
{
"epoch": 1.8592471358428804,
"grad_norm": 0.3213866651058197,
"learning_rate": 2.4285108555160577e-05,
"loss": 3.712,
"step": 71
},
{
"epoch": 1.8854337152209493,
"grad_norm": 0.34646686911582947,
"learning_rate": 2.3180194846605367e-05,
"loss": 3.7291,
"step": 72
},
{
"epoch": 1.9116202945990182,
"grad_norm": 0.3628241717815399,
"learning_rate": 2.2112381607484417e-05,
"loss": 3.8293,
"step": 73
},
{
"epoch": 1.9378068739770868,
"grad_norm": 0.37866660952568054,
"learning_rate": 2.1082913860268765e-05,
"loss": 3.8291,
"step": 74
},
{
"epoch": 1.9639934533551555,
"grad_norm": 0.4907776415348053,
"learning_rate": 2.0092991918301108e-05,
"loss": 3.8753,
"step": 75
},
{
"epoch": 1.9639934533551555,
"eval_loss": 3.767355442047119,
"eval_runtime": 0.2893,
"eval_samples_per_second": 172.844,
"eval_steps_per_second": 44.939,
"step": 75
}
],
"logging_steps": 1,
"max_steps": 95,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 25,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 1,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.70605865336832e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}