k1h0's picture
Upload folder using huggingface_hub
2229846 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9963898916967509,
"eval_steps": 500,
"global_step": 69,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01444043321299639,
"grad_norm": 0.9324310421943665,
"learning_rate": 4.9974091841168195e-05,
"loss": 0.8402,
"num_input_tokens_seen": 2097152,
"step": 1
},
{
"epoch": 0.02888086642599278,
"grad_norm": 0.736538290977478,
"learning_rate": 4.9896421063288286e-05,
"loss": 0.7686,
"num_input_tokens_seen": 4194304,
"step": 2
},
{
"epoch": 0.04332129963898917,
"grad_norm": 0.6694996953010559,
"learning_rate": 4.976714865090827e-05,
"loss": 0.7563,
"num_input_tokens_seen": 6291456,
"step": 3
},
{
"epoch": 0.05776173285198556,
"grad_norm": 0.5700488090515137,
"learning_rate": 4.958654254084355e-05,
"loss": 0.7139,
"num_input_tokens_seen": 8388608,
"step": 4
},
{
"epoch": 0.07220216606498195,
"grad_norm": 0.47626349329948425,
"learning_rate": 4.9354977066836986e-05,
"loss": 0.6793,
"num_input_tokens_seen": 10485760,
"step": 5
},
{
"epoch": 0.08664259927797834,
"grad_norm": 0.4310459494590759,
"learning_rate": 4.907293218369499e-05,
"loss": 0.6631,
"num_input_tokens_seen": 12582912,
"step": 6
},
{
"epoch": 0.10108303249097472,
"grad_norm": 0.41724854707717896,
"learning_rate": 4.874099247250798e-05,
"loss": 0.64,
"num_input_tokens_seen": 14680064,
"step": 7
},
{
"epoch": 0.11552346570397112,
"grad_norm": 0.37938904762268066,
"learning_rate": 4.835984592901678e-05,
"loss": 0.6145,
"num_input_tokens_seen": 16777216,
"step": 8
},
{
"epoch": 0.1299638989169675,
"grad_norm": 0.30510303378105164,
"learning_rate": 4.793028253763633e-05,
"loss": 0.5997,
"num_input_tokens_seen": 18874368,
"step": 9
},
{
"epoch": 0.1444043321299639,
"grad_norm": 0.17869406938552856,
"learning_rate": 4.74531926340924e-05,
"loss": 0.5896,
"num_input_tokens_seen": 20971520,
"step": 10
},
{
"epoch": 0.1588447653429603,
"grad_norm": 0.1084759458899498,
"learning_rate": 4.6929565060064864e-05,
"loss": 0.6025,
"num_input_tokens_seen": 23068672,
"step": 11
},
{
"epoch": 0.17328519855595667,
"grad_norm": 0.09031977504491806,
"learning_rate": 4.6360485113662216e-05,
"loss": 0.5644,
"num_input_tokens_seen": 25165824,
"step": 12
},
{
"epoch": 0.18772563176895307,
"grad_norm": 0.08057376742362976,
"learning_rate": 4.574713229997563e-05,
"loss": 0.5558,
"num_input_tokens_seen": 27262976,
"step": 13
},
{
"epoch": 0.20216606498194944,
"grad_norm": 0.0670362040400505,
"learning_rate": 4.509077788637446e-05,
"loss": 0.5705,
"num_input_tokens_seen": 29360128,
"step": 14
},
{
"epoch": 0.21660649819494585,
"grad_norm": 0.06539376825094223,
"learning_rate": 4.43927822676105e-05,
"loss": 0.5694,
"num_input_tokens_seen": 31457280,
"step": 15
},
{
"epoch": 0.23104693140794225,
"grad_norm": 0.05930742993950844,
"learning_rate": 4.365459214619214e-05,
"loss": 0.559,
"num_input_tokens_seen": 33554432,
"step": 16
},
{
"epoch": 0.24548736462093862,
"grad_norm": 0.054464634507894516,
"learning_rate": 4.2877737533872485e-05,
"loss": 0.5628,
"num_input_tokens_seen": 35651584,
"step": 17
},
{
"epoch": 0.259927797833935,
"grad_norm": 0.053172189742326736,
"learning_rate": 4.206382858046636e-05,
"loss": 0.5553,
"num_input_tokens_seen": 37748736,
"step": 18
},
{
"epoch": 0.2743682310469314,
"grad_norm": 0.04865848645567894,
"learning_rate": 4.12145522365689e-05,
"loss": 0.5401,
"num_input_tokens_seen": 39845888,
"step": 19
},
{
"epoch": 0.2888086642599278,
"grad_norm": 0.04852156713604927,
"learning_rate": 4.033166875709291e-05,
"loss": 0.575,
"num_input_tokens_seen": 41943040,
"step": 20
},
{
"epoch": 0.30324909747292417,
"grad_norm": 0.046296387910842896,
"learning_rate": 3.941700805287168e-05,
"loss": 0.5398,
"num_input_tokens_seen": 44040192,
"step": 21
},
{
"epoch": 0.3176895306859206,
"grad_norm": 0.04640813171863556,
"learning_rate": 3.8472465897889394e-05,
"loss": 0.5389,
"num_input_tokens_seen": 46137344,
"step": 22
},
{
"epoch": 0.33212996389891697,
"grad_norm": 0.043006811290979385,
"learning_rate": 3.7500000000000003e-05,
"loss": 0.5451,
"num_input_tokens_seen": 48234496,
"step": 23
},
{
"epoch": 0.34657039711191334,
"grad_norm": 0.04406141862273216,
"learning_rate": 3.6501625943278805e-05,
"loss": 0.5558,
"num_input_tokens_seen": 50331648,
"step": 24
},
{
"epoch": 0.36101083032490977,
"grad_norm": 0.03994145616889,
"learning_rate": 3.547941301041661e-05,
"loss": 0.5403,
"num_input_tokens_seen": 52428800,
"step": 25
},
{
"epoch": 0.37545126353790614,
"grad_norm": 0.039348017424345016,
"learning_rate": 3.443547989381536e-05,
"loss": 0.5405,
"num_input_tokens_seen": 54525952,
"step": 26
},
{
"epoch": 0.3898916967509025,
"grad_norm": 0.03861572593450546,
"learning_rate": 3.3371990304274656e-05,
"loss": 0.5577,
"num_input_tokens_seen": 56623104,
"step": 27
},
{
"epoch": 0.4043321299638989,
"grad_norm": 0.03878667205572128,
"learning_rate": 3.2291148486370626e-05,
"loss": 0.5345,
"num_input_tokens_seen": 58720256,
"step": 28
},
{
"epoch": 0.4187725631768953,
"grad_norm": 0.03664080426096916,
"learning_rate": 3.11951946498225e-05,
"loss": 0.5484,
"num_input_tokens_seen": 60817408,
"step": 29
},
{
"epoch": 0.4332129963898917,
"grad_norm": 0.036824408918619156,
"learning_rate": 3.008640032631585e-05,
"loss": 0.5485,
"num_input_tokens_seen": 62914560,
"step": 30
},
{
"epoch": 0.44765342960288806,
"grad_norm": 0.037150438874959946,
"learning_rate": 2.8967063661406285e-05,
"loss": 0.5299,
"num_input_tokens_seen": 65011712,
"step": 31
},
{
"epoch": 0.4620938628158845,
"grad_norm": 0.03484778478741646,
"learning_rate": 2.7839504651261872e-05,
"loss": 0.539,
"num_input_tokens_seen": 67108864,
"step": 32
},
{
"epoch": 0.47653429602888087,
"grad_norm": 0.035446375608444214,
"learning_rate": 2.6706060334116777e-05,
"loss": 0.5248,
"num_input_tokens_seen": 69206016,
"step": 33
},
{
"epoch": 0.49097472924187724,
"grad_norm": 0.0350475013256073,
"learning_rate": 2.556907994640264e-05,
"loss": 0.5342,
"num_input_tokens_seen": 71303168,
"step": 34
},
{
"epoch": 0.5054151624548736,
"grad_norm": 0.036621786653995514,
"learning_rate": 2.4430920053597356e-05,
"loss": 0.5431,
"num_input_tokens_seen": 73400320,
"step": 35
},
{
"epoch": 0.51985559566787,
"grad_norm": 0.0347721092402935,
"learning_rate": 2.329393966588323e-05,
"loss": 0.5471,
"num_input_tokens_seen": 75497472,
"step": 36
},
{
"epoch": 0.5342960288808665,
"grad_norm": 0.03457929939031601,
"learning_rate": 2.2160495348738123e-05,
"loss": 0.542,
"num_input_tokens_seen": 77594624,
"step": 37
},
{
"epoch": 0.5487364620938628,
"grad_norm": 0.035683248192071915,
"learning_rate": 2.1032936338593718e-05,
"loss": 0.542,
"num_input_tokens_seen": 79691776,
"step": 38
},
{
"epoch": 0.5631768953068592,
"grad_norm": 0.03531257063150406,
"learning_rate": 1.991359967368416e-05,
"loss": 0.542,
"num_input_tokens_seen": 81788928,
"step": 39
},
{
"epoch": 0.5776173285198556,
"grad_norm": 0.034907545894384384,
"learning_rate": 1.8804805350177505e-05,
"loss": 0.5425,
"num_input_tokens_seen": 83886080,
"step": 40
},
{
"epoch": 0.592057761732852,
"grad_norm": 0.03420661389827728,
"learning_rate": 1.7708851513629377e-05,
"loss": 0.5482,
"num_input_tokens_seen": 85983232,
"step": 41
},
{
"epoch": 0.6064981949458483,
"grad_norm": 0.03401198983192444,
"learning_rate": 1.6628009695725346e-05,
"loss": 0.5598,
"num_input_tokens_seen": 88080384,
"step": 42
},
{
"epoch": 0.6209386281588448,
"grad_norm": 0.03434673324227333,
"learning_rate": 1.5564520106184644e-05,
"loss": 0.5402,
"num_input_tokens_seen": 90177536,
"step": 43
},
{
"epoch": 0.6353790613718412,
"grad_norm": 0.033791348338127136,
"learning_rate": 1.4520586989583406e-05,
"loss": 0.5413,
"num_input_tokens_seen": 92274688,
"step": 44
},
{
"epoch": 0.6498194945848376,
"grad_norm": 0.03330031782388687,
"learning_rate": 1.3498374056721197e-05,
"loss": 0.556,
"num_input_tokens_seen": 94371840,
"step": 45
},
{
"epoch": 0.6642599277978339,
"grad_norm": 0.03298752009868622,
"learning_rate": 1.2500000000000006e-05,
"loss": 0.5341,
"num_input_tokens_seen": 96468992,
"step": 46
},
{
"epoch": 0.6787003610108303,
"grad_norm": 0.03363870084285736,
"learning_rate": 1.1527534102110612e-05,
"loss": 0.5436,
"num_input_tokens_seen": 98566144,
"step": 47
},
{
"epoch": 0.6931407942238267,
"grad_norm": 0.032934173941612244,
"learning_rate": 1.0582991947128324e-05,
"loss": 0.5393,
"num_input_tokens_seen": 100663296,
"step": 48
},
{
"epoch": 0.7075812274368231,
"grad_norm": 0.03438662365078926,
"learning_rate": 9.668331242907089e-06,
"loss": 0.5598,
"num_input_tokens_seen": 102760448,
"step": 49
},
{
"epoch": 0.7220216606498195,
"grad_norm": 0.03351249918341637,
"learning_rate": 8.785447763431101e-06,
"loss": 0.5329,
"num_input_tokens_seen": 104857600,
"step": 50
},
{
"epoch": 0.7364620938628159,
"grad_norm": 0.0330129936337471,
"learning_rate": 7.936171419533653e-06,
"loss": 0.5384,
"num_input_tokens_seen": 106954752,
"step": 51
},
{
"epoch": 0.7509025270758123,
"grad_norm": 0.03433903306722641,
"learning_rate": 7.122262466127514e-06,
"loss": 0.5447,
"num_input_tokens_seen": 109051904,
"step": 52
},
{
"epoch": 0.7653429602888087,
"grad_norm": 0.03314002603292465,
"learning_rate": 6.3454078538078635e-06,
"loss": 0.5291,
"num_input_tokens_seen": 111149056,
"step": 53
},
{
"epoch": 0.779783393501805,
"grad_norm": 0.032993488013744354,
"learning_rate": 5.607217732389503e-06,
"loss": 0.5258,
"num_input_tokens_seen": 113246208,
"step": 54
},
{
"epoch": 0.7942238267148014,
"grad_norm": 0.032629404217004776,
"learning_rate": 4.9092221136255444e-06,
"loss": 0.5508,
"num_input_tokens_seen": 115343360,
"step": 55
},
{
"epoch": 0.8086642599277978,
"grad_norm": 0.03205695375800133,
"learning_rate": 4.252867700024374e-06,
"loss": 0.5439,
"num_input_tokens_seen": 117440512,
"step": 56
},
{
"epoch": 0.8231046931407943,
"grad_norm": 0.03285016119480133,
"learning_rate": 3.6395148863377858e-06,
"loss": 0.5261,
"num_input_tokens_seen": 119537664,
"step": 57
},
{
"epoch": 0.8375451263537906,
"grad_norm": 0.035063523799180984,
"learning_rate": 3.0704349399351435e-06,
"loss": 0.5518,
"num_input_tokens_seen": 121634816,
"step": 58
},
{
"epoch": 0.851985559566787,
"grad_norm": 0.03116844967007637,
"learning_rate": 2.5468073659076e-06,
"loss": 0.5444,
"num_input_tokens_seen": 123731968,
"step": 59
},
{
"epoch": 0.8664259927797834,
"grad_norm": 0.03255166485905647,
"learning_rate": 2.0697174623636794e-06,
"loss": 0.5469,
"num_input_tokens_seen": 125829120,
"step": 60
},
{
"epoch": 0.8808664259927798,
"grad_norm": 0.03249025344848633,
"learning_rate": 1.6401540709832242e-06,
"loss": 0.5335,
"num_input_tokens_seen": 127926272,
"step": 61
},
{
"epoch": 0.8953068592057761,
"grad_norm": 0.03506583720445633,
"learning_rate": 1.2590075274920205e-06,
"loss": 0.5367,
"num_input_tokens_seen": 130023424,
"step": 62
},
{
"epoch": 0.9097472924187726,
"grad_norm": 0.03293128311634064,
"learning_rate": 9.270678163050217e-07,
"loss": 0.551,
"num_input_tokens_seen": 132120576,
"step": 63
},
{
"epoch": 0.924187725631769,
"grad_norm": 0.03347219526767731,
"learning_rate": 6.450229331630253e-07,
"loss": 0.5597,
"num_input_tokens_seen": 134217728,
"step": 64
},
{
"epoch": 0.9386281588447654,
"grad_norm": 0.03135489672422409,
"learning_rate": 4.134574591564494e-07,
"loss": 0.5369,
"num_input_tokens_seen": 136314880,
"step": 65
},
{
"epoch": 0.9530685920577617,
"grad_norm": 0.03153960779309273,
"learning_rate": 2.3285134909173112e-07,
"loss": 0.5451,
"num_input_tokens_seen": 138412032,
"step": 66
},
{
"epoch": 0.9675090252707581,
"grad_norm": 0.03191647306084633,
"learning_rate": 1.0357893671171792e-07,
"loss": 0.5237,
"num_input_tokens_seen": 140509184,
"step": 67
},
{
"epoch": 0.9819494584837545,
"grad_norm": 0.03295959159731865,
"learning_rate": 2.590815883181108e-08,
"loss": 0.5604,
"num_input_tokens_seen": 142606336,
"step": 68
},
{
"epoch": 0.9963898916967509,
"grad_norm": 0.03191553056240082,
"learning_rate": 0.0,
"loss": 0.534,
"num_input_tokens_seen": 144703488,
"step": 69
},
{
"epoch": 0.9963898916967509,
"num_input_tokens_seen": 144703488,
"step": 69,
"total_flos": 5.635565866281075e+18,
"train_loss": 0.5665888682655666,
"train_runtime": 10913.0245,
"train_samples_per_second": 3.247,
"train_steps_per_second": 0.006
}
],
"logging_steps": 1,
"max_steps": 69,
"num_input_tokens_seen": 144703488,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.635565866281075e+18,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}