|
{ |
|
"best_metric": 0.3894089162349701, |
|
"best_model_checkpoint": "prostate-mri-T2w-v05/checkpoint-6223", |
|
"epoch": 7.0, |
|
"eval_steps": 500, |
|
"global_step": 6223, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.025871766029246346, |
|
"grad_norm": 62.908626556396484, |
|
"learning_rate": 2.5871766029246344e-07, |
|
"loss": 0.9082, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.05174353205849269, |
|
"grad_norm": 17.856958389282227, |
|
"learning_rate": 5.174353205849269e-07, |
|
"loss": 0.9712, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.07761529808773904, |
|
"grad_norm": 14.738483428955078, |
|
"learning_rate": 7.761529808773904e-07, |
|
"loss": 0.7274, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.10348706411698538, |
|
"grad_norm": 11.495190620422363, |
|
"learning_rate": 1.0348706411698538e-06, |
|
"loss": 0.5668, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.12935883014623173, |
|
"grad_norm": 14.365918159484863, |
|
"learning_rate": 1.2935883014623174e-06, |
|
"loss": 0.5766, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.15523059617547807, |
|
"grad_norm": 4.656925678253174, |
|
"learning_rate": 1.5523059617547808e-06, |
|
"loss": 0.5059, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.18110236220472442, |
|
"grad_norm": 9.39604663848877, |
|
"learning_rate": 1.8110236220472441e-06, |
|
"loss": 0.4405, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.20697412823397077, |
|
"grad_norm": 9.823530197143555, |
|
"learning_rate": 2.0697412823397075e-06, |
|
"loss": 0.3319, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.23284589426321708, |
|
"grad_norm": 10.687150955200195, |
|
"learning_rate": 2.328458942632171e-06, |
|
"loss": 0.4602, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.25871766029246346, |
|
"grad_norm": 1.1277879476547241, |
|
"learning_rate": 2.5871766029246347e-06, |
|
"loss": 0.3568, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.2845894263217098, |
|
"grad_norm": 0.47847771644592285, |
|
"learning_rate": 2.845894263217098e-06, |
|
"loss": 0.6281, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.31046119235095615, |
|
"grad_norm": 0.5296943187713623, |
|
"learning_rate": 3.1046119235095615e-06, |
|
"loss": 0.3346, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.3363329583802025, |
|
"grad_norm": 11.51565170288086, |
|
"learning_rate": 3.3633295838020245e-06, |
|
"loss": 0.5892, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.36220472440944884, |
|
"grad_norm": 18.972782135009766, |
|
"learning_rate": 3.6220472440944883e-06, |
|
"loss": 0.7544, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.3880764904386952, |
|
"grad_norm": 10.762083053588867, |
|
"learning_rate": 3.880764904386952e-06, |
|
"loss": 0.6232, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.41394825646794153, |
|
"grad_norm": 11.391585350036621, |
|
"learning_rate": 4.139482564679415e-06, |
|
"loss": 0.4856, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.4398200224971879, |
|
"grad_norm": 1.076779842376709, |
|
"learning_rate": 4.3982002249718785e-06, |
|
"loss": 0.3987, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.46569178852643417, |
|
"grad_norm": 11.604365348815918, |
|
"learning_rate": 4.656917885264342e-06, |
|
"loss": 0.5967, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.4915635545556805, |
|
"grad_norm": 10.939277648925781, |
|
"learning_rate": 4.915635545556806e-06, |
|
"loss": 0.3712, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.5174353205849269, |
|
"grad_norm": 10.872827529907227, |
|
"learning_rate": 5.1743532058492695e-06, |
|
"loss": 0.4234, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.5433070866141733, |
|
"grad_norm": 13.358455657958984, |
|
"learning_rate": 5.433070866141733e-06, |
|
"loss": 0.6415, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.5691788526434196, |
|
"grad_norm": 0.5069867968559265, |
|
"learning_rate": 5.691788526434196e-06, |
|
"loss": 0.3304, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.595050618672666, |
|
"grad_norm": 11.854174613952637, |
|
"learning_rate": 5.95050618672666e-06, |
|
"loss": 0.2956, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.6209223847019123, |
|
"grad_norm": 0.6237831711769104, |
|
"learning_rate": 6.209223847019123e-06, |
|
"loss": 0.5922, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.6467941507311586, |
|
"grad_norm": 18.399309158325195, |
|
"learning_rate": 6.467941507311586e-06, |
|
"loss": 0.5529, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.672665916760405, |
|
"grad_norm": 1.0723528861999512, |
|
"learning_rate": 6.726659167604049e-06, |
|
"loss": 0.5464, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.6985376827896513, |
|
"grad_norm": 10.857867240905762, |
|
"learning_rate": 6.985376827896513e-06, |
|
"loss": 0.5258, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.7244094488188977, |
|
"grad_norm": 11.417038917541504, |
|
"learning_rate": 7.244094488188977e-06, |
|
"loss": 0.4599, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.750281214848144, |
|
"grad_norm": 12.366191864013672, |
|
"learning_rate": 7.50281214848144e-06, |
|
"loss": 0.5278, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.7761529808773904, |
|
"grad_norm": 59.656211853027344, |
|
"learning_rate": 7.761529808773903e-06, |
|
"loss": 0.4802, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.8020247469066367, |
|
"grad_norm": 1.6525479555130005, |
|
"learning_rate": 8.020247469066367e-06, |
|
"loss": 0.3855, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.8278965129358831, |
|
"grad_norm": 11.62271499633789, |
|
"learning_rate": 8.27896512935883e-06, |
|
"loss": 0.5377, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.8537682789651294, |
|
"grad_norm": 11.242469787597656, |
|
"learning_rate": 8.537682789651294e-06, |
|
"loss": 0.6707, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.8796400449943758, |
|
"grad_norm": 1.673377513885498, |
|
"learning_rate": 8.796400449943757e-06, |
|
"loss": 0.434, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.905511811023622, |
|
"grad_norm": 0.3748597800731659, |
|
"learning_rate": 9.055118110236222e-06, |
|
"loss": 0.4709, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.9313835770528683, |
|
"grad_norm": 10.111214637756348, |
|
"learning_rate": 9.313835770528684e-06, |
|
"loss": 0.5416, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.9572553430821147, |
|
"grad_norm": 11.05431842803955, |
|
"learning_rate": 9.572553430821147e-06, |
|
"loss": 0.4895, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.983127109111361, |
|
"grad_norm": 0.5102259516716003, |
|
"learning_rate": 9.831271091113612e-06, |
|
"loss": 0.6304, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.7857142857142857, |
|
"eval_auc": 0.8011167033658384, |
|
"eval_f1": 0.0, |
|
"eval_loss": 0.7771126627922058, |
|
"eval_precision": 0.0, |
|
"eval_recall": 0.0, |
|
"eval_runtime": 2.7441, |
|
"eval_samples_per_second": 173.466, |
|
"eval_steps_per_second": 21.865, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 1.0089988751406074, |
|
"grad_norm": 13.26599407196045, |
|
"learning_rate": 1.0089988751406076e-05, |
|
"loss": 0.3247, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 1.0348706411698538, |
|
"grad_norm": 11.029207229614258, |
|
"learning_rate": 1.0348706411698539e-05, |
|
"loss": 0.6095, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.0607424071991, |
|
"grad_norm": 0.07530930638313293, |
|
"learning_rate": 1.0607424071991e-05, |
|
"loss": 0.1472, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 1.0866141732283465, |
|
"grad_norm": 0.6263224482536316, |
|
"learning_rate": 1.0866141732283466e-05, |
|
"loss": 0.6615, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 1.1124859392575928, |
|
"grad_norm": 9.975605964660645, |
|
"learning_rate": 1.1124859392575929e-05, |
|
"loss": 0.4082, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 1.1383577052868392, |
|
"grad_norm": 10.850635528564453, |
|
"learning_rate": 1.1383577052868393e-05, |
|
"loss": 0.4151, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 1.1642294713160855, |
|
"grad_norm": 10.857563972473145, |
|
"learning_rate": 1.1642294713160854e-05, |
|
"loss": 0.4441, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 1.190101237345332, |
|
"grad_norm": 4.742630958557129, |
|
"learning_rate": 1.190101237345332e-05, |
|
"loss": 0.7853, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 1.2159730033745781, |
|
"grad_norm": 0.09981119632720947, |
|
"learning_rate": 1.2159730033745783e-05, |
|
"loss": 0.2316, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 1.2418447694038246, |
|
"grad_norm": 0.37830597162246704, |
|
"learning_rate": 1.2418447694038246e-05, |
|
"loss": 0.5479, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 1.2677165354330708, |
|
"grad_norm": 1.0691417455673218, |
|
"learning_rate": 1.2677165354330708e-05, |
|
"loss": 0.342, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 1.2935883014623173, |
|
"grad_norm": 10.283402442932129, |
|
"learning_rate": 1.2935883014623173e-05, |
|
"loss": 0.591, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.3194600674915635, |
|
"grad_norm": 2.4355926513671875, |
|
"learning_rate": 1.3194600674915636e-05, |
|
"loss": 0.486, |
|
"step": 1173 |
|
}, |
|
{ |
|
"epoch": 1.34533183352081, |
|
"grad_norm": 7.244462966918945, |
|
"learning_rate": 1.3453318335208098e-05, |
|
"loss": 0.5206, |
|
"step": 1196 |
|
}, |
|
{ |
|
"epoch": 1.3712035995500562, |
|
"grad_norm": 9.888749122619629, |
|
"learning_rate": 1.3712035995500563e-05, |
|
"loss": 0.4504, |
|
"step": 1219 |
|
}, |
|
{ |
|
"epoch": 1.3970753655793025, |
|
"grad_norm": 3.3459055423736572, |
|
"learning_rate": 1.3970753655793026e-05, |
|
"loss": 0.4415, |
|
"step": 1242 |
|
}, |
|
{ |
|
"epoch": 1.422947131608549, |
|
"grad_norm": 0.175747811794281, |
|
"learning_rate": 1.4229471316085491e-05, |
|
"loss": 0.2943, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 1.4488188976377954, |
|
"grad_norm": 4.791985988616943, |
|
"learning_rate": 1.4488188976377953e-05, |
|
"loss": 0.566, |
|
"step": 1288 |
|
}, |
|
{ |
|
"epoch": 1.4746906636670416, |
|
"grad_norm": 12.70669174194336, |
|
"learning_rate": 1.4746906636670415e-05, |
|
"loss": 0.4687, |
|
"step": 1311 |
|
}, |
|
{ |
|
"epoch": 1.5005624296962878, |
|
"grad_norm": 0.04389326646924019, |
|
"learning_rate": 1.500562429696288e-05, |
|
"loss": 0.2629, |
|
"step": 1334 |
|
}, |
|
{ |
|
"epoch": 1.5264341957255343, |
|
"grad_norm": 4.369255542755127, |
|
"learning_rate": 1.5264341957255343e-05, |
|
"loss": 0.472, |
|
"step": 1357 |
|
}, |
|
{ |
|
"epoch": 1.5523059617547807, |
|
"grad_norm": 4.454276084899902, |
|
"learning_rate": 1.5523059617547807e-05, |
|
"loss": 0.4307, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.578177727784027, |
|
"grad_norm": 1.1173179149627686, |
|
"learning_rate": 1.578177727784027e-05, |
|
"loss": 0.3431, |
|
"step": 1403 |
|
}, |
|
{ |
|
"epoch": 1.6040494938132732, |
|
"grad_norm": 3.0332882404327393, |
|
"learning_rate": 1.6040494938132734e-05, |
|
"loss": 0.2545, |
|
"step": 1426 |
|
}, |
|
{ |
|
"epoch": 1.6299212598425197, |
|
"grad_norm": 10.521413803100586, |
|
"learning_rate": 1.6299212598425197e-05, |
|
"loss": 0.6555, |
|
"step": 1449 |
|
}, |
|
{ |
|
"epoch": 1.6557930258717661, |
|
"grad_norm": 3.8012301921844482, |
|
"learning_rate": 1.655793025871766e-05, |
|
"loss": 0.37, |
|
"step": 1472 |
|
}, |
|
{ |
|
"epoch": 1.6816647919010124, |
|
"grad_norm": 0.05941835045814514, |
|
"learning_rate": 1.6816647919010127e-05, |
|
"loss": 0.3091, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 1.7075365579302586, |
|
"grad_norm": 0.29451972246170044, |
|
"learning_rate": 1.7075365579302587e-05, |
|
"loss": 0.3511, |
|
"step": 1518 |
|
}, |
|
{ |
|
"epoch": 1.733408323959505, |
|
"grad_norm": 3.2587544918060303, |
|
"learning_rate": 1.733408323959505e-05, |
|
"loss": 0.6164, |
|
"step": 1541 |
|
}, |
|
{ |
|
"epoch": 1.7592800899887515, |
|
"grad_norm": 3.4113192558288574, |
|
"learning_rate": 1.7592800899887514e-05, |
|
"loss": 0.3817, |
|
"step": 1564 |
|
}, |
|
{ |
|
"epoch": 1.7851518560179978, |
|
"grad_norm": 11.975400924682617, |
|
"learning_rate": 1.7851518560179977e-05, |
|
"loss": 0.3421, |
|
"step": 1587 |
|
}, |
|
{ |
|
"epoch": 1.811023622047244, |
|
"grad_norm": 1.5508025884628296, |
|
"learning_rate": 1.8110236220472444e-05, |
|
"loss": 0.6197, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.8368953880764904, |
|
"grad_norm": 0.24332457780838013, |
|
"learning_rate": 1.8368953880764904e-05, |
|
"loss": 0.4227, |
|
"step": 1633 |
|
}, |
|
{ |
|
"epoch": 1.862767154105737, |
|
"grad_norm": 0.10154075175523758, |
|
"learning_rate": 1.8627671541057367e-05, |
|
"loss": 0.4287, |
|
"step": 1656 |
|
}, |
|
{ |
|
"epoch": 1.8886389201349831, |
|
"grad_norm": 10.398853302001953, |
|
"learning_rate": 1.8886389201349834e-05, |
|
"loss": 0.3014, |
|
"step": 1679 |
|
}, |
|
{ |
|
"epoch": 1.9145106861642294, |
|
"grad_norm": 0.8489835858345032, |
|
"learning_rate": 1.9145106861642294e-05, |
|
"loss": 0.3982, |
|
"step": 1702 |
|
}, |
|
{ |
|
"epoch": 1.9403824521934758, |
|
"grad_norm": 0.8622331619262695, |
|
"learning_rate": 1.940382452193476e-05, |
|
"loss": 0.5771, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.9662542182227223, |
|
"grad_norm": 2.847560405731201, |
|
"learning_rate": 1.9662542182227224e-05, |
|
"loss": 0.4758, |
|
"step": 1748 |
|
}, |
|
{ |
|
"epoch": 1.9921259842519685, |
|
"grad_norm": 0.5839571952819824, |
|
"learning_rate": 1.9921259842519684e-05, |
|
"loss": 0.2732, |
|
"step": 1771 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.7857142857142857, |
|
"eval_auc": 0.7965817342979973, |
|
"eval_f1": 0.0, |
|
"eval_loss": 0.9842509627342224, |
|
"eval_precision": 0.0, |
|
"eval_recall": 0.0, |
|
"eval_runtime": 2.74, |
|
"eval_samples_per_second": 173.72, |
|
"eval_steps_per_second": 21.897, |
|
"step": 1778 |
|
}, |
|
{ |
|
"epoch": 2.0179977502812148, |
|
"grad_norm": 0.23481443524360657, |
|
"learning_rate": 2.017997750281215e-05, |
|
"loss": 0.2917, |
|
"step": 1794 |
|
}, |
|
{ |
|
"epoch": 2.043869516310461, |
|
"grad_norm": 0.1679057627916336, |
|
"learning_rate": 2.043869516310461e-05, |
|
"loss": 0.5162, |
|
"step": 1817 |
|
}, |
|
{ |
|
"epoch": 2.0697412823397077, |
|
"grad_norm": 5.455684185028076, |
|
"learning_rate": 2.0697412823397078e-05, |
|
"loss": 0.5789, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 2.095613048368954, |
|
"grad_norm": 7.542656421661377, |
|
"learning_rate": 2.095613048368954e-05, |
|
"loss": 0.5314, |
|
"step": 1863 |
|
}, |
|
{ |
|
"epoch": 2.1214848143982, |
|
"grad_norm": 2.186872720718384, |
|
"learning_rate": 2.1214848143982e-05, |
|
"loss": 0.3816, |
|
"step": 1886 |
|
}, |
|
{ |
|
"epoch": 2.1473565804274464, |
|
"grad_norm": 0.19364489614963531, |
|
"learning_rate": 2.1473565804274468e-05, |
|
"loss": 0.336, |
|
"step": 1909 |
|
}, |
|
{ |
|
"epoch": 2.173228346456693, |
|
"grad_norm": 0.20303942263126373, |
|
"learning_rate": 2.173228346456693e-05, |
|
"loss": 0.421, |
|
"step": 1932 |
|
}, |
|
{ |
|
"epoch": 2.1991001124859393, |
|
"grad_norm": 0.5988159775733948, |
|
"learning_rate": 2.1991001124859395e-05, |
|
"loss": 0.4587, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 2.2249718785151855, |
|
"grad_norm": 7.407308101654053, |
|
"learning_rate": 2.2249718785151858e-05, |
|
"loss": 0.6277, |
|
"step": 1978 |
|
}, |
|
{ |
|
"epoch": 2.2508436445444318, |
|
"grad_norm": 2.7972147464752197, |
|
"learning_rate": 2.2508436445444318e-05, |
|
"loss": 0.4025, |
|
"step": 2001 |
|
}, |
|
{ |
|
"epoch": 2.2767154105736784, |
|
"grad_norm": 14.71440601348877, |
|
"learning_rate": 2.2767154105736785e-05, |
|
"loss": 0.4362, |
|
"step": 2024 |
|
}, |
|
{ |
|
"epoch": 2.3025871766029247, |
|
"grad_norm": 0.6409655213356018, |
|
"learning_rate": 2.302587176602925e-05, |
|
"loss": 0.4014, |
|
"step": 2047 |
|
}, |
|
{ |
|
"epoch": 2.328458942632171, |
|
"grad_norm": 0.0657491534948349, |
|
"learning_rate": 2.328458942632171e-05, |
|
"loss": 0.3942, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 2.354330708661417, |
|
"grad_norm": 0.9892528653144836, |
|
"learning_rate": 2.3543307086614175e-05, |
|
"loss": 0.5464, |
|
"step": 2093 |
|
}, |
|
{ |
|
"epoch": 2.380202474690664, |
|
"grad_norm": 0.6635844111442566, |
|
"learning_rate": 2.380202474690664e-05, |
|
"loss": 0.4767, |
|
"step": 2116 |
|
}, |
|
{ |
|
"epoch": 2.40607424071991, |
|
"grad_norm": 7.151478290557861, |
|
"learning_rate": 2.4060742407199102e-05, |
|
"loss": 0.4637, |
|
"step": 2139 |
|
}, |
|
{ |
|
"epoch": 2.4319460067491563, |
|
"grad_norm": 0.7382416129112244, |
|
"learning_rate": 2.4319460067491565e-05, |
|
"loss": 0.3082, |
|
"step": 2162 |
|
}, |
|
{ |
|
"epoch": 2.4578177727784025, |
|
"grad_norm": 9.286602973937988, |
|
"learning_rate": 2.4578177727784025e-05, |
|
"loss": 0.5282, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 2.483689538807649, |
|
"grad_norm": 4.055200099945068, |
|
"learning_rate": 2.4836895388076492e-05, |
|
"loss": 0.457, |
|
"step": 2208 |
|
}, |
|
{ |
|
"epoch": 2.5095613048368954, |
|
"grad_norm": 0.12764699757099152, |
|
"learning_rate": 2.5095613048368956e-05, |
|
"loss": 0.3318, |
|
"step": 2231 |
|
}, |
|
{ |
|
"epoch": 2.5354330708661417, |
|
"grad_norm": 2.9867072105407715, |
|
"learning_rate": 2.5354330708661416e-05, |
|
"loss": 0.3542, |
|
"step": 2254 |
|
}, |
|
{ |
|
"epoch": 2.5613048368953883, |
|
"grad_norm": 4.703469753265381, |
|
"learning_rate": 2.5613048368953886e-05, |
|
"loss": 0.4748, |
|
"step": 2277 |
|
}, |
|
{ |
|
"epoch": 2.5871766029246346, |
|
"grad_norm": 9.484485626220703, |
|
"learning_rate": 2.5871766029246346e-05, |
|
"loss": 0.4135, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.613048368953881, |
|
"grad_norm": 18.093868255615234, |
|
"learning_rate": 2.613048368953881e-05, |
|
"loss": 0.4656, |
|
"step": 2323 |
|
}, |
|
{ |
|
"epoch": 2.638920134983127, |
|
"grad_norm": 0.21779067814350128, |
|
"learning_rate": 2.6389201349831272e-05, |
|
"loss": 0.3057, |
|
"step": 2346 |
|
}, |
|
{ |
|
"epoch": 2.6647919010123733, |
|
"grad_norm": 0.7615957260131836, |
|
"learning_rate": 2.6647919010123736e-05, |
|
"loss": 0.4652, |
|
"step": 2369 |
|
}, |
|
{ |
|
"epoch": 2.69066366704162, |
|
"grad_norm": 0.35096365213394165, |
|
"learning_rate": 2.6906636670416196e-05, |
|
"loss": 0.3751, |
|
"step": 2392 |
|
}, |
|
{ |
|
"epoch": 2.716535433070866, |
|
"grad_norm": 0.5412973761558533, |
|
"learning_rate": 2.7165354330708666e-05, |
|
"loss": 0.1823, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 2.7424071991001124, |
|
"grad_norm": 1.9220446348190308, |
|
"learning_rate": 2.7424071991001126e-05, |
|
"loss": 0.6485, |
|
"step": 2438 |
|
}, |
|
{ |
|
"epoch": 2.7682789651293587, |
|
"grad_norm": 11.08309268951416, |
|
"learning_rate": 2.768278965129359e-05, |
|
"loss": 0.4777, |
|
"step": 2461 |
|
}, |
|
{ |
|
"epoch": 2.794150731158605, |
|
"grad_norm": 0.0344441793859005, |
|
"learning_rate": 2.7941507311586053e-05, |
|
"loss": 0.2201, |
|
"step": 2484 |
|
}, |
|
{ |
|
"epoch": 2.8200224971878516, |
|
"grad_norm": 0.9013801217079163, |
|
"learning_rate": 2.8200224971878513e-05, |
|
"loss": 0.471, |
|
"step": 2507 |
|
}, |
|
{ |
|
"epoch": 2.845894263217098, |
|
"grad_norm": 1.6073826551437378, |
|
"learning_rate": 2.8458942632170983e-05, |
|
"loss": 0.5535, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 2.871766029246344, |
|
"grad_norm": 3.7488861083984375, |
|
"learning_rate": 2.8717660292463443e-05, |
|
"loss": 0.525, |
|
"step": 2553 |
|
}, |
|
{ |
|
"epoch": 2.8976377952755907, |
|
"grad_norm": 0.2803422808647156, |
|
"learning_rate": 2.8976377952755906e-05, |
|
"loss": 0.4241, |
|
"step": 2576 |
|
}, |
|
{ |
|
"epoch": 2.923509561304837, |
|
"grad_norm": 2.339632749557495, |
|
"learning_rate": 2.923509561304837e-05, |
|
"loss": 0.3422, |
|
"step": 2599 |
|
}, |
|
{ |
|
"epoch": 2.949381327334083, |
|
"grad_norm": 0.10544496774673462, |
|
"learning_rate": 2.949381327334083e-05, |
|
"loss": 0.2698, |
|
"step": 2622 |
|
}, |
|
{ |
|
"epoch": 2.9752530933633294, |
|
"grad_norm": 5.151175498962402, |
|
"learning_rate": 2.97525309336333e-05, |
|
"loss": 0.5485, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.7815126050420168, |
|
"eval_auc": 0.7921778337003251, |
|
"eval_f1": 0.0, |
|
"eval_loss": 0.7073759436607361, |
|
"eval_precision": 0.0, |
|
"eval_recall": 0.0, |
|
"eval_runtime": 2.7321, |
|
"eval_samples_per_second": 174.227, |
|
"eval_steps_per_second": 21.961, |
|
"step": 2667 |
|
}, |
|
{ |
|
"epoch": 3.001124859392576, |
|
"grad_norm": 7.475840091705322, |
|
"learning_rate": 3.001124859392576e-05, |
|
"loss": 0.2998, |
|
"step": 2668 |
|
}, |
|
{ |
|
"epoch": 3.0269966254218224, |
|
"grad_norm": 0.36270514130592346, |
|
"learning_rate": 3.0269966254218223e-05, |
|
"loss": 0.3207, |
|
"step": 2691 |
|
}, |
|
{ |
|
"epoch": 3.0528683914510686, |
|
"grad_norm": 3.655841112136841, |
|
"learning_rate": 3.052868391451069e-05, |
|
"loss": 0.376, |
|
"step": 2714 |
|
}, |
|
{ |
|
"epoch": 3.078740157480315, |
|
"grad_norm": 0.28142863512039185, |
|
"learning_rate": 3.078740157480315e-05, |
|
"loss": 0.3376, |
|
"step": 2737 |
|
}, |
|
{ |
|
"epoch": 3.1046119235095615, |
|
"grad_norm": 0.1056133359670639, |
|
"learning_rate": 3.1046119235095613e-05, |
|
"loss": 0.43, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 3.1304836895388077, |
|
"grad_norm": 10.104964256286621, |
|
"learning_rate": 3.130483689538808e-05, |
|
"loss": 0.4603, |
|
"step": 2783 |
|
}, |
|
{ |
|
"epoch": 3.156355455568054, |
|
"grad_norm": 8.072303771972656, |
|
"learning_rate": 3.156355455568054e-05, |
|
"loss": 0.3976, |
|
"step": 2806 |
|
}, |
|
{ |
|
"epoch": 3.1822272215973, |
|
"grad_norm": 10.715200424194336, |
|
"learning_rate": 3.1822272215973004e-05, |
|
"loss": 0.4361, |
|
"step": 2829 |
|
}, |
|
{ |
|
"epoch": 3.208098987626547, |
|
"grad_norm": 1.3668488264083862, |
|
"learning_rate": 3.208098987626547e-05, |
|
"loss": 0.2668, |
|
"step": 2852 |
|
}, |
|
{ |
|
"epoch": 3.233970753655793, |
|
"grad_norm": 1.127758502960205, |
|
"learning_rate": 3.233970753655794e-05, |
|
"loss": 0.3974, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 3.2598425196850394, |
|
"grad_norm": 0.872011125087738, |
|
"learning_rate": 3.2598425196850394e-05, |
|
"loss": 0.3557, |
|
"step": 2898 |
|
}, |
|
{ |
|
"epoch": 3.2857142857142856, |
|
"grad_norm": 15.174445152282715, |
|
"learning_rate": 3.285714285714286e-05, |
|
"loss": 0.4542, |
|
"step": 2921 |
|
}, |
|
{ |
|
"epoch": 3.3115860517435323, |
|
"grad_norm": 3.535675048828125, |
|
"learning_rate": 3.311586051743532e-05, |
|
"loss": 0.3665, |
|
"step": 2944 |
|
}, |
|
{ |
|
"epoch": 3.3374578177727785, |
|
"grad_norm": 0.1963585466146469, |
|
"learning_rate": 3.3374578177727784e-05, |
|
"loss": 0.4275, |
|
"step": 2967 |
|
}, |
|
{ |
|
"epoch": 3.3633295838020247, |
|
"grad_norm": 1.3127784729003906, |
|
"learning_rate": 3.3633295838020254e-05, |
|
"loss": 0.4884, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 3.389201349831271, |
|
"grad_norm": 1.3405988216400146, |
|
"learning_rate": 3.389201349831271e-05, |
|
"loss": 0.2996, |
|
"step": 3013 |
|
}, |
|
{ |
|
"epoch": 3.415073115860517, |
|
"grad_norm": 9.742510795593262, |
|
"learning_rate": 3.4150731158605174e-05, |
|
"loss": 0.4061, |
|
"step": 3036 |
|
}, |
|
{ |
|
"epoch": 3.440944881889764, |
|
"grad_norm": 0.6711751222610474, |
|
"learning_rate": 3.440944881889764e-05, |
|
"loss": 0.3684, |
|
"step": 3059 |
|
}, |
|
{ |
|
"epoch": 3.46681664791901, |
|
"grad_norm": 9.26884937286377, |
|
"learning_rate": 3.46681664791901e-05, |
|
"loss": 0.4557, |
|
"step": 3082 |
|
}, |
|
{ |
|
"epoch": 3.4926884139482564, |
|
"grad_norm": 11.18705940246582, |
|
"learning_rate": 3.492688413948257e-05, |
|
"loss": 0.3843, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 3.518560179977503, |
|
"grad_norm": 0.6143022775650024, |
|
"learning_rate": 3.518560179977503e-05, |
|
"loss": 0.2909, |
|
"step": 3128 |
|
}, |
|
{ |
|
"epoch": 3.5444319460067493, |
|
"grad_norm": 0.13090083003044128, |
|
"learning_rate": 3.544431946006749e-05, |
|
"loss": 0.3688, |
|
"step": 3151 |
|
}, |
|
{ |
|
"epoch": 3.5703037120359955, |
|
"grad_norm": 8.838688850402832, |
|
"learning_rate": 3.5703037120359954e-05, |
|
"loss": 0.4144, |
|
"step": 3174 |
|
}, |
|
{ |
|
"epoch": 3.5961754780652417, |
|
"grad_norm": 8.904288291931152, |
|
"learning_rate": 3.596175478065242e-05, |
|
"loss": 0.3711, |
|
"step": 3197 |
|
}, |
|
{ |
|
"epoch": 3.622047244094488, |
|
"grad_norm": 0.6677899360656738, |
|
"learning_rate": 3.622047244094489e-05, |
|
"loss": 0.5376, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 3.6479190101237347, |
|
"grad_norm": 2.987168073654175, |
|
"learning_rate": 3.647919010123735e-05, |
|
"loss": 0.3642, |
|
"step": 3243 |
|
}, |
|
{ |
|
"epoch": 3.673790776152981, |
|
"grad_norm": 9.179577827453613, |
|
"learning_rate": 3.673790776152981e-05, |
|
"loss": 0.6663, |
|
"step": 3266 |
|
}, |
|
{ |
|
"epoch": 3.699662542182227, |
|
"grad_norm": 1.0456788539886475, |
|
"learning_rate": 3.699662542182227e-05, |
|
"loss": 0.478, |
|
"step": 3289 |
|
}, |
|
{ |
|
"epoch": 3.725534308211474, |
|
"grad_norm": 8.08246898651123, |
|
"learning_rate": 3.7255343082114735e-05, |
|
"loss": 0.4601, |
|
"step": 3312 |
|
}, |
|
{ |
|
"epoch": 3.75140607424072, |
|
"grad_norm": 8.33056926727295, |
|
"learning_rate": 3.7514060742407205e-05, |
|
"loss": 0.4343, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 3.7772778402699663, |
|
"grad_norm": 2.994161367416382, |
|
"learning_rate": 3.777277840269967e-05, |
|
"loss": 0.3037, |
|
"step": 3358 |
|
}, |
|
{ |
|
"epoch": 3.8031496062992125, |
|
"grad_norm": 0.5124155282974243, |
|
"learning_rate": 3.8031496062992125e-05, |
|
"loss": 0.4441, |
|
"step": 3381 |
|
}, |
|
{ |
|
"epoch": 3.8290213723284587, |
|
"grad_norm": 9.554545402526855, |
|
"learning_rate": 3.829021372328459e-05, |
|
"loss": 0.4561, |
|
"step": 3404 |
|
}, |
|
{ |
|
"epoch": 3.8548931383577054, |
|
"grad_norm": 2.4750254154205322, |
|
"learning_rate": 3.854893138357705e-05, |
|
"loss": 0.5598, |
|
"step": 3427 |
|
}, |
|
{ |
|
"epoch": 3.8807649043869517, |
|
"grad_norm": 0.18071837723255157, |
|
"learning_rate": 3.880764904386952e-05, |
|
"loss": 0.392, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 3.906636670416198, |
|
"grad_norm": 0.9068869948387146, |
|
"learning_rate": 3.9066366704161985e-05, |
|
"loss": 0.4358, |
|
"step": 3473 |
|
}, |
|
{ |
|
"epoch": 3.9325084364454446, |
|
"grad_norm": 2.832122325897217, |
|
"learning_rate": 3.932508436445445e-05, |
|
"loss": 0.3503, |
|
"step": 3496 |
|
}, |
|
{ |
|
"epoch": 3.958380202474691, |
|
"grad_norm": 8.372138023376465, |
|
"learning_rate": 3.9583802024746905e-05, |
|
"loss": 0.3665, |
|
"step": 3519 |
|
}, |
|
{ |
|
"epoch": 3.984251968503937, |
|
"grad_norm": 0.4025736451148987, |
|
"learning_rate": 3.984251968503937e-05, |
|
"loss": 0.3274, |
|
"step": 3542 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.7815126050420168, |
|
"eval_auc": 0.8276973891160743, |
|
"eval_f1": 0.08771929824561403, |
|
"eval_loss": 0.6576350927352905, |
|
"eval_precision": 0.4166666666666667, |
|
"eval_recall": 0.049019607843137254, |
|
"eval_runtime": 2.7248, |
|
"eval_samples_per_second": 174.693, |
|
"eval_steps_per_second": 22.02, |
|
"step": 3556 |
|
}, |
|
{ |
|
"epoch": 4.010123734533184, |
|
"grad_norm": 11.520211219787598, |
|
"learning_rate": 4.010123734533184e-05, |
|
"loss": 0.3699, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 4.0359955005624295, |
|
"grad_norm": 3.6669209003448486, |
|
"learning_rate": 4.03599550056243e-05, |
|
"loss": 0.4811, |
|
"step": 3588 |
|
}, |
|
{ |
|
"epoch": 4.061867266591676, |
|
"grad_norm": 6.646744251251221, |
|
"learning_rate": 4.0618672665916766e-05, |
|
"loss": 0.3609, |
|
"step": 3611 |
|
}, |
|
{ |
|
"epoch": 4.087739032620922, |
|
"grad_norm": 0.03924195468425751, |
|
"learning_rate": 4.087739032620922e-05, |
|
"loss": 0.3344, |
|
"step": 3634 |
|
}, |
|
{ |
|
"epoch": 4.113610798650169, |
|
"grad_norm": 3.2102909088134766, |
|
"learning_rate": 4.1136107986501686e-05, |
|
"loss": 0.6133, |
|
"step": 3657 |
|
}, |
|
{ |
|
"epoch": 4.139482564679415, |
|
"grad_norm": 8.028974533081055, |
|
"learning_rate": 4.1394825646794156e-05, |
|
"loss": 0.322, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 4.165354330708661, |
|
"grad_norm": 0.7968009114265442, |
|
"learning_rate": 4.165354330708662e-05, |
|
"loss": 0.337, |
|
"step": 3703 |
|
}, |
|
{ |
|
"epoch": 4.191226096737908, |
|
"grad_norm": 8.355441093444824, |
|
"learning_rate": 4.191226096737908e-05, |
|
"loss": 0.3879, |
|
"step": 3726 |
|
}, |
|
{ |
|
"epoch": 4.2170978627671545, |
|
"grad_norm": 0.24170738458633423, |
|
"learning_rate": 4.217097862767154e-05, |
|
"loss": 0.3252, |
|
"step": 3749 |
|
}, |
|
{ |
|
"epoch": 4.2429696287964, |
|
"grad_norm": 5.390313625335693, |
|
"learning_rate": 4.2429696287964e-05, |
|
"loss": 0.507, |
|
"step": 3772 |
|
}, |
|
{ |
|
"epoch": 4.268841394825647, |
|
"grad_norm": 5.854102611541748, |
|
"learning_rate": 4.268841394825647e-05, |
|
"loss": 0.3849, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 4.294713160854893, |
|
"grad_norm": 8.469975471496582, |
|
"learning_rate": 4.2947131608548936e-05, |
|
"loss": 0.315, |
|
"step": 3818 |
|
}, |
|
{ |
|
"epoch": 4.320584926884139, |
|
"grad_norm": 3.3237178325653076, |
|
"learning_rate": 4.32058492688414e-05, |
|
"loss": 0.7087, |
|
"step": 3841 |
|
}, |
|
{ |
|
"epoch": 4.346456692913386, |
|
"grad_norm": 1.0640898942947388, |
|
"learning_rate": 4.346456692913386e-05, |
|
"loss": 0.4298, |
|
"step": 3864 |
|
}, |
|
{ |
|
"epoch": 4.372328458942632, |
|
"grad_norm": 0.4618189334869385, |
|
"learning_rate": 4.372328458942632e-05, |
|
"loss": 0.3469, |
|
"step": 3887 |
|
}, |
|
{ |
|
"epoch": 4.398200224971879, |
|
"grad_norm": 0.6110893487930298, |
|
"learning_rate": 4.398200224971879e-05, |
|
"loss": 0.4085, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 4.424071991001125, |
|
"grad_norm": 8.723979949951172, |
|
"learning_rate": 4.424071991001125e-05, |
|
"loss": 0.263, |
|
"step": 3933 |
|
}, |
|
{ |
|
"epoch": 4.449943757030371, |
|
"grad_norm": 19.94253158569336, |
|
"learning_rate": 4.4499437570303716e-05, |
|
"loss": 0.3318, |
|
"step": 3956 |
|
}, |
|
{ |
|
"epoch": 4.475815523059618, |
|
"grad_norm": 0.9224587082862854, |
|
"learning_rate": 4.475815523059618e-05, |
|
"loss": 0.462, |
|
"step": 3979 |
|
}, |
|
{ |
|
"epoch": 4.5016872890888635, |
|
"grad_norm": 9.54886531829834, |
|
"learning_rate": 4.5016872890888636e-05, |
|
"loss": 0.4471, |
|
"step": 4002 |
|
}, |
|
{ |
|
"epoch": 4.52755905511811, |
|
"grad_norm": 3.35776686668396, |
|
"learning_rate": 4.52755905511811e-05, |
|
"loss": 0.4742, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 4.553430821147357, |
|
"grad_norm": 7.887851715087891, |
|
"learning_rate": 4.553430821147357e-05, |
|
"loss": 0.4714, |
|
"step": 4048 |
|
}, |
|
{ |
|
"epoch": 4.579302587176603, |
|
"grad_norm": 3.327909469604492, |
|
"learning_rate": 4.5793025871766033e-05, |
|
"loss": 0.6452, |
|
"step": 4071 |
|
}, |
|
{ |
|
"epoch": 4.605174353205849, |
|
"grad_norm": 1.3243390321731567, |
|
"learning_rate": 4.60517435320585e-05, |
|
"loss": 0.3745, |
|
"step": 4094 |
|
}, |
|
{ |
|
"epoch": 4.631046119235096, |
|
"grad_norm": 0.7224089503288269, |
|
"learning_rate": 4.631046119235096e-05, |
|
"loss": 0.2628, |
|
"step": 4117 |
|
}, |
|
{ |
|
"epoch": 4.656917885264342, |
|
"grad_norm": 0.017284361645579338, |
|
"learning_rate": 4.656917885264342e-05, |
|
"loss": 0.317, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 4.6827896512935885, |
|
"grad_norm": 14.2166166305542, |
|
"learning_rate": 4.682789651293589e-05, |
|
"loss": 0.464, |
|
"step": 4163 |
|
}, |
|
{ |
|
"epoch": 4.708661417322834, |
|
"grad_norm": 0.49798861145973206, |
|
"learning_rate": 4.708661417322835e-05, |
|
"loss": 0.3336, |
|
"step": 4186 |
|
}, |
|
{ |
|
"epoch": 4.734533183352081, |
|
"grad_norm": 1.5816570520401, |
|
"learning_rate": 4.7345331833520814e-05, |
|
"loss": 0.423, |
|
"step": 4209 |
|
}, |
|
{ |
|
"epoch": 4.760404949381328, |
|
"grad_norm": 4.431649208068848, |
|
"learning_rate": 4.760404949381328e-05, |
|
"loss": 0.4474, |
|
"step": 4232 |
|
}, |
|
{ |
|
"epoch": 4.786276715410573, |
|
"grad_norm": 7.487512588500977, |
|
"learning_rate": 4.7862767154105734e-05, |
|
"loss": 0.32, |
|
"step": 4255 |
|
}, |
|
{ |
|
"epoch": 4.81214848143982, |
|
"grad_norm": 3.283292293548584, |
|
"learning_rate": 4.8121484814398204e-05, |
|
"loss": 0.4767, |
|
"step": 4278 |
|
}, |
|
{ |
|
"epoch": 4.838020247469066, |
|
"grad_norm": 7.209409236907959, |
|
"learning_rate": 4.838020247469067e-05, |
|
"loss": 0.2593, |
|
"step": 4301 |
|
}, |
|
{ |
|
"epoch": 4.863892013498313, |
|
"grad_norm": 7.526238441467285, |
|
"learning_rate": 4.863892013498313e-05, |
|
"loss": 0.3552, |
|
"step": 4324 |
|
}, |
|
{ |
|
"epoch": 4.889763779527559, |
|
"grad_norm": 0.18729792535305023, |
|
"learning_rate": 4.8897637795275594e-05, |
|
"loss": 0.3581, |
|
"step": 4347 |
|
}, |
|
{ |
|
"epoch": 4.915635545556805, |
|
"grad_norm": 0.21440748870372772, |
|
"learning_rate": 4.915635545556805e-05, |
|
"loss": 0.2921, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 4.941507311586052, |
|
"grad_norm": 12.171503067016602, |
|
"learning_rate": 4.941507311586052e-05, |
|
"loss": 0.5248, |
|
"step": 4393 |
|
}, |
|
{ |
|
"epoch": 4.967379077615298, |
|
"grad_norm": 0.14398285746574402, |
|
"learning_rate": 4.9673790776152984e-05, |
|
"loss": 0.4253, |
|
"step": 4416 |
|
}, |
|
{ |
|
"epoch": 4.993250843644544, |
|
"grad_norm": 2.316239595413208, |
|
"learning_rate": 4.993250843644545e-05, |
|
"loss": 0.4106, |
|
"step": 4439 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.7815126050420168, |
|
"eval_auc": 0.8033579742057251, |
|
"eval_f1": 0.0, |
|
"eval_loss": 0.50748610496521, |
|
"eval_precision": 0.0, |
|
"eval_recall": 0.0, |
|
"eval_runtime": 2.7592, |
|
"eval_samples_per_second": 172.516, |
|
"eval_steps_per_second": 21.746, |
|
"step": 4445 |
|
}, |
|
{ |
|
"epoch": 5.019122609673791, |
|
"grad_norm": 4.349539279937744, |
|
"learning_rate": 4.997875265591801e-05, |
|
"loss": 0.3046, |
|
"step": 4462 |
|
}, |
|
{ |
|
"epoch": 5.0449943757030375, |
|
"grad_norm": 0.3448341190814972, |
|
"learning_rate": 4.995000624921885e-05, |
|
"loss": 0.3477, |
|
"step": 4485 |
|
}, |
|
{ |
|
"epoch": 5.070866141732283, |
|
"grad_norm": 0.6105608940124512, |
|
"learning_rate": 4.9921259842519685e-05, |
|
"loss": 0.5053, |
|
"step": 4508 |
|
}, |
|
{ |
|
"epoch": 5.09673790776153, |
|
"grad_norm": 8.691744804382324, |
|
"learning_rate": 4.989251343582053e-05, |
|
"loss": 0.3679, |
|
"step": 4531 |
|
}, |
|
{ |
|
"epoch": 5.122609673790776, |
|
"grad_norm": 2.42826771736145, |
|
"learning_rate": 4.9863767029121364e-05, |
|
"loss": 0.2206, |
|
"step": 4554 |
|
}, |
|
{ |
|
"epoch": 5.1484814398200225, |
|
"grad_norm": 2.756155490875244, |
|
"learning_rate": 4.98350206224222e-05, |
|
"loss": 0.6075, |
|
"step": 4577 |
|
}, |
|
{ |
|
"epoch": 5.174353205849269, |
|
"grad_norm": 0.6143556833267212, |
|
"learning_rate": 4.9806274215723035e-05, |
|
"loss": 0.2272, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 5.200224971878515, |
|
"grad_norm": 0.17892147600650787, |
|
"learning_rate": 4.977752780902387e-05, |
|
"loss": 0.5256, |
|
"step": 4623 |
|
}, |
|
{ |
|
"epoch": 5.226096737907762, |
|
"grad_norm": 0.12368874251842499, |
|
"learning_rate": 4.9748781402324714e-05, |
|
"loss": 0.5569, |
|
"step": 4646 |
|
}, |
|
{ |
|
"epoch": 5.251968503937007, |
|
"grad_norm": 0.048263806849718094, |
|
"learning_rate": 4.972003499562555e-05, |
|
"loss": 0.2298, |
|
"step": 4669 |
|
}, |
|
{ |
|
"epoch": 5.277840269966254, |
|
"grad_norm": 4.790956020355225, |
|
"learning_rate": 4.9691288588926386e-05, |
|
"loss": 0.5415, |
|
"step": 4692 |
|
}, |
|
{ |
|
"epoch": 5.303712035995501, |
|
"grad_norm": 5.684449672698975, |
|
"learning_rate": 4.966254218222722e-05, |
|
"loss": 0.3086, |
|
"step": 4715 |
|
}, |
|
{ |
|
"epoch": 5.329583802024747, |
|
"grad_norm": 0.4559657871723175, |
|
"learning_rate": 4.9633795775528064e-05, |
|
"loss": 0.3575, |
|
"step": 4738 |
|
}, |
|
{ |
|
"epoch": 5.355455568053993, |
|
"grad_norm": 7.768259048461914, |
|
"learning_rate": 4.96050493688289e-05, |
|
"loss": 0.4601, |
|
"step": 4761 |
|
}, |
|
{ |
|
"epoch": 5.38132733408324, |
|
"grad_norm": 0.44012290239334106, |
|
"learning_rate": 4.9576302962129736e-05, |
|
"loss": 0.3265, |
|
"step": 4784 |
|
}, |
|
{ |
|
"epoch": 5.407199100112486, |
|
"grad_norm": 0.22265122830867767, |
|
"learning_rate": 4.954755655543057e-05, |
|
"loss": 0.6152, |
|
"step": 4807 |
|
}, |
|
{ |
|
"epoch": 5.433070866141732, |
|
"grad_norm": 4.463795185089111, |
|
"learning_rate": 4.951881014873141e-05, |
|
"loss": 0.3046, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 5.458942632170979, |
|
"grad_norm": 5.145118236541748, |
|
"learning_rate": 4.949006374203225e-05, |
|
"loss": 0.4773, |
|
"step": 4853 |
|
}, |
|
{ |
|
"epoch": 5.484814398200225, |
|
"grad_norm": 0.24177011847496033, |
|
"learning_rate": 4.9461317335333087e-05, |
|
"loss": 0.2634, |
|
"step": 4876 |
|
}, |
|
{ |
|
"epoch": 5.5106861642294716, |
|
"grad_norm": 0.8902063369750977, |
|
"learning_rate": 4.943257092863392e-05, |
|
"loss": 0.3362, |
|
"step": 4899 |
|
}, |
|
{ |
|
"epoch": 5.536557930258717, |
|
"grad_norm": 0.39589497447013855, |
|
"learning_rate": 4.940382452193476e-05, |
|
"loss": 0.6168, |
|
"step": 4922 |
|
}, |
|
{ |
|
"epoch": 5.562429696287964, |
|
"grad_norm": 1.736325979232788, |
|
"learning_rate": 4.9375078115235594e-05, |
|
"loss": 0.5466, |
|
"step": 4945 |
|
}, |
|
{ |
|
"epoch": 5.588301462317211, |
|
"grad_norm": 2.3112049102783203, |
|
"learning_rate": 4.934633170853644e-05, |
|
"loss": 0.3546, |
|
"step": 4968 |
|
}, |
|
{ |
|
"epoch": 5.6141732283464565, |
|
"grad_norm": 0.4776885509490967, |
|
"learning_rate": 4.931758530183727e-05, |
|
"loss": 0.3755, |
|
"step": 4991 |
|
}, |
|
{ |
|
"epoch": 5.640044994375703, |
|
"grad_norm": 0.09375002235174179, |
|
"learning_rate": 4.928883889513811e-05, |
|
"loss": 0.2044, |
|
"step": 5014 |
|
}, |
|
{ |
|
"epoch": 5.665916760404949, |
|
"grad_norm": 11.319817543029785, |
|
"learning_rate": 4.9260092488438945e-05, |
|
"loss": 0.5937, |
|
"step": 5037 |
|
}, |
|
{ |
|
"epoch": 5.691788526434196, |
|
"grad_norm": 1.1792347431182861, |
|
"learning_rate": 4.923134608173979e-05, |
|
"loss": 0.3856, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 5.717660292463442, |
|
"grad_norm": 6.615360736846924, |
|
"learning_rate": 4.920259967504062e-05, |
|
"loss": 0.3669, |
|
"step": 5083 |
|
}, |
|
{ |
|
"epoch": 5.743532058492688, |
|
"grad_norm": 6.705723762512207, |
|
"learning_rate": 4.917385326834146e-05, |
|
"loss": 0.2221, |
|
"step": 5106 |
|
}, |
|
{ |
|
"epoch": 5.769403824521935, |
|
"grad_norm": 2.312495470046997, |
|
"learning_rate": 4.9145106861642295e-05, |
|
"loss": 0.4308, |
|
"step": 5129 |
|
}, |
|
{ |
|
"epoch": 5.7952755905511815, |
|
"grad_norm": 11.482613563537598, |
|
"learning_rate": 4.911636045494313e-05, |
|
"loss": 0.3644, |
|
"step": 5152 |
|
}, |
|
{ |
|
"epoch": 5.821147356580427, |
|
"grad_norm": 4.5011982917785645, |
|
"learning_rate": 4.9087614048243973e-05, |
|
"loss": 0.5318, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 5.847019122609674, |
|
"grad_norm": 3.1337974071502686, |
|
"learning_rate": 4.905886764154481e-05, |
|
"loss": 0.4274, |
|
"step": 5198 |
|
}, |
|
{ |
|
"epoch": 5.872890888638921, |
|
"grad_norm": 0.16391399502754211, |
|
"learning_rate": 4.9030121234845645e-05, |
|
"loss": 0.3304, |
|
"step": 5221 |
|
}, |
|
{ |
|
"epoch": 5.898762654668166, |
|
"grad_norm": 0.630847692489624, |
|
"learning_rate": 4.900137482814649e-05, |
|
"loss": 0.4878, |
|
"step": 5244 |
|
}, |
|
{ |
|
"epoch": 5.924634420697413, |
|
"grad_norm": 3.9900641441345215, |
|
"learning_rate": 4.897262842144732e-05, |
|
"loss": 0.3846, |
|
"step": 5267 |
|
}, |
|
{ |
|
"epoch": 5.950506186726659, |
|
"grad_norm": 5.799670219421387, |
|
"learning_rate": 4.894388201474816e-05, |
|
"loss": 0.3469, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 5.9763779527559056, |
|
"grad_norm": 0.49845901131629944, |
|
"learning_rate": 4.8915135608048996e-05, |
|
"loss": 0.3733, |
|
"step": 5313 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.7857142857142857, |
|
"eval_auc": 0.7628709237705779, |
|
"eval_f1": 0.0, |
|
"eval_loss": 0.6193206310272217, |
|
"eval_precision": 0.0, |
|
"eval_recall": 0.0, |
|
"eval_runtime": 2.7283, |
|
"eval_samples_per_second": 174.467, |
|
"eval_steps_per_second": 21.992, |
|
"step": 5334 |
|
}, |
|
{ |
|
"epoch": 6.002249718785152, |
|
"grad_norm": 0.5700758695602417, |
|
"learning_rate": 4.888638920134984e-05, |
|
"loss": 0.3025, |
|
"step": 5336 |
|
}, |
|
{ |
|
"epoch": 6.028121484814398, |
|
"grad_norm": 0.3606770932674408, |
|
"learning_rate": 4.8857642794650674e-05, |
|
"loss": 0.383, |
|
"step": 5359 |
|
}, |
|
{ |
|
"epoch": 6.053993250843645, |
|
"grad_norm": 0.136641263961792, |
|
"learning_rate": 4.88288963879515e-05, |
|
"loss": 0.2687, |
|
"step": 5382 |
|
}, |
|
{ |
|
"epoch": 6.0798650168728905, |
|
"grad_norm": 0.3220977187156677, |
|
"learning_rate": 4.8800149981252346e-05, |
|
"loss": 0.4189, |
|
"step": 5405 |
|
}, |
|
{ |
|
"epoch": 6.105736782902137, |
|
"grad_norm": 0.8333466053009033, |
|
"learning_rate": 4.877140357455318e-05, |
|
"loss": 0.3677, |
|
"step": 5428 |
|
}, |
|
{ |
|
"epoch": 6.131608548931384, |
|
"grad_norm": 0.08014432340860367, |
|
"learning_rate": 4.8742657167854025e-05, |
|
"loss": 0.3624, |
|
"step": 5451 |
|
}, |
|
{ |
|
"epoch": 6.15748031496063, |
|
"grad_norm": 8.192437171936035, |
|
"learning_rate": 4.871391076115486e-05, |
|
"loss": 0.4036, |
|
"step": 5474 |
|
}, |
|
{ |
|
"epoch": 6.183352080989876, |
|
"grad_norm": 6.411347389221191, |
|
"learning_rate": 4.8685164354455696e-05, |
|
"loss": 0.3529, |
|
"step": 5497 |
|
}, |
|
{ |
|
"epoch": 6.209223847019123, |
|
"grad_norm": 0.006956384517252445, |
|
"learning_rate": 4.865641794775653e-05, |
|
"loss": 0.1559, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 6.235095613048369, |
|
"grad_norm": 0.1893380731344223, |
|
"learning_rate": 4.862767154105737e-05, |
|
"loss": 0.5716, |
|
"step": 5543 |
|
}, |
|
{ |
|
"epoch": 6.2609673790776155, |
|
"grad_norm": 0.47687458992004395, |
|
"learning_rate": 4.859892513435821e-05, |
|
"loss": 0.399, |
|
"step": 5566 |
|
}, |
|
{ |
|
"epoch": 6.286839145106861, |
|
"grad_norm": 4.109427452087402, |
|
"learning_rate": 4.857017872765905e-05, |
|
"loss": 0.3858, |
|
"step": 5589 |
|
}, |
|
{ |
|
"epoch": 6.312710911136108, |
|
"grad_norm": 0.6130828857421875, |
|
"learning_rate": 4.854143232095988e-05, |
|
"loss": 0.3434, |
|
"step": 5612 |
|
}, |
|
{ |
|
"epoch": 6.338582677165355, |
|
"grad_norm": 2.936263084411621, |
|
"learning_rate": 4.851268591426072e-05, |
|
"loss": 0.7265, |
|
"step": 5635 |
|
}, |
|
{ |
|
"epoch": 6.3644544431946, |
|
"grad_norm": 0.4069950580596924, |
|
"learning_rate": 4.848393950756156e-05, |
|
"loss": 0.2387, |
|
"step": 5658 |
|
}, |
|
{ |
|
"epoch": 6.390326209223847, |
|
"grad_norm": 3.4261677265167236, |
|
"learning_rate": 4.84551931008624e-05, |
|
"loss": 0.4388, |
|
"step": 5681 |
|
}, |
|
{ |
|
"epoch": 6.416197975253094, |
|
"grad_norm": 4.583181381225586, |
|
"learning_rate": 4.842644669416323e-05, |
|
"loss": 0.321, |
|
"step": 5704 |
|
}, |
|
{ |
|
"epoch": 6.44206974128234, |
|
"grad_norm": 4.256218433380127, |
|
"learning_rate": 4.839770028746407e-05, |
|
"loss": 0.3261, |
|
"step": 5727 |
|
}, |
|
{ |
|
"epoch": 6.467941507311586, |
|
"grad_norm": 1.0955240726470947, |
|
"learning_rate": 4.8368953880764905e-05, |
|
"loss": 0.3753, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 6.493813273340832, |
|
"grad_norm": 7.2450361251831055, |
|
"learning_rate": 4.834020747406575e-05, |
|
"loss": 0.2428, |
|
"step": 5773 |
|
}, |
|
{ |
|
"epoch": 6.519685039370079, |
|
"grad_norm": 5.181588172912598, |
|
"learning_rate": 4.831146106736658e-05, |
|
"loss": 0.3341, |
|
"step": 5796 |
|
}, |
|
{ |
|
"epoch": 6.545556805399325, |
|
"grad_norm": 2.217822790145874, |
|
"learning_rate": 4.828271466066742e-05, |
|
"loss": 0.4004, |
|
"step": 5819 |
|
}, |
|
{ |
|
"epoch": 6.571428571428571, |
|
"grad_norm": 0.3265409469604492, |
|
"learning_rate": 4.8253968253968255e-05, |
|
"loss": 0.4519, |
|
"step": 5842 |
|
}, |
|
{ |
|
"epoch": 6.597300337457818, |
|
"grad_norm": 3.4193084239959717, |
|
"learning_rate": 4.822522184726909e-05, |
|
"loss": 0.4347, |
|
"step": 5865 |
|
}, |
|
{ |
|
"epoch": 6.6231721034870645, |
|
"grad_norm": 3.1898536682128906, |
|
"learning_rate": 4.8196475440569934e-05, |
|
"loss": 0.2508, |
|
"step": 5888 |
|
}, |
|
{ |
|
"epoch": 6.64904386951631, |
|
"grad_norm": 5.070316314697266, |
|
"learning_rate": 4.816772903387077e-05, |
|
"loss": 0.4112, |
|
"step": 5911 |
|
}, |
|
{ |
|
"epoch": 6.674915635545557, |
|
"grad_norm": 0.23132477700710297, |
|
"learning_rate": 4.8138982627171605e-05, |
|
"loss": 0.2206, |
|
"step": 5934 |
|
}, |
|
{ |
|
"epoch": 6.700787401574803, |
|
"grad_norm": 4.237873554229736, |
|
"learning_rate": 4.811023622047244e-05, |
|
"loss": 0.5504, |
|
"step": 5957 |
|
}, |
|
{ |
|
"epoch": 6.7266591676040495, |
|
"grad_norm": 0.857448160648346, |
|
"learning_rate": 4.808148981377328e-05, |
|
"loss": 0.3815, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 6.752530933633296, |
|
"grad_norm": 5.193442344665527, |
|
"learning_rate": 4.805274340707412e-05, |
|
"loss": 0.3129, |
|
"step": 6003 |
|
}, |
|
{ |
|
"epoch": 6.778402699662542, |
|
"grad_norm": 4.156959056854248, |
|
"learning_rate": 4.8023997000374956e-05, |
|
"loss": 0.4625, |
|
"step": 6026 |
|
}, |
|
{ |
|
"epoch": 6.804274465691789, |
|
"grad_norm": 0.4680914878845215, |
|
"learning_rate": 4.799525059367579e-05, |
|
"loss": 0.2069, |
|
"step": 6049 |
|
}, |
|
{ |
|
"epoch": 6.830146231721034, |
|
"grad_norm": 5.006599426269531, |
|
"learning_rate": 4.796650418697663e-05, |
|
"loss": 0.4084, |
|
"step": 6072 |
|
}, |
|
{ |
|
"epoch": 6.856017997750281, |
|
"grad_norm": 6.840053081512451, |
|
"learning_rate": 4.793775778027747e-05, |
|
"loss": 0.2994, |
|
"step": 6095 |
|
}, |
|
{ |
|
"epoch": 6.881889763779528, |
|
"grad_norm": 0.11373735964298248, |
|
"learning_rate": 4.7909011373578306e-05, |
|
"loss": 0.2934, |
|
"step": 6118 |
|
}, |
|
{ |
|
"epoch": 6.907761529808774, |
|
"grad_norm": 5.152923107147217, |
|
"learning_rate": 4.788026496687914e-05, |
|
"loss": 0.2687, |
|
"step": 6141 |
|
}, |
|
{ |
|
"epoch": 6.93363329583802, |
|
"grad_norm": 0.021206054836511612, |
|
"learning_rate": 4.785151856017998e-05, |
|
"loss": 0.3019, |
|
"step": 6164 |
|
}, |
|
{ |
|
"epoch": 6.959505061867267, |
|
"grad_norm": 0.06403312087059021, |
|
"learning_rate": 4.7822772153480814e-05, |
|
"loss": 0.2471, |
|
"step": 6187 |
|
}, |
|
{ |
|
"epoch": 6.985376827896513, |
|
"grad_norm": 3.1448378562927246, |
|
"learning_rate": 4.7794025746781657e-05, |
|
"loss": 0.6191, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.819327731092437, |
|
"eval_auc": 0.8453785257418476, |
|
"eval_f1": 0.4625, |
|
"eval_loss": 0.3894089162349701, |
|
"eval_precision": 0.6379310344827587, |
|
"eval_recall": 0.3627450980392157, |
|
"eval_runtime": 2.7066, |
|
"eval_samples_per_second": 175.867, |
|
"eval_steps_per_second": 22.168, |
|
"step": 6223 |
|
} |
|
], |
|
"logging_steps": 23, |
|
"max_steps": 44450, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 50, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 20, |
|
"early_stopping_threshold": 0.01 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.0174156175154545e+19, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|