|
{ |
|
"best_metric": 0.6728110599078341, |
|
"best_model_checkpoint": "convnextv2-base-22k-224-finetuned-tekno24-highdata-90/checkpoint-1553", |
|
"epoch": 29.724770642201836, |
|
"eval_steps": 500, |
|
"global_step": 2430, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.12232415902140673, |
|
"grad_norm": 15.331552505493164, |
|
"learning_rate": 2.05761316872428e-06, |
|
"loss": 1.3954, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.24464831804281345, |
|
"grad_norm": 13.410693168640137, |
|
"learning_rate": 4.11522633744856e-06, |
|
"loss": 1.3826, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.3669724770642202, |
|
"grad_norm": 14.4244966506958, |
|
"learning_rate": 5.967078189300412e-06, |
|
"loss": 1.3692, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.4892966360856269, |
|
"grad_norm": 13.09442138671875, |
|
"learning_rate": 8.02469135802469e-06, |
|
"loss": 1.3596, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.6116207951070336, |
|
"grad_norm": 14.526571273803711, |
|
"learning_rate": 1.008230452674897e-05, |
|
"loss": 1.3519, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.7339449541284404, |
|
"grad_norm": 21.34126091003418, |
|
"learning_rate": 1.2139917695473252e-05, |
|
"loss": 1.3338, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.8562691131498471, |
|
"grad_norm": 17.52889060974121, |
|
"learning_rate": 1.3991769547325103e-05, |
|
"loss": 1.3236, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.9785932721712538, |
|
"grad_norm": 17.072633743286133, |
|
"learning_rate": 1.604938271604938e-05, |
|
"loss": 1.3277, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.9908256880733946, |
|
"eval_accuracy": 0.4147465437788018, |
|
"eval_f1": 0.32804846883348265, |
|
"eval_loss": 1.2870396375656128, |
|
"eval_precision": 0.3713606931606389, |
|
"eval_recall": 0.4147465437788018, |
|
"eval_runtime": 2.1883, |
|
"eval_samples_per_second": 99.163, |
|
"eval_steps_per_second": 6.398, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 1.1009174311926606, |
|
"grad_norm": 22.833921432495117, |
|
"learning_rate": 1.8106995884773663e-05, |
|
"loss": 1.3091, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.2232415902140672, |
|
"grad_norm": 10.650703430175781, |
|
"learning_rate": 2.016460905349794e-05, |
|
"loss": 1.3123, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.345565749235474, |
|
"grad_norm": 9.893655776977539, |
|
"learning_rate": 2.2222222222222223e-05, |
|
"loss": 1.2944, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.4678899082568808, |
|
"grad_norm": 54.886756896972656, |
|
"learning_rate": 2.3868312757201647e-05, |
|
"loss": 1.2842, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.5902140672782874, |
|
"grad_norm": 16.818227767944336, |
|
"learning_rate": 2.5925925925925925e-05, |
|
"loss": 1.2572, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.7125382262996942, |
|
"grad_norm": 9.368868827819824, |
|
"learning_rate": 2.7983539094650207e-05, |
|
"loss": 1.2222, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.834862385321101, |
|
"grad_norm": 12.728253364562988, |
|
"learning_rate": 3.0041152263374488e-05, |
|
"loss": 1.2346, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.9571865443425076, |
|
"grad_norm": 13.562764167785645, |
|
"learning_rate": 3.209876543209876e-05, |
|
"loss": 1.2024, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.9938837920489296, |
|
"eval_accuracy": 0.47465437788018433, |
|
"eval_f1": 0.3906684622928812, |
|
"eval_loss": 1.0890263319015503, |
|
"eval_precision": 0.49441907661085743, |
|
"eval_recall": 0.47465437788018433, |
|
"eval_runtime": 2.1621, |
|
"eval_samples_per_second": 100.363, |
|
"eval_steps_per_second": 6.475, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 2.079510703363914, |
|
"grad_norm": 10.129838943481445, |
|
"learning_rate": 3.4156378600823045e-05, |
|
"loss": 1.2231, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.2018348623853212, |
|
"grad_norm": 10.82052230834961, |
|
"learning_rate": 3.6213991769547327e-05, |
|
"loss": 1.2082, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.324159021406728, |
|
"grad_norm": 15.055948257446289, |
|
"learning_rate": 3.82716049382716e-05, |
|
"loss": 1.1827, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.4464831804281344, |
|
"grad_norm": 14.08010196685791, |
|
"learning_rate": 4.032921810699588e-05, |
|
"loss": 1.1564, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.5688073394495414, |
|
"grad_norm": 19.530576705932617, |
|
"learning_rate": 4.2386831275720165e-05, |
|
"loss": 1.1508, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.691131498470948, |
|
"grad_norm": 17.710391998291016, |
|
"learning_rate": 4.4444444444444447e-05, |
|
"loss": 1.2407, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.8134556574923546, |
|
"grad_norm": 19.007694244384766, |
|
"learning_rate": 4.650205761316873e-05, |
|
"loss": 1.1333, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.9357798165137616, |
|
"grad_norm": 11.404845237731934, |
|
"learning_rate": 4.855967078189301e-05, |
|
"loss": 1.2067, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.996941896024465, |
|
"eval_accuracy": 0.543778801843318, |
|
"eval_f1": 0.49651048113106194, |
|
"eval_loss": 1.060142993927002, |
|
"eval_precision": 0.5083802142557903, |
|
"eval_recall": 0.543778801843318, |
|
"eval_runtime": 2.1941, |
|
"eval_samples_per_second": 98.904, |
|
"eval_steps_per_second": 6.381, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 3.058103975535168, |
|
"grad_norm": 7.823658466339111, |
|
"learning_rate": 4.993141289437586e-05, |
|
"loss": 1.1512, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 3.180428134556575, |
|
"grad_norm": 7.272976875305176, |
|
"learning_rate": 4.970278920896205e-05, |
|
"loss": 1.2064, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 3.302752293577982, |
|
"grad_norm": 8.453497886657715, |
|
"learning_rate": 4.9474165523548245e-05, |
|
"loss": 1.1758, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 3.4250764525993884, |
|
"grad_norm": 6.063063144683838, |
|
"learning_rate": 4.924554183813443e-05, |
|
"loss": 1.1481, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 3.547400611620795, |
|
"grad_norm": 11.22547435760498, |
|
"learning_rate": 4.901691815272063e-05, |
|
"loss": 1.1818, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 3.669724770642202, |
|
"grad_norm": 8.580599784851074, |
|
"learning_rate": 4.8788294467306815e-05, |
|
"loss": 1.1345, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 3.7920489296636086, |
|
"grad_norm": 16.203880310058594, |
|
"learning_rate": 4.855967078189301e-05, |
|
"loss": 1.1451, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 3.914373088685015, |
|
"grad_norm": 7.94223165512085, |
|
"learning_rate": 4.83310470964792e-05, |
|
"loss": 1.206, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.5391705069124424, |
|
"eval_f1": 0.515930279628824, |
|
"eval_loss": 1.0142838954925537, |
|
"eval_precision": 0.5180022054326838, |
|
"eval_recall": 0.5391705069124424, |
|
"eval_runtime": 2.1844, |
|
"eval_samples_per_second": 99.339, |
|
"eval_steps_per_second": 6.409, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 4.036697247706422, |
|
"grad_norm": 7.936139106750488, |
|
"learning_rate": 4.810242341106539e-05, |
|
"loss": 1.1749, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 4.159021406727828, |
|
"grad_norm": 6.124804496765137, |
|
"learning_rate": 4.787379972565158e-05, |
|
"loss": 1.1473, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 4.281345565749236, |
|
"grad_norm": 9.12104320526123, |
|
"learning_rate": 4.764517604023777e-05, |
|
"loss": 1.1809, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 4.4036697247706424, |
|
"grad_norm": 6.582239627838135, |
|
"learning_rate": 4.741655235482396e-05, |
|
"loss": 1.1728, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 4.525993883792049, |
|
"grad_norm": 7.613648891448975, |
|
"learning_rate": 4.718792866941015e-05, |
|
"loss": 1.1315, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 4.648318042813456, |
|
"grad_norm": 11.376667976379395, |
|
"learning_rate": 4.6959304983996346e-05, |
|
"loss": 1.1481, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 4.770642201834862, |
|
"grad_norm": 6.794886112213135, |
|
"learning_rate": 4.6730681298582534e-05, |
|
"loss": 1.0927, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 4.892966360856269, |
|
"grad_norm": 7.452205657958984, |
|
"learning_rate": 4.650205761316873e-05, |
|
"loss": 1.1049, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 4.990825688073395, |
|
"eval_accuracy": 0.576036866359447, |
|
"eval_f1": 0.5450544928742728, |
|
"eval_loss": 0.968787431716919, |
|
"eval_precision": 0.5467442569565754, |
|
"eval_recall": 0.576036866359447, |
|
"eval_runtime": 2.1812, |
|
"eval_samples_per_second": 99.488, |
|
"eval_steps_per_second": 6.419, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 5.015290519877676, |
|
"grad_norm": 6.982997417449951, |
|
"learning_rate": 4.6273433927754916e-05, |
|
"loss": 1.1165, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 5.137614678899083, |
|
"grad_norm": 7.05279541015625, |
|
"learning_rate": 4.604481024234111e-05, |
|
"loss": 1.0755, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 5.259938837920489, |
|
"grad_norm": 9.04000186920166, |
|
"learning_rate": 4.58161865569273e-05, |
|
"loss": 1.1144, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 5.382262996941896, |
|
"grad_norm": 5.530284881591797, |
|
"learning_rate": 4.558756287151349e-05, |
|
"loss": 1.088, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 5.504587155963303, |
|
"grad_norm": 11.77010440826416, |
|
"learning_rate": 4.535893918609968e-05, |
|
"loss": 1.0805, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 5.626911314984709, |
|
"grad_norm": 9.45374870300293, |
|
"learning_rate": 4.513031550068587e-05, |
|
"loss": 1.1098, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 5.749235474006117, |
|
"grad_norm": 9.706932067871094, |
|
"learning_rate": 4.4901691815272064e-05, |
|
"loss": 1.0726, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 5.871559633027523, |
|
"grad_norm": 9.147931098937988, |
|
"learning_rate": 4.467306812985825e-05, |
|
"loss": 1.0934, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 5.99388379204893, |
|
"grad_norm": 6.612918376922607, |
|
"learning_rate": 4.4444444444444447e-05, |
|
"loss": 1.0931, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 5.99388379204893, |
|
"eval_accuracy": 0.5622119815668203, |
|
"eval_f1": 0.5562044726083856, |
|
"eval_loss": 1.0351340770721436, |
|
"eval_precision": 0.5939199765927876, |
|
"eval_recall": 0.5622119815668203, |
|
"eval_runtime": 2.1678, |
|
"eval_samples_per_second": 100.104, |
|
"eval_steps_per_second": 6.458, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 6.116207951070336, |
|
"grad_norm": 10.358129501342773, |
|
"learning_rate": 4.4215820759030634e-05, |
|
"loss": 1.0469, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 6.238532110091743, |
|
"grad_norm": 5.924838542938232, |
|
"learning_rate": 4.398719707361683e-05, |
|
"loss": 1.0814, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 6.36085626911315, |
|
"grad_norm": 7.012035369873047, |
|
"learning_rate": 4.3758573388203024e-05, |
|
"loss": 1.1091, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 6.483180428134556, |
|
"grad_norm": 6.860468864440918, |
|
"learning_rate": 4.352994970278921e-05, |
|
"loss": 1.091, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 6.605504587155964, |
|
"grad_norm": 6.236204624176025, |
|
"learning_rate": 4.3301326017375406e-05, |
|
"loss": 1.1033, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 6.72782874617737, |
|
"grad_norm": 11.546574592590332, |
|
"learning_rate": 4.3072702331961594e-05, |
|
"loss": 1.067, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 6.850152905198777, |
|
"grad_norm": 10.157209396362305, |
|
"learning_rate": 4.284407864654779e-05, |
|
"loss": 1.054, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 6.972477064220183, |
|
"grad_norm": 7.421886444091797, |
|
"learning_rate": 4.261545496113398e-05, |
|
"loss": 1.0752, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 6.996941896024465, |
|
"eval_accuracy": 0.5898617511520737, |
|
"eval_f1": 0.5592428280600323, |
|
"eval_loss": 0.9370155930519104, |
|
"eval_precision": 0.5730448203692986, |
|
"eval_recall": 0.5898617511520737, |
|
"eval_runtime": 2.1864, |
|
"eval_samples_per_second": 99.25, |
|
"eval_steps_per_second": 6.403, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 7.09480122324159, |
|
"grad_norm": 8.062494277954102, |
|
"learning_rate": 4.2386831275720165e-05, |
|
"loss": 1.0298, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 7.217125382262997, |
|
"grad_norm": 7.170814514160156, |
|
"learning_rate": 4.215820759030636e-05, |
|
"loss": 1.0451, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 7.339449541284404, |
|
"grad_norm": 7.2061662673950195, |
|
"learning_rate": 4.192958390489255e-05, |
|
"loss": 1.0586, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 7.461773700305811, |
|
"grad_norm": 5.97224760055542, |
|
"learning_rate": 4.170096021947874e-05, |
|
"loss": 1.0543, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 7.584097859327217, |
|
"grad_norm": 7.366073131561279, |
|
"learning_rate": 4.147233653406493e-05, |
|
"loss": 1.0903, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 7.706422018348624, |
|
"grad_norm": 6.186879634857178, |
|
"learning_rate": 4.1243712848651125e-05, |
|
"loss": 1.0919, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 7.82874617737003, |
|
"grad_norm": 8.208704948425293, |
|
"learning_rate": 4.101508916323731e-05, |
|
"loss": 1.1254, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 7.951070336391437, |
|
"grad_norm": 13.736848831176758, |
|
"learning_rate": 4.078646547782351e-05, |
|
"loss": 1.03, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.576036866359447, |
|
"eval_f1": 0.5509975224553819, |
|
"eval_loss": 0.9416872262954712, |
|
"eval_precision": 0.5414011052041612, |
|
"eval_recall": 0.576036866359447, |
|
"eval_runtime": 2.1964, |
|
"eval_samples_per_second": 98.796, |
|
"eval_steps_per_second": 6.374, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 8.073394495412844, |
|
"grad_norm": 7.8366546630859375, |
|
"learning_rate": 4.0557841792409695e-05, |
|
"loss": 1.0581, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 8.19571865443425, |
|
"grad_norm": 6.876219272613525, |
|
"learning_rate": 4.032921810699588e-05, |
|
"loss": 1.04, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 8.318042813455657, |
|
"grad_norm": 7.931809902191162, |
|
"learning_rate": 4.010059442158208e-05, |
|
"loss": 1.0319, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 8.440366972477065, |
|
"grad_norm": 11.545510292053223, |
|
"learning_rate": 3.9871970736168266e-05, |
|
"loss": 1.0027, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 8.562691131498472, |
|
"grad_norm": 7.059225559234619, |
|
"learning_rate": 3.964334705075446e-05, |
|
"loss": 1.0164, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 8.685015290519878, |
|
"grad_norm": 6.870743751525879, |
|
"learning_rate": 3.941472336534065e-05, |
|
"loss": 1.0063, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 8.807339449541285, |
|
"grad_norm": 7.4264631271362305, |
|
"learning_rate": 3.918609967992684e-05, |
|
"loss": 1.0359, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 8.929663608562691, |
|
"grad_norm": 8.050666809082031, |
|
"learning_rate": 3.895747599451303e-05, |
|
"loss": 0.988, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 8.990825688073395, |
|
"eval_accuracy": 0.5990783410138248, |
|
"eval_f1": 0.5772079574092954, |
|
"eval_loss": 0.8942155838012695, |
|
"eval_precision": 0.5818760258445811, |
|
"eval_recall": 0.5990783410138248, |
|
"eval_runtime": 2.178, |
|
"eval_samples_per_second": 99.631, |
|
"eval_steps_per_second": 6.428, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 9.051987767584098, |
|
"grad_norm": 6.86763858795166, |
|
"learning_rate": 3.8728852309099226e-05, |
|
"loss": 0.9927, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 9.174311926605505, |
|
"grad_norm": 7.786592483520508, |
|
"learning_rate": 3.8500228623685414e-05, |
|
"loss": 0.976, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 9.296636085626911, |
|
"grad_norm": 9.186685562133789, |
|
"learning_rate": 3.82716049382716e-05, |
|
"loss": 1.0239, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 9.418960244648318, |
|
"grad_norm": 8.555779457092285, |
|
"learning_rate": 3.8042981252857796e-05, |
|
"loss": 0.9526, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 9.541284403669724, |
|
"grad_norm": 9.215676307678223, |
|
"learning_rate": 3.7814357567443984e-05, |
|
"loss": 1.0258, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 9.663608562691131, |
|
"grad_norm": 7.625224590301514, |
|
"learning_rate": 3.758573388203018e-05, |
|
"loss": 1.0042, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 9.785932721712538, |
|
"grad_norm": 9.265899658203125, |
|
"learning_rate": 3.7357110196616373e-05, |
|
"loss": 1.008, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 9.908256880733944, |
|
"grad_norm": 7.059952735900879, |
|
"learning_rate": 3.712848651120257e-05, |
|
"loss": 0.9692, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 9.99388379204893, |
|
"eval_accuracy": 0.6082949308755761, |
|
"eval_f1": 0.5936874933582452, |
|
"eval_loss": 0.9091479182243347, |
|
"eval_precision": 0.5980988666361812, |
|
"eval_recall": 0.6082949308755761, |
|
"eval_runtime": 2.1969, |
|
"eval_samples_per_second": 98.777, |
|
"eval_steps_per_second": 6.373, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 10.030581039755353, |
|
"grad_norm": 8.655425071716309, |
|
"learning_rate": 3.6899862825788756e-05, |
|
"loss": 1.0144, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 10.15290519877676, |
|
"grad_norm": 7.1922607421875, |
|
"learning_rate": 3.6671239140374944e-05, |
|
"loss": 0.9994, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 10.275229357798166, |
|
"grad_norm": 7.687134265899658, |
|
"learning_rate": 3.644261545496114e-05, |
|
"loss": 0.9879, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 10.397553516819572, |
|
"grad_norm": 9.036182403564453, |
|
"learning_rate": 3.6213991769547327e-05, |
|
"loss": 0.9218, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 10.519877675840979, |
|
"grad_norm": 11.018427848815918, |
|
"learning_rate": 3.598536808413352e-05, |
|
"loss": 1.0079, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 10.642201834862385, |
|
"grad_norm": 9.172164916992188, |
|
"learning_rate": 3.575674439871971e-05, |
|
"loss": 0.9795, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 10.764525993883792, |
|
"grad_norm": 10.210972785949707, |
|
"learning_rate": 3.5528120713305904e-05, |
|
"loss": 0.9675, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 10.886850152905199, |
|
"grad_norm": 9.860912322998047, |
|
"learning_rate": 3.529949702789209e-05, |
|
"loss": 0.9896, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 10.996941896024465, |
|
"eval_accuracy": 0.6036866359447005, |
|
"eval_f1": 0.5904668698923761, |
|
"eval_loss": 0.8689674139022827, |
|
"eval_precision": 0.5936795435008436, |
|
"eval_recall": 0.6036866359447005, |
|
"eval_runtime": 2.1973, |
|
"eval_samples_per_second": 98.756, |
|
"eval_steps_per_second": 6.371, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 11.009174311926605, |
|
"grad_norm": 9.831244468688965, |
|
"learning_rate": 3.5070873342478286e-05, |
|
"loss": 0.9261, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 11.131498470948012, |
|
"grad_norm": 11.262879371643066, |
|
"learning_rate": 3.4842249657064474e-05, |
|
"loss": 0.9191, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 11.253822629969418, |
|
"grad_norm": 9.582681655883789, |
|
"learning_rate": 3.461362597165066e-05, |
|
"loss": 0.9026, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 11.376146788990825, |
|
"grad_norm": 9.12176513671875, |
|
"learning_rate": 3.438500228623686e-05, |
|
"loss": 0.9882, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 11.498470948012232, |
|
"grad_norm": 10.87016773223877, |
|
"learning_rate": 3.4156378600823045e-05, |
|
"loss": 0.9379, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 11.62079510703364, |
|
"grad_norm": 8.557551383972168, |
|
"learning_rate": 3.392775491540924e-05, |
|
"loss": 0.9312, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 11.743119266055047, |
|
"grad_norm": 8.361717224121094, |
|
"learning_rate": 3.369913122999543e-05, |
|
"loss": 0.9085, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 11.865443425076453, |
|
"grad_norm": 7.925179958343506, |
|
"learning_rate": 3.347050754458162e-05, |
|
"loss": 1.0008, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 11.98776758409786, |
|
"grad_norm": 9.740839004516602, |
|
"learning_rate": 3.324188385916781e-05, |
|
"loss": 0.9479, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_accuracy": 0.6405529953917051, |
|
"eval_f1": 0.6268359763707001, |
|
"eval_loss": 0.8705018162727356, |
|
"eval_precision": 0.6307341542199126, |
|
"eval_recall": 0.6405529953917051, |
|
"eval_runtime": 2.1984, |
|
"eval_samples_per_second": 98.71, |
|
"eval_steps_per_second": 6.368, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 12.110091743119266, |
|
"grad_norm": 9.437625885009766, |
|
"learning_rate": 3.3013260173754005e-05, |
|
"loss": 0.8172, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 12.232415902140673, |
|
"grad_norm": 8.812919616699219, |
|
"learning_rate": 3.278463648834019e-05, |
|
"loss": 0.8996, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 12.35474006116208, |
|
"grad_norm": 8.750353813171387, |
|
"learning_rate": 3.255601280292638e-05, |
|
"loss": 0.9655, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 12.477064220183486, |
|
"grad_norm": 7.671780586242676, |
|
"learning_rate": 3.2327389117512575e-05, |
|
"loss": 0.9139, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 12.599388379204893, |
|
"grad_norm": 9.657299041748047, |
|
"learning_rate": 3.209876543209876e-05, |
|
"loss": 0.9269, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 12.7217125382263, |
|
"grad_norm": 9.090998649597168, |
|
"learning_rate": 3.187014174668496e-05, |
|
"loss": 0.8867, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 12.844036697247706, |
|
"grad_norm": 9.977676391601562, |
|
"learning_rate": 3.1641518061271146e-05, |
|
"loss": 0.89, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 12.966360856269112, |
|
"grad_norm": 10.456171035766602, |
|
"learning_rate": 3.141289437585734e-05, |
|
"loss": 0.898, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 12.990825688073395, |
|
"eval_accuracy": 0.6497695852534562, |
|
"eval_f1": 0.6439767898106951, |
|
"eval_loss": 0.8568853735923767, |
|
"eval_precision": 0.6465264959675692, |
|
"eval_recall": 0.6497695852534562, |
|
"eval_runtime": 2.1871, |
|
"eval_samples_per_second": 99.216, |
|
"eval_steps_per_second": 6.401, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 13.08868501529052, |
|
"grad_norm": 9.029878616333008, |
|
"learning_rate": 3.118427069044353e-05, |
|
"loss": 0.9008, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 13.211009174311927, |
|
"grad_norm": 11.082198143005371, |
|
"learning_rate": 3.095564700502972e-05, |
|
"loss": 0.8678, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 13.333333333333334, |
|
"grad_norm": 8.106987953186035, |
|
"learning_rate": 3.072702331961592e-05, |
|
"loss": 0.8724, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 13.45565749235474, |
|
"grad_norm": 9.417874336242676, |
|
"learning_rate": 3.0498399634202106e-05, |
|
"loss": 0.8435, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 13.577981651376147, |
|
"grad_norm": 10.039156913757324, |
|
"learning_rate": 3.0269775948788297e-05, |
|
"loss": 0.7984, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 13.700305810397554, |
|
"grad_norm": 10.906753540039062, |
|
"learning_rate": 3.0041152263374488e-05, |
|
"loss": 0.8578, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 13.82262996941896, |
|
"grad_norm": 9.698974609375, |
|
"learning_rate": 2.981252857796068e-05, |
|
"loss": 0.8699, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 13.944954128440367, |
|
"grad_norm": 9.868678092956543, |
|
"learning_rate": 2.958390489254687e-05, |
|
"loss": 0.9101, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 13.99388379204893, |
|
"eval_accuracy": 0.6129032258064516, |
|
"eval_f1": 0.6090804206647874, |
|
"eval_loss": 0.8736193180084229, |
|
"eval_precision": 0.6179152160718983, |
|
"eval_recall": 0.6129032258064516, |
|
"eval_runtime": 2.1919, |
|
"eval_samples_per_second": 99.002, |
|
"eval_steps_per_second": 6.387, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 14.067278287461773, |
|
"grad_norm": 8.79443645477295, |
|
"learning_rate": 2.9355281207133062e-05, |
|
"loss": 0.8439, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 14.18960244648318, |
|
"grad_norm": 9.417186737060547, |
|
"learning_rate": 2.9126657521719253e-05, |
|
"loss": 0.8763, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 14.311926605504587, |
|
"grad_norm": 11.611254692077637, |
|
"learning_rate": 2.8898033836305445e-05, |
|
"loss": 0.8305, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 14.434250764525993, |
|
"grad_norm": 10.025824546813965, |
|
"learning_rate": 2.8669410150891636e-05, |
|
"loss": 0.7743, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 14.5565749235474, |
|
"grad_norm": 10.79056453704834, |
|
"learning_rate": 2.8440786465477824e-05, |
|
"loss": 0.8414, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 14.678899082568808, |
|
"grad_norm": 8.00452709197998, |
|
"learning_rate": 2.8212162780064015e-05, |
|
"loss": 0.8934, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 14.801223241590215, |
|
"grad_norm": 11.359430313110352, |
|
"learning_rate": 2.7983539094650207e-05, |
|
"loss": 0.8844, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 14.923547400611621, |
|
"grad_norm": 10.003506660461426, |
|
"learning_rate": 2.7754915409236398e-05, |
|
"loss": 0.8431, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 14.996941896024465, |
|
"eval_accuracy": 0.6451612903225806, |
|
"eval_f1": 0.6418579664744979, |
|
"eval_loss": 0.8684141039848328, |
|
"eval_precision": 0.6447143927349024, |
|
"eval_recall": 0.6451612903225806, |
|
"eval_runtime": 2.1833, |
|
"eval_samples_per_second": 99.392, |
|
"eval_steps_per_second": 6.412, |
|
"step": 1226 |
|
}, |
|
{ |
|
"epoch": 15.045871559633028, |
|
"grad_norm": 8.596139907836914, |
|
"learning_rate": 2.752629172382259e-05, |
|
"loss": 0.8224, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 15.168195718654435, |
|
"grad_norm": 11.239164352416992, |
|
"learning_rate": 2.729766803840878e-05, |
|
"loss": 0.8407, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 15.290519877675841, |
|
"grad_norm": 13.581149101257324, |
|
"learning_rate": 2.7069044352994972e-05, |
|
"loss": 0.7794, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 15.412844036697248, |
|
"grad_norm": 10.948393821716309, |
|
"learning_rate": 2.6840420667581163e-05, |
|
"loss": 0.8199, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 15.535168195718654, |
|
"grad_norm": 9.145750999450684, |
|
"learning_rate": 2.6611796982167354e-05, |
|
"loss": 0.7257, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 15.65749235474006, |
|
"grad_norm": 11.599444389343262, |
|
"learning_rate": 2.6383173296753542e-05, |
|
"loss": 0.7847, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 15.779816513761467, |
|
"grad_norm": 11.259026527404785, |
|
"learning_rate": 2.6154549611339734e-05, |
|
"loss": 0.8404, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 15.902140672782874, |
|
"grad_norm": 11.746826171875, |
|
"learning_rate": 2.5925925925925925e-05, |
|
"loss": 0.8187, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_accuracy": 0.6221198156682027, |
|
"eval_f1": 0.6198809621231663, |
|
"eval_loss": 0.9032032489776611, |
|
"eval_precision": 0.6206784143654135, |
|
"eval_recall": 0.6221198156682027, |
|
"eval_runtime": 2.1909, |
|
"eval_samples_per_second": 99.047, |
|
"eval_steps_per_second": 6.39, |
|
"step": 1308 |
|
}, |
|
{ |
|
"epoch": 16.02446483180428, |
|
"grad_norm": 9.55320930480957, |
|
"learning_rate": 2.5697302240512116e-05, |
|
"loss": 0.7986, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 16.146788990825687, |
|
"grad_norm": 10.655790328979492, |
|
"learning_rate": 2.5468678555098308e-05, |
|
"loss": 0.7589, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 16.269113149847094, |
|
"grad_norm": 10.707428932189941, |
|
"learning_rate": 2.52400548696845e-05, |
|
"loss": 0.7315, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 16.3914373088685, |
|
"grad_norm": 12.252084732055664, |
|
"learning_rate": 2.501143118427069e-05, |
|
"loss": 0.7892, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 16.513761467889907, |
|
"grad_norm": 9.307701110839844, |
|
"learning_rate": 2.4782807498856885e-05, |
|
"loss": 0.7698, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 16.636085626911314, |
|
"grad_norm": 10.571149826049805, |
|
"learning_rate": 2.4554183813443073e-05, |
|
"loss": 0.7547, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 16.75840978593272, |
|
"grad_norm": 11.456555366516113, |
|
"learning_rate": 2.4325560128029264e-05, |
|
"loss": 0.7723, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 16.88073394495413, |
|
"grad_norm": 11.944501876831055, |
|
"learning_rate": 2.4096936442615455e-05, |
|
"loss": 0.7614, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 16.990825688073393, |
|
"eval_accuracy": 0.6359447004608295, |
|
"eval_f1": 0.6305311465399625, |
|
"eval_loss": 0.9012994170188904, |
|
"eval_precision": 0.6433955302551128, |
|
"eval_recall": 0.6359447004608295, |
|
"eval_runtime": 2.1782, |
|
"eval_samples_per_second": 99.624, |
|
"eval_steps_per_second": 6.427, |
|
"step": 1389 |
|
}, |
|
{ |
|
"epoch": 17.003058103975537, |
|
"grad_norm": 11.058562278747559, |
|
"learning_rate": 2.3868312757201647e-05, |
|
"loss": 0.7566, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 17.125382262996943, |
|
"grad_norm": 12.137784004211426, |
|
"learning_rate": 2.3639689071787838e-05, |
|
"loss": 0.6842, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 17.24770642201835, |
|
"grad_norm": 12.046640396118164, |
|
"learning_rate": 2.341106538637403e-05, |
|
"loss": 0.7126, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 17.370030581039757, |
|
"grad_norm": 13.35624885559082, |
|
"learning_rate": 2.318244170096022e-05, |
|
"loss": 0.721, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 17.492354740061163, |
|
"grad_norm": 10.566619873046875, |
|
"learning_rate": 2.2953818015546412e-05, |
|
"loss": 0.7424, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 17.61467889908257, |
|
"grad_norm": 11.238773345947266, |
|
"learning_rate": 2.2725194330132603e-05, |
|
"loss": 0.7375, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 17.737003058103976, |
|
"grad_norm": 11.583772659301758, |
|
"learning_rate": 2.2496570644718794e-05, |
|
"loss": 0.7177, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 17.859327217125383, |
|
"grad_norm": 11.983872413635254, |
|
"learning_rate": 2.2267946959304986e-05, |
|
"loss": 0.7393, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 17.98165137614679, |
|
"grad_norm": 9.852224349975586, |
|
"learning_rate": 2.2039323273891177e-05, |
|
"loss": 0.725, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 17.99388379204893, |
|
"eval_accuracy": 0.5990783410138248, |
|
"eval_f1": 0.5975030533834751, |
|
"eval_loss": 0.9701642394065857, |
|
"eval_precision": 0.6072043560801691, |
|
"eval_recall": 0.5990783410138248, |
|
"eval_runtime": 2.1744, |
|
"eval_samples_per_second": 99.798, |
|
"eval_steps_per_second": 6.439, |
|
"step": 1471 |
|
}, |
|
{ |
|
"epoch": 18.103975535168196, |
|
"grad_norm": 10.969733238220215, |
|
"learning_rate": 2.1810699588477368e-05, |
|
"loss": 0.6728, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 18.226299694189603, |
|
"grad_norm": 11.671346664428711, |
|
"learning_rate": 2.158207590306356e-05, |
|
"loss": 0.6974, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 18.34862385321101, |
|
"grad_norm": 8.975910186767578, |
|
"learning_rate": 2.135345221764975e-05, |
|
"loss": 0.6503, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 18.470948012232416, |
|
"grad_norm": 13.026418685913086, |
|
"learning_rate": 2.1124828532235942e-05, |
|
"loss": 0.6879, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 18.593272171253822, |
|
"grad_norm": 15.196861267089844, |
|
"learning_rate": 2.089620484682213e-05, |
|
"loss": 0.6912, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 18.71559633027523, |
|
"grad_norm": 8.060693740844727, |
|
"learning_rate": 2.066758116140832e-05, |
|
"loss": 0.7287, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 18.837920489296636, |
|
"grad_norm": 10.259760856628418, |
|
"learning_rate": 2.0438957475994513e-05, |
|
"loss": 0.6657, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 18.960244648318042, |
|
"grad_norm": 10.083440780639648, |
|
"learning_rate": 2.0210333790580704e-05, |
|
"loss": 0.6938, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 18.996941896024463, |
|
"eval_accuracy": 0.6728110599078341, |
|
"eval_f1": 0.6659537595021466, |
|
"eval_loss": 0.9598256349563599, |
|
"eval_precision": 0.683992662820187, |
|
"eval_recall": 0.6728110599078341, |
|
"eval_runtime": 2.1821, |
|
"eval_samples_per_second": 99.445, |
|
"eval_steps_per_second": 6.416, |
|
"step": 1553 |
|
}, |
|
{ |
|
"epoch": 19.08256880733945, |
|
"grad_norm": 13.321717262268066, |
|
"learning_rate": 1.9981710105166895e-05, |
|
"loss": 0.6849, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 19.204892966360855, |
|
"grad_norm": 9.200907707214355, |
|
"learning_rate": 1.9753086419753087e-05, |
|
"loss": 0.6397, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 19.327217125382262, |
|
"grad_norm": 12.469816207885742, |
|
"learning_rate": 1.9524462734339278e-05, |
|
"loss": 0.6554, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 19.44954128440367, |
|
"grad_norm": 10.692110061645508, |
|
"learning_rate": 1.929583904892547e-05, |
|
"loss": 0.656, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 19.571865443425075, |
|
"grad_norm": 11.429883003234863, |
|
"learning_rate": 1.906721536351166e-05, |
|
"loss": 0.6959, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 19.69418960244648, |
|
"grad_norm": 10.087126731872559, |
|
"learning_rate": 1.8838591678097852e-05, |
|
"loss": 0.6632, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 19.81651376146789, |
|
"grad_norm": 16.846824645996094, |
|
"learning_rate": 1.8609967992684043e-05, |
|
"loss": 0.6307, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 19.938837920489295, |
|
"grad_norm": 12.659772872924805, |
|
"learning_rate": 1.8381344307270234e-05, |
|
"loss": 0.6761, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_accuracy": 0.6082949308755761, |
|
"eval_f1": 0.6112076266773331, |
|
"eval_loss": 0.9886434078216553, |
|
"eval_precision": 0.6242090052989822, |
|
"eval_recall": 0.6082949308755761, |
|
"eval_runtime": 2.1998, |
|
"eval_samples_per_second": 98.647, |
|
"eval_steps_per_second": 6.364, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 20.061162079510705, |
|
"grad_norm": 11.294504165649414, |
|
"learning_rate": 1.8152720621856426e-05, |
|
"loss": 0.6726, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 20.18348623853211, |
|
"grad_norm": 10.856670379638672, |
|
"learning_rate": 1.7924096936442617e-05, |
|
"loss": 0.626, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 20.30581039755352, |
|
"grad_norm": 13.345258712768555, |
|
"learning_rate": 1.769547325102881e-05, |
|
"loss": 0.6446, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 20.428134556574925, |
|
"grad_norm": 9.459267616271973, |
|
"learning_rate": 1.7466849565615e-05, |
|
"loss": 0.6125, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 20.55045871559633, |
|
"grad_norm": 14.87169361114502, |
|
"learning_rate": 1.723822588020119e-05, |
|
"loss": 0.575, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 20.672782874617738, |
|
"grad_norm": 14.521281242370605, |
|
"learning_rate": 1.700960219478738e-05, |
|
"loss": 0.6999, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 20.795107033639145, |
|
"grad_norm": 12.60116195678711, |
|
"learning_rate": 1.678097850937357e-05, |
|
"loss": 0.5862, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 20.91743119266055, |
|
"grad_norm": 10.943224906921387, |
|
"learning_rate": 1.655235482395976e-05, |
|
"loss": 0.5865, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 20.990825688073393, |
|
"eval_accuracy": 0.6497695852534562, |
|
"eval_f1": 0.6428003832696357, |
|
"eval_loss": 0.9367409348487854, |
|
"eval_precision": 0.6431866677852993, |
|
"eval_recall": 0.6497695852534562, |
|
"eval_runtime": 2.1726, |
|
"eval_samples_per_second": 99.881, |
|
"eval_steps_per_second": 6.444, |
|
"step": 1716 |
|
}, |
|
{ |
|
"epoch": 21.039755351681958, |
|
"grad_norm": 11.365513801574707, |
|
"learning_rate": 1.6323731138545953e-05, |
|
"loss": 0.6488, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 21.162079510703364, |
|
"grad_norm": 8.919206619262695, |
|
"learning_rate": 1.6095107453132147e-05, |
|
"loss": 0.6201, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 21.28440366972477, |
|
"grad_norm": 12.377992630004883, |
|
"learning_rate": 1.586648376771834e-05, |
|
"loss": 0.5806, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 21.406727828746178, |
|
"grad_norm": 12.146440505981445, |
|
"learning_rate": 1.563786008230453e-05, |
|
"loss": 0.5506, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 21.529051987767584, |
|
"grad_norm": 11.200637817382812, |
|
"learning_rate": 1.5409236396890718e-05, |
|
"loss": 0.5667, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 21.65137614678899, |
|
"grad_norm": 14.65993595123291, |
|
"learning_rate": 1.5180612711476911e-05, |
|
"loss": 0.5936, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 21.773700305810397, |
|
"grad_norm": 14.228142738342285, |
|
"learning_rate": 1.49519890260631e-05, |
|
"loss": 0.5791, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 21.896024464831804, |
|
"grad_norm": 12.14127254486084, |
|
"learning_rate": 1.4723365340649292e-05, |
|
"loss": 0.5857, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 21.99388379204893, |
|
"eval_accuracy": 0.631336405529954, |
|
"eval_f1": 0.6321523990020061, |
|
"eval_loss": 0.9693921208381653, |
|
"eval_precision": 0.6331329862312, |
|
"eval_recall": 0.631336405529954, |
|
"eval_runtime": 2.1772, |
|
"eval_samples_per_second": 99.669, |
|
"eval_steps_per_second": 6.43, |
|
"step": 1798 |
|
}, |
|
{ |
|
"epoch": 22.01834862385321, |
|
"grad_norm": 8.9346342086792, |
|
"learning_rate": 1.4494741655235483e-05, |
|
"loss": 0.553, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 22.140672782874617, |
|
"grad_norm": 15.74846363067627, |
|
"learning_rate": 1.4266117969821674e-05, |
|
"loss": 0.5693, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 22.262996941896024, |
|
"grad_norm": 17.549697875976562, |
|
"learning_rate": 1.4037494284407866e-05, |
|
"loss": 0.5955, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 22.38532110091743, |
|
"grad_norm": 12.935697555541992, |
|
"learning_rate": 1.3808870598994055e-05, |
|
"loss": 0.5184, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 22.507645259938837, |
|
"grad_norm": 14.531293869018555, |
|
"learning_rate": 1.3580246913580247e-05, |
|
"loss": 0.5873, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 22.629969418960243, |
|
"grad_norm": 11.71330451965332, |
|
"learning_rate": 1.3351623228166438e-05, |
|
"loss": 0.538, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 22.75229357798165, |
|
"grad_norm": 11.269133567810059, |
|
"learning_rate": 1.312299954275263e-05, |
|
"loss": 0.5819, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 22.874617737003057, |
|
"grad_norm": 9.762799263000488, |
|
"learning_rate": 1.2894375857338819e-05, |
|
"loss": 0.5612, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 22.996941896024463, |
|
"grad_norm": 9.858174324035645, |
|
"learning_rate": 1.2665752171925014e-05, |
|
"loss": 0.556, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 22.996941896024463, |
|
"eval_accuracy": 0.6359447004608295, |
|
"eval_f1": 0.629638753332476, |
|
"eval_loss": 1.021164059638977, |
|
"eval_precision": 0.6574294916995147, |
|
"eval_recall": 0.6359447004608295, |
|
"eval_runtime": 2.1738, |
|
"eval_samples_per_second": 99.824, |
|
"eval_steps_per_second": 6.44, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 23.119266055045873, |
|
"grad_norm": 18.760498046875, |
|
"learning_rate": 1.2437128486511203e-05, |
|
"loss": 0.5428, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 23.24159021406728, |
|
"grad_norm": 12.281718254089355, |
|
"learning_rate": 1.2208504801097394e-05, |
|
"loss": 0.5557, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 23.363914373088686, |
|
"grad_norm": 11.702547073364258, |
|
"learning_rate": 1.1979881115683586e-05, |
|
"loss": 0.547, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 23.486238532110093, |
|
"grad_norm": 9.12270450592041, |
|
"learning_rate": 1.1751257430269777e-05, |
|
"loss": 0.4957, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 23.6085626911315, |
|
"grad_norm": 12.279609680175781, |
|
"learning_rate": 1.1522633744855968e-05, |
|
"loss": 0.5025, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 23.730886850152906, |
|
"grad_norm": 12.506765365600586, |
|
"learning_rate": 1.129401005944216e-05, |
|
"loss": 0.5863, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 23.853211009174313, |
|
"grad_norm": 13.289384841918945, |
|
"learning_rate": 1.106538637402835e-05, |
|
"loss": 0.5533, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 23.97553516819572, |
|
"grad_norm": 12.350972175598145, |
|
"learning_rate": 1.083676268861454e-05, |
|
"loss": 0.4871, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_accuracy": 0.5944700460829493, |
|
"eval_f1": 0.587885178704199, |
|
"eval_loss": 1.032782793045044, |
|
"eval_precision": 0.5950577616931705, |
|
"eval_recall": 0.5944700460829493, |
|
"eval_runtime": 2.183, |
|
"eval_samples_per_second": 99.404, |
|
"eval_steps_per_second": 6.413, |
|
"step": 1962 |
|
}, |
|
{ |
|
"epoch": 24.097859327217126, |
|
"grad_norm": 9.447975158691406, |
|
"learning_rate": 1.0608139003200732e-05, |
|
"loss": 0.4881, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 24.220183486238533, |
|
"grad_norm": 11.92766284942627, |
|
"learning_rate": 1.0379515317786923e-05, |
|
"loss": 0.5066, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 24.34250764525994, |
|
"grad_norm": 11.550183296203613, |
|
"learning_rate": 1.0150891632373114e-05, |
|
"loss": 0.4547, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 24.464831804281346, |
|
"grad_norm": 10.570334434509277, |
|
"learning_rate": 9.922267946959306e-06, |
|
"loss": 0.504, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 24.587155963302752, |
|
"grad_norm": 12.56131362915039, |
|
"learning_rate": 9.693644261545497e-06, |
|
"loss": 0.4765, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 24.70948012232416, |
|
"grad_norm": 10.12960433959961, |
|
"learning_rate": 9.465020576131688e-06, |
|
"loss": 0.4862, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 24.831804281345565, |
|
"grad_norm": 14.820987701416016, |
|
"learning_rate": 9.236396890717878e-06, |
|
"loss": 0.4748, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 24.954128440366972, |
|
"grad_norm": 13.520421981811523, |
|
"learning_rate": 9.00777320530407e-06, |
|
"loss": 0.5254, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 24.990825688073393, |
|
"eval_accuracy": 0.5944700460829493, |
|
"eval_f1": 0.5917457657549822, |
|
"eval_loss": 1.0132023096084595, |
|
"eval_precision": 0.5968091048736209, |
|
"eval_recall": 0.5944700460829493, |
|
"eval_runtime": 2.1697, |
|
"eval_samples_per_second": 100.012, |
|
"eval_steps_per_second": 6.452, |
|
"step": 2043 |
|
}, |
|
{ |
|
"epoch": 25.07645259938838, |
|
"grad_norm": 11.470428466796875, |
|
"learning_rate": 8.77914951989026e-06, |
|
"loss": 0.4879, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 25.198776758409785, |
|
"grad_norm": 12.04905891418457, |
|
"learning_rate": 8.550525834476454e-06, |
|
"loss": 0.4719, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 25.321100917431192, |
|
"grad_norm": 14.935491561889648, |
|
"learning_rate": 8.321902149062643e-06, |
|
"loss": 0.4989, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 25.4434250764526, |
|
"grad_norm": 13.884613037109375, |
|
"learning_rate": 8.093278463648834e-06, |
|
"loss": 0.4718, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 25.565749235474005, |
|
"grad_norm": 14.0990629196167, |
|
"learning_rate": 7.864654778235026e-06, |
|
"loss": 0.4833, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 25.68807339449541, |
|
"grad_norm": 15.498763084411621, |
|
"learning_rate": 7.636031092821217e-06, |
|
"loss": 0.4719, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 25.810397553516818, |
|
"grad_norm": 17.733688354492188, |
|
"learning_rate": 7.4074074074074075e-06, |
|
"loss": 0.5214, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 25.932721712538225, |
|
"grad_norm": 12.92156982421875, |
|
"learning_rate": 7.178783721993598e-06, |
|
"loss": 0.5054, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 25.99388379204893, |
|
"eval_accuracy": 0.5944700460829493, |
|
"eval_f1": 0.5943911153011756, |
|
"eval_loss": 1.038478136062622, |
|
"eval_precision": 0.5987861045709855, |
|
"eval_recall": 0.5944700460829493, |
|
"eval_runtime": 2.2069, |
|
"eval_samples_per_second": 98.326, |
|
"eval_steps_per_second": 6.344, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 26.05504587155963, |
|
"grad_norm": 12.785786628723145, |
|
"learning_rate": 6.950160036579791e-06, |
|
"loss": 0.5009, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 26.17737003058104, |
|
"grad_norm": 12.252779006958008, |
|
"learning_rate": 6.721536351165981e-06, |
|
"loss": 0.4896, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 26.299694189602448, |
|
"grad_norm": 8.970731735229492, |
|
"learning_rate": 6.492912665752173e-06, |
|
"loss": 0.4409, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 26.422018348623855, |
|
"grad_norm": 14.119239807128906, |
|
"learning_rate": 6.264288980338363e-06, |
|
"loss": 0.4685, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 26.54434250764526, |
|
"grad_norm": 13.817703247070312, |
|
"learning_rate": 6.0356652949245544e-06, |
|
"loss": 0.423, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 26.666666666666668, |
|
"grad_norm": 11.813767433166504, |
|
"learning_rate": 5.807041609510746e-06, |
|
"loss": 0.4338, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 26.788990825688074, |
|
"grad_norm": 10.11328411102295, |
|
"learning_rate": 5.578417924096937e-06, |
|
"loss": 0.4317, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 26.91131498470948, |
|
"grad_norm": 12.501155853271484, |
|
"learning_rate": 5.3497942386831275e-06, |
|
"loss": 0.4706, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 26.996941896024463, |
|
"eval_accuracy": 0.6036866359447005, |
|
"eval_f1": 0.5983142033282286, |
|
"eval_loss": 1.0625784397125244, |
|
"eval_precision": 0.6100467719653185, |
|
"eval_recall": 0.6036866359447005, |
|
"eval_runtime": 2.1672, |
|
"eval_samples_per_second": 100.128, |
|
"eval_steps_per_second": 6.46, |
|
"step": 2207 |
|
}, |
|
{ |
|
"epoch": 27.033639143730888, |
|
"grad_norm": 12.064492225646973, |
|
"learning_rate": 5.121170553269319e-06, |
|
"loss": 0.4762, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 27.155963302752294, |
|
"grad_norm": 12.347169876098633, |
|
"learning_rate": 4.89254686785551e-06, |
|
"loss": 0.4493, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 27.2782874617737, |
|
"grad_norm": 12.065176010131836, |
|
"learning_rate": 4.663923182441701e-06, |
|
"loss": 0.4341, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 27.400611620795107, |
|
"grad_norm": 12.744647979736328, |
|
"learning_rate": 4.435299497027892e-06, |
|
"loss": 0.4373, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 27.522935779816514, |
|
"grad_norm": 14.091771125793457, |
|
"learning_rate": 4.206675811614083e-06, |
|
"loss": 0.4286, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 27.64525993883792, |
|
"grad_norm": 11.384113311767578, |
|
"learning_rate": 3.9780521262002744e-06, |
|
"loss": 0.4338, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 27.767584097859327, |
|
"grad_norm": 12.185340881347656, |
|
"learning_rate": 3.7494284407864657e-06, |
|
"loss": 0.4482, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 27.889908256880734, |
|
"grad_norm": 10.328293800354004, |
|
"learning_rate": 3.5208047553726566e-06, |
|
"loss": 0.418, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_accuracy": 0.5806451612903226, |
|
"eval_f1": 0.5774461884155286, |
|
"eval_loss": 1.053132176399231, |
|
"eval_precision": 0.5829576016453093, |
|
"eval_recall": 0.5806451612903226, |
|
"eval_runtime": 2.1931, |
|
"eval_samples_per_second": 98.948, |
|
"eval_steps_per_second": 6.384, |
|
"step": 2289 |
|
}, |
|
{ |
|
"epoch": 28.01223241590214, |
|
"grad_norm": 13.196370124816895, |
|
"learning_rate": 3.2921810699588483e-06, |
|
"loss": 0.4523, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 28.134556574923547, |
|
"grad_norm": 12.735855102539062, |
|
"learning_rate": 3.063557384545039e-06, |
|
"loss": 0.3987, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 28.256880733944953, |
|
"grad_norm": 13.810506820678711, |
|
"learning_rate": 2.83493369913123e-06, |
|
"loss": 0.4244, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 28.37920489296636, |
|
"grad_norm": 14.41588306427002, |
|
"learning_rate": 2.6063100137174214e-06, |
|
"loss": 0.4746, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 28.501529051987767, |
|
"grad_norm": 13.99752426147461, |
|
"learning_rate": 2.3776863283036123e-06, |
|
"loss": 0.4344, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 28.623853211009173, |
|
"grad_norm": 13.782439231872559, |
|
"learning_rate": 2.1490626428898036e-06, |
|
"loss": 0.4018, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 28.74617737003058, |
|
"grad_norm": 10.79996109008789, |
|
"learning_rate": 1.920438957475995e-06, |
|
"loss": 0.3779, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 28.868501529051986, |
|
"grad_norm": 12.254060745239258, |
|
"learning_rate": 1.6918152720621857e-06, |
|
"loss": 0.4671, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 28.990825688073393, |
|
"grad_norm": 12.211686134338379, |
|
"learning_rate": 1.4631915866483768e-06, |
|
"loss": 0.455, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 28.990825688073393, |
|
"eval_accuracy": 0.6082949308755761, |
|
"eval_f1": 0.6039494237985392, |
|
"eval_loss": 1.0340049266815186, |
|
"eval_precision": 0.615058251488602, |
|
"eval_recall": 0.6082949308755761, |
|
"eval_runtime": 2.21, |
|
"eval_samples_per_second": 98.191, |
|
"eval_steps_per_second": 6.335, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 29.1131498470948, |
|
"grad_norm": 12.376998901367188, |
|
"learning_rate": 1.234567901234568e-06, |
|
"loss": 0.405, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 29.235474006116206, |
|
"grad_norm": 13.317730903625488, |
|
"learning_rate": 1.0059442158207592e-06, |
|
"loss": 0.3908, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 29.357798165137616, |
|
"grad_norm": 14.004073143005371, |
|
"learning_rate": 7.773205304069502e-07, |
|
"loss": 0.4006, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 29.480122324159023, |
|
"grad_norm": 11.803933143615723, |
|
"learning_rate": 5.486968449931413e-07, |
|
"loss": 0.4255, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 29.60244648318043, |
|
"grad_norm": 13.087397575378418, |
|
"learning_rate": 3.200731595793324e-07, |
|
"loss": 0.4354, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 29.724770642201836, |
|
"grad_norm": 11.753482818603516, |
|
"learning_rate": 9.144947416552355e-08, |
|
"loss": 0.4414, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 29.724770642201836, |
|
"eval_accuracy": 0.6129032258064516, |
|
"eval_f1": 0.6087354657244137, |
|
"eval_loss": 1.0280050039291382, |
|
"eval_precision": 0.6161055592910982, |
|
"eval_recall": 0.6129032258064516, |
|
"eval_runtime": 2.2362, |
|
"eval_samples_per_second": 97.04, |
|
"eval_steps_per_second": 6.261, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 29.724770642201836, |
|
"step": 2430, |
|
"total_flos": 1.2301888231484006e+19, |
|
"train_loss": 0.8217329954414211, |
|
"train_runtime": 6060.3586, |
|
"train_samples_per_second": 25.865, |
|
"train_steps_per_second": 0.401 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2430, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 30, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.2301888231484006e+19, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|