|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9000969932104753, |
|
"eval_steps": 557, |
|
"global_step": 5568, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.015033947623666343, |
|
"grad_norm": 63.938453674316406, |
|
"learning_rate": 4.741379310344828e-07, |
|
"loss": 11.7791, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.030067895247332686, |
|
"grad_norm": 28.60324478149414, |
|
"learning_rate": 9.752155172413793e-07, |
|
"loss": 10.5226, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.04510184287099903, |
|
"grad_norm": 17.869054794311523, |
|
"learning_rate": 1.476293103448276e-06, |
|
"loss": 9.3586, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.06013579049466537, |
|
"grad_norm": 18.835189819335938, |
|
"learning_rate": 1.977370689655172e-06, |
|
"loss": 8.4548, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.07516973811833172, |
|
"grad_norm": 55.06333541870117, |
|
"learning_rate": 2.4784482758620692e-06, |
|
"loss": 8.3449, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_nli-pairs_loss": 7.450161933898926, |
|
"eval_nli-pairs_runtime": 14.8176, |
|
"eval_nli-pairs_samples_per_second": 101.231, |
|
"eval_nli-pairs_steps_per_second": 3.172, |
|
"eval_sts-test_pearson_cosine": 0.3696416595298566, |
|
"eval_sts-test_pearson_dot": 0.13300273461809956, |
|
"eval_sts-test_pearson_euclidean": 0.3836877008752716, |
|
"eval_sts-test_pearson_manhattan": 0.41477338080789633, |
|
"eval_sts-test_pearson_max": 0.41477338080789633, |
|
"eval_sts-test_spearman_cosine": 0.3694134524358256, |
|
"eval_sts-test_spearman_dot": 0.12058818576425179, |
|
"eval_sts-test_spearman_euclidean": 0.38615620458370276, |
|
"eval_sts-test_spearman_manhattan": 0.4121304311152658, |
|
"eval_sts-test_spearman_max": 0.4121304311152658, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_vitaminc-pairs_loss": 5.403733730316162, |
|
"eval_vitaminc-pairs_runtime": 14.4427, |
|
"eval_vitaminc-pairs_samples_per_second": 69.17, |
|
"eval_vitaminc-pairs_steps_per_second": 2.216, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_qnli-contrastive_loss": 8.863805770874023, |
|
"eval_qnli-contrastive_runtime": 4.835, |
|
"eval_qnli-contrastive_samples_per_second": 310.238, |
|
"eval_qnli-contrastive_steps_per_second": 9.721, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_scitail-pairs-qa_loss": 4.2781147956848145, |
|
"eval_scitail-pairs-qa_runtime": 5.761, |
|
"eval_scitail-pairs-qa_samples_per_second": 130.186, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.166, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_scitail-pairs-pos_loss": 4.930431365966797, |
|
"eval_scitail-pairs-pos_runtime": 15.2161, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.699, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.695, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_xsum-pairs_loss": 4.778657913208008, |
|
"eval_xsum-pairs_runtime": 3.0397, |
|
"eval_xsum-pairs_samples_per_second": 65.795, |
|
"eval_xsum-pairs_steps_per_second": 2.303, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_compression-pairs_loss": 3.489774465560913, |
|
"eval_compression-pairs_runtime": 1.2758, |
|
"eval_compression-pairs_samples_per_second": 391.924, |
|
"eval_compression-pairs_steps_per_second": 12.542, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_sciq_pairs_loss": 7.707857131958008, |
|
"eval_sciq_pairs_runtime": 28.932, |
|
"eval_sciq_pairs_samples_per_second": 20.185, |
|
"eval_sciq_pairs_steps_per_second": 0.657, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_qasc_pairs_loss": 12.320230484008789, |
|
"eval_qasc_pairs_runtime": 5.2561, |
|
"eval_qasc_pairs_samples_per_second": 77.434, |
|
"eval_qasc_pairs_steps_per_second": 2.473, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_openbookqa_pairs_loss": 7.607065677642822, |
|
"eval_openbookqa_pairs_runtime": 1.189, |
|
"eval_openbookqa_pairs_samples_per_second": 115.224, |
|
"eval_openbookqa_pairs_steps_per_second": 4.205, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_msmarco_pairs_loss": 8.519609451293945, |
|
"eval_msmarco_pairs_runtime": 19.7371, |
|
"eval_msmarco_pairs_samples_per_second": 50.666, |
|
"eval_msmarco_pairs_steps_per_second": 1.621, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_nq_pairs_loss": 8.43966007232666, |
|
"eval_nq_pairs_runtime": 37.2943, |
|
"eval_nq_pairs_samples_per_second": 26.814, |
|
"eval_nq_pairs_steps_per_second": 0.858, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_trivia_pairs_loss": 8.689275741577148, |
|
"eval_trivia_pairs_runtime": 64.6902, |
|
"eval_trivia_pairs_samples_per_second": 15.458, |
|
"eval_trivia_pairs_steps_per_second": 0.495, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_quora_pairs_loss": 4.54755163192749, |
|
"eval_quora_pairs_runtime": 3.7394, |
|
"eval_quora_pairs_samples_per_second": 133.71, |
|
"eval_quora_pairs_steps_per_second": 4.279, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_gooaq_pairs_loss": 7.57412052154541, |
|
"eval_gooaq_pairs_runtime": 13.3648, |
|
"eval_gooaq_pairs_samples_per_second": 74.824, |
|
"eval_gooaq_pairs_steps_per_second": 2.394, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09020368574199807, |
|
"grad_norm": 21.490495681762695, |
|
"learning_rate": 2.9795258620689654e-06, |
|
"loss": 7.4107, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.1052376333656644, |
|
"grad_norm": 35.23164367675781, |
|
"learning_rate": 3.480603448275862e-06, |
|
"loss": 7.0111, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.12027158098933075, |
|
"grad_norm": 28.60922622680664, |
|
"learning_rate": 3.981681034482759e-06, |
|
"loss": 5.9488, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.13530552861299708, |
|
"grad_norm": 22.0839786529541, |
|
"learning_rate": 4.482758620689656e-06, |
|
"loss": 6.3051, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.15033947623666344, |
|
"grad_norm": 34.64767837524414, |
|
"learning_rate": 4.983836206896552e-06, |
|
"loss": 5.2056, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.16537342386032977, |
|
"grad_norm": 13.605194091796875, |
|
"learning_rate": 5.484913793103448e-06, |
|
"loss": 5.0794, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_nli-pairs_loss": 4.693160057067871, |
|
"eval_nli-pairs_runtime": 14.776, |
|
"eval_nli-pairs_samples_per_second": 101.516, |
|
"eval_nli-pairs_steps_per_second": 3.181, |
|
"eval_sts-test_pearson_cosine": 0.6672572282189017, |
|
"eval_sts-test_pearson_dot": 0.5110207654364226, |
|
"eval_sts-test_pearson_euclidean": 0.65934512843155, |
|
"eval_sts-test_pearson_manhattan": 0.6615633016003417, |
|
"eval_sts-test_pearson_max": 0.6672572282189017, |
|
"eval_sts-test_spearman_cosine": 0.6364241996636473, |
|
"eval_sts-test_spearman_dot": 0.4920702168931027, |
|
"eval_sts-test_spearman_euclidean": 0.638848044048013, |
|
"eval_sts-test_spearman_manhattan": 0.6407402326117282, |
|
"eval_sts-test_spearman_max": 0.6407402326117282, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_vitaminc-pairs_loss": 3.5585296154022217, |
|
"eval_vitaminc-pairs_runtime": 14.3661, |
|
"eval_vitaminc-pairs_samples_per_second": 69.538, |
|
"eval_vitaminc-pairs_steps_per_second": 2.227, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_qnli-contrastive_loss": 11.759393692016602, |
|
"eval_qnli-contrastive_runtime": 4.7771, |
|
"eval_qnli-contrastive_samples_per_second": 313.999, |
|
"eval_qnli-contrastive_steps_per_second": 9.839, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_scitail-pairs-qa_loss": 1.3339837789535522, |
|
"eval_scitail-pairs-qa_runtime": 5.8445, |
|
"eval_scitail-pairs-qa_samples_per_second": 128.326, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.106, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_scitail-pairs-pos_loss": 2.508178234100342, |
|
"eval_scitail-pairs-pos_runtime": 15.12, |
|
"eval_scitail-pairs-pos_samples_per_second": 86.244, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.712, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_xsum-pairs_loss": 2.727797508239746, |
|
"eval_xsum-pairs_runtime": 3.0382, |
|
"eval_xsum-pairs_samples_per_second": 65.828, |
|
"eval_xsum-pairs_steps_per_second": 2.304, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_compression-pairs_loss": 1.769970178604126, |
|
"eval_compression-pairs_runtime": 1.3017, |
|
"eval_compression-pairs_samples_per_second": 384.105, |
|
"eval_compression-pairs_steps_per_second": 12.291, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_sciq_pairs_loss": 6.772756576538086, |
|
"eval_sciq_pairs_runtime": 29.0556, |
|
"eval_sciq_pairs_samples_per_second": 20.099, |
|
"eval_sciq_pairs_steps_per_second": 0.654, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_qasc_pairs_loss": 10.429397583007812, |
|
"eval_qasc_pairs_runtime": 5.2915, |
|
"eval_qasc_pairs_samples_per_second": 76.916, |
|
"eval_qasc_pairs_steps_per_second": 2.457, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_openbookqa_pairs_loss": 5.441987037658691, |
|
"eval_openbookqa_pairs_runtime": 1.1725, |
|
"eval_openbookqa_pairs_samples_per_second": 116.843, |
|
"eval_openbookqa_pairs_steps_per_second": 4.264, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_msmarco_pairs_loss": 4.586305618286133, |
|
"eval_msmarco_pairs_runtime": 19.8051, |
|
"eval_msmarco_pairs_samples_per_second": 50.492, |
|
"eval_msmarco_pairs_steps_per_second": 1.616, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_nq_pairs_loss": 5.193580150604248, |
|
"eval_nq_pairs_runtime": 37.2293, |
|
"eval_nq_pairs_samples_per_second": 26.861, |
|
"eval_nq_pairs_steps_per_second": 0.86, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_trivia_pairs_loss": 5.120319366455078, |
|
"eval_trivia_pairs_runtime": 64.3021, |
|
"eval_trivia_pairs_samples_per_second": 15.552, |
|
"eval_trivia_pairs_steps_per_second": 0.498, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_quora_pairs_loss": 1.4382946491241455, |
|
"eval_quora_pairs_runtime": 3.7168, |
|
"eval_quora_pairs_samples_per_second": 134.525, |
|
"eval_quora_pairs_steps_per_second": 4.305, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_gooaq_pairs_loss": 3.7513082027435303, |
|
"eval_gooaq_pairs_runtime": 13.2594, |
|
"eval_gooaq_pairs_samples_per_second": 75.418, |
|
"eval_gooaq_pairs_steps_per_second": 2.413, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.18040737148399613, |
|
"grad_norm": 39.661067962646484, |
|
"learning_rate": 5.9859913793103445e-06, |
|
"loss": 4.362, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 0.19544131910766246, |
|
"grad_norm": 34.55929946899414, |
|
"learning_rate": 6.487068965517242e-06, |
|
"loss": 4.3245, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 0.2104752667313288, |
|
"grad_norm": 25.775623321533203, |
|
"learning_rate": 6.988146551724138e-06, |
|
"loss": 4.359, |
|
"step": 1302 |
|
}, |
|
{ |
|
"epoch": 0.22550921435499516, |
|
"grad_norm": 57.751529693603516, |
|
"learning_rate": 7.489224137931035e-06, |
|
"loss": 4.0488, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.2405431619786615, |
|
"grad_norm": 33.448768615722656, |
|
"learning_rate": 7.990301724137931e-06, |
|
"loss": 4.0499, |
|
"step": 1488 |
|
}, |
|
{ |
|
"epoch": 0.2555771096023278, |
|
"grad_norm": 29.5295467376709, |
|
"learning_rate": 8.491379310344827e-06, |
|
"loss": 3.7256, |
|
"step": 1581 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_nli-pairs_loss": 3.6348764896392822, |
|
"eval_nli-pairs_runtime": 14.761, |
|
"eval_nli-pairs_samples_per_second": 101.619, |
|
"eval_nli-pairs_steps_per_second": 3.184, |
|
"eval_sts-test_pearson_cosine": 0.6972871286019413, |
|
"eval_sts-test_pearson_dot": 0.5667394576060958, |
|
"eval_sts-test_pearson_euclidean": 0.6946767725059286, |
|
"eval_sts-test_pearson_manhattan": 0.7031948219809795, |
|
"eval_sts-test_pearson_max": 0.7031948219809795, |
|
"eval_sts-test_spearman_cosine": 0.6666257059052387, |
|
"eval_sts-test_spearman_dot": 0.5430589355737897, |
|
"eval_sts-test_spearman_euclidean": 0.6727104999188346, |
|
"eval_sts-test_spearman_manhattan": 0.6812486382298032, |
|
"eval_sts-test_spearman_max": 0.6812486382298032, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_vitaminc-pairs_loss": 2.811436653137207, |
|
"eval_vitaminc-pairs_runtime": 14.4779, |
|
"eval_vitaminc-pairs_samples_per_second": 69.002, |
|
"eval_vitaminc-pairs_steps_per_second": 2.21, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_qnli-contrastive_loss": 9.688385009765625, |
|
"eval_qnli-contrastive_runtime": 4.8234, |
|
"eval_qnli-contrastive_samples_per_second": 310.984, |
|
"eval_qnli-contrastive_steps_per_second": 9.744, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_scitail-pairs-qa_loss": 0.8966398239135742, |
|
"eval_scitail-pairs-qa_runtime": 5.8274, |
|
"eval_scitail-pairs-qa_samples_per_second": 128.702, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.118, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_scitail-pairs-pos_loss": 1.9770935773849487, |
|
"eval_scitail-pairs-pos_runtime": 15.5498, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.859, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.637, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_xsum-pairs_loss": 2.1290199756622314, |
|
"eval_xsum-pairs_runtime": 3.0544, |
|
"eval_xsum-pairs_samples_per_second": 65.479, |
|
"eval_xsum-pairs_steps_per_second": 2.292, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_compression-pairs_loss": 1.2427340745925903, |
|
"eval_compression-pairs_runtime": 1.3179, |
|
"eval_compression-pairs_samples_per_second": 379.389, |
|
"eval_compression-pairs_steps_per_second": 12.14, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_sciq_pairs_loss": 6.424961090087891, |
|
"eval_sciq_pairs_runtime": 29.2749, |
|
"eval_sciq_pairs_samples_per_second": 19.949, |
|
"eval_sciq_pairs_steps_per_second": 0.649, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_qasc_pairs_loss": 9.239206314086914, |
|
"eval_qasc_pairs_runtime": 5.2408, |
|
"eval_qasc_pairs_samples_per_second": 77.66, |
|
"eval_qasc_pairs_steps_per_second": 2.481, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_openbookqa_pairs_loss": 4.784058570861816, |
|
"eval_openbookqa_pairs_runtime": 1.2088, |
|
"eval_openbookqa_pairs_samples_per_second": 113.332, |
|
"eval_openbookqa_pairs_steps_per_second": 4.136, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_msmarco_pairs_loss": 3.655301094055176, |
|
"eval_msmarco_pairs_runtime": 19.7273, |
|
"eval_msmarco_pairs_samples_per_second": 50.691, |
|
"eval_msmarco_pairs_steps_per_second": 1.622, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_nq_pairs_loss": 4.114762783050537, |
|
"eval_nq_pairs_runtime": 37.3924, |
|
"eval_nq_pairs_samples_per_second": 26.743, |
|
"eval_nq_pairs_steps_per_second": 0.856, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_trivia_pairs_loss": 4.019989967346191, |
|
"eval_trivia_pairs_runtime": 64.3836, |
|
"eval_trivia_pairs_samples_per_second": 15.532, |
|
"eval_trivia_pairs_steps_per_second": 0.497, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_quora_pairs_loss": 1.002946138381958, |
|
"eval_quora_pairs_runtime": 3.7091, |
|
"eval_quora_pairs_samples_per_second": 134.803, |
|
"eval_quora_pairs_steps_per_second": 4.314, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_gooaq_pairs_loss": 2.87351131439209, |
|
"eval_gooaq_pairs_runtime": 13.2151, |
|
"eval_gooaq_pairs_samples_per_second": 75.671, |
|
"eval_gooaq_pairs_steps_per_second": 2.421, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27061105722599416, |
|
"grad_norm": 142.8954315185547, |
|
"learning_rate": 8.992456896551725e-06, |
|
"loss": 3.6978, |
|
"step": 1674 |
|
}, |
|
{ |
|
"epoch": 0.28564500484966054, |
|
"grad_norm": 35.244972229003906, |
|
"learning_rate": 9.493534482758622e-06, |
|
"loss": 3.9547, |
|
"step": 1767 |
|
}, |
|
{ |
|
"epoch": 0.3006789524733269, |
|
"grad_norm": 24.055269241333008, |
|
"learning_rate": 9.994612068965518e-06, |
|
"loss": 4.162, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.3157129000969932, |
|
"grad_norm": 39.16152572631836, |
|
"learning_rate": 1.0495689655172414e-05, |
|
"loss": 3.739, |
|
"step": 1953 |
|
}, |
|
{ |
|
"epoch": 0.33074684772065954, |
|
"grad_norm": 30.219701766967773, |
|
"learning_rate": 1.099676724137931e-05, |
|
"loss": 4.2688, |
|
"step": 2046 |
|
}, |
|
{ |
|
"epoch": 0.3457807953443259, |
|
"grad_norm": 35.78736114501953, |
|
"learning_rate": 1.1497844827586206e-05, |
|
"loss": 3.3462, |
|
"step": 2139 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_nli-pairs_loss": 3.1179466247558594, |
|
"eval_nli-pairs_runtime": 14.8963, |
|
"eval_nli-pairs_samples_per_second": 100.696, |
|
"eval_nli-pairs_steps_per_second": 3.155, |
|
"eval_sts-test_pearson_cosine": 0.7110231765083281, |
|
"eval_sts-test_pearson_dot": 0.5911922048259769, |
|
"eval_sts-test_pearson_euclidean": 0.7113733036920562, |
|
"eval_sts-test_pearson_manhattan": 0.7195266505149516, |
|
"eval_sts-test_pearson_max": 0.7195266505149516, |
|
"eval_sts-test_spearman_cosine": 0.6844829069963456, |
|
"eval_sts-test_spearman_dot": 0.567786285795314, |
|
"eval_sts-test_spearman_euclidean": 0.6897367727395519, |
|
"eval_sts-test_spearman_manhattan": 0.6977703060020186, |
|
"eval_sts-test_spearman_max": 0.6977703060020186, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_vitaminc-pairs_loss": 2.4076764583587646, |
|
"eval_vitaminc-pairs_runtime": 14.3487, |
|
"eval_vitaminc-pairs_samples_per_second": 69.623, |
|
"eval_vitaminc-pairs_steps_per_second": 2.23, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_qnli-contrastive_loss": 6.651412010192871, |
|
"eval_qnli-contrastive_runtime": 4.759, |
|
"eval_qnli-contrastive_samples_per_second": 315.189, |
|
"eval_qnli-contrastive_steps_per_second": 9.876, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_scitail-pairs-qa_loss": 0.6640239357948303, |
|
"eval_scitail-pairs-qa_runtime": 5.8178, |
|
"eval_scitail-pairs-qa_samples_per_second": 128.916, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.125, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_scitail-pairs-pos_loss": 1.6727737188339233, |
|
"eval_scitail-pairs-pos_runtime": 15.3144, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.149, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.677, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_xsum-pairs_loss": 1.7385025024414062, |
|
"eval_xsum-pairs_runtime": 3.0465, |
|
"eval_xsum-pairs_samples_per_second": 65.649, |
|
"eval_xsum-pairs_steps_per_second": 2.298, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_compression-pairs_loss": 1.0110334157943726, |
|
"eval_compression-pairs_runtime": 1.2863, |
|
"eval_compression-pairs_samples_per_second": 388.706, |
|
"eval_compression-pairs_steps_per_second": 12.439, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_sciq_pairs_loss": 6.164596080780029, |
|
"eval_sciq_pairs_runtime": 28.909, |
|
"eval_sciq_pairs_samples_per_second": 20.201, |
|
"eval_sciq_pairs_steps_per_second": 0.657, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_qasc_pairs_loss": 7.898203372955322, |
|
"eval_qasc_pairs_runtime": 5.2211, |
|
"eval_qasc_pairs_samples_per_second": 77.953, |
|
"eval_qasc_pairs_steps_per_second": 2.49, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_openbookqa_pairs_loss": 4.36992073059082, |
|
"eval_openbookqa_pairs_runtime": 1.1844, |
|
"eval_openbookqa_pairs_samples_per_second": 115.667, |
|
"eval_openbookqa_pairs_steps_per_second": 4.221, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_msmarco_pairs_loss": 3.067087411880493, |
|
"eval_msmarco_pairs_runtime": 19.7205, |
|
"eval_msmarco_pairs_samples_per_second": 50.709, |
|
"eval_msmarco_pairs_steps_per_second": 1.623, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_nq_pairs_loss": 3.3717246055603027, |
|
"eval_nq_pairs_runtime": 37.1177, |
|
"eval_nq_pairs_samples_per_second": 26.941, |
|
"eval_nq_pairs_steps_per_second": 0.862, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_trivia_pairs_loss": 3.4189703464508057, |
|
"eval_trivia_pairs_runtime": 64.257, |
|
"eval_trivia_pairs_samples_per_second": 15.563, |
|
"eval_trivia_pairs_steps_per_second": 0.498, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_quora_pairs_loss": 0.8419531583786011, |
|
"eval_quora_pairs_runtime": 3.7448, |
|
"eval_quora_pairs_samples_per_second": 133.518, |
|
"eval_quora_pairs_steps_per_second": 4.273, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_gooaq_pairs_loss": 2.4530532360076904, |
|
"eval_gooaq_pairs_runtime": 13.3141, |
|
"eval_gooaq_pairs_samples_per_second": 75.108, |
|
"eval_gooaq_pairs_steps_per_second": 2.403, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.36081474296799226, |
|
"grad_norm": 33.25382614135742, |
|
"learning_rate": 1.1998922413793104e-05, |
|
"loss": 3.8272, |
|
"step": 2232 |
|
}, |
|
{ |
|
"epoch": 0.3758486905916586, |
|
"grad_norm": 22.22249984741211, |
|
"learning_rate": 1.25e-05, |
|
"loss": 3.4532, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.39088263821532493, |
|
"grad_norm": 32.02521514892578, |
|
"learning_rate": 1.3001077586206897e-05, |
|
"loss": 3.7391, |
|
"step": 2418 |
|
}, |
|
{ |
|
"epoch": 0.40591658583899126, |
|
"grad_norm": 84.23457336425781, |
|
"learning_rate": 1.3502155172413793e-05, |
|
"loss": 3.4451, |
|
"step": 2511 |
|
}, |
|
{ |
|
"epoch": 0.4209505334626576, |
|
"grad_norm": 82.23103332519531, |
|
"learning_rate": 1.400323275862069e-05, |
|
"loss": 3.2086, |
|
"step": 2604 |
|
}, |
|
{ |
|
"epoch": 0.435984481086324, |
|
"grad_norm": 128.5042266845703, |
|
"learning_rate": 1.4504310344827587e-05, |
|
"loss": 3.153, |
|
"step": 2697 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_nli-pairs_loss": 2.7247567176818848, |
|
"eval_nli-pairs_runtime": 14.7363, |
|
"eval_nli-pairs_samples_per_second": 101.79, |
|
"eval_nli-pairs_steps_per_second": 3.189, |
|
"eval_sts-test_pearson_cosine": 0.7313032856033084, |
|
"eval_sts-test_pearson_dot": 0.639231944179888, |
|
"eval_sts-test_pearson_euclidean": 0.7299296327875673, |
|
"eval_sts-test_pearson_manhattan": 0.7376485877411852, |
|
"eval_sts-test_pearson_max": 0.7376485877411852, |
|
"eval_sts-test_spearman_cosine": 0.713852235440012, |
|
"eval_sts-test_spearman_dot": 0.6141091584560591, |
|
"eval_sts-test_spearman_euclidean": 0.7125938749269385, |
|
"eval_sts-test_spearman_manhattan": 0.7209089187255475, |
|
"eval_sts-test_spearman_max": 0.7209089187255475, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_vitaminc-pairs_loss": 2.1636030673980713, |
|
"eval_vitaminc-pairs_runtime": 14.3573, |
|
"eval_vitaminc-pairs_samples_per_second": 69.582, |
|
"eval_vitaminc-pairs_steps_per_second": 2.229, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_qnli-contrastive_loss": 6.1082000732421875, |
|
"eval_qnli-contrastive_runtime": 4.8812, |
|
"eval_qnli-contrastive_samples_per_second": 307.299, |
|
"eval_qnli-contrastive_steps_per_second": 9.629, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_scitail-pairs-qa_loss": 0.47009098529815674, |
|
"eval_scitail-pairs-qa_runtime": 5.8135, |
|
"eval_scitail-pairs-qa_samples_per_second": 129.011, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.128, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_scitail-pairs-pos_loss": 1.38162362575531, |
|
"eval_scitail-pairs-pos_runtime": 15.0892, |
|
"eval_scitail-pairs-pos_samples_per_second": 86.42, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.717, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_xsum-pairs_loss": 1.4567902088165283, |
|
"eval_xsum-pairs_runtime": 3.0536, |
|
"eval_xsum-pairs_samples_per_second": 65.496, |
|
"eval_xsum-pairs_steps_per_second": 2.292, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_compression-pairs_loss": 0.8824976682662964, |
|
"eval_compression-pairs_runtime": 1.2907, |
|
"eval_compression-pairs_samples_per_second": 387.382, |
|
"eval_compression-pairs_steps_per_second": 12.396, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_sciq_pairs_loss": 5.971975326538086, |
|
"eval_sciq_pairs_runtime": 28.8816, |
|
"eval_sciq_pairs_samples_per_second": 20.22, |
|
"eval_sciq_pairs_steps_per_second": 0.658, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_qasc_pairs_loss": 7.14296817779541, |
|
"eval_qasc_pairs_runtime": 5.2619, |
|
"eval_qasc_pairs_samples_per_second": 77.348, |
|
"eval_qasc_pairs_steps_per_second": 2.471, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_openbookqa_pairs_loss": 3.890052080154419, |
|
"eval_openbookqa_pairs_runtime": 1.1822, |
|
"eval_openbookqa_pairs_samples_per_second": 115.884, |
|
"eval_openbookqa_pairs_steps_per_second": 4.229, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_msmarco_pairs_loss": 2.6956124305725098, |
|
"eval_msmarco_pairs_runtime": 19.7685, |
|
"eval_msmarco_pairs_samples_per_second": 50.586, |
|
"eval_msmarco_pairs_steps_per_second": 1.619, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_nq_pairs_loss": 2.9799299240112305, |
|
"eval_nq_pairs_runtime": 37.2854, |
|
"eval_nq_pairs_samples_per_second": 26.82, |
|
"eval_nq_pairs_steps_per_second": 0.858, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_trivia_pairs_loss": 3.0525760650634766, |
|
"eval_trivia_pairs_runtime": 64.286, |
|
"eval_trivia_pairs_samples_per_second": 15.555, |
|
"eval_trivia_pairs_steps_per_second": 0.498, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_quora_pairs_loss": 0.7019712924957275, |
|
"eval_quora_pairs_runtime": 3.759, |
|
"eval_quora_pairs_samples_per_second": 133.013, |
|
"eval_quora_pairs_steps_per_second": 4.256, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_gooaq_pairs_loss": 2.1218321323394775, |
|
"eval_gooaq_pairs_runtime": 13.3843, |
|
"eval_gooaq_pairs_samples_per_second": 74.715, |
|
"eval_gooaq_pairs_steps_per_second": 2.391, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.4510184287099903, |
|
"grad_norm": 7.123088836669922, |
|
"learning_rate": 1.5e-05, |
|
"loss": 2.9093, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.46605237633365665, |
|
"grad_norm": 26.165264129638672, |
|
"learning_rate": 1.5501077586206898e-05, |
|
"loss": 2.6223, |
|
"step": 2883 |
|
}, |
|
{ |
|
"epoch": 0.481086323957323, |
|
"grad_norm": 7.543350696563721, |
|
"learning_rate": 1.6002155172413795e-05, |
|
"loss": 2.8362, |
|
"step": 2976 |
|
}, |
|
{ |
|
"epoch": 0.4961202715809893, |
|
"grad_norm": 26.791278839111328, |
|
"learning_rate": 1.6503232758620687e-05, |
|
"loss": 2.6383, |
|
"step": 3069 |
|
}, |
|
{ |
|
"epoch": 0.5111542192046556, |
|
"grad_norm": 24.906707763671875, |
|
"learning_rate": 1.7004310344827587e-05, |
|
"loss": 2.6265, |
|
"step": 3162 |
|
}, |
|
{ |
|
"epoch": 0.526188166828322, |
|
"grad_norm": 17.12710189819336, |
|
"learning_rate": 1.7505387931034483e-05, |
|
"loss": 2.8718, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_nli-pairs_loss": 2.4381556510925293, |
|
"eval_nli-pairs_runtime": 14.7362, |
|
"eval_nli-pairs_samples_per_second": 101.79, |
|
"eval_nli-pairs_steps_per_second": 3.189, |
|
"eval_sts-test_pearson_cosine": 0.7355921455984917, |
|
"eval_sts-test_pearson_dot": 0.6441608711789287, |
|
"eval_sts-test_pearson_euclidean": 0.7293547151744354, |
|
"eval_sts-test_pearson_manhattan": 0.7409243217735156, |
|
"eval_sts-test_pearson_max": 0.7409243217735156, |
|
"eval_sts-test_spearman_cosine": 0.7175266639957124, |
|
"eval_sts-test_spearman_dot": 0.6242096516614428, |
|
"eval_sts-test_spearman_euclidean": 0.7122336979289147, |
|
"eval_sts-test_spearman_manhattan": 0.7250251572437655, |
|
"eval_sts-test_spearman_max": 0.7250251572437655, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_vitaminc-pairs_loss": 1.8586883544921875, |
|
"eval_vitaminc-pairs_runtime": 14.319, |
|
"eval_vitaminc-pairs_samples_per_second": 69.767, |
|
"eval_vitaminc-pairs_steps_per_second": 2.235, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_qnli-contrastive_loss": 3.9992873668670654, |
|
"eval_qnli-contrastive_runtime": 4.8306, |
|
"eval_qnli-contrastive_samples_per_second": 310.523, |
|
"eval_qnli-contrastive_steps_per_second": 9.73, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_scitail-pairs-qa_loss": 0.3319751024246216, |
|
"eval_scitail-pairs-qa_runtime": 5.7118, |
|
"eval_scitail-pairs-qa_samples_per_second": 131.308, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.202, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_scitail-pairs-pos_loss": 1.288506269454956, |
|
"eval_scitail-pairs-pos_runtime": 15.2101, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.732, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.696, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_xsum-pairs_loss": 1.3039193153381348, |
|
"eval_xsum-pairs_runtime": 3.0719, |
|
"eval_xsum-pairs_samples_per_second": 65.106, |
|
"eval_xsum-pairs_steps_per_second": 2.279, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_compression-pairs_loss": 0.6815783977508545, |
|
"eval_compression-pairs_runtime": 1.2859, |
|
"eval_compression-pairs_samples_per_second": 388.839, |
|
"eval_compression-pairs_steps_per_second": 12.443, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_sciq_pairs_loss": 5.809567451477051, |
|
"eval_sciq_pairs_runtime": 29.0991, |
|
"eval_sciq_pairs_samples_per_second": 20.069, |
|
"eval_sciq_pairs_steps_per_second": 0.653, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_qasc_pairs_loss": 6.919505596160889, |
|
"eval_qasc_pairs_runtime": 5.2734, |
|
"eval_qasc_pairs_samples_per_second": 77.179, |
|
"eval_qasc_pairs_steps_per_second": 2.465, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_openbookqa_pairs_loss": 3.8856022357940674, |
|
"eval_openbookqa_pairs_runtime": 1.1676, |
|
"eval_openbookqa_pairs_samples_per_second": 117.334, |
|
"eval_openbookqa_pairs_steps_per_second": 4.282, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_msmarco_pairs_loss": 2.367374897003174, |
|
"eval_msmarco_pairs_runtime": 19.5819, |
|
"eval_msmarco_pairs_samples_per_second": 51.068, |
|
"eval_msmarco_pairs_steps_per_second": 1.634, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_nq_pairs_loss": 2.5974488258361816, |
|
"eval_nq_pairs_runtime": 37.1479, |
|
"eval_nq_pairs_samples_per_second": 26.919, |
|
"eval_nq_pairs_steps_per_second": 0.861, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_trivia_pairs_loss": 2.8283140659332275, |
|
"eval_trivia_pairs_runtime": 63.9593, |
|
"eval_trivia_pairs_samples_per_second": 15.635, |
|
"eval_trivia_pairs_steps_per_second": 0.5, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_quora_pairs_loss": 0.5709843039512634, |
|
"eval_quora_pairs_runtime": 3.6911, |
|
"eval_quora_pairs_samples_per_second": 135.462, |
|
"eval_quora_pairs_steps_per_second": 4.335, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_gooaq_pairs_loss": 1.9126713275909424, |
|
"eval_gooaq_pairs_runtime": 13.0893, |
|
"eval_gooaq_pairs_samples_per_second": 76.398, |
|
"eval_gooaq_pairs_steps_per_second": 2.445, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5412221144519883, |
|
"grad_norm": 25.740751266479492, |
|
"learning_rate": 1.800646551724138e-05, |
|
"loss": 2.4683, |
|
"step": 3348 |
|
}, |
|
{ |
|
"epoch": 0.5562560620756547, |
|
"grad_norm": 26.303068161010742, |
|
"learning_rate": 1.850754310344828e-05, |
|
"loss": 2.3363, |
|
"step": 3441 |
|
}, |
|
{ |
|
"epoch": 0.5712900096993211, |
|
"grad_norm": 29.08409881591797, |
|
"learning_rate": 1.9008620689655172e-05, |
|
"loss": 2.474, |
|
"step": 3534 |
|
}, |
|
{ |
|
"epoch": 0.5863239573229874, |
|
"grad_norm": 39.23952865600586, |
|
"learning_rate": 1.9509698275862068e-05, |
|
"loss": 2.4987, |
|
"step": 3627 |
|
}, |
|
{ |
|
"epoch": 0.6013579049466538, |
|
"grad_norm": 28.61606216430664, |
|
"learning_rate": 2.0010775862068968e-05, |
|
"loss": 2.4353, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.61639185257032, |
|
"grad_norm": 21.921194076538086, |
|
"learning_rate": 2.0511853448275864e-05, |
|
"loss": 2.4929, |
|
"step": 3813 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_nli-pairs_loss": 2.255383253097534, |
|
"eval_nli-pairs_runtime": 14.6431, |
|
"eval_nli-pairs_samples_per_second": 102.438, |
|
"eval_nli-pairs_steps_per_second": 3.21, |
|
"eval_sts-test_pearson_cosine": 0.7362034907049473, |
|
"eval_sts-test_pearson_dot": 0.6436366129416246, |
|
"eval_sts-test_pearson_euclidean": 0.732771735900544, |
|
"eval_sts-test_pearson_manhattan": 0.7403401803178591, |
|
"eval_sts-test_pearson_max": 0.7403401803178591, |
|
"eval_sts-test_spearman_cosine": 0.7249586438571696, |
|
"eval_sts-test_spearman_dot": 0.6183909485601774, |
|
"eval_sts-test_spearman_euclidean": 0.7179356791740029, |
|
"eval_sts-test_spearman_manhattan": 0.7276798444958615, |
|
"eval_sts-test_spearman_max": 0.7276798444958615, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_vitaminc-pairs_loss": 1.743114948272705, |
|
"eval_vitaminc-pairs_runtime": 14.2467, |
|
"eval_vitaminc-pairs_samples_per_second": 70.121, |
|
"eval_vitaminc-pairs_steps_per_second": 2.246, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_qnli-contrastive_loss": 3.645235776901245, |
|
"eval_qnli-contrastive_runtime": 4.7587, |
|
"eval_qnli-contrastive_samples_per_second": 315.213, |
|
"eval_qnli-contrastive_steps_per_second": 9.877, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_scitail-pairs-qa_loss": 0.2727784514427185, |
|
"eval_scitail-pairs-qa_runtime": 5.8386, |
|
"eval_scitail-pairs-qa_samples_per_second": 128.455, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.111, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_scitail-pairs-pos_loss": 1.1768519878387451, |
|
"eval_scitail-pairs-pos_runtime": 15.1046, |
|
"eval_scitail-pairs-pos_samples_per_second": 86.331, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.714, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_xsum-pairs_loss": 1.086603045463562, |
|
"eval_xsum-pairs_runtime": 3.03, |
|
"eval_xsum-pairs_samples_per_second": 66.006, |
|
"eval_xsum-pairs_steps_per_second": 2.31, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_compression-pairs_loss": 0.5864972472190857, |
|
"eval_compression-pairs_runtime": 1.2863, |
|
"eval_compression-pairs_samples_per_second": 388.702, |
|
"eval_compression-pairs_steps_per_second": 12.438, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_sciq_pairs_loss": 5.628969669342041, |
|
"eval_sciq_pairs_runtime": 28.8592, |
|
"eval_sciq_pairs_samples_per_second": 20.236, |
|
"eval_sciq_pairs_steps_per_second": 0.658, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_qasc_pairs_loss": 6.505127429962158, |
|
"eval_qasc_pairs_runtime": 5.215, |
|
"eval_qasc_pairs_samples_per_second": 78.045, |
|
"eval_qasc_pairs_steps_per_second": 2.493, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_openbookqa_pairs_loss": 3.6648342609405518, |
|
"eval_openbookqa_pairs_runtime": 1.1816, |
|
"eval_openbookqa_pairs_samples_per_second": 115.942, |
|
"eval_openbookqa_pairs_steps_per_second": 4.231, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_msmarco_pairs_loss": 2.1173150539398193, |
|
"eval_msmarco_pairs_runtime": 19.6479, |
|
"eval_msmarco_pairs_samples_per_second": 50.896, |
|
"eval_msmarco_pairs_steps_per_second": 1.629, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_nq_pairs_loss": 2.3446831703186035, |
|
"eval_nq_pairs_runtime": 37.0755, |
|
"eval_nq_pairs_samples_per_second": 26.972, |
|
"eval_nq_pairs_steps_per_second": 0.863, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_trivia_pairs_loss": 2.3582851886749268, |
|
"eval_trivia_pairs_runtime": 64.1047, |
|
"eval_trivia_pairs_samples_per_second": 15.599, |
|
"eval_trivia_pairs_steps_per_second": 0.499, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_quora_pairs_loss": 0.5091556310653687, |
|
"eval_quora_pairs_runtime": 3.7211, |
|
"eval_quora_pairs_samples_per_second": 134.37, |
|
"eval_quora_pairs_steps_per_second": 4.3, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6302942127384417, |
|
"eval_gooaq_pairs_loss": 1.697973608970642, |
|
"eval_gooaq_pairs_runtime": 13.3237, |
|
"eval_gooaq_pairs_samples_per_second": 75.054, |
|
"eval_gooaq_pairs_steps_per_second": 2.402, |
|
"step": 3899 |
|
}, |
|
{ |
|
"epoch": 0.6314258001939864, |
|
"grad_norm": 21.32211685180664, |
|
"learning_rate": 2.101293103448276e-05, |
|
"loss": 2.4941, |
|
"step": 3906 |
|
}, |
|
{ |
|
"epoch": 0.6464597478176528, |
|
"grad_norm": 7.4477925300598145, |
|
"learning_rate": 2.1514008620689653e-05, |
|
"loss": 2.4004, |
|
"step": 3999 |
|
}, |
|
{ |
|
"epoch": 0.6614936954413191, |
|
"grad_norm": 28.622692108154297, |
|
"learning_rate": 2.2015086206896553e-05, |
|
"loss": 2.3529, |
|
"step": 4092 |
|
}, |
|
{ |
|
"epoch": 0.6765276430649855, |
|
"grad_norm": 22.51656150817871, |
|
"learning_rate": 2.251616379310345e-05, |
|
"loss": 2.1828, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 0.6915615906886518, |
|
"grad_norm": 16.42972755432129, |
|
"learning_rate": 2.3017241379310345e-05, |
|
"loss": 2.1688, |
|
"step": 4278 |
|
}, |
|
{ |
|
"epoch": 0.7065955383123181, |
|
"grad_norm": 9.458460807800293, |
|
"learning_rate": 2.351831896551724e-05, |
|
"loss": 2.1722, |
|
"step": 4371 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_nli-pairs_loss": 2.082688808441162, |
|
"eval_nli-pairs_runtime": 14.7547, |
|
"eval_nli-pairs_samples_per_second": 101.662, |
|
"eval_nli-pairs_steps_per_second": 3.185, |
|
"eval_sts-test_pearson_cosine": 0.7512760123852994, |
|
"eval_sts-test_pearson_dot": 0.6565372535148624, |
|
"eval_sts-test_pearson_euclidean": 0.7429357726720194, |
|
"eval_sts-test_pearson_manhattan": 0.7519893029331556, |
|
"eval_sts-test_pearson_max": 0.7519893029331556, |
|
"eval_sts-test_spearman_cosine": 0.7374054618465373, |
|
"eval_sts-test_spearman_dot": 0.6325403342150556, |
|
"eval_sts-test_spearman_euclidean": 0.7271706354542181, |
|
"eval_sts-test_spearman_manhattan": 0.7380975025260701, |
|
"eval_sts-test_spearman_max": 0.7380975025260701, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_vitaminc-pairs_loss": 1.5221832990646362, |
|
"eval_vitaminc-pairs_runtime": 14.2853, |
|
"eval_vitaminc-pairs_samples_per_second": 69.932, |
|
"eval_vitaminc-pairs_steps_per_second": 2.24, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_qnli-contrastive_loss": 3.0361263751983643, |
|
"eval_qnli-contrastive_runtime": 4.808, |
|
"eval_qnli-contrastive_samples_per_second": 311.983, |
|
"eval_qnli-contrastive_steps_per_second": 9.775, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_scitail-pairs-qa_loss": 0.22337602078914642, |
|
"eval_scitail-pairs-qa_runtime": 5.7385, |
|
"eval_scitail-pairs-qa_samples_per_second": 130.697, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.182, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_scitail-pairs-pos_loss": 1.0112017393112183, |
|
"eval_scitail-pairs-pos_runtime": 15.1901, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.845, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.699, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_xsum-pairs_loss": 0.9723415970802307, |
|
"eval_xsum-pairs_runtime": 3.0382, |
|
"eval_xsum-pairs_samples_per_second": 65.829, |
|
"eval_xsum-pairs_steps_per_second": 2.304, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_compression-pairs_loss": 0.5026127696037292, |
|
"eval_compression-pairs_runtime": 1.2886, |
|
"eval_compression-pairs_samples_per_second": 388.022, |
|
"eval_compression-pairs_steps_per_second": 12.417, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_sciq_pairs_loss": 5.491844654083252, |
|
"eval_sciq_pairs_runtime": 28.7711, |
|
"eval_sciq_pairs_samples_per_second": 20.298, |
|
"eval_sciq_pairs_steps_per_second": 0.66, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_qasc_pairs_loss": 6.287917613983154, |
|
"eval_qasc_pairs_runtime": 5.2088, |
|
"eval_qasc_pairs_samples_per_second": 78.136, |
|
"eval_qasc_pairs_steps_per_second": 2.496, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_openbookqa_pairs_loss": 3.57511568069458, |
|
"eval_openbookqa_pairs_runtime": 1.1702, |
|
"eval_openbookqa_pairs_samples_per_second": 117.075, |
|
"eval_openbookqa_pairs_steps_per_second": 4.273, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_msmarco_pairs_loss": 1.9658100605010986, |
|
"eval_msmarco_pairs_runtime": 19.6025, |
|
"eval_msmarco_pairs_samples_per_second": 51.014, |
|
"eval_msmarco_pairs_steps_per_second": 1.632, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_nq_pairs_loss": 2.010303258895874, |
|
"eval_nq_pairs_runtime": 37.0777, |
|
"eval_nq_pairs_samples_per_second": 26.97, |
|
"eval_nq_pairs_steps_per_second": 0.863, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_trivia_pairs_loss": 2.2463245391845703, |
|
"eval_trivia_pairs_runtime": 64.2182, |
|
"eval_trivia_pairs_samples_per_second": 15.572, |
|
"eval_trivia_pairs_steps_per_second": 0.498, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_quora_pairs_loss": 0.4209747612476349, |
|
"eval_quora_pairs_runtime": 3.7187, |
|
"eval_quora_pairs_samples_per_second": 134.456, |
|
"eval_quora_pairs_steps_per_second": 4.303, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7203362431296476, |
|
"eval_gooaq_pairs_loss": 1.5004125833511353, |
|
"eval_gooaq_pairs_runtime": 13.2045, |
|
"eval_gooaq_pairs_samples_per_second": 75.732, |
|
"eval_gooaq_pairs_steps_per_second": 2.423, |
|
"step": 4456 |
|
}, |
|
{ |
|
"epoch": 0.7216294859359845, |
|
"grad_norm": 38.99781799316406, |
|
"learning_rate": 2.4019396551724138e-05, |
|
"loss": 2.2422, |
|
"step": 4464 |
|
}, |
|
{ |
|
"epoch": 0.7366634335596508, |
|
"grad_norm": 1.7083271741867065, |
|
"learning_rate": 2.4520474137931034e-05, |
|
"loss": 2.4258, |
|
"step": 4557 |
|
}, |
|
{ |
|
"epoch": 0.7516973811833172, |
|
"grad_norm": 26.088481903076172, |
|
"learning_rate": 2.5021551724137933e-05, |
|
"loss": 2.1841, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.7667313288069835, |
|
"grad_norm": 30.17441749572754, |
|
"learning_rate": 2.552262931034483e-05, |
|
"loss": 2.0512, |
|
"step": 4743 |
|
}, |
|
{ |
|
"epoch": 0.7817652764306499, |
|
"grad_norm": 19.184717178344727, |
|
"learning_rate": 2.6023706896551723e-05, |
|
"loss": 2.1915, |
|
"step": 4836 |
|
}, |
|
{ |
|
"epoch": 0.7967992240543162, |
|
"grad_norm": 13.038640975952148, |
|
"learning_rate": 2.6524784482758622e-05, |
|
"loss": 2.0704, |
|
"step": 4929 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_nli-pairs_loss": 1.975083589553833, |
|
"eval_nli-pairs_runtime": 14.7438, |
|
"eval_nli-pairs_samples_per_second": 101.738, |
|
"eval_nli-pairs_steps_per_second": 3.188, |
|
"eval_sts-test_pearson_cosine": 0.7529761489757181, |
|
"eval_sts-test_pearson_dot": 0.6377829563094783, |
|
"eval_sts-test_pearson_euclidean": 0.7513689535698412, |
|
"eval_sts-test_pearson_manhattan": 0.7579433693255461, |
|
"eval_sts-test_pearson_max": 0.7579433693255461, |
|
"eval_sts-test_spearman_cosine": 0.7450311313653241, |
|
"eval_sts-test_spearman_dot": 0.6167543293489108, |
|
"eval_sts-test_spearman_euclidean": 0.7382401626273042, |
|
"eval_sts-test_spearman_manhattan": 0.7467817396021367, |
|
"eval_sts-test_spearman_max": 0.7467817396021367, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_vitaminc-pairs_loss": 1.3757116794586182, |
|
"eval_vitaminc-pairs_runtime": 14.3406, |
|
"eval_vitaminc-pairs_samples_per_second": 69.662, |
|
"eval_vitaminc-pairs_steps_per_second": 2.231, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_qnli-contrastive_loss": 3.0041842460632324, |
|
"eval_qnli-contrastive_runtime": 4.8321, |
|
"eval_qnli-contrastive_samples_per_second": 310.427, |
|
"eval_qnli-contrastive_steps_per_second": 9.727, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_scitail-pairs-qa_loss": 0.2100004106760025, |
|
"eval_scitail-pairs-qa_runtime": 5.7923, |
|
"eval_scitail-pairs-qa_samples_per_second": 129.481, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.143, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_scitail-pairs-pos_loss": 0.9822331070899963, |
|
"eval_scitail-pairs-pos_runtime": 15.3361, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.028, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.673, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_xsum-pairs_loss": 0.9295198321342468, |
|
"eval_xsum-pairs_runtime": 3.0414, |
|
"eval_xsum-pairs_samples_per_second": 65.759, |
|
"eval_xsum-pairs_steps_per_second": 2.302, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_compression-pairs_loss": 0.45234963297843933, |
|
"eval_compression-pairs_runtime": 1.2825, |
|
"eval_compression-pairs_samples_per_second": 389.864, |
|
"eval_compression-pairs_steps_per_second": 12.476, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_sciq_pairs_loss": 5.314521789550781, |
|
"eval_sciq_pairs_runtime": 29.1947, |
|
"eval_sciq_pairs_samples_per_second": 20.004, |
|
"eval_sciq_pairs_steps_per_second": 0.651, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_qasc_pairs_loss": 6.56479024887085, |
|
"eval_qasc_pairs_runtime": 5.2907, |
|
"eval_qasc_pairs_samples_per_second": 76.927, |
|
"eval_qasc_pairs_steps_per_second": 2.457, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_openbookqa_pairs_loss": 3.3964192867279053, |
|
"eval_openbookqa_pairs_runtime": 1.1691, |
|
"eval_openbookqa_pairs_samples_per_second": 117.184, |
|
"eval_openbookqa_pairs_steps_per_second": 4.277, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_msmarco_pairs_loss": 1.7340222597122192, |
|
"eval_msmarco_pairs_runtime": 19.7221, |
|
"eval_msmarco_pairs_samples_per_second": 50.705, |
|
"eval_msmarco_pairs_steps_per_second": 1.623, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_nq_pairs_loss": 1.8559346199035645, |
|
"eval_nq_pairs_runtime": 37.0328, |
|
"eval_nq_pairs_samples_per_second": 27.003, |
|
"eval_nq_pairs_steps_per_second": 0.864, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_trivia_pairs_loss": 2.1573712825775146, |
|
"eval_trivia_pairs_runtime": 64.2779, |
|
"eval_trivia_pairs_samples_per_second": 15.557, |
|
"eval_trivia_pairs_steps_per_second": 0.498, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_quora_pairs_loss": 0.40095773339271545, |
|
"eval_quora_pairs_runtime": 3.7934, |
|
"eval_quora_pairs_samples_per_second": 131.809, |
|
"eval_quora_pairs_steps_per_second": 4.218, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8103782735208536, |
|
"eval_gooaq_pairs_loss": 1.3385512828826904, |
|
"eval_gooaq_pairs_runtime": 13.3898, |
|
"eval_gooaq_pairs_samples_per_second": 74.684, |
|
"eval_gooaq_pairs_steps_per_second": 2.39, |
|
"step": 5013 |
|
}, |
|
{ |
|
"epoch": 0.8118331716779825, |
|
"grad_norm": 11.097302436828613, |
|
"learning_rate": 2.702586206896552e-05, |
|
"loss": 2.1681, |
|
"step": 5022 |
|
}, |
|
{ |
|
"epoch": 0.8268671193016489, |
|
"grad_norm": 22.10071563720703, |
|
"learning_rate": 2.7526939655172415e-05, |
|
"loss": 2.2056, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 0.8419010669253152, |
|
"grad_norm": 19.801000595092773, |
|
"learning_rate": 2.802801724137931e-05, |
|
"loss": 1.9344, |
|
"step": 5208 |
|
}, |
|
{ |
|
"epoch": 0.8569350145489816, |
|
"grad_norm": 12.516477584838867, |
|
"learning_rate": 2.8529094827586207e-05, |
|
"loss": 2.0812, |
|
"step": 5301 |
|
}, |
|
{ |
|
"epoch": 0.871968962172648, |
|
"grad_norm": 29.498964309692383, |
|
"learning_rate": 2.9030172413793103e-05, |
|
"loss": 1.8509, |
|
"step": 5394 |
|
}, |
|
{ |
|
"epoch": 0.8870029097963142, |
|
"grad_norm": 23.26563835144043, |
|
"learning_rate": 2.953125e-05, |
|
"loss": 2.0887, |
|
"step": 5487 |
|
} |
|
], |
|
"logging_steps": 93, |
|
"max_steps": 18558, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 1856, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|