|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.6000646621403168, |
|
"eval_steps": 557, |
|
"global_step": 3712, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.015033947623666343, |
|
"grad_norm": 63.938453674316406, |
|
"learning_rate": 4.741379310344828e-07, |
|
"loss": 11.7791, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.030067895247332686, |
|
"grad_norm": 28.60324478149414, |
|
"learning_rate": 9.752155172413793e-07, |
|
"loss": 10.5226, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.04510184287099903, |
|
"grad_norm": 17.869054794311523, |
|
"learning_rate": 1.476293103448276e-06, |
|
"loss": 9.3586, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.06013579049466537, |
|
"grad_norm": 18.835189819335938, |
|
"learning_rate": 1.977370689655172e-06, |
|
"loss": 8.4548, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.07516973811833172, |
|
"grad_norm": 55.06333541870117, |
|
"learning_rate": 2.4784482758620692e-06, |
|
"loss": 8.3449, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_nli-pairs_loss": 7.450161933898926, |
|
"eval_nli-pairs_runtime": 14.8176, |
|
"eval_nli-pairs_samples_per_second": 101.231, |
|
"eval_nli-pairs_steps_per_second": 3.172, |
|
"eval_sts-test_pearson_cosine": 0.3696416595298566, |
|
"eval_sts-test_pearson_dot": 0.13300273461809956, |
|
"eval_sts-test_pearson_euclidean": 0.3836877008752716, |
|
"eval_sts-test_pearson_manhattan": 0.41477338080789633, |
|
"eval_sts-test_pearson_max": 0.41477338080789633, |
|
"eval_sts-test_spearman_cosine": 0.3694134524358256, |
|
"eval_sts-test_spearman_dot": 0.12058818576425179, |
|
"eval_sts-test_spearman_euclidean": 0.38615620458370276, |
|
"eval_sts-test_spearman_manhattan": 0.4121304311152658, |
|
"eval_sts-test_spearman_max": 0.4121304311152658, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_vitaminc-pairs_loss": 5.403733730316162, |
|
"eval_vitaminc-pairs_runtime": 14.4427, |
|
"eval_vitaminc-pairs_samples_per_second": 69.17, |
|
"eval_vitaminc-pairs_steps_per_second": 2.216, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_qnli-contrastive_loss": 8.863805770874023, |
|
"eval_qnli-contrastive_runtime": 4.835, |
|
"eval_qnli-contrastive_samples_per_second": 310.238, |
|
"eval_qnli-contrastive_steps_per_second": 9.721, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_scitail-pairs-qa_loss": 4.2781147956848145, |
|
"eval_scitail-pairs-qa_runtime": 5.761, |
|
"eval_scitail-pairs-qa_samples_per_second": 130.186, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.166, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_scitail-pairs-pos_loss": 4.930431365966797, |
|
"eval_scitail-pairs-pos_runtime": 15.2161, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.699, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.695, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_xsum-pairs_loss": 4.778657913208008, |
|
"eval_xsum-pairs_runtime": 3.0397, |
|
"eval_xsum-pairs_samples_per_second": 65.795, |
|
"eval_xsum-pairs_steps_per_second": 2.303, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_compression-pairs_loss": 3.489774465560913, |
|
"eval_compression-pairs_runtime": 1.2758, |
|
"eval_compression-pairs_samples_per_second": 391.924, |
|
"eval_compression-pairs_steps_per_second": 12.542, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_sciq_pairs_loss": 7.707857131958008, |
|
"eval_sciq_pairs_runtime": 28.932, |
|
"eval_sciq_pairs_samples_per_second": 20.185, |
|
"eval_sciq_pairs_steps_per_second": 0.657, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_qasc_pairs_loss": 12.320230484008789, |
|
"eval_qasc_pairs_runtime": 5.2561, |
|
"eval_qasc_pairs_samples_per_second": 77.434, |
|
"eval_qasc_pairs_steps_per_second": 2.473, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_openbookqa_pairs_loss": 7.607065677642822, |
|
"eval_openbookqa_pairs_runtime": 1.189, |
|
"eval_openbookqa_pairs_samples_per_second": 115.224, |
|
"eval_openbookqa_pairs_steps_per_second": 4.205, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_msmarco_pairs_loss": 8.519609451293945, |
|
"eval_msmarco_pairs_runtime": 19.7371, |
|
"eval_msmarco_pairs_samples_per_second": 50.666, |
|
"eval_msmarco_pairs_steps_per_second": 1.621, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_nq_pairs_loss": 8.43966007232666, |
|
"eval_nq_pairs_runtime": 37.2943, |
|
"eval_nq_pairs_samples_per_second": 26.814, |
|
"eval_nq_pairs_steps_per_second": 0.858, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_trivia_pairs_loss": 8.689275741577148, |
|
"eval_trivia_pairs_runtime": 64.6902, |
|
"eval_trivia_pairs_samples_per_second": 15.458, |
|
"eval_trivia_pairs_steps_per_second": 0.495, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_quora_pairs_loss": 4.54755163192749, |
|
"eval_quora_pairs_runtime": 3.7394, |
|
"eval_quora_pairs_samples_per_second": 133.71, |
|
"eval_quora_pairs_steps_per_second": 4.279, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09004203039120595, |
|
"eval_gooaq_pairs_loss": 7.57412052154541, |
|
"eval_gooaq_pairs_runtime": 13.3648, |
|
"eval_gooaq_pairs_samples_per_second": 74.824, |
|
"eval_gooaq_pairs_steps_per_second": 2.394, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09020368574199807, |
|
"grad_norm": 21.490495681762695, |
|
"learning_rate": 2.9795258620689654e-06, |
|
"loss": 7.4107, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.1052376333656644, |
|
"grad_norm": 35.23164367675781, |
|
"learning_rate": 3.480603448275862e-06, |
|
"loss": 7.0111, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.12027158098933075, |
|
"grad_norm": 28.60922622680664, |
|
"learning_rate": 3.981681034482759e-06, |
|
"loss": 5.9488, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.13530552861299708, |
|
"grad_norm": 22.0839786529541, |
|
"learning_rate": 4.482758620689656e-06, |
|
"loss": 6.3051, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.15033947623666344, |
|
"grad_norm": 34.64767837524414, |
|
"learning_rate": 4.983836206896552e-06, |
|
"loss": 5.2056, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.16537342386032977, |
|
"grad_norm": 13.605194091796875, |
|
"learning_rate": 5.484913793103448e-06, |
|
"loss": 5.0794, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_nli-pairs_loss": 4.693160057067871, |
|
"eval_nli-pairs_runtime": 14.776, |
|
"eval_nli-pairs_samples_per_second": 101.516, |
|
"eval_nli-pairs_steps_per_second": 3.181, |
|
"eval_sts-test_pearson_cosine": 0.6672572282189017, |
|
"eval_sts-test_pearson_dot": 0.5110207654364226, |
|
"eval_sts-test_pearson_euclidean": 0.65934512843155, |
|
"eval_sts-test_pearson_manhattan": 0.6615633016003417, |
|
"eval_sts-test_pearson_max": 0.6672572282189017, |
|
"eval_sts-test_spearman_cosine": 0.6364241996636473, |
|
"eval_sts-test_spearman_dot": 0.4920702168931027, |
|
"eval_sts-test_spearman_euclidean": 0.638848044048013, |
|
"eval_sts-test_spearman_manhattan": 0.6407402326117282, |
|
"eval_sts-test_spearman_max": 0.6407402326117282, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_vitaminc-pairs_loss": 3.5585296154022217, |
|
"eval_vitaminc-pairs_runtime": 14.3661, |
|
"eval_vitaminc-pairs_samples_per_second": 69.538, |
|
"eval_vitaminc-pairs_steps_per_second": 2.227, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_qnli-contrastive_loss": 11.759393692016602, |
|
"eval_qnli-contrastive_runtime": 4.7771, |
|
"eval_qnli-contrastive_samples_per_second": 313.999, |
|
"eval_qnli-contrastive_steps_per_second": 9.839, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_scitail-pairs-qa_loss": 1.3339837789535522, |
|
"eval_scitail-pairs-qa_runtime": 5.8445, |
|
"eval_scitail-pairs-qa_samples_per_second": 128.326, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.106, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_scitail-pairs-pos_loss": 2.508178234100342, |
|
"eval_scitail-pairs-pos_runtime": 15.12, |
|
"eval_scitail-pairs-pos_samples_per_second": 86.244, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.712, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_xsum-pairs_loss": 2.727797508239746, |
|
"eval_xsum-pairs_runtime": 3.0382, |
|
"eval_xsum-pairs_samples_per_second": 65.828, |
|
"eval_xsum-pairs_steps_per_second": 2.304, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_compression-pairs_loss": 1.769970178604126, |
|
"eval_compression-pairs_runtime": 1.3017, |
|
"eval_compression-pairs_samples_per_second": 384.105, |
|
"eval_compression-pairs_steps_per_second": 12.291, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_sciq_pairs_loss": 6.772756576538086, |
|
"eval_sciq_pairs_runtime": 29.0556, |
|
"eval_sciq_pairs_samples_per_second": 20.099, |
|
"eval_sciq_pairs_steps_per_second": 0.654, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_qasc_pairs_loss": 10.429397583007812, |
|
"eval_qasc_pairs_runtime": 5.2915, |
|
"eval_qasc_pairs_samples_per_second": 76.916, |
|
"eval_qasc_pairs_steps_per_second": 2.457, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_openbookqa_pairs_loss": 5.441987037658691, |
|
"eval_openbookqa_pairs_runtime": 1.1725, |
|
"eval_openbookqa_pairs_samples_per_second": 116.843, |
|
"eval_openbookqa_pairs_steps_per_second": 4.264, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_msmarco_pairs_loss": 4.586305618286133, |
|
"eval_msmarco_pairs_runtime": 19.8051, |
|
"eval_msmarco_pairs_samples_per_second": 50.492, |
|
"eval_msmarco_pairs_steps_per_second": 1.616, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_nq_pairs_loss": 5.193580150604248, |
|
"eval_nq_pairs_runtime": 37.2293, |
|
"eval_nq_pairs_samples_per_second": 26.861, |
|
"eval_nq_pairs_steps_per_second": 0.86, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_trivia_pairs_loss": 5.120319366455078, |
|
"eval_trivia_pairs_runtime": 64.3021, |
|
"eval_trivia_pairs_samples_per_second": 15.552, |
|
"eval_trivia_pairs_steps_per_second": 0.498, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_quora_pairs_loss": 1.4382946491241455, |
|
"eval_quora_pairs_runtime": 3.7168, |
|
"eval_quora_pairs_samples_per_second": 134.525, |
|
"eval_quora_pairs_steps_per_second": 4.305, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.1800840607824119, |
|
"eval_gooaq_pairs_loss": 3.7513082027435303, |
|
"eval_gooaq_pairs_runtime": 13.2594, |
|
"eval_gooaq_pairs_samples_per_second": 75.418, |
|
"eval_gooaq_pairs_steps_per_second": 2.413, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.18040737148399613, |
|
"grad_norm": 39.661067962646484, |
|
"learning_rate": 5.9859913793103445e-06, |
|
"loss": 4.362, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 0.19544131910766246, |
|
"grad_norm": 34.55929946899414, |
|
"learning_rate": 6.487068965517242e-06, |
|
"loss": 4.3245, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 0.2104752667313288, |
|
"grad_norm": 25.775623321533203, |
|
"learning_rate": 6.988146551724138e-06, |
|
"loss": 4.359, |
|
"step": 1302 |
|
}, |
|
{ |
|
"epoch": 0.22550921435499516, |
|
"grad_norm": 57.751529693603516, |
|
"learning_rate": 7.489224137931035e-06, |
|
"loss": 4.0488, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.2405431619786615, |
|
"grad_norm": 33.448768615722656, |
|
"learning_rate": 7.990301724137931e-06, |
|
"loss": 4.0499, |
|
"step": 1488 |
|
}, |
|
{ |
|
"epoch": 0.2555771096023278, |
|
"grad_norm": 29.5295467376709, |
|
"learning_rate": 8.491379310344827e-06, |
|
"loss": 3.7256, |
|
"step": 1581 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_nli-pairs_loss": 3.6348764896392822, |
|
"eval_nli-pairs_runtime": 14.761, |
|
"eval_nli-pairs_samples_per_second": 101.619, |
|
"eval_nli-pairs_steps_per_second": 3.184, |
|
"eval_sts-test_pearson_cosine": 0.6972871286019413, |
|
"eval_sts-test_pearson_dot": 0.5667394576060958, |
|
"eval_sts-test_pearson_euclidean": 0.6946767725059286, |
|
"eval_sts-test_pearson_manhattan": 0.7031948219809795, |
|
"eval_sts-test_pearson_max": 0.7031948219809795, |
|
"eval_sts-test_spearman_cosine": 0.6666257059052387, |
|
"eval_sts-test_spearman_dot": 0.5430589355737897, |
|
"eval_sts-test_spearman_euclidean": 0.6727104999188346, |
|
"eval_sts-test_spearman_manhattan": 0.6812486382298032, |
|
"eval_sts-test_spearman_max": 0.6812486382298032, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_vitaminc-pairs_loss": 2.811436653137207, |
|
"eval_vitaminc-pairs_runtime": 14.4779, |
|
"eval_vitaminc-pairs_samples_per_second": 69.002, |
|
"eval_vitaminc-pairs_steps_per_second": 2.21, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_qnli-contrastive_loss": 9.688385009765625, |
|
"eval_qnli-contrastive_runtime": 4.8234, |
|
"eval_qnli-contrastive_samples_per_second": 310.984, |
|
"eval_qnli-contrastive_steps_per_second": 9.744, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_scitail-pairs-qa_loss": 0.8966398239135742, |
|
"eval_scitail-pairs-qa_runtime": 5.8274, |
|
"eval_scitail-pairs-qa_samples_per_second": 128.702, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.118, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_scitail-pairs-pos_loss": 1.9770935773849487, |
|
"eval_scitail-pairs-pos_runtime": 15.5498, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.859, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.637, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_xsum-pairs_loss": 2.1290199756622314, |
|
"eval_xsum-pairs_runtime": 3.0544, |
|
"eval_xsum-pairs_samples_per_second": 65.479, |
|
"eval_xsum-pairs_steps_per_second": 2.292, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_compression-pairs_loss": 1.2427340745925903, |
|
"eval_compression-pairs_runtime": 1.3179, |
|
"eval_compression-pairs_samples_per_second": 379.389, |
|
"eval_compression-pairs_steps_per_second": 12.14, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_sciq_pairs_loss": 6.424961090087891, |
|
"eval_sciq_pairs_runtime": 29.2749, |
|
"eval_sciq_pairs_samples_per_second": 19.949, |
|
"eval_sciq_pairs_steps_per_second": 0.649, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_qasc_pairs_loss": 9.239206314086914, |
|
"eval_qasc_pairs_runtime": 5.2408, |
|
"eval_qasc_pairs_samples_per_second": 77.66, |
|
"eval_qasc_pairs_steps_per_second": 2.481, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_openbookqa_pairs_loss": 4.784058570861816, |
|
"eval_openbookqa_pairs_runtime": 1.2088, |
|
"eval_openbookqa_pairs_samples_per_second": 113.332, |
|
"eval_openbookqa_pairs_steps_per_second": 4.136, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_msmarco_pairs_loss": 3.655301094055176, |
|
"eval_msmarco_pairs_runtime": 19.7273, |
|
"eval_msmarco_pairs_samples_per_second": 50.691, |
|
"eval_msmarco_pairs_steps_per_second": 1.622, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_nq_pairs_loss": 4.114762783050537, |
|
"eval_nq_pairs_runtime": 37.3924, |
|
"eval_nq_pairs_samples_per_second": 26.743, |
|
"eval_nq_pairs_steps_per_second": 0.856, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_trivia_pairs_loss": 4.019989967346191, |
|
"eval_trivia_pairs_runtime": 64.3836, |
|
"eval_trivia_pairs_samples_per_second": 15.532, |
|
"eval_trivia_pairs_steps_per_second": 0.497, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_quora_pairs_loss": 1.002946138381958, |
|
"eval_quora_pairs_runtime": 3.7091, |
|
"eval_quora_pairs_samples_per_second": 134.803, |
|
"eval_quora_pairs_steps_per_second": 4.314, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27012609117361785, |
|
"eval_gooaq_pairs_loss": 2.87351131439209, |
|
"eval_gooaq_pairs_runtime": 13.2151, |
|
"eval_gooaq_pairs_samples_per_second": 75.671, |
|
"eval_gooaq_pairs_steps_per_second": 2.421, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.27061105722599416, |
|
"grad_norm": 142.8954315185547, |
|
"learning_rate": 8.992456896551725e-06, |
|
"loss": 3.6978, |
|
"step": 1674 |
|
}, |
|
{ |
|
"epoch": 0.28564500484966054, |
|
"grad_norm": 35.244972229003906, |
|
"learning_rate": 9.493534482758622e-06, |
|
"loss": 3.9547, |
|
"step": 1767 |
|
}, |
|
{ |
|
"epoch": 0.3006789524733269, |
|
"grad_norm": 24.055269241333008, |
|
"learning_rate": 9.994612068965518e-06, |
|
"loss": 4.162, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.3157129000969932, |
|
"grad_norm": 39.16152572631836, |
|
"learning_rate": 1.0495689655172414e-05, |
|
"loss": 3.739, |
|
"step": 1953 |
|
}, |
|
{ |
|
"epoch": 0.33074684772065954, |
|
"grad_norm": 30.219701766967773, |
|
"learning_rate": 1.099676724137931e-05, |
|
"loss": 4.2688, |
|
"step": 2046 |
|
}, |
|
{ |
|
"epoch": 0.3457807953443259, |
|
"grad_norm": 35.78736114501953, |
|
"learning_rate": 1.1497844827586206e-05, |
|
"loss": 3.3462, |
|
"step": 2139 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_nli-pairs_loss": 3.1179466247558594, |
|
"eval_nli-pairs_runtime": 14.8963, |
|
"eval_nli-pairs_samples_per_second": 100.696, |
|
"eval_nli-pairs_steps_per_second": 3.155, |
|
"eval_sts-test_pearson_cosine": 0.7110231765083281, |
|
"eval_sts-test_pearson_dot": 0.5911922048259769, |
|
"eval_sts-test_pearson_euclidean": 0.7113733036920562, |
|
"eval_sts-test_pearson_manhattan": 0.7195266505149516, |
|
"eval_sts-test_pearson_max": 0.7195266505149516, |
|
"eval_sts-test_spearman_cosine": 0.6844829069963456, |
|
"eval_sts-test_spearman_dot": 0.567786285795314, |
|
"eval_sts-test_spearman_euclidean": 0.6897367727395519, |
|
"eval_sts-test_spearman_manhattan": 0.6977703060020186, |
|
"eval_sts-test_spearman_max": 0.6977703060020186, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_vitaminc-pairs_loss": 2.4076764583587646, |
|
"eval_vitaminc-pairs_runtime": 14.3487, |
|
"eval_vitaminc-pairs_samples_per_second": 69.623, |
|
"eval_vitaminc-pairs_steps_per_second": 2.23, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_qnli-contrastive_loss": 6.651412010192871, |
|
"eval_qnli-contrastive_runtime": 4.759, |
|
"eval_qnli-contrastive_samples_per_second": 315.189, |
|
"eval_qnli-contrastive_steps_per_second": 9.876, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_scitail-pairs-qa_loss": 0.6640239357948303, |
|
"eval_scitail-pairs-qa_runtime": 5.8178, |
|
"eval_scitail-pairs-qa_samples_per_second": 128.916, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.125, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_scitail-pairs-pos_loss": 1.6727737188339233, |
|
"eval_scitail-pairs-pos_runtime": 15.3144, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.149, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.677, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_xsum-pairs_loss": 1.7385025024414062, |
|
"eval_xsum-pairs_runtime": 3.0465, |
|
"eval_xsum-pairs_samples_per_second": 65.649, |
|
"eval_xsum-pairs_steps_per_second": 2.298, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_compression-pairs_loss": 1.0110334157943726, |
|
"eval_compression-pairs_runtime": 1.2863, |
|
"eval_compression-pairs_samples_per_second": 388.706, |
|
"eval_compression-pairs_steps_per_second": 12.439, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_sciq_pairs_loss": 6.164596080780029, |
|
"eval_sciq_pairs_runtime": 28.909, |
|
"eval_sciq_pairs_samples_per_second": 20.201, |
|
"eval_sciq_pairs_steps_per_second": 0.657, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_qasc_pairs_loss": 7.898203372955322, |
|
"eval_qasc_pairs_runtime": 5.2211, |
|
"eval_qasc_pairs_samples_per_second": 77.953, |
|
"eval_qasc_pairs_steps_per_second": 2.49, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_openbookqa_pairs_loss": 4.36992073059082, |
|
"eval_openbookqa_pairs_runtime": 1.1844, |
|
"eval_openbookqa_pairs_samples_per_second": 115.667, |
|
"eval_openbookqa_pairs_steps_per_second": 4.221, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_msmarco_pairs_loss": 3.067087411880493, |
|
"eval_msmarco_pairs_runtime": 19.7205, |
|
"eval_msmarco_pairs_samples_per_second": 50.709, |
|
"eval_msmarco_pairs_steps_per_second": 1.623, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_nq_pairs_loss": 3.3717246055603027, |
|
"eval_nq_pairs_runtime": 37.1177, |
|
"eval_nq_pairs_samples_per_second": 26.941, |
|
"eval_nq_pairs_steps_per_second": 0.862, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_trivia_pairs_loss": 3.4189703464508057, |
|
"eval_trivia_pairs_runtime": 64.257, |
|
"eval_trivia_pairs_samples_per_second": 15.563, |
|
"eval_trivia_pairs_steps_per_second": 0.498, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_quora_pairs_loss": 0.8419531583786011, |
|
"eval_quora_pairs_runtime": 3.7448, |
|
"eval_quora_pairs_samples_per_second": 133.518, |
|
"eval_quora_pairs_steps_per_second": 4.273, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.3601681215648238, |
|
"eval_gooaq_pairs_loss": 2.4530532360076904, |
|
"eval_gooaq_pairs_runtime": 13.3141, |
|
"eval_gooaq_pairs_samples_per_second": 75.108, |
|
"eval_gooaq_pairs_steps_per_second": 2.403, |
|
"step": 2228 |
|
}, |
|
{ |
|
"epoch": 0.36081474296799226, |
|
"grad_norm": 33.25382614135742, |
|
"learning_rate": 1.1998922413793104e-05, |
|
"loss": 3.8272, |
|
"step": 2232 |
|
}, |
|
{ |
|
"epoch": 0.3758486905916586, |
|
"grad_norm": 22.22249984741211, |
|
"learning_rate": 1.25e-05, |
|
"loss": 3.4532, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.39088263821532493, |
|
"grad_norm": 32.02521514892578, |
|
"learning_rate": 1.3001077586206897e-05, |
|
"loss": 3.7391, |
|
"step": 2418 |
|
}, |
|
{ |
|
"epoch": 0.40591658583899126, |
|
"grad_norm": 84.23457336425781, |
|
"learning_rate": 1.3502155172413793e-05, |
|
"loss": 3.4451, |
|
"step": 2511 |
|
}, |
|
{ |
|
"epoch": 0.4209505334626576, |
|
"grad_norm": 82.23103332519531, |
|
"learning_rate": 1.400323275862069e-05, |
|
"loss": 3.2086, |
|
"step": 2604 |
|
}, |
|
{ |
|
"epoch": 0.435984481086324, |
|
"grad_norm": 128.5042266845703, |
|
"learning_rate": 1.4504310344827587e-05, |
|
"loss": 3.153, |
|
"step": 2697 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_nli-pairs_loss": 2.7247567176818848, |
|
"eval_nli-pairs_runtime": 14.7363, |
|
"eval_nli-pairs_samples_per_second": 101.79, |
|
"eval_nli-pairs_steps_per_second": 3.189, |
|
"eval_sts-test_pearson_cosine": 0.7313032856033084, |
|
"eval_sts-test_pearson_dot": 0.639231944179888, |
|
"eval_sts-test_pearson_euclidean": 0.7299296327875673, |
|
"eval_sts-test_pearson_manhattan": 0.7376485877411852, |
|
"eval_sts-test_pearson_max": 0.7376485877411852, |
|
"eval_sts-test_spearman_cosine": 0.713852235440012, |
|
"eval_sts-test_spearman_dot": 0.6141091584560591, |
|
"eval_sts-test_spearman_euclidean": 0.7125938749269385, |
|
"eval_sts-test_spearman_manhattan": 0.7209089187255475, |
|
"eval_sts-test_spearman_max": 0.7209089187255475, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_vitaminc-pairs_loss": 2.1636030673980713, |
|
"eval_vitaminc-pairs_runtime": 14.3573, |
|
"eval_vitaminc-pairs_samples_per_second": 69.582, |
|
"eval_vitaminc-pairs_steps_per_second": 2.229, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_qnli-contrastive_loss": 6.1082000732421875, |
|
"eval_qnli-contrastive_runtime": 4.8812, |
|
"eval_qnli-contrastive_samples_per_second": 307.299, |
|
"eval_qnli-contrastive_steps_per_second": 9.629, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_scitail-pairs-qa_loss": 0.47009098529815674, |
|
"eval_scitail-pairs-qa_runtime": 5.8135, |
|
"eval_scitail-pairs-qa_samples_per_second": 129.011, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.128, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_scitail-pairs-pos_loss": 1.38162362575531, |
|
"eval_scitail-pairs-pos_runtime": 15.0892, |
|
"eval_scitail-pairs-pos_samples_per_second": 86.42, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.717, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_xsum-pairs_loss": 1.4567902088165283, |
|
"eval_xsum-pairs_runtime": 3.0536, |
|
"eval_xsum-pairs_samples_per_second": 65.496, |
|
"eval_xsum-pairs_steps_per_second": 2.292, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_compression-pairs_loss": 0.8824976682662964, |
|
"eval_compression-pairs_runtime": 1.2907, |
|
"eval_compression-pairs_samples_per_second": 387.382, |
|
"eval_compression-pairs_steps_per_second": 12.396, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_sciq_pairs_loss": 5.971975326538086, |
|
"eval_sciq_pairs_runtime": 28.8816, |
|
"eval_sciq_pairs_samples_per_second": 20.22, |
|
"eval_sciq_pairs_steps_per_second": 0.658, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_qasc_pairs_loss": 7.14296817779541, |
|
"eval_qasc_pairs_runtime": 5.2619, |
|
"eval_qasc_pairs_samples_per_second": 77.348, |
|
"eval_qasc_pairs_steps_per_second": 2.471, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_openbookqa_pairs_loss": 3.890052080154419, |
|
"eval_openbookqa_pairs_runtime": 1.1822, |
|
"eval_openbookqa_pairs_samples_per_second": 115.884, |
|
"eval_openbookqa_pairs_steps_per_second": 4.229, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_msmarco_pairs_loss": 2.6956124305725098, |
|
"eval_msmarco_pairs_runtime": 19.7685, |
|
"eval_msmarco_pairs_samples_per_second": 50.586, |
|
"eval_msmarco_pairs_steps_per_second": 1.619, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_nq_pairs_loss": 2.9799299240112305, |
|
"eval_nq_pairs_runtime": 37.2854, |
|
"eval_nq_pairs_samples_per_second": 26.82, |
|
"eval_nq_pairs_steps_per_second": 0.858, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_trivia_pairs_loss": 3.0525760650634766, |
|
"eval_trivia_pairs_runtime": 64.286, |
|
"eval_trivia_pairs_samples_per_second": 15.555, |
|
"eval_trivia_pairs_steps_per_second": 0.498, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_quora_pairs_loss": 0.7019712924957275, |
|
"eval_quora_pairs_runtime": 3.759, |
|
"eval_quora_pairs_samples_per_second": 133.013, |
|
"eval_quora_pairs_steps_per_second": 4.256, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.45021015195602976, |
|
"eval_gooaq_pairs_loss": 2.1218321323394775, |
|
"eval_gooaq_pairs_runtime": 13.3843, |
|
"eval_gooaq_pairs_samples_per_second": 74.715, |
|
"eval_gooaq_pairs_steps_per_second": 2.391, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.4510184287099903, |
|
"grad_norm": 7.123088836669922, |
|
"learning_rate": 1.5e-05, |
|
"loss": 2.9093, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.46605237633365665, |
|
"grad_norm": 26.165264129638672, |
|
"learning_rate": 1.5501077586206898e-05, |
|
"loss": 2.6223, |
|
"step": 2883 |
|
}, |
|
{ |
|
"epoch": 0.481086323957323, |
|
"grad_norm": 7.543350696563721, |
|
"learning_rate": 1.6002155172413795e-05, |
|
"loss": 2.8362, |
|
"step": 2976 |
|
}, |
|
{ |
|
"epoch": 0.4961202715809893, |
|
"grad_norm": 26.791278839111328, |
|
"learning_rate": 1.6503232758620687e-05, |
|
"loss": 2.6383, |
|
"step": 3069 |
|
}, |
|
{ |
|
"epoch": 0.5111542192046556, |
|
"grad_norm": 24.906707763671875, |
|
"learning_rate": 1.7004310344827587e-05, |
|
"loss": 2.6265, |
|
"step": 3162 |
|
}, |
|
{ |
|
"epoch": 0.526188166828322, |
|
"grad_norm": 17.12710189819336, |
|
"learning_rate": 1.7505387931034483e-05, |
|
"loss": 2.8718, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_nli-pairs_loss": 2.4381556510925293, |
|
"eval_nli-pairs_runtime": 14.7362, |
|
"eval_nli-pairs_samples_per_second": 101.79, |
|
"eval_nli-pairs_steps_per_second": 3.189, |
|
"eval_sts-test_pearson_cosine": 0.7355921455984917, |
|
"eval_sts-test_pearson_dot": 0.6441608711789287, |
|
"eval_sts-test_pearson_euclidean": 0.7293547151744354, |
|
"eval_sts-test_pearson_manhattan": 0.7409243217735156, |
|
"eval_sts-test_pearson_max": 0.7409243217735156, |
|
"eval_sts-test_spearman_cosine": 0.7175266639957124, |
|
"eval_sts-test_spearman_dot": 0.6242096516614428, |
|
"eval_sts-test_spearman_euclidean": 0.7122336979289147, |
|
"eval_sts-test_spearman_manhattan": 0.7250251572437655, |
|
"eval_sts-test_spearman_max": 0.7250251572437655, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_vitaminc-pairs_loss": 1.8586883544921875, |
|
"eval_vitaminc-pairs_runtime": 14.319, |
|
"eval_vitaminc-pairs_samples_per_second": 69.767, |
|
"eval_vitaminc-pairs_steps_per_second": 2.235, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_qnli-contrastive_loss": 3.9992873668670654, |
|
"eval_qnli-contrastive_runtime": 4.8306, |
|
"eval_qnli-contrastive_samples_per_second": 310.523, |
|
"eval_qnli-contrastive_steps_per_second": 9.73, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_scitail-pairs-qa_loss": 0.3319751024246216, |
|
"eval_scitail-pairs-qa_runtime": 5.7118, |
|
"eval_scitail-pairs-qa_samples_per_second": 131.308, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.202, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_scitail-pairs-pos_loss": 1.288506269454956, |
|
"eval_scitail-pairs-pos_runtime": 15.2101, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.732, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.696, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_xsum-pairs_loss": 1.3039193153381348, |
|
"eval_xsum-pairs_runtime": 3.0719, |
|
"eval_xsum-pairs_samples_per_second": 65.106, |
|
"eval_xsum-pairs_steps_per_second": 2.279, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_compression-pairs_loss": 0.6815783977508545, |
|
"eval_compression-pairs_runtime": 1.2859, |
|
"eval_compression-pairs_samples_per_second": 388.839, |
|
"eval_compression-pairs_steps_per_second": 12.443, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_sciq_pairs_loss": 5.809567451477051, |
|
"eval_sciq_pairs_runtime": 29.0991, |
|
"eval_sciq_pairs_samples_per_second": 20.069, |
|
"eval_sciq_pairs_steps_per_second": 0.653, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_qasc_pairs_loss": 6.919505596160889, |
|
"eval_qasc_pairs_runtime": 5.2734, |
|
"eval_qasc_pairs_samples_per_second": 77.179, |
|
"eval_qasc_pairs_steps_per_second": 2.465, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_openbookqa_pairs_loss": 3.8856022357940674, |
|
"eval_openbookqa_pairs_runtime": 1.1676, |
|
"eval_openbookqa_pairs_samples_per_second": 117.334, |
|
"eval_openbookqa_pairs_steps_per_second": 4.282, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_msmarco_pairs_loss": 2.367374897003174, |
|
"eval_msmarco_pairs_runtime": 19.5819, |
|
"eval_msmarco_pairs_samples_per_second": 51.068, |
|
"eval_msmarco_pairs_steps_per_second": 1.634, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_nq_pairs_loss": 2.5974488258361816, |
|
"eval_nq_pairs_runtime": 37.1479, |
|
"eval_nq_pairs_samples_per_second": 26.919, |
|
"eval_nq_pairs_steps_per_second": 0.861, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_trivia_pairs_loss": 2.8283140659332275, |
|
"eval_trivia_pairs_runtime": 63.9593, |
|
"eval_trivia_pairs_samples_per_second": 15.635, |
|
"eval_trivia_pairs_steps_per_second": 0.5, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_quora_pairs_loss": 0.5709843039512634, |
|
"eval_quora_pairs_runtime": 3.6911, |
|
"eval_quora_pairs_samples_per_second": 135.462, |
|
"eval_quora_pairs_steps_per_second": 4.335, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5402521823472357, |
|
"eval_gooaq_pairs_loss": 1.9126713275909424, |
|
"eval_gooaq_pairs_runtime": 13.0893, |
|
"eval_gooaq_pairs_samples_per_second": 76.398, |
|
"eval_gooaq_pairs_steps_per_second": 2.445, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 0.5412221144519883, |
|
"grad_norm": 25.740751266479492, |
|
"learning_rate": 1.800646551724138e-05, |
|
"loss": 2.4683, |
|
"step": 3348 |
|
}, |
|
{ |
|
"epoch": 0.5562560620756547, |
|
"grad_norm": 26.303068161010742, |
|
"learning_rate": 1.850754310344828e-05, |
|
"loss": 2.3363, |
|
"step": 3441 |
|
}, |
|
{ |
|
"epoch": 0.5712900096993211, |
|
"grad_norm": 29.08409881591797, |
|
"learning_rate": 1.9008620689655172e-05, |
|
"loss": 2.474, |
|
"step": 3534 |
|
}, |
|
{ |
|
"epoch": 0.5863239573229874, |
|
"grad_norm": 39.23952865600586, |
|
"learning_rate": 1.9509698275862068e-05, |
|
"loss": 2.4987, |
|
"step": 3627 |
|
} |
|
], |
|
"logging_steps": 93, |
|
"max_steps": 18558, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 1856, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|