File size: 3,788 Bytes
7a3f53c 5ce0ed3 d8e5c78 47ae815 52abf14 c147493 a83b41f 97457c2 9f3c5a1 2d982a5 fdd1188 322f2e5 31d7a8e 9868c35 92494a1 2920471 ce64d36 cf5672a 834e83e 79b62d4 c2c7c3e 2f07ccc 34bad96 ce81692 6416240 8cb0747 ce12e9d 80b99be de50a1a cec9357 518fc09 ef95ceb 4b8330c 4f21930 035a68a e810ad4 2fbb178 15e5473 bb822d6 f5d8743 e270abd 3be0f93 71b04e5 6dbe30d 5fc3146 93e48ed 98df9fb d71c824 74b225e 287e0fa 612062a 5b2122e e137093 2f93ecd 91e8085 5b0f282 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 |
Started at: 19:59:36 nb-bert-base, 0.001, 320 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Started at: 08:43:48 nb-bert-base, 0.001, 320 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 0.5151217855513096 - MAE: 0.5709627828340156 Validation loss : 0.2530486115387508 - MAE: 0.3794793018729891 Epoch: 1 Training loss: 0.22132941260933875 - MAE: 0.36308932887849105 Validation loss : 0.18178956849234446 - MAE: 0.33422993724658395 Epoch: 2 Training loss: 0.1807804986834526 - MAE: 0.32391827749935836 Validation loss : 0.1691616369145257 - MAE: 0.3194961844436014 Epoch: 3 Training loss: 0.16955089792609215 - MAE: 0.3127070982966003 Validation loss : 0.16161458726440156 - MAE: 0.3090284268632521 Epoch: 4 Training loss: 0.1653532862663269 - MAE: 0.3086919556242476 Validation loss : 0.15732865567718232 - MAE: 0.30360814254621993 Epoch: 5 Training loss: 0.16227756589651107 - MAE: 0.3042661441997587 Validation loss : 0.15460806446416037 - MAE: 0.30049916526709397 Epoch: 6 Training loss: 0.16006522327661515 - MAE: 0.30157349356756313 Validation loss : 0.1526533163019589 - MAE: 0.2983025067294122 Epoch: 7 Training loss: 0.15598474368453025 - MAE: 0.2972633061454769 Validation loss : 0.15159909852913447 - MAE: 0.29780595918171027 Epoch: 8 Training loss: 0.15715683475136757 - MAE: 0.2993698693050316 Validation loss : 0.14978873516832078 - MAE: 0.2951820444435869 Epoch: 9 Training loss: 0.15433909222483636 - MAE: 0.2960977264887328 Validation loss : 0.14908447116613388 - MAE: 0.2946113529152755 Epoch: 10 Training loss: 0.15214167535305023 - MAE: 0.29483903058219846 Validation loss : 0.14805126083748682 - MAE: 0.2932608487909774 Epoch: 11 Training loss: 0.15106371901929377 - MAE: 0.2920793530721409 Validation loss : 0.1474460448537554 - MAE: 0.29279081904305354 Epoch: 12 Training loss: 0.15020511336624623 - MAE: 0.2906949591110362 Validation loss : 0.1477930418082646 - MAE: 0.29405441929054094 Epoch: 13 Training loss: 0.14813920147717 - MAE: 0.2906090941705027 Validation loss : 0.1457444993512971 - MAE: 0.29027402522304774 Epoch: 14 Training loss: 0.1487877056002617 - MAE: 0.2888525660923103 Validation loss : 0.14652987463133677 - MAE: 0.29252290711365386 Epoch: 15 Training loss: 0.1479304190725088 - MAE: 0.2892559876816411 Validation loss : 0.14556479028293065 - MAE: 0.29082692460944776 Epoch: 16 |