File size: 2,698 Bytes
ec2fdb1 3844811 7c59103 5394486 03608e5 d750750 fdfc99a 197cf7a eb40cdf 5bdcf16 6fc84af 94b2230 b24877b caf95c2 bf265ca 535a9c1 7411a43 d34e089 c5f1408 63f1c02 ceccc29 5743c78 ae1f38f a5fbc89 d93c03a 9c9e463 4e74cb5 6a43f6e 7c84c42 acf7a03 c015ae6 e277f2b 90b8862 ac55872 de81d07 b5a527c 4af90b7 65dc7d4 2c9a3f4 d3c07b5 9d9a5c9 a417a75 ac8f475 cfc9567 5403a4b 66602f7 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 |
Started at: 09:20:17 nb-bert-base, 0.001, 512 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 0.6738183131584754 - MAE: 0.671559791949798 Validation loss : 0.3496964991092682 - MAE: 0.481417017739273 Epoch: 1 Training loss: 0.2604064620458163 - MAE: 0.3969383760031138 Validation loss : 0.22093867957592012 - MAE: 0.35313969813939866 Epoch: 2 Training loss: 0.20208562452059525 - MAE: 0.3434821083990275 Validation loss : 0.1883617550134659 - MAE: 0.3379014120649767 Epoch: 3 Training loss: 0.1782165696987739 - MAE: 0.31999390375888026 Validation loss : 0.16966865658760072 - MAE: 0.31801389648176803 Epoch: 4 Training loss: 0.16990482119413522 - MAE: 0.31338787739566665 Validation loss : 0.1638330489397049 - MAE: 0.3106934884794657 Epoch: 5 Training loss: 0.1681439051261315 - MAE: 0.31001855013012786 Validation loss : 0.16361375153064728 - MAE: 0.3107548141570412 Epoch: 6 Training loss: 0.16585055337502405 - MAE: 0.3068083279056544 Validation loss : 0.15926897823810576 - MAE: 0.30532344804199296 Epoch: 7 Training loss: 0.1640594372382531 - MAE: 0.30565590681538174 Validation loss : 0.15682196021080017 - MAE: 0.3021364879085161 Epoch: 8 Training loss: 0.15877233560268694 - MAE: 0.30019124717441487 Validation loss : 0.15556727051734925 - MAE: 0.3009043710813451 Epoch: 9 Training loss: 0.157441843014497 - MAE: 0.29782973353451203 Validation loss : 0.15489959716796875 - MAE: 0.3003483373637247 Epoch: 10 Training loss: 0.1569079951598094 - MAE: 0.29850384416372977 Validation loss : 0.15268584191799164 - MAE: 0.2973265470132604 Epoch: 11 Training loss: 0.15540967537806585 - MAE: 0.29631087713396775 Validation loss : 0.1516599327325821 - MAE: 0.29605310392714956 Epoch: 12 Training loss: 0.15601101173804358 - MAE: 0.29642744149228534 Validation loss : 0.1511812061071396 - MAE: 0.2954877034441389 Epoch: 13 Training loss: 0.15504613060217637 - MAE: 0.295577142279247 Validation loss : 0.15018589794635773 - MAE: 0.29435485554038643 Epoch: 14 |