File size: 2,499 Bytes
a8ac709 3859c25 6529817 4514b14 6e0a8f6 14f529a a72986b 1a69ea7 f297e0c 442fc24 f4c2df2 5317f9f c31f45e 5dc0118 78abf23 84557e7 2a1a3e9 a5e300e 95ea73f c7c2512 cba1cef 5fb793b bf159f7 863e0fc e7b94b7 a9a1d34 ca6032c 1723a4c 0557470 00b4de5 10d55ad c5ee41a 25bcbbf c0aab7a 30791a3 95fb2c8 3ffe382 4c3228c 6e11297 82848a1 f76b568 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 |
Started at: 12:25:48 nb-bert-base, 5e-06, 256 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 0.6139226149428975 - MAE: 0.6256533578877125 Validation loss : 0.38856891581886693 - MAE: 0.5055659858766396 Epoch: 1 Training loss: 0.35679552988572555 - MAE: 0.47769355546848036 Validation loss : 0.2988674891622443 - MAE: 0.427160547565342 Epoch: 2 Training loss: 0.2514720949259671 - MAE: 0.3919834602387325 Validation loss : 0.20192746033794 - MAE: 0.35617145796178046 Epoch: 3 Training loss: 0.18519783670252019 - MAE: 0.34110099718701625 Validation loss : 0.16307617410233147 - MAE: 0.31849970460916255 Epoch: 4 Training loss: 0.15608852465044368 - MAE: 0.3137766481088091 Validation loss : 0.14197467973357752 - MAE: 0.29514997623851247 Epoch: 5 Training loss: 0.13567890524864196 - MAE: 0.29194648927171835 Validation loss : 0.12267970411401045 - MAE: 0.27265562083225603 Epoch: 6 Training loss: 0.11870894933288748 - MAE: 0.2718382958457312 Validation loss : 0.11097796692659981 - MAE: 0.25758373825453007 Epoch: 7 Training loss: 0.10573411136865615 - MAE: 0.25600807072103077 Validation loss : 0.1004621064976642 - MAE: 0.24238915549512713 Epoch: 8 Training loss: 0.095039548386227 - MAE: 0.24105853745370692 Validation loss : 0.08920641753234361 - MAE: 0.22744885245923066 Epoch: 9 Training loss: 0.08564553965221752 - MAE: 0.2290994268022599 Validation loss : 0.08278850426799372 - MAE: 0.21588505087327514 Epoch: 10 Training loss: 0.07813239084048705 - MAE: 0.21902914877383883 Validation loss : 0.07834271557236973 - MAE: 0.20735584144221048 Epoch: 11 Training loss: 0.07202434621073983 - MAE: 0.20906077481733826 Validation loss : 0.07437812458527715 - MAE: 0.20084316074607808 Epoch: 12 Training loss: 0.06499892032959244 - MAE: 0.19809261164608147 |