File size: 2,890 Bytes
6a0262b 91dd1ea c10d212 5e0f2a7 a66f788 c8fa7de 74fd1b5 0b3ae78 6f51af3 06b5032 de16fed 28b8ba2 7c5a45e bece3c4 8c64a1e 0808e54 0c4a126 7227538 3821e2f fe2cf47 e65ae93 a5cc751 99d643d e4ba947 c9ec556 a96e2e0 2ce3c51 02607a8 270aaa7 b393a05 b6f827d 1627b8c 25cd78a fc07372 03beba3 3340cfa b73265c 626fdfd e05a46e e73f52f 5c15c38 3494494 1ae31f6 3baadc9 177fdc1 d11ad39 907cfb7 624da24 c9c8e81 93de992 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 |
Started at: 12:15:47 nb-bert-base, 0.001, 128 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 0.31982501298189164 - MAE: 0.4287338867738713 Validation loss : 0.16746729612350464 - MAE: 0.31536104399438897 Epoch: 1 Training loss: 0.16691179797053338 - MAE: 0.3121238207010245 Validation loss : 0.15773158739594853 - MAE: 0.3046954138541735 Epoch: 2 Training loss: 0.15689186841249467 - MAE: 0.3008464796601783 Validation loss : 0.15233191146570094 - MAE: 0.29779621490764496 Epoch: 3 Training loss: 0.15314741998910905 - MAE: 0.296528327871902 Validation loss : 0.1497987894450917 - MAE: 0.29459793118874833 Epoch: 4 Training loss: 0.14834036141633988 - MAE: 0.2910206240007747 Validation loss : 0.1474542622180546 - MAE: 0.2915430406480437 Epoch: 5 Training loss: 0.14879375964403152 - MAE: 0.2913913004246226 Validation loss : 0.14557460432543473 - MAE: 0.2891730110373066 Epoch: 6 Training loss: 0.14662521675229073 - MAE: 0.2886886095421301 Validation loss : 0.14476492475060856 - MAE: 0.28795673596095495 Epoch: 7 Training loss: 0.14347930431365966 - MAE: 0.2869532495194668 Validation loss : 0.1444830390460351 - MAE: 0.2884910243549571 Epoch: 8 Training loss: 0.14114041782915593 - MAE: 0.2831716332297515 Validation loss : 0.14459675462806926 - MAE: 0.28917789765778673 Epoch: 9 Training loss: 0.1433439274132252 - MAE: 0.2865429012477346 Validation loss : 0.14251315286930868 - MAE: 0.2853422469518363 Epoch: 10 Training loss: 0.1406331904232502 - MAE: 0.28132702192119613 Validation loss : 0.14170135689132354 - MAE: 0.2838292713935197 Epoch: 11 Training loss: 0.14126647457480432 - MAE: 0.28410774431585517 Validation loss : 0.1421863292069996 - MAE: 0.2853570444611879 Epoch: 12 Training loss: 0.13998315826058388 - MAE: 0.282938143440139 Validation loss : 0.1432714159874355 - MAE: 0.2874122980672003 Epoch: 13 Training loss: 0.1395747384428978 - MAE: 0.2831718714019129 Validation loss : 0.14076389372348785 - MAE: 0.283541297177118 Epoch: 14 Training loss: 0.139395814538002 - MAE: 0.2820615278357802 Validation loss : 0.14124228147899404 - MAE: 0.2840655921472793 Epoch: 15 Training loss: 0.13893909364938736 - MAE: 0.2813219249611208 |