File size: 2,626 Bytes
3061553 f9a6296 ed40b96 d847b76 2b0079d 8f23bf5 7dc4346 29a25f0 d3979b1 e1995fc 5aef70d bc84955 96c9b1d bb9cf08 0344f2f 87706e2 9e1f333 c92e465 4496332 05511dc 2c1426f bf3863a c22e3cd b418b7f 5736721 c73281f b1711d4 f1aff99 44f1488 a8476f8 3de698e 0eacc4b 190fd55 3ffd147 b4abbeb 7ab99dd cd6115e 786c61d fd51ae9 e0f96d0 e2b0e67 1833b58 a8e4579 b2206b0 5a42d52 87383ac 94efde3 361c95d eba74ef efef819 6d5b096 5ddb172 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 |
Started at: 06:25:02 norbert2, 0.001, 128 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.26594408303499223 - MAE: 0.3978314918997288 Validation loss : 0.19398287113975077 - MAE: 0.34193473061730845 Epoch: 1 Training loss: 0.1824344827234745 - MAE: 0.3272282950583859 Validation loss : 0.18646611360942617 - MAE: 0.3356285057272865 Epoch: 2 Training loss: 0.1719924859702587 - MAE: 0.31573345724688817 Validation loss : 0.17564336341970108 - MAE: 0.3236708021458511 Epoch: 3 Training loss: 0.16539704591035842 - MAE: 0.30983962574381724 Validation loss : 0.16587493612485774 - MAE: 0.3130272370269703 Epoch: 4 Training loss: 0.1597985264658928 - MAE: 0.30454135879504085 Validation loss : 0.1627409287235316 - MAE: 0.30946326864776774 Epoch: 5 Training loss: 0.153805871643126 - MAE: 0.29754943904831743 Validation loss : 0.16067567380035624 - MAE: 0.3053273121402249 Epoch: 6 Training loss: 0.15524272337555886 - MAE: 0.29936721330158883 Validation loss : 0.16136379583793528 - MAE: 0.30574541067519767 Epoch: 7 Training loss: 0.15196644343435764 - MAE: 0.29472622361245104 Validation loss : 0.1603953365017386 - MAE: 0.3040597750065931 Epoch: 8 Training loss: 0.1517229761183262 - MAE: 0.2955280447325332 Validation loss : 0.15818107917028315 - MAE: 0.30266661202354717 Epoch: 9 Training loss: 0.14952938295900822 - MAE: 0.2923810918291241 Validation loss : 0.15911468775833354 - MAE: 0.30159275869333224 Epoch: 10 Training loss: 0.1476191408187151 - MAE: 0.29077664893054256 Validation loss : 0.16191191182417028 - MAE: 0.3025654413929292 Epoch: 11 Training loss: 0.14881269738078118 - MAE: 0.29319386249020507 Validation loss : 0.15962355101809783 - MAE: 0.30117699112236357 Epoch: 12 Training loss: 0.1451718310266733 - MAE: 0.28961566673102823 Validation loss : 0.16168749376254923 - MAE: 0.30352713335972603 Epoch: 13 Training loss: 0.14698604956269265 - MAE: 0.2906234678272731 Validation loss : 0.15626758424674764 - MAE: 0.29897112160121386 Epoch: 14 Training loss: 0.14597914703190326 - MAE: 0.28922028553247936 Validation loss : 0.15734932308687882 - MAE: 0.2993024581905405 Epoch: 15 Training loss: 0.14458707965910433 - MAE: 0.28938238036868486 Validation loss : 0.15550065084415324 - MAE: 0.2983666009213386 Epoch: 16 |