File size: 2,906 Bytes
a4a2f82 cfc770f 9c992a6 28bf9f1 75dcb42 21a5baf eb54f30 3246e3a e48c1ef b6961bd 22b1dea 162a8d5 0139966 7318a0e 5c80e3a 0cd3df7 a183143 aa7f186 7b30979 02d35ac 0c1e4b0 8934653 6182ef8 42ff635 0476b6f aa6643a cdf1648 a24764c 334ef94 3670160 44dc5da b7f6c96 4cf3332 11f98aa 5dee1cb cd8bf05 70e03b2 8598079 15f04fc 2a1ca5c 9b89216 f54d93b 175cb4d ddf7829 7fbd4c4 ca0f8b6 336a1b8 48207fb 72556db 3def4df 2cfd04f 84450a3 5ae04e4 8dae9e7 69e5a6c b232346 6e7ada7 eeb7192 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 |
Started at: 06:49:33 norbert2, 0.001, 256 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.31302871882915495 - MAE: 0.43550123211776 Validation loss : 0.19499443140294817 - MAE: 0.3375275454882718 Epoch: 1 Training loss: 0.20615242600440978 - MAE: 0.34713061837107256 Validation loss : 0.1856598754723867 - MAE: 0.3336331175439254 Epoch: 2 Training loss: 0.1905288130044937 - MAE: 0.331755109482026 Validation loss : 0.17539048691590628 - MAE: 0.3215592502543141 Epoch: 3 Training loss: 0.18015263855457306 - MAE: 0.3219696294630834 Validation loss : 0.17270306249459585 - MAE: 0.3160431837363882 Epoch: 4 Training loss: 0.17445907533168792 - MAE: 0.3157378477701357 Validation loss : 0.16871446867783865 - MAE: 0.31174037115932135 Epoch: 5 Training loss: 0.16842794716358184 - MAE: 0.31048144075869144 Validation loss : 0.16465580628977883 - MAE: 0.30827743898959675 Epoch: 6 Training loss: 0.1638428008556366 - MAE: 0.3059258622414145 Validation loss : 0.16216961708333757 - MAE: 0.3055697565557946 Epoch: 7 Training loss: 0.16214038848876952 - MAE: 0.30366597705137366 Validation loss : 0.16221882071759966 - MAE: 0.30516110512925027 Epoch: 8 Training loss: 0.1599144631624222 - MAE: 0.3020778099929018 Validation loss : 0.1606932199663586 - MAE: 0.30484856126917326 Epoch: 9 Training loss: 0.15885521054267884 - MAE: 0.29955762107383804 Validation loss : 0.15996411111619738 - MAE: 0.3046552257937225 Epoch: 10 Training loss: 0.15486705422401428 - MAE: 0.29813364601763576 Validation loss : 0.15921063390043047 - MAE: 0.3030789867279675 Epoch: 11 Training loss: 0.15471409738063813 - MAE: 0.2971227495552318 Validation loss : 0.15626052849822575 - MAE: 0.30025851520982877 Epoch: 12 Training loss: 0.15294371783733368 - MAE: 0.2958215080920924 Validation loss : 0.15606091254287296 - MAE: 0.29940732752482685 Epoch: 13 Training loss: 0.15222665548324585 - MAE: 0.2957011197318725 Validation loss : 0.15529203911622366 - MAE: 0.29927462986033415 Epoch: 14 Training loss: 0.1514916968345642 - MAE: 0.29329306132812527 Validation loss : 0.15512731340196398 - MAE: 0.29836073878752045 Epoch: 15 Training loss: 0.14988289177417755 - MAE: 0.29146307981803693 Validation loss : 0.1547977460755242 - MAE: 0.2984126944784066 Epoch: 16 Training loss: 0.14967645615339278 - MAE: 0.29069896681020774 Validation loss : 0.1542857653564877 - MAE: 0.299382354318533 Epoch: 17 Training loss: 0.14684118330478668 - MAE: 0.2883821342488331 Validation loss : 0.1548976716068056 - MAE: 0.29951980386692484 Stopped after 17 epochs. |