File size: 961 Bytes
07d0fda 0877d51 11eee91 e089ced 88a30b5 09f616b 570dc53 f47a613 50e788e 2ef7680 4368dc7 fd34521 d0ce911 7783f1b 96e9bb1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
Started at: 13:43:22 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'afb829e3d0b861bd5f8cda6522b32ca0b097d7eb'}, {}) Epoch: 0 Training loss: 2.4582972388518485 - MSE: 1.516153895348859 Validation loss : 2.348844211548567 - MSE: 1.485455886438558 Epoch: 1 Training loss: 2.4582972388518485 - MSE: 1.516153895348859 Validation loss : 2.348844211548567 - MSE: 1.485455886438558 Epoch: 2 Training loss: 2.4582972388518485 - MSE: 1.516153895348859 Validation loss : 2.348844211548567 - MSE: 1.485455886438558 Epoch: 3 Training loss: 2.4582972388518485 - MSE: 1.516153895348859 Validation loss : 2.348844211548567 - MSE: 1.485455886438558 Epoch: 4 |