File size: 3,197 Bytes
c80765a 57c5f87 fe885cb c3db1af fb9d943 a96bb15 426374a 33034e8 e4d3891 80986d2 165f615 db01149 ba6de8a 62ec26b fb06b70 7095b4c 9f1cdf5 8c747d1 ec83c1e 42c6cf5 efeae8b 8ec0d6f dd8dd1f 7ebcd20 bd2c140 bc6d7ec d03a806 dd28607 9ed2b39 5a31ea4 7dbf6b5 eef46ef 16a4945 0d7e410 8602263 aced0b3 b7cd71a e35c421 f63161c 7403d46 a98bea1 28f1d7b 4143126 4d5f2f6 7e91ee5 2850126 fc5d92d 63a3f60 809248e f06c467 c4140f4 d92075f f345587 f515eb6 c6df857 b76c8e8 46f9763 03bf202 77c2f3e 3c13c79 80f0412 5ee1fef aa31b2b 6f4c06d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 |
Started at: 11:52:44 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'afb829e3d0b861bd5f8cda6522b32ca0b097d7eb'}, {}) Epoch: 0 Training loss: 0.19169998451283105 - MSE: 0.3224822171590252 Validation loss : 0.1576854281593114 - MSE: 0.3052585763866773 Epoch: 1 Training loss: 0.170721908933238 - MSE: 0.3106345883186025 Validation loss : 0.14974205638282 - MSE: 0.30164828824445067 Epoch: 2 Training loss: 0.16939535588026047 - MSE: 0.31008999561987205 Validation loss : 0.15034097363241017 - MSE: 0.30232355450380055 Epoch: 3 Training loss: 0.1694463299293267 - MSE: 0.3101087405316537 Validation loss : 0.15061823884025216 - MSE: 0.3026353720119914 Epoch: 4 Training loss: 0.16941545233130456 - MSE: 0.310085018434928 Validation loss : 0.15079936920665205 - MSE: 0.30283750300395695 Epoch: 5 Training loss: 0.16935620966710543 - MSE: 0.31004690146169944 Validation loss : 0.15084448363631964 - MSE: 0.3028875614759272 Epoch: 6 Training loss: 0.16949052591072886 - MSE: 0.3101781524663373 Validation loss : 0.15096625522710383 - MSE: 0.30302316902771054 Epoch: 7 Training loss: 0.16949893361643742 - MSE: 0.310188779241269 Validation loss : 0.15100383502431214 - MSE: 0.30306482790138034 Epoch: 8 Training loss: 0.1694979808440334 - MSE: 0.3101921144601653 Validation loss : 0.15103243081830442 - MSE: 0.30309663850448487 Epoch: 9 Training loss: 0.1694973421724219 - MSE: 0.3101948706230913 Validation loss : 0.1510547660291195 - MSE: 0.303121549986372 Epoch: 10 Training loss: 0.16949688792228698 - MSE: 0.31019701891325113 Validation loss : 0.15107234381139278 - MSE: 0.3031413168423569 Epoch: 11 Training loss: 0.16949651574617938 - MSE: 0.3101986945843064 Validation loss : 0.15108662215061486 - MSE: 0.3031575035779497 Epoch: 12 Training loss: 0.16949623556513535 - MSE: 0.3102000477910796 Validation loss : 0.1510984308551997 - MSE: 0.30317088319088725 Epoch: 13 Training loss: 0.16949602914483924 - MSE: 0.3102011731046972 Validation loss : 0.15110818133689463 - MSE: 0.30318192262302546 Epoch: 14 Training loss: 0.16949585684035953 - MSE: 0.31020210549164046 Validation loss : 0.15111643797717988 - MSE: 0.30319126495260207 Epoch: 15 Training loss: 0.1694957421798455 - MSE: 0.3102031148284955 Validation loss : 0.15112340752966702 - MSE: 0.30319914952951876 Epoch: 16 Training loss: 0.1694956130103061 - MSE: 0.3102041142706642 Validation loss : 0.15112938289530575 - MSE: 0.3032059101165032 Epoch: 17 Training loss: 0.16949550995701237 - MSE: 0.31020503356981505 Validation loss : 0.1511345561593771 - MSE: 0.30321175591188876 Epoch: 18 Training loss: 0.16949546227329657 - MSE: 0.31020586811368406 Validation loss : 0.15113910171203315 - MSE: 0.3032168940185329 Epoch: 19 Training loss: 0.1694954017275258 - MSE: 0.3102065734353139 Validation loss : 0.15114298881962895 - MSE: 0.30322128473881094 Epoch: 20 Training loss: 0.16949532439834195 - MSE: 0.3102071649483396 |