File size: 3,434 Bytes
5552b94 ae2fd9c a175971 da2905e 4e0388f 7ee0095 1032667 68e9b01 4d8ecee d24808e 0e9b655 91cd278 1674edc dccfaf7 b9ad253 667a823 7796f67 5cff094 cbf1e6b b5f1c41 c034201 cd862b5 176932e c5ce206 4e27620 9d367dc 1364cc1 b55770b a713748 cfe319b 192d6f2 dc1c95e 75a8cc8 eda653b d3d602f ddf2d92 73f6309 60bafca 9afeddf 30452f0 3527ea1 41502e2 79d49ef e92f396 56c80eb 47115c7 14599ab 2a07878 42b860c b2a9c86 9c6556d e03f879 f84dcfb c3f630f 416e590 791d227 9c5cc55 494fd45 821dcce 87d4837 832a220 c775220 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 |
Started at: 11:40:24 nb-bert-base, 0.001, 512 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 0.6829635271659265 - MAE: 0.6693431870059429 Validation loss : 0.5015756666660309 - MAE: 0.5997545175999092 Epoch: 1 Training loss: 0.26068564561697155 - MAE: 0.40026651290975057 Validation loss : 0.2343240737915039 - MAE: 0.3639362216460188 Epoch: 2 Training loss: 0.20825687165443713 - MAE: 0.34524354469083296 Validation loss : 0.20746740698814392 - MAE: 0.35860255150045284 Epoch: 3 Training loss: 0.1804403788768328 - MAE: 0.324681643823719 Validation loss : 0.17072068750858307 - MAE: 0.3124492505853874 Epoch: 4 Training loss: 0.17333761086830726 - MAE: 0.3132351201497551 Validation loss : 0.17095611095428467 - MAE: 0.31916315360789926 Epoch: 5 Training loss: 0.16724205017089844 - MAE: 0.3070826757836269 Validation loss : 0.16323553621768952 - MAE: 0.30865708568301353 Epoch: 6 Training loss: 0.16199684257690722 - MAE: 0.3026997098344876 Validation loss : 0.1614771842956543 - MAE: 0.3067986414590729 Epoch: 7 Training loss: 0.16401460537543663 - MAE: 0.3037438752420999 Validation loss : 0.15994466841220856 - MAE: 0.30503234812758634 Epoch: 8 Training loss: 0.16340345602769119 - MAE: 0.30281380735994723 Validation loss : 0.1587683379650116 - MAE: 0.3037075319625505 Epoch: 9 Training loss: 0.16126641172629136 - MAE: 0.3015211312899261 Validation loss : 0.15727587938308715 - MAE: 0.30159571389132384 Epoch: 10 Training loss: 0.16030011841884026 - MAE: 0.29977143854425936 Validation loss : 0.1565701872110367 - MAE: 0.30090318825320855 Epoch: 11 Training loss: 0.15990765392780304 - MAE: 0.3001817094076697 Validation loss : 0.15565772354602814 - MAE: 0.29973778065557 Epoch: 12 Training loss: 0.15747033861967233 - MAE: 0.2969911949201895 Validation loss : 0.15501231849193572 - MAE: 0.2991009715387324 Epoch: 13 Training loss: 0.15866013902884263 - MAE: 0.2974925745491436 Validation loss : 0.15455043613910674 - MAE: 0.298642265264826 Epoch: 14 Training loss: 0.15624549755683312 - MAE: 0.29534914888328495 Validation loss : 0.15433281362056733 - MAE: 0.2986108357413379 Epoch: 15 Training loss: 0.15603324541678795 - MAE: 0.2951994368480578 Validation loss : 0.1534305989742279 - MAE: 0.2972280074655698 Epoch: 16 Training loss: 0.15464195150595444 - MAE: 0.2940797652071446 Validation loss : 0.15260524451732635 - MAE: 0.29575102870976666 Epoch: 17 Training loss: 0.1543641434266017 - MAE: 0.2931998174008073 Validation loss : 0.15271952152252197 - MAE: 0.29645292921958 Epoch: 18 Training loss: 0.15505598829342768 - MAE: 0.29452743047415114 Validation loss : 0.15231834948062897 - MAE: 0.29597031674905394 Epoch: 19 Training loss: 0.1531683344107408 - MAE: 0.2934012395219041 |