File size: 3,301 Bytes
bbb722c 4c157ac 2bde72e 3081378 16c7808 3cc8443 9d9ea54 debea42 4bd8d4e 25b4faa a29f512 36bda71 745b97f 2f3d716 2cb0d6f eb83d32 5aeb413 686fa7c 45248f0 4a5b452 c2d5a7f bb137ee 8fee9fc c6b6989 41fde32 f7a96cd 3c7fcf3 b8dd206 15fc6d8 9cb438d 10c488e 20e2cf6 94c9d98 4e97b3f 87790cb c915edc a1a8ba3 d8d14d9 994bffd 2414614 a6bb4cc dfa8f17 13835ec 5f49bea 820f0eb e35aeaf 33393bd f0a5ed6 378a5b0 80e2b3a 59badb3 80e0f04 6558f71 6322791 77fe01d a412ff7 6977ec9 205f721 ed6906f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 |
Started at: 13:11:04 nb-bert-base, 0.001, 256 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 0.47113645792007447 - MAE: 0.5372690642533006 Validation loss : 0.25360237227545845 - MAE: 0.38100557323888096 Epoch: 1 Training loss: 0.1951601755619049 - MAE: 0.3381085981919985 Validation loss : 0.16845231420463985 - MAE: 0.3146699028671109 Epoch: 2 Training loss: 0.1700725769996643 - MAE: 0.3132894074707096 Validation loss : 0.16252560251288944 - MAE: 0.3101315378266 Epoch: 3 Training loss: 0.16299866735935212 - MAE: 0.30571749229812295 Validation loss : 0.1575031735830837 - MAE: 0.3036086169337241 Epoch: 4 Training loss: 0.16002290964126586 - MAE: 0.3030143814677838 Validation loss : 0.1555153015587065 - MAE: 0.30161708462789844 Epoch: 5 Training loss: 0.15610405176877976 - MAE: 0.2988610182826646 Validation loss : 0.15254790584246317 - MAE: 0.2976352880446658 Epoch: 6 Training loss: 0.15590642750263214 - MAE: 0.2972598594587939 Validation loss : 0.15147964821921456 - MAE: 0.29633871076747587 Epoch: 7 Training loss: 0.15364289730787278 - MAE: 0.2946944064180254 Validation loss : 0.1510375224881702 - MAE: 0.29637298396301465 Epoch: 8 Training loss: 0.1531897461414337 - MAE: 0.29468431902613157 Validation loss : 0.1495788519581159 - MAE: 0.29459826933056066 Epoch: 9 Training loss: 0.15212808579206466 - MAE: 0.2927512017766453 Validation loss : 0.14972183108329773 - MAE: 0.2947980987493768 Epoch: 10 Training loss: 0.15067319005727767 - MAE: 0.29223722020557824 Validation loss : 0.14739697674910227 - MAE: 0.29118543288491733 Epoch: 11 Training loss: 0.14746269524097444 - MAE: 0.28925764259432013 Validation loss : 0.14714962244033813 - MAE: 0.2916633442698583 Epoch: 12 Training loss: 0.1482395714521408 - MAE: 0.29003915570395034 Validation loss : 0.14708403746287027 - MAE: 0.2910518509892996 Epoch: 13 Training loss: 0.1483614018559456 - MAE: 0.29189581430634054 Validation loss : 0.14528870417012107 - MAE: 0.28829864578681347 Epoch: 14 Training loss: 0.14468369722366334 - MAE: 0.2842549207700623 Validation loss : 0.14488287270069122 - MAE: 0.28816803717811224 Epoch: 15 Training loss: 0.14738886833190917 - MAE: 0.2893857530495054 Validation loss : 0.14360415935516357 - MAE: 0.2859131647420325 Epoch: 16 Training loss: 0.14493520855903624 - MAE: 0.2866658051391056 Validation loss : 0.14300978928804398 - MAE: 0.2846472568063103 Epoch: 17 Training loss: 0.1460772478580475 - MAE: 0.2878771283563147 Validation loss : 0.14235875838332707 - MAE: 0.2838446738966764 Epoch: 18 Training loss: 0.14529186487197876 - MAE: 0.286218467370668 |