File size: 2,701 Bytes
89e79dc ec771a3 f445a1f f7ea47e 21cea55 02ab522 6d2bb95 87f214e 7186a39 79fe858 b751190 9e10751 4656908 6dcdf15 bb71c16 3d71114 0edc455 b2d70ee 127064a fe2d0fb f20fa87 589edf8 b139430 768bceb 414b734 6eff4b2 10cf495 f2bddd3 3b0320d d06925b 1d6421d 7a56191 e0011f5 c96cf14 3a7b0b5 f1a40e6 06b2fb5 4ad597f af94459 0f10186 fe02582 447f202 1972bbe 87ff732 031b935 ff20ffd |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 |
Started at: 15:32:56 nb-bert-base, 1e-06, 128 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 0.7517342874407769 - MAE: 0.7057451002280275 Validation loss : 0.2215285485281664 - MAE: 0.36895833487830076 Epoch: 1 Training loss: 0.19639551222324372 - MAE: 0.3396230939000504 Validation loss : 0.19613290621953852 - MAE: 0.3407071633363561 Epoch: 2 Training loss: 0.189819974899292 - MAE: 0.33152794044204265 Validation loss : 0.1862245491322349 - MAE: 0.3306322601904093 Epoch: 3 Training loss: 0.18293176978826522 - MAE: 0.32441511104916815 Validation loss : 0.17851299398085652 - MAE: 0.3229232679845315 Epoch: 4 Training loss: 0.17267104253172874 - MAE: 0.31568550798758915 Validation loss : 0.1724623275153777 - MAE: 0.3165547082446247 Epoch: 5 Training loss: 0.16829490080475806 - MAE: 0.3120916956077127 Validation loss : 0.16718720863847172 - MAE: 0.3110304108857384 Epoch: 6 Training loss: 0.16253711566329002 - MAE: 0.3063455763538055 Validation loss : 0.16258587469072902 - MAE: 0.30655345038718496 Epoch: 7 Training loss: 0.16034438580274582 - MAE: 0.30328251732911143 Validation loss : 0.1597431661451564 - MAE: 0.30170444947052183 Epoch: 8 Training loss: 0.1557214939594269 - MAE: 0.2978233031704816 Validation loss : 0.15588842639151743 - MAE: 0.2980465650755997 Epoch: 9 Training loss: 0.15138985186815263 - MAE: 0.29315482496406786 Validation loss : 0.15272060474928686 - MAE: 0.2938781084993315 Epoch: 10 Training loss: 0.14734334856271744 - MAE: 0.28864492743138265 Validation loss : 0.14985116001437693 - MAE: 0.29023116106870367 Epoch: 11 Training loss: 0.1457201571762562 - MAE: 0.2867967428580185 Validation loss : 0.1477173132931485 - MAE: 0.2867585527915968 Epoch: 12 Training loss: 0.14163992531597613 - MAE: 0.2835520938095877 Validation loss : 0.14506700503475525 - MAE: 0.2834552494116601 Epoch: 13 Training loss: 0.13645638167858123 - MAE: 0.2770969076622815 Validation loss : 0.14318845815518322 - MAE: 0.28016682881578747 Epoch: 14 |