temporary0-0name commited on
Commit
96fa43c
·
1 Parent(s): fb499d0

End of training

Browse files
Files changed (1) hide show
  1. README.md +23 -23
README.md CHANGED
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the wikitext dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.1400
21
 
22
  ## Model description
23
 
@@ -37,38 +37,38 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0001
40
- - train_batch_size: 64
41
- - eval_batch_size: 64
42
  - seed: 42
43
  - gradient_accumulation_steps: 8
44
- - total_train_batch_size: 512
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: cosine
47
  - lr_scheduler_warmup_steps: 100
48
- - num_epochs: 10
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
- | 9.1576 | 0.55 | 50 | 7.9518 |
55
- | 7.1585 | 1.1 | 100 | 6.6554 |
56
- | 6.4795 | 1.65 | 150 | 6.2877 |
57
- | 6.1004 | 2.19 | 200 | 5.8841 |
58
- | 5.3975 | 2.74 | 250 | 4.3378 |
59
- | 3.2884 | 3.29 | 300 | 2.1826 |
60
- | 1.7833 | 3.84 | 350 | 1.1134 |
61
- | 1.0053 | 4.39 | 400 | 0.6347 |
62
- | 0.6362 | 4.94 | 450 | 0.4108 |
63
- | 0.4388 | 5.49 | 500 | 0.2961 |
64
- | 0.3388 | 6.04 | 550 | 0.2316 |
65
- | 0.2713 | 6.58 | 600 | 0.1930 |
66
- | 0.235 | 7.13 | 650 | 0.1695 |
67
- | 0.2103 | 7.68 | 700 | 0.1550 |
68
- | 0.1953 | 8.23 | 750 | 0.1466 |
69
- | 0.1876 | 8.78 | 800 | 0.1422 |
70
- | 0.1834 | 9.33 | 850 | 0.1403 |
71
- | 0.1812 | 9.88 | 900 | 0.1400 |
72
 
73
 
74
  ### Framework versions
 
17
 
18
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the wikitext dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.2449
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0001
40
+ - train_batch_size: 32
41
+ - eval_batch_size: 32
42
  - seed: 42
43
  - gradient_accumulation_steps: 8
44
+ - total_train_batch_size: 256
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: cosine
47
  - lr_scheduler_warmup_steps: 100
48
+ - num_epochs: 5
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
+ | 9.1704 | 0.27 | 50 | 8.0057 |
55
+ | 7.2118 | 0.55 | 100 | 6.6834 |
56
+ | 6.5244 | 0.82 | 150 | 6.3491 |
57
+ | 6.2201 | 1.1 | 200 | 6.0229 |
58
+ | 5.7189 | 1.37 | 250 | 5.1311 |
59
+ | 4.1268 | 1.65 | 300 | 2.9582 |
60
+ | 2.4963 | 1.92 | 350 | 1.7429 |
61
+ | 1.5611 | 2.2 | 400 | 1.0743 |
62
+ | 1.0537 | 2.47 | 450 | 0.7155 |
63
+ | 0.7665 | 2.75 | 500 | 0.5189 |
64
+ | 0.5947 | 3.02 | 550 | 0.4061 |
65
+ | 0.4782 | 3.29 | 600 | 0.3396 |
66
+ | 0.4161 | 3.57 | 650 | 0.2976 |
67
+ | 0.3785 | 3.84 | 700 | 0.2718 |
68
+ | 0.3491 | 4.12 | 750 | 0.2567 |
69
+ | 0.3319 | 4.39 | 800 | 0.2488 |
70
+ | 0.3286 | 4.67 | 850 | 0.2455 |
71
+ | 0.326 | 4.94 | 900 | 0.2449 |
72
 
73
 
74
  ### Framework versions