JuwonOh commited on
Commit
23c0bf2
·
1 Parent(s): cbac854

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -3
README.md CHANGED
@@ -14,7 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
- - Loss: 0.4919
18
 
19
  ## Model description
20
 
@@ -42,16 +42,19 @@ The following hyperparameters were used during training:
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_steps: 1000
45
- - num_epochs: 10
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
49
 
 
 
 
50
 
51
 
52
  ### Framework versions
53
 
54
  - Transformers 4.26.1
55
- - Pytorch 1.7.1+cu110
56
  - Datasets 2.10.1
57
  - Tokenizers 0.13.2
 
14
 
15
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 0.0078
18
 
19
  ## Model description
20
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_steps: 1000
45
+ - num_epochs: 100
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
49
 
50
+ | Training Loss | Epoch | Step | Validation Loss |
51
+ |:-------------:|:-----:|:----:|:---------------:|
52
+ | 0.3681 | 74.63 | 5000 | 0.0080 |
53
 
54
 
55
  ### Framework versions
56
 
57
  - Transformers 4.26.1
58
+ - Pytorch 1.10.1+cu111
59
  - Datasets 2.10.1
60
  - Tokenizers 0.13.2