natanea commited on
Commit
14c6431
·
verified ·
1 Parent(s): aa1c464

Model save

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the multi_news dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 2.1968
22
 
23
  ## Model description
24
 
@@ -38,11 +38,11 @@ More information needed
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 2e-05
41
- - train_batch_size: 2
42
- - eval_batch_size: 2
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
- - total_train_batch_size: 4
46
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
48
  - num_epochs: 1
@@ -52,7 +52,7 @@ The following hyperparameters were used during training:
52
 
53
  | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:-----:|:----:|:---------------:|
55
- | 2.3197 | 1.0 | 1250 | 2.1968 |
56
 
57
 
58
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the multi_news dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.9843
22
 
23
  ## Model description
24
 
 
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 2e-05
41
+ - train_batch_size: 4
42
+ - eval_batch_size: 4
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
+ - total_train_batch_size: 8
46
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
48
  - num_epochs: 1
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:-----:|:----:|:---------------:|
55
+ | 2.0612 | 1.0 | 2500 | 1.9843 |
56
 
57
 
58
  ### Framework versions