zera09 commited on
Commit
d60a84f
1 Parent(s): 3dc7e42

End of training

Browse files
Files changed (2) hide show
  1. README.md +114 -0
  2. generation_config.json +7 -0
README.md ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: google/long-t5-tglobal-base
5
+ tags:
6
+ - generated_from_trainer
7
+ metrics:
8
+ - rouge
9
+ model-index:
10
+ - name: long_t5_4
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # long_t5_4
18
+
19
+ This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 3.0847
22
+ - Rouge1: 0.5303
23
+ - Rouge2: 0.3398
24
+ - Rougel: 0.477
25
+ - Rougelsum: 0.477
26
+ - Gen Len: 31.974
27
+
28
+ ## Model description
29
+
30
+ More information needed
31
+
32
+ ## Intended uses & limitations
33
+
34
+ More information needed
35
+
36
+ ## Training and evaluation data
37
+
38
+ More information needed
39
+
40
+ ## Training procedure
41
+
42
+ ### Training hyperparameters
43
+
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 0.0001
46
+ - train_batch_size: 8
47
+ - eval_batch_size: 8
48
+ - seed: 42
49
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
+ - lr_scheduler_type: linear
51
+ - num_epochs: 50
52
+
53
+ ### Training results
54
+
55
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
56
+ |:-------------:|:-----:|:-----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
57
+ | 2.0147 | 1.0 | 1000 | 1.5675 | 0.4907 | 0.3059 | 0.4453 | 0.4454 | 25.7975 |
58
+ | 1.7618 | 2.0 | 2000 | 1.5138 | 0.5037 | 0.3169 | 0.4578 | 0.458 | 26.608 |
59
+ | 1.5904 | 3.0 | 3000 | 1.5015 | 0.5091 | 0.3239 | 0.4645 | 0.4648 | 25.5405 |
60
+ | 1.4555 | 4.0 | 4000 | 1.5083 | 0.5183 | 0.3335 | 0.4727 | 0.4732 | 26.777 |
61
+ | 1.3579 | 5.0 | 5000 | 1.5363 | 0.5205 | 0.3353 | 0.4743 | 0.4744 | 27.916 |
62
+ | 1.2345 | 6.0 | 6000 | 1.5543 | 0.5193 | 0.338 | 0.4772 | 0.4769 | 25.6475 |
63
+ | 1.1663 | 7.0 | 7000 | 1.5570 | 0.5299 | 0.3449 | 0.4837 | 0.4837 | 26.9075 |
64
+ | 1.0754 | 8.0 | 8000 | 1.5953 | 0.5289 | 0.3422 | 0.4804 | 0.4804 | 29.1995 |
65
+ | 0.9901 | 9.0 | 9000 | 1.6392 | 0.5333 | 0.3443 | 0.483 | 0.4831 | 28.9815 |
66
+ | 0.9321 | 10.0 | 10000 | 1.6641 | 0.5269 | 0.3361 | 0.4764 | 0.4765 | 28.8695 |
67
+ | 0.87 | 11.0 | 11000 | 1.7062 | 0.5299 | 0.3409 | 0.4793 | 0.4794 | 29.366 |
68
+ | 0.8062 | 12.0 | 12000 | 1.7558 | 0.5287 | 0.342 | 0.4794 | 0.4798 | 29.29 |
69
+ | 0.7595 | 13.0 | 13000 | 1.8033 | 0.5256 | 0.3402 | 0.4784 | 0.4783 | 29.204 |
70
+ | 0.7195 | 14.0 | 14000 | 1.8229 | 0.5293 | 0.3425 | 0.4802 | 0.4803 | 30.156 |
71
+ | 0.668 | 15.0 | 15000 | 1.8817 | 0.5288 | 0.3421 | 0.4791 | 0.4792 | 30.1525 |
72
+ | 0.6283 | 16.0 | 16000 | 1.9278 | 0.5294 | 0.3404 | 0.478 | 0.4778 | 29.942 |
73
+ | 0.5957 | 17.0 | 17000 | 1.9536 | 0.5312 | 0.3416 | 0.4807 | 0.4809 | 29.525 |
74
+ | 0.5496 | 18.0 | 18000 | 2.0396 | 0.5309 | 0.3403 | 0.4788 | 0.479 | 30.359 |
75
+ | 0.5208 | 19.0 | 19000 | 2.0539 | 0.5312 | 0.3442 | 0.4813 | 0.481 | 30.173 |
76
+ | 0.491 | 20.0 | 20000 | 2.0836 | 0.5297 | 0.3395 | 0.4794 | 0.4792 | 29.554 |
77
+ | 0.4522 | 21.0 | 21000 | 2.1548 | 0.5282 | 0.3396 | 0.4751 | 0.4753 | 31.565 |
78
+ | 0.4339 | 22.0 | 22000 | 2.2076 | 0.5264 | 0.338 | 0.476 | 0.476 | 30.0425 |
79
+ | 0.4095 | 23.0 | 23000 | 2.2331 | 0.5258 | 0.3366 | 0.4751 | 0.475 | 31.307 |
80
+ | 0.3818 | 24.0 | 24000 | 2.3036 | 0.5275 | 0.3371 | 0.4756 | 0.4753 | 31.8185 |
81
+ | 0.362 | 25.0 | 25000 | 2.3462 | 0.529 | 0.3374 | 0.4739 | 0.4741 | 32.9885 |
82
+ | 0.3414 | 26.0 | 26000 | 2.3989 | 0.5335 | 0.3444 | 0.482 | 0.4819 | 30.4255 |
83
+ | 0.3188 | 27.0 | 27000 | 2.4419 | 0.5257 | 0.3367 | 0.4745 | 0.4744 | 30.6095 |
84
+ | 0.2976 | 28.0 | 28000 | 2.4965 | 0.5256 | 0.3336 | 0.4702 | 0.4701 | 33.6375 |
85
+ | 0.2896 | 29.0 | 29000 | 2.4841 | 0.5254 | 0.3341 | 0.4725 | 0.4725 | 32.7325 |
86
+ | 0.2702 | 30.0 | 30000 | 2.5704 | 0.5298 | 0.3399 | 0.4775 | 0.4778 | 31.307 |
87
+ | 0.2583 | 31.0 | 31000 | 2.6376 | 0.5306 | 0.3411 | 0.4773 | 0.4774 | 31.0695 |
88
+ | 0.2472 | 32.0 | 32000 | 2.6134 | 0.5266 | 0.3376 | 0.4729 | 0.473 | 32.3075 |
89
+ | 0.2361 | 33.0 | 33000 | 2.6922 | 0.5294 | 0.3391 | 0.4763 | 0.4764 | 31.5785 |
90
+ | 0.2242 | 34.0 | 34000 | 2.7246 | 0.5292 | 0.3383 | 0.4745 | 0.4747 | 32.823 |
91
+ | 0.2173 | 35.0 | 35000 | 2.7647 | 0.5294 | 0.3386 | 0.4754 | 0.4754 | 32.0915 |
92
+ | 0.2057 | 36.0 | 36000 | 2.7717 | 0.5297 | 0.343 | 0.4781 | 0.4781 | 32.132 |
93
+ | 0.1957 | 37.0 | 37000 | 2.8077 | 0.5257 | 0.3372 | 0.4729 | 0.4728 | 32.147 |
94
+ | 0.1895 | 38.0 | 38000 | 2.8661 | 0.5268 | 0.3375 | 0.4733 | 0.4734 | 32.156 |
95
+ | 0.1818 | 39.0 | 39000 | 2.8841 | 0.5272 | 0.3388 | 0.4747 | 0.475 | 31.3275 |
96
+ | 0.1749 | 40.0 | 40000 | 2.9060 | 0.5278 | 0.3395 | 0.4752 | 0.4751 | 31.835 |
97
+ | 0.1705 | 41.0 | 41000 | 2.9260 | 0.5262 | 0.3365 | 0.4729 | 0.4732 | 32.3635 |
98
+ | 0.163 | 42.0 | 42000 | 2.9924 | 0.5284 | 0.3383 | 0.4754 | 0.4754 | 31.4935 |
99
+ | 0.163 | 43.0 | 43000 | 2.9798 | 0.5299 | 0.3403 | 0.4762 | 0.4765 | 31.8165 |
100
+ | 0.1583 | 44.0 | 44000 | 2.9919 | 0.5291 | 0.3397 | 0.4755 | 0.4759 | 31.6065 |
101
+ | 0.1537 | 45.0 | 45000 | 3.0308 | 0.5281 | 0.3381 | 0.4748 | 0.4749 | 31.447 |
102
+ | 0.1493 | 46.0 | 46000 | 3.0491 | 0.5287 | 0.339 | 0.4753 | 0.4755 | 31.944 |
103
+ | 0.1437 | 47.0 | 47000 | 3.0595 | 0.5282 | 0.3383 | 0.4744 | 0.4746 | 31.833 |
104
+ | 0.1437 | 48.0 | 48000 | 3.0804 | 0.5307 | 0.3401 | 0.477 | 0.4771 | 31.837 |
105
+ | 0.1435 | 49.0 | 49000 | 3.0782 | 0.5312 | 0.3406 | 0.4772 | 0.4772 | 31.798 |
106
+ | 0.1392 | 50.0 | 50000 | 3.0847 | 0.5303 | 0.3398 | 0.477 | 0.477 | 31.974 |
107
+
108
+
109
+ ### Framework versions
110
+
111
+ - Transformers 4.45.1
112
+ - Pytorch 2.2.1
113
+ - Datasets 3.0.1
114
+ - Tokenizers 0.20.0
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "decoder_start_token_id": 0,
3
+ "eos_token_id": 1,
4
+ "max_length": 100,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.45.1"
7
+ }